CN113055715A - Processing method, device and equipment for playing resources and storage medium - Google Patents

Processing method, device and equipment for playing resources and storage medium Download PDF

Info

Publication number
CN113055715A
CN113055715A CN201911384286.7A CN201911384286A CN113055715A CN 113055715 A CN113055715 A CN 113055715A CN 201911384286 A CN201911384286 A CN 201911384286A CN 113055715 A CN113055715 A CN 113055715A
Authority
CN
China
Prior art keywords
playing
information
resources
display end
user
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201911384286.7A
Other languages
Chinese (zh)
Inventor
桑建
陈庆勇
魏远伦
马玉涛
吴毅
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
China Mobile Communications Group Co Ltd
China Mobile Chengdu ICT Co Ltd
Original Assignee
China Mobile Communications Group Co Ltd
China Mobile Chengdu ICT Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by China Mobile Communications Group Co Ltd, China Mobile Chengdu ICT Co Ltd filed Critical China Mobile Communications Group Co Ltd
Priority to CN201911384286.7A priority Critical patent/CN113055715A/en
Publication of CN113055715A publication Critical patent/CN113055715A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/239Interfacing the upstream path of the transmission network, e.g. prioritizing client content requests
    • H04N21/2393Interfacing the upstream path of the transmission network, e.g. prioritizing client content requests involving handling client requests
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/044Recurrent networks, e.g. Hopfield networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/02Feature extraction for speech recognition; Selection of recognition unit
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/08Speech classification or search
    • G10L15/16Speech classification or search using artificial neural networks
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/03Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 characterised by the type of extracted parameters
    • G10L25/24Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 characterised by the type of extracted parameters the extracted parameters being the cepstrum
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/234Processing of video elementary streams, e.g. splicing of video streams, manipulating MPEG-4 scene graphs
    • H04N21/2347Processing of video elementary streams, e.g. splicing of video streams, manipulating MPEG-4 scene graphs involving video stream encryption
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/25Management operations performed by the server for facilitating the content distribution or administrating data related to end-users or client devices, e.g. end-user or client device authentication, learning user preferences for recommending movies
    • H04N21/254Management at additional data server, e.g. shopping server, rights management server
    • H04N21/2541Rights Management
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/25Management operations performed by the server for facilitating the content distribution or administrating data related to end-users or client devices, e.g. end-user or client device authentication, learning user preferences for recommending movies
    • H04N21/258Client or end-user data management, e.g. managing client capabilities, user preferences or demographics, processing of multiple end-users preferences to derive collaborative data
    • H04N21/25866Management of end-user data
    • H04N21/25875Management of end-user data involving end-user authentication
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream, rendering scenes according to MPEG-4 scene graphs
    • H04N21/4405Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream, rendering scenes according to MPEG-4 scene graphs involving video stream decryption
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/441Acquiring end-user identification, e.g. using personal code sent by the remote control or by inserting a card
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/45Management operations performed by the client for facilitating the reception of or the interaction with the content or administrating data related to the end-user or to the client device itself, e.g. learning user preferences for recommending movies, resolving scheduling conflicts
    • H04N21/462Content or additional data management, e.g. creating a master electronic program guide from data received from the Internet and a Head-end, controlling the complexity of a video stream by scaling the resolution or bit-rate based on the client capabilities
    • H04N21/4627Rights management associated to the content
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/02Feature extraction for speech recognition; Selection of recognition unit
    • G10L2015/025Phonemes, fenemes or fenones being the recognition units
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • G10L2015/223Execution procedure of a spoken command

Abstract

The embodiment of the invention provides a processing method, a device, equipment and a storage medium of playing resources, wherein the method comprises the following steps: receiving application playing information sent by a display end, wherein the application playing information comprises display end information; identifying the authority of the display terminal according to the information of the display terminal to obtain an authority identification result; acquiring a playing resource corresponding to the display end from the cloud server according to the permission identification result and the application playing information; and when the received operation information input by the user is the control playing information, controlling the display end to play the playing resource according to the control playing information. The invention can control the display end according to the operation information input by the user, and carry out authentication and authority updating on the playing resources in real time under the unattended condition, thereby improving the use experience of the user.

Description

Processing method, device and equipment for playing resources and storage medium
Technical Field
The present invention relates to the field of intelligent control technologies, and in particular, to a method, an apparatus, a device, and a computer-readable storage medium for processing a playback resource.
Background
When playing files, a traditional exhibition hall needs workers to copy the files to be played in advance, meanwhile, wind control standby personnel wait, if a plurality of files need to be played, the traditional exhibition hall can only play the files in sequence, the played files cannot be authenticated and authority can not be updated in real time, if the played files need to be replaced or data needs to be modified midway, multiple people are needed to cooperate, and manpower and time are greatly consumed.
Disclosure of Invention
The embodiment of the invention provides a processing method, a processing device, processing equipment and a computer readable storage medium for playing resources, which can control a display end according to operation information input by a user, authenticate and update the authority of the playing resources in real time under the unattended condition, and improve the use experience of the user.
In a first aspect, a method for processing a playback resource is provided, where the method includes: receiving application playing information sent by a display end, wherein the application playing information comprises display end information; identifying the authority of the display terminal according to the information of the display terminal to obtain an authority identification result; acquiring a playing resource corresponding to the display end from the cloud server according to the permission identification result and the application playing information; and when the received operation information input by the user is the control playing information, controlling the display end to play the playing resource according to the control playing information.
In some implementations of the first aspect, when the received operation information input by the user is control light information, the light of the display end is controlled according to the control light information.
In some implementations of the first aspect, the application playback information, the operation information, and the playback resources are stored encrypted.
In some implementations of the first aspect, the received operation information input by the user is input by the user through a keyboard and/or voice.
In some implementations of the first aspect, the operational information input by the user via speech is recognized according to an LSTM + CTC neural network.
In some implementations of the first aspect, before controlling the display end to play the playback resource according to the control playback information, the method includes: and adjusting the playing sequence of the playing resources and/or modifying the playing resources according to the control playing information.
In a second aspect, there is provided an apparatus for processing a playback resource, the apparatus comprising: the communication module is used for receiving application playing information sent by the display end, and the application playing information comprises display end information; the authority identification module is used for identifying the authority of the display terminal according to the information of the display terminal to obtain an authority identification result; the acquisition module is used for acquiring the playing resources corresponding to the display end from the cloud server according to the authority identification result and the application playing information; and the processing module is used for controlling the display end to play the playing resources according to the control playing information when the received operation information input by the user is the control playing information.
In some implementations of the second aspect, the communication module includes: at least one of a Bluetooth communication module, an infrared communication module, a WIFI communication module and a Zigbee communication module.
In some implementation manners of the second aspect, the display device further includes a light module, configured to control light of the display end according to the control light information when the received operation information input by the user is the control light information.
In some implementations of the second aspect, the processing module is further configured to identify operational information input by the user via speech according to the LSTM + CTC neural network.
In some realizations of the second aspect, before controlling the display end to play the playback resources according to the control playback information: the processing module is further configured to adjust a playing sequence of the playing resources and/or modify the playing resources according to the playing control information.
In a third aspect, a device for processing a playback resource is provided, the device comprising: a processor and a memory storing computer program instructions; the processor, when executing the computer program instructions, implements the method of processing a playback resource of the first aspect or some realizations of the first aspect.
In a fourth aspect, a computer-readable storage medium is provided, on which computer program instructions are stored, which, when executed by a processor, implement the processing method of playing resources in the first aspect or some realizations of the first aspect.
The invention relates to the technical field of intelligent control, in particular to a processing method, a device, equipment and a computer readable storage medium of playing resources, wherein the playing application information sent by a display terminal is received, and the playing application information comprises the information of the display terminal; identifying the authority of the display terminal according to the information of the display terminal to obtain an authority identification result; acquiring a playing resource corresponding to the display end from the cloud server according to the permission identification result and the application playing information; when the received operation information input by the user is the control playing information, the display end is controlled to play the playing resources according to the control playing information, the display end can be flexibly controlled, the playing resources are authenticated and authority is updated in real time under the unattended condition, and the use experience of the user is improved.
Drawings
In order to more clearly illustrate the technical solutions of the embodiments of the present invention, the drawings required to be used in the embodiments of the present invention will be briefly described below, and for those skilled in the art, other drawings can be obtained according to the drawings without creative efforts.
Fig. 1 is a schematic flowchart of a processing method for playing back resources according to an embodiment of the present invention;
FIG. 2 is a flowchart illustrating an operation of processing a playback resource according to an embodiment of the present invention;
fig. 3 is a schematic structural diagram of a speech information recognition module according to an embodiment of the present invention;
fig. 4 is a flowchart illustrating a method for recognizing voice information according to an embodiment of the present invention;
FIG. 5 is a block diagram of a long term memory network according to an embodiment of the present invention;
fig. 6 is a schematic structural diagram of a processing apparatus for playing back a resource according to an embodiment of the present invention;
fig. 7 is a schematic structural diagram of a processing device for playing back a resource according to an embodiment of the present invention.
Detailed Description
Features and exemplary embodiments of various aspects of the present invention will be described in detail below, and in order to make objects, technical solutions and advantages of the present invention more apparent, the present invention will be further described in detail below with reference to the accompanying drawings and embodiments. It should be understood that the specific embodiments described herein are merely illustrative of the invention and are not to be construed as limiting the invention. It will be apparent to one skilled in the art that the present invention may be practiced without some of these specific details. The following description of the embodiments is merely intended to provide a better understanding of the present invention by illustrating examples of the present invention.
It is noted that, herein, relational terms such as first and second, and the like may be used solely to distinguish one entity or action from another entity or action without necessarily requiring or implying any actual such relationship or order between such entities or actions. Also, the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, an element defined by the phrase "comprising … …" does not exclude the presence of other identical elements in a process, method, article, or apparatus that comprises the element.
The term "and/or" herein is merely an association describing an associated object, meaning that three relationships may exist, e.g., a and/or B, may mean: a exists alone, A and B exist simultaneously, and B exists alone.
In the embodiment of the present invention, the processing method of the playing resource may be applied to a central control system installed on a central control device (Android or iOS smart phone or pad), and the central control system is in communication connection with a display end and a server end.
The following describes a processing method for playing resources according to an embodiment of the present invention with reference to fig. 1 and fig. 2.
Fig. 1 is a flowchart illustrating a processing method for playing back a resource according to an embodiment of the present invention.
Fig. 2 is a schematic operational flow diagram for processing a playback resource according to an embodiment of the present invention.
The processing method of playing back the assets may include S110-S140.
S110, receiving application playing information sent by the display end, wherein the application playing information comprises display end information.
Specifically, the application playing information sent by the display end includes the relevant attribute of the application playing file and the display end information, the relevant attribute of the application playing file may include the file name, the file type, the thumbnail and the playing address of the application playing file, and the display end information may include the custom name, the IP address, the MAC address and the device type (television, display, etc.) of the display end.
Specifically, the central control system encrypts and stores the received application playing information.
Optionally, in an embodiment, if the file type of the file applying for playing is a video, the application playing information may further include a video format and a video duration; if the file type of the application playing file is PPT, the application playing information can also comprise the PPT page number and the slide show duration.
It should be understood that, before receiving the application playing information sent by the display end, the user firstly logs in the central Control system, registers the High-Level Data Link Control (HDLC) service after successfully logging in, and monitors the access of the display end; if the central control system fails to log in, logging in again. Before sending the application playing information to the central control system, the display terminal needs to log in and establish connection with the central control system in advance, and if the login fails, the display terminal logs in again.
The HDLC service is a bit-oriented data link layer protocol for transmitting data over a synchronous network, and is used to monitor all the exhibition ends searched, thereby controlling all the exhibition ends.
After a user logs in the central control system, the central control system automatically starts to search the display terminal equipment in the current local area network, and after the search is successful, the display terminal equipment is automatically connected through a private protocol.
When the display end is connected with the central control system, the central control system can automatically update the connected display end equipment list and store the current list, and when new display end equipment is connected or disconnected, the central control system can synchronously update the list information in real time.
And S120, identifying the authority of the display terminal according to the information of the display terminal to obtain an authority identification result.
Specifically, the central control system can identify or judge the authority of the display terminal according to the user-defined name, the IP address and the MAC address information of the display terminal in the information of the display terminal.
After the central control system identifies the display end authority, an authority identification result can be returned through the security protocol.
And S130, acquiring the playing resource corresponding to the display end from the cloud server according to the authority identification result and the application playing information.
Specifically, the central control system may obtain the play resource from the cloud server according to the file name, the file type, and the play address of the play application file in the play application information.
The central control system firstly matches a corresponding playing file for the display end according to the obtained display end authority and the application playing information, and then obtains a corresponding playing resource from the cloud server.
The central control system can match corresponding playing files for the display end according to the equipment type of the display end, and can also match corresponding playing carriers (display ends) for the playing files according to the file types of the playing files.
Optionally, in one embodiment, the central control system is visible from the play resources acquired by the cloud server.
Optionally, in an embodiment, when the central control system obtains the corresponding play resource from the cloud server, a high-performance security cache mechanism may be started to perform asynchronous video downloading, and after the downloading is successful, the technical protection such as encoding, encrypting, compressing, scattering and the like may be performed on the file.
And S140, when the received operation information input by the user is the control playing information, controlling the display end to play the playing resource according to the control playing information.
The received operation information input by the user may be input by the user through a keyboard.
Controlling the playback information may include: whether to play circularly, whether to terminate the current play and play time interval.
Specifically, the central control system controls the display end to play videos, pictures, PPT and the like according to the control playing information, and the display end can automatically exit after playing is finished.
Optionally, in an embodiment, when the received operation information input by the user is control light information, light information such as light brightness and light mode of the display end is controlled according to the control light information.
Optionally, in an embodiment, before controlling the display end to play the playback resource according to the control playback information, the method includes: and adjusting the playing sequence of the playing resources and/or modifying the playing resources according to the control playing information.
Optionally, in an embodiment, when a new playing resource needs to be played by skipping a current playing resource, the user may input new operation information, where the operation information includes a command to skip the current playing resource, and the central control system immediately exits the currently playing resource according to the operation information and starts to play the new playing resource; if the playing resources needing to be played are required to be circularly multicast, the playing resources needing to be played can be added into the playing set, the operation information comprises a playing resource circulating multicast instruction, and the central control system can circularly play the playing resources according to the sequence of the added playing files.
It should be understood that when the playing resources are modified according to the control playing information of the user, the central control system, the display end and the cloud server update the modification of the playing resources synchronously in real time.
It should be understood that the central control system stores the received operation information in an encrypted manner.
Alternatively, in one embodiment, the received operation information input by the user may be input by the user through voice, and the central control system identifies the operation information input by the user through a Long Short-Term Memory (LSTM) network and a time-series Classification (CTC) neural network. Fig. 3 is a schematic structural diagram of a speech information recognition module according to an embodiment of the present invention, and as shown in fig. 3, recognizing speech information according to the LSTM + CTC neural network mainly includes the following steps:
step 1, extracting acoustic features from a sound waveform of voice information;
step 2, converting the acoustic features into phonemes of pronunciation according to the acoustic model;
and 3, converting the phonemes into recognition results (texts) according to decoding technologies such as language models.
Fig. 4 is a schematic flow chart of a method for recognizing speech information according to an embodiment of the present invention, as shown in fig. 4:
wherein the language model comprises an LSTM + CTC neural network.
The step 1 comprises the following steps: extracting Mel-Frequency Cepstral Coefficients (MFCC) of an original waveform sound file (WAV) by adopting python _ speed _ features, and calculating a returned characteristic value with 13-dimensional length from data of each frame;
the step 2 comprises the following steps: converting the 13-dimensional length sequence information corresponding to the characteristic values into phoneme classifications, wherein the phoneme classifications correspond to 24+8 × 4 phonemes and respectively correspond to 24 consonant letters and 8 vowel letters in the Chinese pinyin, and each vowel letter has four tones;
the step 3 comprises the following steps: the defined phonemes are passed into a bi-directional LSTM, which outputs an array of 26 data.
Fig. 5 is a schematic diagram of a long-term and short-term memory network according to an embodiment of the present invention, and as shown in fig. 5, the network architecture of the LSTM includes: the convolution kernel mainly extracts the feature values of related phonemes, transmits the feature value information into the pooling layer, and finally outputs an array containing 26 data.
After the calculation of the LSTM, the array containing 26 data passes through a softmax layer to obtain the posterior probability of the phoneme.
The posterior probability of a phoneme can be expressed by formula (1):
Figure BDA0002343131250000071
where t is a certain time, n is the number of phoneme types, k represents the number of phonemes, and the posterior probabilities of all phonemes add up to 1 on one frame of data.
The a posteriori probability of the array containing 26 data and the calculated phonemes is transmitted to CTC (loss function) to obtain the final recognition result.
The training set of CTCs can be represented by equation (2):
s={(x1,z1),(x2,z2),…(xn,zn)} (2)
wherein s is an array containing data, n is the number of training samples, x is an input sample, z is a corresponding label of real output, the input of one sample is a sequence, the output label is also a sequence, and the length of the input sequence is greater than that of the output sequence.
Wherein, for input sample data X with length of T frame, X ═ X (X)1,X2,X3,...,XT),XiFor the result obtained by performing MFCC calculation on the ith frame data in the input sample data X, the data of each frame is a vector with dimension m, i.e. each Xi∈RmR is any real number; obtaining an output sample Z of the input sample data X through a training set of CTC (Z ═ Z)1,Z2,Z3,...,ZT)。
As a specific embodiment, the input sample is a piece of voice data of 'hello' pronunciation, and the acoustic feature X is obtained after MFCC calculation1Obtaining corresponding phoneme information Z according to the acoustic model1,Z1=[n,i,h,a,o](temporarily taking each Pinyin letter as a phoneme), and using the phoneme information Z according to the training set of CTC1The recognition result (text information is "hello") is converted, thereby achieving successful speech recognition.
The processing method of the playing resources of the embodiment of the invention receives the playing application information sent by the display terminal, wherein the playing application information comprises the display terminal information; identifying the authority of the display terminal according to the information of the display terminal to obtain an authority identification result; acquiring a playing resource corresponding to the display end from the cloud server according to the permission identification result and the application playing information; when the received operation information input by the user is the control playing information, the display end is controlled to play the playing resources according to the control playing information, the playing resources are authenticated and authority is updated in real time under the unattended condition, the playing files of the display end can be flexibly controlled according to the operation information, the display efficiency is improved, and the labor cost is reduced.
Fig. 6 is a schematic structural diagram of a processing apparatus for playing back a resource according to an embodiment of the present invention, and as shown in fig. 6, the processing apparatus 200 for playing back a resource in a group may include: the system comprises a communication module 210, a permission identification module 220, an acquisition module 230 and a processing module 240.
The communication module 210 is configured to receive application playing information sent by a display terminal, where the application playing information includes display terminal information; the authority identification module 220 is used for identifying the authority of the display terminal according to the information of the display terminal to obtain an authority identification result; the obtaining module 230 is configured to obtain, according to the permission identification result and the application playing information, a playing resource corresponding to the display end from the cloud server; and the processing module 240 is configured to control the display end to play the playing resource according to the control playing information when the operation information input by the user is the control playing information.
In some embodiments, the communication module 210 includes: at least one of a Bluetooth communication module, an infrared communication module, a WIFI communication module and a Zigbee communication module.
In some embodiments, the display device further comprises a light module for controlling the light of the display end according to the control light information when the received operation information input by the user is the control light information.
In some embodiments, the processing module 240 is further configured to recognize the operation information input by the user through voice according to the LSTM + CTC neural network.
In some embodiments, before controlling the display end to play the play resource according to the control play information: the processing module 240 is further configured to adjust the playing sequence of the playing resources and/or modify the playing resources according to the playing control information.
The processing device for playing resources of the embodiment of the invention receives the application playing information sent by the display terminal, wherein the application playing information comprises the display terminal information; identifying the authority of the display terminal according to the information of the display terminal to obtain an authority identification result; acquiring a playing resource corresponding to the display end from the cloud server according to the permission identification result and the application playing information; when the received operation information input by the user is the control playing information, the display end is controlled to play the playing resources according to the control playing information, the display end can be flexibly controlled, the playing resources are authenticated and authority is updated in real time under the unattended condition, and the use experience of the user is improved.
It can be understood that the processing apparatus 200 for playing back resources according to the embodiment of the present invention may correspond to the execution main body of the processing method for playing back resources in fig. 1 according to the embodiment of the present invention, and specific details of the operation and/or the function of each module/unit of the processing apparatus 200 for playing back resources may refer to the description of the corresponding part in the processing method for playing back resources according to the embodiment of the present invention in fig. 1 and fig. 2, and are not described herein again for brevity.
Fig. 7 is a schematic diagram of a hardware structure of a processing device for playing back resources according to an embodiment of the present invention.
As shown in fig. 7, the processing device 300 for playing back resources in the present embodiment includes an input device 301, an input interface 302, a central processing unit 303, a memory 304, an output interface 305, and an output device 306. The input interface 302, the central processing unit 303, the memory 304, and the output interface 305 are connected to each other through a bus 310, and the input device 301 and the output device 306 are connected to the bus 310 through the input interface 302 and the output interface 305, respectively, and further connected to other components of the information acquisition device 300.
Specifically, the input device 301 receives input information from the outside and transmits the input information to the central processor 303 through the input interface 302; central processor 303 processes the input information based on computer-executable instructions stored in memory 304 to generate output information, stores the output information temporarily or permanently in memory 304, and then transmits the output information to output device 306 through output interface 305; the output device 306 outputs the output information to the outside of the information acquisition device 300 for use by the user.
In one embodiment, the processing device 300 for playing back assets shown in FIG. 7 comprises: a memory 304 for storing programs; a processor 303 for executing the program stored in the memory to perform the method of the embodiments shown in fig. 1-5 provided by the embodiments of the present invention.
An embodiment of the present invention further provides a computer-readable storage medium, where the computer-readable storage medium has computer program instructions stored thereon; which when executed by a processor implement the method of the embodiments of figures 1-5 provided by embodiments of the present invention.
It is to be understood that the invention is not limited to the specific arrangements and instrumentality described above and shown in the drawings. A detailed description of known methods is omitted herein for the sake of brevity. In the above embodiments, several specific steps are described and shown as examples. However, the method processes of the present invention are not limited to the specific steps described and illustrated, and those skilled in the art can make various changes, modifications and additions or change the order between the steps after comprehending the spirit of the present invention.
The functional blocks shown in the above-described structural block diagrams may be implemented as hardware, software, firmware, or a combination thereof. When implemented in hardware, it may be, for example, an electronic Circuit, an Application Specific Integrated Circuit (ASIC), suitable firmware, plug-in, function card, or the like. When implemented in software, the elements of the invention are the programs or code segments used to perform the required tasks. The program or code segments may be stored in a machine-readable medium or transmitted by a data signal carried in a carrier wave over a transmission medium or a communication link. A "machine-readable medium" may include any medium that can store or transfer information. Examples of machine-readable media include electronic circuits, semiconductor Memory devices, Read-Only memories (ROMs), flash memories, erasable ROMs (eroms), floppy disks, CD-ROMs, optical disks, hard disks, fiber optic media, Radio Frequency (RF) links, and so forth. The code segments may be downloaded via computer networks such as the internet, intranet, etc.
It should also be noted that the exemplary embodiments mentioned in this patent describe some methods or systems based on a series of steps or devices. However, the present invention is not limited to the order of the above-described steps, that is, the steps may be performed in the order mentioned in the embodiments, may be performed in an order different from the order in the embodiments, or may be performed simultaneously.
As described above, only the specific embodiments of the present invention are provided, and it can be clearly understood by those skilled in the art that, for convenience and brevity of description, the specific working processes of the system, the module and the unit described above may refer to the corresponding processes in the foregoing method embodiments, and are not described herein again. It should be understood that the scope of the present invention is not limited thereto, and any person skilled in the art can easily conceive various equivalent modifications or substitutions within the technical scope of the present invention, and these modifications or substitutions should be covered within the scope of the present invention.

Claims (13)

1. A method for processing a playback resource, the method comprising:
receiving application playing information sent by a display end, wherein the application playing information comprises display end information;
identifying the authority of the display terminal according to the display terminal information to obtain an authority identification result;
acquiring a playing resource corresponding to the display terminal from a cloud server according to the permission identification result and the application playing information;
and when the received operation information input by the user is control playing information, controlling the display end to play the playing resource according to the control playing information.
2. The method of claim 1, further comprising:
and when the received operation information input by the user is control light information, controlling the light of the display end according to the control light information.
3. The method according to claim 1 or 2, characterized in that the method further comprises:
and encrypting and storing the application playing information, the operation information and the playing resources.
4. The method according to claim 1, wherein the received operation information input by the user is input by the user through a keyboard and/or voice.
5. The method of claim 4, further comprising:
and identifying the operation information input by the user through voice according to the LSTM + CTC neural network.
6. The method according to claim 1, wherein before said controlling said display end to play said playback resources according to said control playback information, said method comprises:
and adjusting the playing sequence of the playing resources and/or modifying the playing resources according to the control playing information.
7. A processing apparatus for playing back a resource, the apparatus comprising:
the communication module is used for receiving application playing information sent by a display end, wherein the application playing information comprises display end information;
the authority identification module is used for identifying the authority of the display terminal according to the display terminal information to obtain an authority identification result;
the acquisition module is used for acquiring the playing resources corresponding to the display end from a cloud server according to the authority identification result and the application playing information;
and the processing module is used for controlling the display end to play the playing resource according to the control playing information when the received operation information input by the user is the control playing information.
8. The apparatus for processing resources as claimed in claim 7, wherein the communication module comprises: at least one of a Bluetooth communication module, an infrared communication module, a WIFI communication module and a Zigbee communication module.
9. The apparatus for processing resources for playing back as claimed in claim 7, wherein said apparatus further comprises:
and the light module is used for controlling the light of the display end according to the control light information when the received operation information input by the user is the control light information.
10. The apparatus for processing playback resources of claim 7, wherein:
the processing module is also used for identifying the operation information input by the user through voice according to the LSTM + CTC neural network.
11. The apparatus for processing playback resources of claim 7, wherein before the controlling the presentation end to play back the playback resources according to the control playback information: the processing module is further configured to adjust a playing sequence of the playing resources and/or modify the playing resources according to the playing control information.
12. A processing device for playing back a resource, the device comprising: a processor and a memory storing computer program instructions;
the processor, when executing the computer instructions, implements the processing method of the playback resource according to any one of claims 1 to 6.
13. A computer-readable storage medium, having computer program instructions stored thereon, which, when executed by a processor, implement the processing method of playback resources according to any one of claims 1-6.
CN201911384286.7A 2019-12-28 2019-12-28 Processing method, device and equipment for playing resources and storage medium Pending CN113055715A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201911384286.7A CN113055715A (en) 2019-12-28 2019-12-28 Processing method, device and equipment for playing resources and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201911384286.7A CN113055715A (en) 2019-12-28 2019-12-28 Processing method, device and equipment for playing resources and storage medium

Publications (1)

Publication Number Publication Date
CN113055715A true CN113055715A (en) 2021-06-29

Family

ID=76507463

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201911384286.7A Pending CN113055715A (en) 2019-12-28 2019-12-28 Processing method, device and equipment for playing resources and storage medium

Country Status (1)

Country Link
CN (1) CN113055715A (en)

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
DE202015001494U1 (en) * 2015-02-26 2015-04-07 Wolfgang Borkner Remote control unit
US20160241891A1 (en) * 2013-09-26 2016-08-18 Masaki Ninomiya Distribution management apparatus, distribution method, and program
CN105911972A (en) * 2016-06-24 2016-08-31 苏州和氏设计营造股份有限公司 Intelligent control system for exhibition hall
CN106328122A (en) * 2016-08-19 2017-01-11 深圳市唯特视科技有限公司 Voice identification method using long-short term memory model recurrent neural network
CN109782706A (en) * 2018-12-18 2019-05-21 清华大学 Exhibition room control system and method, Cloud Server and terminal control equipment
CN110602547A (en) * 2019-10-28 2019-12-20 广州奇川展览有限公司 Exhibition hall multi-user sharing system and method

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20160241891A1 (en) * 2013-09-26 2016-08-18 Masaki Ninomiya Distribution management apparatus, distribution method, and program
DE202015001494U1 (en) * 2015-02-26 2015-04-07 Wolfgang Borkner Remote control unit
CN105911972A (en) * 2016-06-24 2016-08-31 苏州和氏设计营造股份有限公司 Intelligent control system for exhibition hall
CN106328122A (en) * 2016-08-19 2017-01-11 深圳市唯特视科技有限公司 Voice identification method using long-short term memory model recurrent neural network
CN109782706A (en) * 2018-12-18 2019-05-21 清华大学 Exhibition room control system and method, Cloud Server and terminal control equipment
CN110602547A (en) * 2019-10-28 2019-12-20 广州奇川展览有限公司 Exhibition hall multi-user sharing system and method

Similar Documents

Publication Publication Date Title
JP6613347B2 (en) Method and apparatus for pushing information
US10824874B2 (en) Method and apparatus for processing video
CN107492379B (en) Voiceprint creating and registering method and device
CN105654950B (en) Adaptive voice feedback method and device
CN107464554B (en) Method and device for generating speech synthesis model
CN107481720B (en) Explicit voiceprint recognition method and device
CN108022586A (en) Method and apparatus for controlling the page
CN107657017A (en) Method and apparatus for providing voice service
JP6785904B2 (en) Information push method and equipment
US10824664B2 (en) Method and apparatus for providing text push information responsive to a voice query request
CN107748500A (en) Method and apparatus for controlling smart machine
CN111261151B (en) Voice processing method and device, electronic equipment and storage medium
CN108986790A (en) The method and apparatus of voice recognition of contact
CN109271533A (en) A kind of multimedia document retrieval method
CN108877782A (en) Audio recognition method and device
CN108900612A (en) Method and apparatus for pushed information
CN112530408A (en) Method, apparatus, electronic device, and medium for recognizing speech
CN109660865A (en) Make method and device, medium and the electronic equipment of video tab automatically for video
CN108924218A (en) Method and apparatus for pushed information
CN111027291B (en) Method and device for adding mark symbols in text and method and device for training model, and electronic equipment
CN107731229A (en) Method and apparatus for identifying voice
CN111243604B (en) Training method for speaker recognition neural network model supporting multiple awakening words, speaker recognition method and system
CN113220828B (en) Method, device, computer equipment and storage medium for processing intention recognition model
CN112163084B (en) Problem feedback method, device, medium and electronic equipment
US11036996B2 (en) Method and apparatus for determining (raw) video materials for news

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20210629