CN116800725A - Data processing method and device - Google Patents

Data processing method and device Download PDF

Info

Publication number
CN116800725A
CN116800725A CN202310769269.5A CN202310769269A CN116800725A CN 116800725 A CN116800725 A CN 116800725A CN 202310769269 A CN202310769269 A CN 202310769269A CN 116800725 A CN116800725 A CN 116800725A
Authority
CN
China
Prior art keywords
data
data packet
target
text information
information
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202310769269.5A
Other languages
Chinese (zh)
Inventor
谢玉君
彭方振
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Lenovo Beijing Ltd
Original Assignee
Lenovo Beijing Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Lenovo Beijing Ltd filed Critical Lenovo Beijing Ltd
Priority to CN202310769269.5A priority Critical patent/CN116800725A/en
Publication of CN116800725A publication Critical patent/CN116800725A/en
Pending legal-status Critical Current

Links

Landscapes

  • Data Exchanges In Wide-Area Networks (AREA)

Abstract

The application provides a data processing method and a data processing device, wherein in the method, a transmitting end analyzes at least one data packet of target data, obtains text information associated with the data packet, and transmits the data packet of the target data and the text information associated with the data packet to a receiving end. The receiving terminal receives a data packet of target data sent by the sending terminal and text information associated with the data packet; under the condition that the receiving end detects that the target data packet is lost, acquiring target text information associated with the lost data packet; and outputting the data fragments corresponding to the lost data packets in the target data based on the target text information.

Description

Data processing method and device
Technical Field
The present application relates to the field of data processing technologies, and in particular, to a data processing method and apparatus.
Background
In some real-time interaction processes such as online live broadcast, conference and conversation, transmission of data such as audio, video and text is required. These data are transmitted from the sender to the receiver over the network and recombined at the receiver to produce continuous video and audio.
When the data arrives at the receiving end, the data packet is lost due to network delay and the like, so that the situation of blocking or incomplete information receiving and the like in the online interaction process is caused, and interaction of the two parties is affected.
Disclosure of Invention
In view of the above, the present application provides a data processing method and apparatus.
A data processing method, the method being applied to a receiving end, the method comprising:
receiving a data packet of target data sent by a sending end and text information associated with the data packet;
under the condition that the target data packet loss is detected, acquiring target text information associated with the lost data packet;
and outputting the data segment corresponding to the lost data packet in the target data based on the target text information.
The method further comprises the steps of:
obtaining receiving information corresponding to the target data;
under the condition that the receiving information does not meet the preset receiving condition, determining that the target data packet is lost;
wherein the receiving condition is at least one of: the size of the data packet of the target data reaches a preset threshold, the time stamps corresponding to two adjacent data packets in the target data are continuous time stamps, and the data content of the target data is identifiable content.
In the method, the target data is voice data; the target text information comprises voice text content and voiceprint description information associated with the lost data packet;
the outputting, based on the target text information, a data segment corresponding to a lost data packet in the target data includes:
and outputting a first voice fragment corresponding to the lost data packet in the target data generated based on the voice text content and the voiceprint description information.
In the method, the target text information further comprises environment description information;
the outputting, based on the target text information, a data segment corresponding to a lost data packet in the target data includes:
outputting the first speech segment generated based on the speech text content, the voiceprint description information, and the environment description information.
The method further comprises the steps of:
analyzing a second voice segment corresponding to a data packet without packet loss, and obtaining volume information corresponding to the second voice segment;
and adjusting the volume of the first voice fragment based on the volume information.
The method, wherein the environment description information comprises at least one of the following:
environmental audio description information, environmental scene tag information.
In the above method, the target data is video data, and the target text information includes video text content and picture parameter information associated with the lost data packet;
the outputting, based on the target text information, a data segment corresponding to a lost data packet in the target data includes:
and outputting video fragments corresponding to the lost data packets in the target data generated based on the video text content and the picture parameter information.
A data processing method, the method being applied to a transmitting end, the method comprising:
analyzing at least one data packet of target data to obtain text information associated with the data packet, wherein the text information is used for describing a data part in the data packet associated with the text information;
and sending the data packet of the target data to a receiving end and text information associated with the data packet.
A data processing apparatus, the apparatus being applied to a receiving end, the apparatus comprising:
the receiving unit is used for receiving a data packet of target data sent by the sending end and text information associated with the data packet;
the acquisition unit is used for acquiring target text information associated with the lost data packet under the condition that the target data packet is detected to be lost;
and the output unit is used for outputting the data segment corresponding to the lost data packet in the target data based on the target text information.
A data processing apparatus, the apparatus being applied to a transmitting end, the apparatus comprising:
an analysis unit, configured to analyze at least one data packet of target data, and obtain text information associated with the data packet, where the text information is used to describe a data portion in the data packet associated with the text information;
and the sending unit is used for sending the data packet of the target data and text information associated with the data packet to a receiving end.
Drawings
In order to more clearly illustrate the embodiments of the present application or the technical solutions in the related art, the drawings that are required to be used in the embodiments or the related technical descriptions will be briefly described, and it is apparent that the drawings in the following description are only embodiments of the present application, and other drawings may be obtained according to the provided drawings without inventive effort for those skilled in the art.
Fig. 1 is a flowchart of a method for providing a data processing method applied to a receiving end according to an embodiment of the present application;
FIG. 2 is a diagram illustrating an exemplary method of data processing according to an embodiment of the present application;
FIG. 3 is a flowchart of another method of a data processing method applied to a transmitting end according to an embodiment of the present application;
FIG. 4 is a diagram illustrating another exemplary method of data processing according to an embodiment of the present application;
FIG. 5 is a block diagram of a data processing apparatus applied to a receiving end according to an embodiment of the present application;
fig. 6 is a device structure diagram of a data processing device applied to a transmitting end according to an embodiment of the present application.
Detailed Description
The following description of the embodiments of the present application will be made clearly and completely with reference to the accompanying drawings, in which it is apparent that the embodiments described are only some embodiments of the present application, but not all embodiments. All other embodiments, which can be made by those skilled in the art based on the embodiments of the application without making any inventive effort, are intended to be within the scope of the application.
In the present disclosure, relational terms such as first and second, and the like are used solely to distinguish one entity or action from another entity or action without necessarily requiring or implying any actual such relationship or order between such entities or actions, and the terms "comprise," "include," or any other variation thereof, are intended to cover a non-exclusive inclusion such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, an element defined by the phrase "comprising one … …" does not exclude the presence of other like elements in a process, method, article, or apparatus that comprises the element.
The application is operational with numerous general purpose or special purpose computing device environments or configurations. For example: personal computers, server computers, hand-held or portable devices, tablet devices, multiprocessor devices, distributed computing environments that include any of the above devices or devices, and the like.
The embodiment of the application provides a data processing method, which can be applied to a receiving end, wherein a flow chart of the method is shown in fig. 1, and specifically comprises the following steps:
s101: and receiving the data packet of the target data sent by the sending end and text information associated with the data packet.
In the application, the text information of the transmitting end comprises the data content and the data characteristic information of the data packet, and the data content and the data characteristic information are expressed in a text form. The transmitting end transmits text information while transmitting the data packet to the receiving end.
S102: and under the condition that the target data packet loss is detected, acquiring target text information associated with the lost data packet.
Specifically, the receiving end determines whether the packet loss exists according to the receiving information in the process of receiving the data packet. The method for determining whether the receiving end pairs the packet is as follows: obtaining receiving information corresponding to the target data; and determining that the target data packet is lost when the received information does not meet at least one of the following receiving conditions.
Wherein the receiving conditions include:
(1) The size of the data packet of the target data reaches a preset threshold value;
(2) The time stamps corresponding to two adjacent data packets in the target data are continuous time stamps;
(3) The data content of the target data is identifiable content.
In the present application, the manner of determining whether the received information satisfies the receiving condition based on the receiving condition includes at least one of the following:
the method comprises the steps of obtaining the size of a data packet of target data received by a receiving end contained in receiving information; if the size of the data packet of the target data reaches a preset threshold, the received information is represented to meet the receiving condition, otherwise, the received information is represented to not meet the receiving condition.
And obtaining a time stamp corresponding to the last data packet and a time stamp corresponding to the current data packet which are received by the receiving terminal and included in the receiving information, and if the time stamp corresponding to the last data packet and the time stamp corresponding to the current data packet are continuous time stamps, characterizing that the receiving information meets the receiving condition, otherwise, characterizing that the receiving information does not meet the receiving condition. Wherein, the time stamps corresponding to the two data packets are continuous time stamps, which means that the end time stamp of the last received data packet is consistent with the start time stamp of the current received data.
And obtaining description information for describing the data content of the received data packet contained in the receiving information, if the description information represents that the data content of the target data is identifiable content, representing that the receiving information meets the receiving condition, otherwise, representing that the receiving information does not meet the receiving condition. For example: the description information indicates that the data content of the data packet is a messy code or an error, the data packet content of the data packet is unidentifiable content, and if the description information indicates that the data content of the data packet is normal, the data packet content of the data packet is identifiable content.
It can be understood that if there is a missing data portion in the data packet received by the receiving end, or the entire data packet is missed, or the data content of the received data packet cannot be identified, the packet loss of the receiving end is represented.
In the application, after determining that the receiving end loses a packet, the target text information related to the lost packet can be determined according to the packet code of the lost packet, and the packet code of the lost packet can be obtained based on the packet code of the received packet.
S103: and outputting the data fragments corresponding to the lost data packets in the target data based on the target text information.
In the application, the process of outputting the data segment corresponding to the lost data packet is as follows: outputting the data segment corresponding to the lost data packet according to the ending output time of the data part in the first data packet. The first data packet is the last data packet of the lost data packet in the target data.
The data packet includes information such as a header, a start time stamp, a data portion, and an end time stamp. The receiving end analyzes the data packet after receiving the data packet, and outputs a data part in the data packet, wherein the data part is a data fragment corresponding to the data packet.
In the data processing method provided by the embodiment of the application, the transmitting end represents the data content and the data characteristics of the data packet in a text form, obtains the text information associated with the data packet, and simultaneously transmits the text information associated with the data packet when transmitting the data packet of the target data to the receiving end. Because the text information is data in a text form, the occupied bandwidth is smaller in the transmission process, the transmission efficiency is higher, and the text information cannot be failed to be transmitted or delayed to be transmitted due to network delay in the process of transmitting the data packet and the text information to the receiving end by the transmitting end. The bandwidth occupied by the receiving end is larger in the process of receiving the data packet of the target data, and the packet can be lost under the condition of network delay. If the receiving end loses a packet, the lost data packet can be recovered through target text information associated with the lost data packet, so that the blocking and the information loss in the online interaction process caused by network delay are avoided, and the interaction experience is improved.
Referring to fig. 2, in the data processing method according to the embodiment of the present application, a receiving end detects packet loss abnormality during receiving a data packet sent by a sending end, and simultaneously receives text information sent by the sending end, and stores the text information in a local buffer. When the packet loss is detected, the text information of the lost data packet is obtained from the local cache, and if the packet loss is detected before the text information is stored in the local cache, the text information sent by the sending end is directly obtained for data recovery and output. If the data part in the data packet obtained by the receiving end is unidentifiable data, after the data recovery, the recovered data is covered with the originally received data.
Based on the method provided in the above embodiment, the target data in the present application may be voice data. In the process of sending the data packet of the voice data, the sending end can recover the lost data packet according to the text information associated with the data packet.
Specifically, if the target data is voice data, the target text information includes voice text content associated with the lost data packet and voiceprint description information. The specific process of outputting the data segment corresponding to the lost data packet in the target data by the receiving terminal based on the target text information is as follows: and outputting a first voice fragment corresponding to the lost data packet in the target data generated based on the voice text content and the voiceprint description information.
It can be understood that the receiving end synthesizes the first voice segment corresponding to the lost data packet according to the voice text content and the voiceprint description information, and outputs the first voice segment.
In the application, the sending end synchronizes the voice text content and the voiceprint description information to the receiving end in real time in a text form in the process of sending the data packet of the voice data, so as to recover the voice data in time when the receiving end loses the packet.
Alternatively, the receiving end may synthesize the first speech segment using a text-based multi-feature fusion TTS (voice synthesis) model. The multi-feature fusion TTS model comprises an ASR (speech recognition) model, a VPR (voiceprint recognition) model and a feature synthesis model, wherein the ASR model and the VPR model are respectively used for processing the speech text content and voiceprint description information to respectively obtain speech text features and voiceprint features, and the feature synthesis model synthesizes speech fragments according to the speech text features and the voiceprint features.
Further, if the target data is voice data, the target text information may include environment description information in addition to voice text content and voiceprint description information. The environment description information is used for representing the surrounding environment of the transmitting end. Therefore, the specific process of outputting the data segment corresponding to the lost data packet in the target data by the receiving terminal based on the target text information may also be: a first speech segment generated based on the speech text content, the voiceprint description information, and the environment description information is output.
It can be understood that the receiving end synthesizes the first voice segment corresponding to the lost data packet according to the voice text content, the voiceprint description information and the environmental characteristics in the environmental description information, and outputs the first voice segment.
The environment description information includes at least one of environment audio description information and environment scene tag information of voice data. The environment audio description information is used for describing the real environment where the sending end is located, and the receiving end synthesizes the audio of the real environment where the sending end is located in the first voice segment according to the environment audio description information. The environment scene tag information is used for indicating an environment type of an environment where the sending end is currently located, the receiving end can acquire environment audio corresponding to the environment type from a local or cloud end according to the environment type in the process of synthesizing the first voice segment, and superimpose the environment audio acquired from the local or cloud end into the voice segment generated according to the voice text content and the voiceprint description information in the process of synthesizing the first voice segment so as to acquire the first voice frequency band. The environment audio obtained from the local or cloud end is different from the audio of the real environment where the transmitting end is located, for example, the environment type of the environment where the knowledge transmitting end is currently located in the environment scene tag information is street, and the receiving end directly finds the audio of a street pedestrian from the local or cloud end to be added.
Alternatively, the receiving end may synthesize the first speech segment using a text-based multi-feature fusion TTS (voice synthesis) model. The multi-feature fusion TTS model comprises an ASR (speech recognition) model, a VPR (voiceprint recognition) model, a BPR (Bayesian) model and a feature synthesis model, wherein the ASR model, the VPR model and the BPR model are respectively used for processing speech text content, voiceprint description information and environment description information to respectively obtain speech text features, voiceprint features and environment sound features, and the feature synthesis model synthesizes speech fragments according to the speech text features, the voiceprint features and the environment sound features.
In the application, the sending end synchronizes the voice text content, the voiceprint description information and the environment description information to the receiving end in real time in a text form in the process of sending the data packet of the voice data, so as to recover the voice data in time when the receiving end loses the packet.
In the embodiment of the application, in the process of synthesizing the voice fragments output by the receiving end, corresponding audio is added according to the environment of the receiving end, so that the voice fragments output by the receiving end are closer to the real voice fragments before packet loss, the difference of voice output before and after voice is found in the process of receiving voice by a user is avoided, and the user experience is improved.
Based on the method provided in the foregoing embodiment, if the target data is audio data, after the first speech segment is synthesized, the volume of the first speech segment may be adjusted, where a specific volume adjustment manner is: analyzing a second voice segment corresponding to the data packet without packet loss to obtain volume information corresponding to the second voice segment; and adjusting the volume of the first voice fragment based on the volume information.
It can be understood that, according to the second voice segment pair volume information, the volume of the output volume of the second voice segment is determined, and the volume of the first voice segment is adjusted according to the volume, so as to ensure that the volume of the receiving end can be consistent with the volume of the front and rear voice segments when the first voice segment is output, avoid the difference of the front and rear voice output found by the user in the process of receiving the voice, and further improve the user experience.
Based on the method provided in the above embodiment, the target data in the present application may be video data. In the process of sending the data packet of the video data, the sending end can recover the lost data packet according to the text information associated with the data packet.
Specifically, if the target data is video data, the target text information includes video text content associated with the lost data packet and picture parameter information. The specific process of outputting the data segment corresponding to the lost data packet in the target data by the receiving terminal based on the target text information is as follows: and outputting video fragments corresponding to the lost data packets in the target data generated based on the video text content and the picture parameter information.
The picture parameter information at least comprises picture resolution and definition.
It can be understood that the receiving end synthesizes the video clip corresponding to the lost data packet according to the video text content and the picture parameter information, and outputs the video clip.
Alternatively, the receiving end may synthesize the video clip using a text-based multi-feature fusion video synthesis model.
In the application, the transmitting end synchronizes the text content and the picture parameter information of the video to the receiving end in real time in the text form in the process of transmitting the data packet of the video data, so as to recover the video data in time when the receiving end loses the packet.
Corresponding to the method of fig. 1, the embodiment of the present application further provides a data processing method, where the method may be applied to a transmitting end, and a method flowchart of the method is shown in fig. 3, and specifically includes:
s301: and analyzing at least one data packet of the target data to obtain text information associated with the data packet.
Wherein the text information is used to describe the data portion of the data packet associated therewith.
In the present application, the target data may be voice data. If the target data is voice data, the text information comprises voice text content and voiceprint description information of a data packet of the voice data.
Specifically, if the target data is voice data, the sender analyzes a data packet of the target data, and the specific manner of obtaining text information associated with the data packet is as follows: and identifying a voice fragment corresponding to the data part in the data packet, and obtaining voice text content and voiceprint description information according to the voice fragment.
The sending end can recognize the voice text content and the voiceprint description information of the data packet through the audio recognition model. The audio recognition model comprises a real-time voice recognition model and a voiceprint recognition model which are respectively used for recognizing voice text content and voiceprint description information of the voice fragments.
Optionally, the text information may also include environment description information. The environment description information may be used to describe the current environment of the sender. The transmitting end obtains environment description information based on the environment data of the environment to which the transmitting end belongs.
The environment description information includes at least one of environment audio description information and environment scene tag information.
The manner of obtaining the environmental audio description information may be: the sending end recognizes the environmental audio in the voice segment corresponding to the data packet, and obtains the environmental audio description information according to the environmental audio. The transmitting end can recognize the environmental audio in the voice segment through the environmental recognition model in the audio recognition model and output the environmental audio description information corresponding to the environmental audio.
The manner of obtaining the environmental scene tag information may be: the transmitting end starts the camera shooting module to collect the environmental data of the transmitting end, determines the environmental type of the current environment of the transmitting end according to the environmental data, and determines the environmental scene label information according to the environmental type.
In the present application, the target data may be video data. If the target data is video data, the text information contains video text content and picture parameter information.
Specifically, if the target data is video data, the sender analyzes a data packet of the target data, and the specific manner of obtaining text information associated with the data packet is as follows: and identifying video fragments corresponding to the data packets in the video data, and obtaining text information associated with the data packets.
The sending end can identify video text content and picture parameter information of the data packet through the video identification model.
S302: and sending the data packet of the target data to the receiving end and text information associated with the data packet.
In the method provided by the embodiment of the application, the transmitting end analyzes the data packet in the target data, obtains the text information associated with the data packet, simultaneously transmits the data packet and the text information to the receiving end, and the receiving end carries out packet loss anomaly detection in real time, can carry out data recovery according to the text information when detecting packet loss, and outputs the data fragment of the lost data packet.
The text information of the transmitting end includes the data content and the data characteristic information of the data packet, and the data content and the data characteristic information are expressed in text form. The application transmits the data content and the data characteristic information in the text form while transmitting the data packet, and the output efficiency is high because the bandwidth occupied by the text information is smaller, and the missing data fragments can be recovered and output in time according to the text information when the receiving end loses the packet.
Referring to fig. 4, taking the target data as the voice data as an example, the process of the interaction between the sending end and the receiving end may specifically be: in the process of voice interaction, the transmitting end recognizes the data packet of voice data in real time to obtain voice text content, voiceprint description information and environment description information in text form. The transmitting end transmits the data packet and text information containing voice text content, voiceprint description information and environment description information to the receiving end. The receiving end detects whether the packet is lost in real time, and updates the text information to the local cache after receiving the text information. After detecting packet loss, the receiving end obtains text information of the lost data packet, synthesizes a voice fragment based on voice text content, voiceprint description information and environment description information in the text information by applying a text-based multi-feature fusion TTS model, covers the Bao Yin frequency band of the voice fragment, and outputs the voice fragment. Before receiving the voice data, the receiving end can train the text-based multi-feature fusion TTS model in advance, and specifically trains a VSR model, a VPR model, a BPR model and a feature synthesis model in the text-based multi-feature fusion TTS model so as to perform voice synthesis through the text-based multi-feature fusion TTS model.
Based on the method provided by the embodiment, the multi-feature fusion speech synthesis model based on the text, the voiceprint features and the environmental sound features can be obtained through training. In the voice process, voice data of a transmitting end are transmitted to an independent server to perform real-time voice recognition, voiceprint feature recognition and environment feature recognition. And synchronizing the identification contents of the three to a receiving end in real time in a text form. The receiving end detects the voice packet loss condition. When the voice packet loss occurs, the receiving end uses the trained model to perform real-time audio synthesis based on the voice recognition text of the sending end, the locally cached voiceprint features and the environment features. And using synthesized audio to cover the voice signal with packet loss at the receiving end. The application replaces the original voice transmission by text transmission, and has small occupied bandwidth and higher transmission efficiency. The method is independent of historical audio, greatly improves the fault tolerance rate of the packet loss duration, and is suitable for more complex network environments. The voice recognition accuracy is high, the environment characteristics are not lost, and the synthesized audio content is more true and reliable.
Based on the method provided by the embodiment, lost data packets are timely recovered through text information when packet loss occurs, and the problem that network delay causes blocking and information loss in the online interaction process is avoided.
The specific implementation process and derivative manner of the above embodiments are all within the protection scope of the present application.
Corresponding to the method shown in fig. 1, the embodiment of the present application further provides a data processing apparatus, which is used for implementing the method shown in fig. 1, where the data processing apparatus provided in the embodiment of the present application may be applied to a receiving end, and a schematic structural diagram of the data processing apparatus is shown in fig. 5, and specifically includes:
a receiving unit 501, configured to receive a data packet of target data sent by a sending end and text information associated with the data packet;
an obtaining unit 502, configured to obtain, when the target data packet is detected to be lost, target text information associated with the lost data packet;
and an output unit 503, configured to output, based on the target text information, a data segment corresponding to a lost data packet in the target data.
In the data processing device provided by the embodiment of the application, the transmitting end represents the data content and the data characteristics of the data packet in a text form, obtains the text information associated with the data packet, and simultaneously transmits the text information associated with the data packet when transmitting the data packet of the target data to the receiving end. Because the text information is data in a text form, the occupied bandwidth is smaller in the transmission process, the transmission efficiency is higher, and the text information cannot be failed to be transmitted or delayed to be transmitted due to network delay in the process of transmitting the data packet and the text information to the receiving end by the transmitting end. The bandwidth occupied by the receiving end is larger in the process of receiving the data packet of the target data, and the packet can be lost under the condition of network delay. If the receiving end loses a packet, the lost data packet can be recovered through target text information associated with the lost data packet, so that the blocking and the information loss in the online interaction process caused by network delay are avoided, and the interaction experience is improved.
The device provided by the embodiment of the application further comprises:
the packet loss detection unit is used for obtaining the receiving information corresponding to the target data; under the condition that the receiving information does not meet the preset receiving condition, determining that the target data packet is lost;
wherein the receiving condition is at least one of: the size of the data packet of the target data reaches a preset threshold, the time stamps corresponding to two adjacent data packets in the target data are continuous time stamps, and the data content of the target data is identifiable content.
In the device provided by the embodiment of the application, the target data is voice data; the target text information comprises voice text content and voiceprint description information associated with the lost data packet;
the output unit outputs, based on the target text information, a data segment corresponding to a lost data packet in the target data, where the data segment is specifically configured to: and outputting a first voice fragment corresponding to the lost data packet in the target data generated based on the voice text content and the voiceprint description information.
In the device provided by the embodiment of the application, the target text information also comprises environment description information;
the output unit outputs, based on the target text information, a data segment corresponding to a lost data packet in the target data, where the data segment is specifically configured to: outputting the first speech segment generated based on the speech text content, the voiceprint description information, and the environment description information.
The device provided by the embodiment of the application further comprises:
the volume adjusting unit is used for analyzing a second voice fragment corresponding to the data packet without packet loss and obtaining volume information corresponding to the second voice fragment; and adjusting the volume of the first voice fragment based on the volume information.
In the device provided by the embodiment of the application, the environment description information comprises at least one of the following: environmental audio description information, environmental scene tag information.
In the device provided by the embodiment of the application, the target data is video data, and the target text information comprises video text content and picture parameter information associated with the lost data packet;
the output unit outputs, based on the target text information, a data segment corresponding to a lost data packet in the target data, where the data segment is specifically configured to:
and outputting video fragments corresponding to the lost data packets in the target data generated based on the video text content and the picture parameter information.
The specific working process of each unit and subunit in the data processing device applied to the receiving end disclosed in the above embodiment of the present application can be referred to the corresponding content in the data processing method applied to the receiving end disclosed in the above embodiment of the present application, and will not be described herein.
Corresponding to the method shown in fig. 3, the embodiment of the present application further provides a data processing apparatus, which is used for implementing the method shown in fig. 3, where the data processing apparatus provided in the embodiment of the present application may be applied to a transmitting end, and a schematic structural diagram of the data processing apparatus is shown in fig. 6, and specifically includes:
an analysis unit 601, configured to analyze at least one data packet of target data, and obtain text information associated with the data packet, where the text information is used to describe a data portion in the data packet associated with the text information;
and a sending unit 602, configured to send, to a receiving end, a data packet of the target data and text information associated with the data packet.
In the device provided by the embodiment of the application, the transmitting end analyzes the data packet in the target data, obtains the text information related to the data packet, simultaneously transmits the data packet and the text information to the receiving end, and the receiving end detects the abnormal packet loss in real time, can recover the data according to the text information when detecting the packet loss, and outputs the data fragment of the lost data packet.
The specific working process of each unit and subunit in the data processing device applied to the transmitting end disclosed in the above embodiment of the present application can refer to the corresponding content in the data processing method applied to the transmitting end disclosed in the above embodiment of the present application, and will not be described herein.
In this specification, each embodiment is described in a progressive manner, and identical and similar parts of each embodiment are all referred to each other, and each embodiment mainly describes differences from other embodiments. In particular, for a system or system embodiment, since it is substantially similar to a method embodiment, the description is relatively simple, with reference to the description of the method embodiment being made in part. The systems and system embodiments described above are merely illustrative, wherein the elements illustrated as separate elements may or may not be physically separate, and the elements shown as elements may or may not be physical elements, may be located in one place, or may be distributed over a plurality of network elements. Some or all of the modules may be selected according to actual needs to achieve the purpose of the solution of this embodiment. Those of ordinary skill in the art will understand and implement the present application without undue burden.
Those of skill would further appreciate that the elements and algorithm steps of the examples described in connection with the embodiments disclosed herein may be implemented as electronic hardware, computer software, or combinations of both.
To clearly illustrate this interchangeability of hardware and software, various illustrative components and steps have been described above generally in terms of their functionality. Whether such functionality is implemented as hardware or software depends upon the particular application and design constraints imposed on the solution. Skilled artisans may implement the described functionality in varying ways for each particular application, but such implementation decisions should not be interpreted as causing a departure from the scope of the present application.
The previous description of the disclosed embodiments is provided to enable any person skilled in the art to make or use the present application. Various modifications to these embodiments will be readily apparent to those skilled in the art, and the generic principles defined herein may be applied to other embodiments without departing from the spirit or scope of the application. Thus, the present application is not intended to be limited to the embodiments shown herein but is to be accorded the widest scope consistent with the principles and novel features disclosed herein.

Claims (10)

1. A data processing method, the method being applied to a receiving end, the method comprising:
receiving a data packet of target data sent by a sending end and text information associated with the data packet;
under the condition that the target data packet loss is detected, acquiring target text information associated with the lost data packet;
and outputting the data segment corresponding to the lost data packet in the target data based on the target text information.
2. The method of claim 1, further comprising:
obtaining receiving information corresponding to the target data;
under the condition that the receiving information does not meet the preset receiving condition, determining that the target data packet is lost;
wherein the receiving condition is at least one of: the size of the data packet of the target data reaches a preset threshold value; the time stamps corresponding to two adjacent data packets in the target data are continuous time stamps; the data content of the target data is identifiable content.
3. The method of claim 1, the target data being voice data; the target text information comprises voice text content and voiceprint description information associated with the lost data packet;
the outputting, based on the target text information, a data segment corresponding to a lost data packet in the target data includes:
and outputting a first voice fragment corresponding to the lost data packet in the target data generated based on the voice text content and the voiceprint description information.
4. The method of claim 3, the target text information further comprising environment description information;
the outputting, based on the target text information, a data segment corresponding to a lost data packet in the target data includes:
outputting the first speech segment generated based on the speech text content, the voiceprint description information, and the environment description information.
5. The method of claim 4, further comprising:
analyzing a second voice segment corresponding to a data packet without packet loss, and obtaining volume information corresponding to the second voice segment;
and adjusting the volume of the first voice fragment based on the volume information.
6. The method of claim 4, the environment description information comprising at least one of:
environmental audio description information, environmental scene tag information.
7. The method of claim 1, the target data being video data, the target text information comprising video text content associated with the lost data packet and picture parameter information;
the outputting, based on the target text information, a data segment corresponding to a lost data packet in the target data includes:
and outputting video fragments corresponding to the lost data packets in the target data generated based on the video text content and the picture parameter information.
8. A data processing method, the method being applied to a transmitting end, the method comprising:
analyzing at least one data packet of target data to obtain text information associated with the data packet, wherein the text information is used for describing a data part in the data packet associated with the text information;
and sending the data packet of the target data to a receiving end and text information associated with the data packet.
9. A data processing apparatus, the apparatus being applied to a receiving end, the apparatus comprising:
the receiving unit is used for receiving a data packet of target data sent by the sending end and text information associated with the data packet;
the acquisition unit is used for acquiring target text information associated with the lost data packet under the condition that the target data packet is detected to be lost;
and the output unit is used for outputting the data segment corresponding to the lost data packet in the target data based on the target text information.
10. A data processing apparatus, the apparatus being applied to a transmitting end, the apparatus comprising:
an analysis unit, configured to analyze at least one data packet of target data, and obtain text information associated with the data packet, where the text information is used to describe a data portion in the data packet associated with the text information;
and the sending unit is used for sending the data packet of the target data and text information associated with the data packet to a receiving end.
CN202310769269.5A 2023-06-27 2023-06-27 Data processing method and device Pending CN116800725A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202310769269.5A CN116800725A (en) 2023-06-27 2023-06-27 Data processing method and device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202310769269.5A CN116800725A (en) 2023-06-27 2023-06-27 Data processing method and device

Publications (1)

Publication Number Publication Date
CN116800725A true CN116800725A (en) 2023-09-22

Family

ID=88041131

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202310769269.5A Pending CN116800725A (en) 2023-06-27 2023-06-27 Data processing method and device

Country Status (1)

Country Link
CN (1) CN116800725A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117789680A (en) * 2024-02-23 2024-03-29 青岛海尔科技有限公司 Method, device and storage medium for generating multimedia resources based on large model

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117789680A (en) * 2024-02-23 2024-03-29 青岛海尔科技有限公司 Method, device and storage medium for generating multimedia resources based on large model
CN117789680B (en) * 2024-02-23 2024-05-24 青岛海尔科技有限公司 Method, device and storage medium for generating multimedia resources based on large model

Similar Documents

Publication Publication Date Title
CN108305628B (en) Speech recognition method, speech recognition device, computer equipment and storage medium
CN107181675A (en) Service calling method and device
CN116800725A (en) Data processing method and device
US20080225845A1 (en) Methods and Apparatus for Receiving Data in a Packet Network
US8885485B2 (en) Generation of diagnostic data from interarrival times of VoIP packets
US20130091528A1 (en) Video reproduction system, receive terminal, home gateway device, and quality control system
WO2012142797A1 (en) Video monitoring system and method
CN103780610A (en) Network data recovery method based on protocol characteristics
CN107517410B (en) Method and device for evaluating video service quality
CN102932676A (en) Self-adaptive bandwidth transmitting and playing method based on audio and video frequency synchronization
US11553025B2 (en) System and method for interleaved media communication and conversion
CN110691204A (en) Audio and video processing method and device, electronic equipment and storage medium
CN109257610B (en) Media low-delay communication method and system for internet remote education
CN110740289B (en) System and method for acquiring alarm
CN109600341B (en) Instant messaging detection method, equipment and computer storage medium
CN109842821A (en) A kind of method and apparatus of video data transmission
CN111400047A (en) Method for detecting and identifying human face from monitoring video stream through cloud edge cooperation
CN108881119B (en) Method, device and system for video concentration
US9246631B2 (en) Communication devices that encode and transmit data, methods of controlling such communication devices, and computer-readable storage media storing instructions for controlling such communication devices
US9571425B2 (en) Method and apparatus for voice communication based on voice activity detection
CN102625074B (en) Method for transmitting intelligent analyzed data by aid of HD-SDI (high-definition serial digital interface) camera
CN111405230A (en) Conference information processing method and device, electronic equipment and storage medium
CN112639963A (en) Audio acquisition device, audio receiving device and audio processing method
CN108880925B (en) Method and system for monitoring flow
CN102932673B (en) The transmission synthetic method of a kind of vision signal and audio signal, system and device

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination