WO2021190372A1 - Video file processing method and device, and watermark extraction method and device - Google Patents

Video file processing method and device, and watermark extraction method and device Download PDF

Info

Publication number
WO2021190372A1
WO2021190372A1 PCT/CN2021/081259 CN2021081259W WO2021190372A1 WO 2021190372 A1 WO2021190372 A1 WO 2021190372A1 CN 2021081259 W CN2021081259 W CN 2021081259W WO 2021190372 A1 WO2021190372 A1 WO 2021190372A1
Authority
WO
WIPO (PCT)
Prior art keywords
video
watermark
watermark information
data
audio
Prior art date
Application number
PCT/CN2021/081259
Other languages
French (fr)
Chinese (zh)
Inventor
刘永亮
杨锐
Original Assignee
阿里巴巴集团控股有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 阿里巴巴集团控股有限公司 filed Critical 阿里巴巴集团控股有限公司
Publication of WO2021190372A1 publication Critical patent/WO2021190372A1/en

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/233Processing of audio elementary streams
    • H04N21/2335Processing of audio elementary streams involving reformatting operations of audio signals, e.g. by converting from one coding standard to another
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/233Processing of audio elementary streams
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/234Processing of video elementary streams, e.g. splicing of video streams, manipulating MPEG-4 scene graphs
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/234Processing of video elementary streams, e.g. splicing of video streams, manipulating MPEG-4 scene graphs
    • H04N21/23418Processing of video elementary streams, e.g. splicing of video streams, manipulating MPEG-4 scene graphs involving operations for analysing video streams, e.g. detecting features or characteristics
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/234Processing of video elementary streams, e.g. splicing of video streams, manipulating MPEG-4 scene graphs
    • H04N21/2343Processing of video elementary streams, e.g. splicing of video streams, manipulating MPEG-4 scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/235Processing of additional data, e.g. scrambling of additional data or processing content descriptors
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/235Processing of additional data, e.g. scrambling of additional data or processing content descriptors
    • H04N21/2355Processing of additional data, e.g. scrambling of additional data or processing content descriptors involving reformatting operations of additional data, e.g. HTML pages
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/435Processing of additional data, e.g. decrypting of additional data, reconstructing software from modules extracted from the transport stream
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/435Processing of additional data, e.g. decrypting of additional data, reconstructing software from modules extracted from the transport stream
    • H04N21/4355Processing of additional data, e.g. decrypting of additional data, reconstructing software from modules extracted from the transport stream involving reformatting operations of additional data, e.g. HTML pages on a television screen
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/439Processing of audio elementary streams
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/439Processing of audio elementary streams
    • H04N21/4398Processing of audio elementary streams involving reformatting operations of audio signals
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream, rendering scenes according to MPEG-4 scene graphs
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream, rendering scenes according to MPEG-4 scene graphs
    • H04N21/44008Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream, rendering scenes according to MPEG-4 scene graphs involving operations for analysing video streams, e.g. detecting features or characteristics in the video stream
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream, rendering scenes according to MPEG-4 scene graphs
    • H04N21/4402Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream, rendering scenes according to MPEG-4 scene graphs involving reformatting operations of video signals for household redistribution, storage or real-time display
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/83Generation or processing of protective or descriptive data associated with content; Content structuring
    • H04N21/835Generation of protective data, e.g. certificates
    • H04N21/8358Generation of protective data, e.g. certificates involving watermark
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/83Generation or processing of protective or descriptive data associated with content; Content structuring
    • H04N21/845Structuring of content, e.g. decomposing content into time segments
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/83Generation or processing of protective or descriptive data associated with content; Content structuring
    • H04N21/845Structuring of content, e.g. decomposing content into time segments
    • H04N21/8456Structuring of content, e.g. decomposing content into time segments by decomposing the content in the time domain, e.g. in time segments

Definitions

  • the present disclosure relates to the technical field of digital media processing, and in particular to a method and device for video file processing and watermark extraction.
  • the present disclosure provides a video file processing and corresponding watermark extraction method.
  • This method utilizes the temporal and spatial relevance of the video sequence and audio data contained in the video, and adds mutually related audio and video watermarks to the video file, thereby improving the watermark coverage of the video file.
  • the robustness of embedded watermarks against malicious tampering is improved.
  • a video file processing method including: acquiring video data and audio data of the video file; embedding first watermark information in the video data; Associating and embedding the second watermark information; and obtaining the watermark-embedded video file.
  • a method for extracting a video watermark including: obtaining the watermark-embedded video file according to the first aspect; and extracting video data and audio data from the watermark-embedded video file ; Extracting the first watermark information embedded in the video data; and extracting the second watermark information embedded in the audio data.
  • a method for extracting a streaming media watermark including: acquiring streaming media data embedded with a watermark, the streaming media data being generated from the watermark-embedded video file described in the first aspect; Extracting video data and audio data from the watermark-embedded streaming media data; extracting the first watermark information embedded in the video data; and extracting the second watermark information embedded in the audio data.
  • a streaming media data processing method including: acquiring video data and audio data of the streaming media data; embedding first watermark information into the video data; The second watermark information is associated and embedded in the data; and the watermark-embedded streaming media data is obtained.
  • a video file processing device including: a video parsing unit for acquiring video data and audio data in the video file; a video watermark embedding unit for sending the video to the video
  • the first watermark information is embedded in the data
  • the audio watermark embedding unit is used to associate and embed the second watermark information into the audio data
  • the video mixing unit is used to mix the video data embedded in the first watermark information and the second watermark information.
  • the audio data of the watermark information is used to obtain a watermark embedded video file.
  • a video watermark extraction device including: a video parsing unit for obtaining video data and audio data in a watermark-embedded video file according to the first aspect; video watermark The extraction unit is used to extract the first watermark information from the video data; and the audio watermark extraction unit is used to extract and embed the second watermark information from the audio data.
  • a computing device including: a processor; and a memory on which executable code is stored.
  • the processor is caused to execute the first To the method described in the fourth aspect.
  • a non-transitory machine-readable storage medium having executable code stored thereon.
  • the executable code When executed by a processor of an electronic device, the processor executes the above-mentioned One to the method described in the fourth aspect.
  • This application embeds audio watermark and video watermark in the video file at the same time.
  • the two do not interfere with each other and complement each other.
  • the audio and video watermark extraction information can be adaptively fused, thereby greatly improving the robustness of the video file watermark, especially Fight against malicious editing attacks on video content.
  • Fig. 1 shows a schematic flowchart of a video file processing method according to the present application
  • Figure 2 shows an example of embedding watermark information for video sequence and audio code stream respectively
  • Figure 3 shows an example of the joint watermark embedding process according to the present application
  • Figure 4 shows a schematic flowchart of a video watermark extraction method according to an embodiment of the present application
  • Figure 5 shows an example of a joint watermark extraction process according to the present application
  • Figure 6 shows a schematic diagram of a brief scheme of the watermark embedding and extraction operations of this application
  • Fig. 7 shows a schematic structural diagram of a computing device that can be used to implement the above video processing and watermark extraction method according to an embodiment of the present application.
  • this application proposes an audio-video joint watermarking scheme.
  • the different data content (video sequence and audio data) contained in the video has a strong correlation in time and space and restricts each other.
  • Most of the existing video watermarking algorithms only do embedding and extraction processing on the video code stream, without using the correlation between audio and video, or without considering the existence of audio data at all.
  • This application introduces the correlation between audio and video data into the video watermarking process to make the video watermark more accurate and robust, so that copyright protection and digital content authentication can achieve better results.
  • FIG. 1 shows a schematic flowchart of a video file processing method according to the present application.
  • video file refers to a file that usually includes both video and audio information.
  • video files include both image content and sound content.
  • step S110 the video data and audio data of the video file are acquired.
  • existing tools can be used to separate audio and video.
  • a video file parsing tool can be used to operate on the video file to be embedded with the watermark to extract video data (usually a video sequence) and audio data (usually an audio stream).
  • step S120 the first watermark information is embedded in the video data.
  • step S130 the second watermark information is associated and embedded into the audio data.
  • the steps of inserting the video watermark and the audio watermark can also be performed at the same time, or the audio watermark can be inserted first.
  • the above-mentioned correlation can be used to perform correlation verification of audio and video watermarks, thereby improving the anti-tampering ability of the watermarks.
  • the above-mentioned association may be a time association, that is, the time when the second watermark information is embedded in the audio data may be associated with the time when the first watermark information is embedded in the video data.
  • the above-mentioned association may also be a content association, that is, the content of the second watermark information embedded in the audio data may be associated with the content of the first watermark information embedded in the video data.
  • the time and content of the audio and video watermarking may have a predetermined association relationship, so as to facilitate subsequent mutual verification of the extracted audio and video watermark information.
  • the above-mentioned relevance in time and content may be the same as each other.
  • the length of watermark data often needs to involve continuous addition within a period of time, and there is a difference in embedding capacity between video sequences and audio streams of the same duration (that is, it may only take 7 seconds for a video sequence to complete the embedding of 64 bits, but may require The 10-second audio code stream can only complete the 64-bit embedding), so the same time may be the same initial time of embedding.
  • the start time of embedding the second watermark information into the audio data may be the same as the start time of embedding the first watermark information into the video data. Therefore, when the double watermark is extracted, it is convenient to align and verify each other of the audio and video watermarks.
  • the watermark content of audio and video may be related to each other based on a certain mapping relationship, rather than being completely the same.
  • the audio and video have the same watermark content.
  • the content of the second watermark information embedded in the audio data may be the same as the content of the first watermark information embedded in the video data.
  • embedding the first watermark information into the video data may include: embedding a plurality of first watermark information into the video data at a first predetermined time interval.
  • embedding the associated second watermark information into the audio data includes: embedding a plurality of second watermark information into the audio data at a second predetermined time interval.
  • the predetermined time interval may be a non-uniform time interval, for example, embedding one every 10 seconds in the first minute and embedding one every 15 seconds in the next minute, it is still preferable to perform the watermark insertion at a uniform predetermined time interval to improve the watermarking performance. Anti-tampering ability.
  • each of the plurality of first watermark information may include: first ranking data and first watermark data.
  • each second watermark information in the plurality of second watermark information may include: second ranking data and second watermark data.
  • the actually added watermark information may include the sort code and the watermark information itself, and it is repeated at a certain time interval.
  • Figure 2 shows an example of embedding watermark information for a video sequence and an audio code stream respectively.
  • the same watermark information can be repeatedly inserted in the video sequence and audio code stream aligned on the time axis at a time interval of 10 seconds.
  • the watermark information inserted in each time interval includes synchronization code and single-period watermark information.
  • the "synchronization code” can be a code used in audio or video to sort and count the added single-period watermark information.
  • the single-period watermark information added at the beginning of the first time interval (ie, the 0th second) is "000...0000", at the beginning of the second time interval (ie, the 10th second)
  • the added single-period watermark information can be increased by 1 to become “000->0001”
  • the single-period watermark information added at the beginning of the third time interval ie, the 20th second
  • the "synchronization" of the "synchronization code” may refer to the synchronization between the video sequence and the audio code stream. That is, in the subsequent watermark extraction stage, the video sequence and the audio code stream can be aligned in time with the help of the recovered synchronization code.
  • single-period watermark information can refer to an identification code used to uniquely represent the identity of the video or the identity of the video producer (or the identity of the video publisher), for example, the "48-bit complete cycle of one period" shown in Figure 2 "Watermark information" is thus distinguished from all watermark-related information inserted in the entire audio or video (including the ever-increasing synchronization code and the watermark identification code that has been repeated thereafter).
  • both the synchronization code and the single-period watermark information are binary data represented by 0 and 1, which is the same as the data system in the actual audio and video processing.
  • the watermarking interval is repeated. It should be no less than 10 seconds long.
  • the watermark insertion is continuously performed at a predetermined interval of 10 seconds in this example, it should be understood that in other embodiments, the watermark insertion may also be performed at a longer time interval, for example, once every 20 seconds. .
  • each embedded synchronization code and single-period watermark information can also have a different number of bits, for example, synchronization code 32 bits, watermark information 32 bits, etc., and the overall length of synchronization code and watermark information can also be less than 64 bits. Other lengths.
  • step S140 the watermark-embedded video file can be obtained.
  • a video sequence and audio data containing the same watermark information can be stream-mixed to obtain a video file containing a double watermark.
  • the video file with the double watermark embedded above can be released later, and when necessary, the watermark is extracted and restored based on a predetermined method.
  • different audio and video watermark embedding methods can be selected to insert watermark data for audio and video respectively, as long as the embedding has relevance.
  • embedding the first watermark information into the video data may include: adding the first watermark information to a non-significant area of a video frame in the video data.
  • video key frames can be extracted, and content analysis can be performed to select non-significant areas for adding watermark information.
  • embedding the second watermark information into the audio data includes: adding the second watermark information to the auditory insensitive area of the audio frame in the audio data.
  • the audio is divided into frames to select areas that are not auditorily sensitive to the frequency band in each frame of audio for adding watermark information.
  • embedding the first watermark information into the video data may include: embedding the first watermark information by adjusting the energy relationship between adjacent regions of the video frame transform domain in the video data.
  • embedding the first watermark information may include: selecting a series of specific video frames in a video sequence, and the video data is a video sequence; And adjust the energy relationship of the adjacent regions of the transform domain of the series of specific video frames, and embed the constituent bit information of the first watermark information one by one.
  • selecting a series of specific video frames in the video sequence may include: selecting video key frames in the video sequence. For example, one bit of the 64-bit information is embedded in each key frame, and finally the synchronization code and the single-period watermark information are embedded in the 64 consecutive key frames.
  • embedding the associated second watermark information into the audio data may include: adjusting the energy relationship of adjacent audio frames, and embedding the second watermark information.
  • adjusting the energy relationship of adjacent audio frames, and embedding the second watermark information may include: adjusting the energy relationship of adjacent frequency bands of the series of adjacent audio frames, and embedding the composition of the second watermark information one by one Bit information.
  • the energy relationship between two adjacent frequency bands can be adjusted, and one bit of the 64-bit information can be embedded.
  • the synchronization code and the single-period watermark information can be embedded between 65 consecutive adjacent frames.
  • FIG 3 shows an example of the joint watermark embedding process according to the present application.
  • the video file to be watermarked you can first use the video file analysis tool to extract the video sequence and audio data separately. Subsequently, as shown in Figure 3, the process is divided into two parallel branches.
  • the video key frames are extracted, and content analysis is performed to select non-salient areas.
  • synchronization codes and watermark bits can be embedded based on the energy relationship of adjacent blocks to obtain a video sequence with watermarked information.
  • DCT Discrete Cosine Transform
  • a selected non-salient area of a key frame is embedded with one bit, so that 64 consecutive key frames contain 64-bit synchronization code and single-period watermark information. It is convenient for the subsequent extraction process to locate the watermark information through key frame extraction, non-significant region selection and adjacent region energy relationship search.
  • DCT Discrete Cosine Transform
  • the adjustment of the energy relationship, the embedding of watermark information for example, select a hearing-insensitive area between two adjacent frames and embed a bit, so that 65 consecutive frames contain a 64-bit synchronization code and a single cycle Watermark information to facilitate the subsequent extraction process to locate the watermark information through audio framing, auditory insensitive area selection, and energy relationship search.
  • the video sequence and audio data containing the same watermark information can be stream-mixed to obtain a video file with double watermarks.
  • the video file with the above-mentioned audio and video double watermark can be released for viewing and use.
  • the video files may be tampered with and released as pseudo-original videos for a second time.
  • the watermark extraction operation can be performed on these videos to clarify the identity of the original publisher of the video.
  • FIG. 4 shows a schematic flowchart of a video watermark extraction method according to an embodiment of the present application.
  • step S410 a video file embedded with a watermark is obtained.
  • the video file with embedded watermark may be the processed video file described above in conjunction with FIG. 1 to FIG. 3, and the video file may be a video file with embedded audio and video double watermarks.
  • the embedding of the watermark also needs to be able to be extracted even when the video is attacked and tampered with in many cases. For this reason, what is obtained in step S410 may be a tampered video file with embedded watermark.
  • step S420 video data and audio data are extracted from the watermark-embedded video file.
  • existing tools can be used to separate audio and video.
  • a video file parsing tool can be used to operate on the video file to be watermarked to extract video data (usually a video sequence) and audio data (usually an audio stream).
  • step 430 the first watermark information embedded in the video data is extracted.
  • step S440 the second watermark information embedded in the audio data is extracted. It should be understood that the above extraction steps of the video watermark and the audio watermark may also be performed at the same time, or the audio watermark may be extracted first.
  • the extracted watermark can use the relevance when the previous watermark is added to perform mutual verification of audio and video watermarks, so as to improve the anti-tampering ability of the watermark.
  • the above-mentioned relevance can be a time relevance or a content relevance.
  • the above-mentioned relevance in time and content may be the same as each other.
  • the video file to be embedded can be segmented according to the time axis.
  • the video sequence and audio data at the same time point are embedded with the same watermark information to facilitate the extraction of the video watermark information and audio watermark information at the same time point.
  • the watermark extraction method of the present application may further include generating the extracted watermark of the video file according to the extracted first watermark information and the second watermark information.
  • the generation of the final extracted watermark can be determined based on the correlation between the audio and video watermarks obtained in advance.
  • a weighted summation may be performed on the watermark data included in each of the first watermark information and the second watermark information To generate the extracted watermark of the video file.
  • the weight of the watermark data included in each of the first watermark information and the second watermark information may be adjusted according to the degree of confidence.
  • the added first watermark information may include multiple sets of watermark data containing the first ranking data and the first watermark data
  • the added second watermark information may include Multiple sets of watermark data of the second sorted data and the second watermark data.
  • extracting the first watermark information embedded in the video data may include: determining subsequent first watermark data based on the extracted first ranking data
  • extracting the second watermark information embedded in the video data may include: Based on the extracted second ranking data, the subsequent second watermark data is determined.
  • the existence of watermark information can be further located by locating the sorted data that is easier to distinguish. For example, by finding the synchronization code "000...0001" as shown in FIG. 2 to locate the single-cycle synchronization code immediately following it.
  • extracting the first watermark information embedded in the video data includes: determining the video frame and/or video area that contains the first watermark information in the video data, and extracting the second watermark information embedded in the audio data.
  • the watermark information includes: determining an audio frame and/or audio area in the audio data that contains the second watermark information.
  • the determination of the above-mentioned area can be reversed based on the area when the watermark is embedded. For example, selecting a non-salient area of a key video frame, and/or selecting a frequency band auditory insensitive area in an adjacent audio frame.
  • the extraction of watermark bits can also be reversed based on the embedding algorithm.
  • the embedded watermark is an "energy watermark”
  • the constituent bit information of the first watermark information that conforms to a predetermined energy relationship can be extracted from a series of determined video frames and/or video regions; and the extracted constituent bit information Combined into the first watermark information.
  • the constituent bit information of the second watermark information that conforms to a predetermined energy relationship can be extracted from a series of determined audio frames and/or audio regions; and the extracted constituent bit information can be combined into all The second watermark information.
  • FIG. 5 shows an example of a joint watermark extraction process according to the present application.
  • the watermark extraction process in FIG. 5 can be regarded as the corresponding operation of the watermark embedding process in FIG. 3.
  • the synchronization code and watermark bits are extracted from the video sequence based on the energy relationship of the neighboring blocks, and the watermark information wm1 is obtained.
  • the watermark information wm1 can be regarded as the single-period watermark information added to the video sequence in the example in FIG. 2.
  • the audio can be divided into frames, and an area that is insensitive to hearing in the frequency band of each frame of audio can be selected. Subsequently, the synchronization code and watermark bits are extracted from the audio based on the energy relationship of the adjacent frequency bands, and the watermark information wm2 is obtained.
  • the watermark information wm1 can be regarded as the single-period watermark information added to the audio code stream in the example in FIG. 2.
  • the synchronization code can be used to align the audio watermark information and the video watermark, and the watermark information of the two sources after the alignment is weighted and added.
  • the weights a1 and a2 are adaptively adjusted according to whether the current time axis position can successfully extract the watermark.
  • FIG. 6 shows a schematic diagram of a brief scheme of the watermark embedding and extraction operations of this application.
  • the video file to be protected is segmented according to the time axis, and the video sequence and audio data at the same time point are embedded with the same watermark information.
  • the figure shows the watermark information as 0 or 1 embedded at the beginning of each segment.
  • there is a difference in embedding capacity that is, it may only take 7 seconds for a video sequence to complete the embedding of 64 bits, while it may take 10 seconds for an audio code stream.
  • the watermark extraction operation can be performed when the copyright information of the video needs to be proved.
  • A is a video file to be embedded with a watermark (such as a film and television work).
  • a joint watermark can be embedded for it.
  • the video file analysis tool is used to extract the video sequence and audio data separately from the video file to be embedded with the watermark.
  • video key frames can be extracted, and the embedding area can be selected based on the key and content analysis.
  • a video watermarking algorithm is selected to embed the synchronization code and watermark bits in the video sequence (that is, the embedding area selected in the previous step) to obtain a video sequence with watermarked information.
  • the audio can be divided into frames, and the embedded area in each frame of audio can be selected based on the key. Select an audio watermarking algorithm to embed the synchronization code and watermark bits in the audio to obtain the audio with watermarked information. Finally, the video sequence containing the same watermark information is mixed with the audio data to obtain a video file containing double watermarks.
  • the extraction process of the joint watermark may include operations corresponding to the above embedding process.
  • the video file analysis tool can also be used to extract video sequences and audio data from watermarked video files.
  • video key frames can be extracted, and the extraction area can be selected based on the key and content analysis.
  • the audio branch the audio can be divided into frames, and the extraction area of each frame of audio can be selected based on the key.
  • the synchronization code can be used to align the audio watermark information and the video watermark, and the watermark information of the two sources after the alignment is weighted and added, and the weights a1 and a2 are adaptive according to the confidence that the current time axis position contains the audio and video watermark. Adjustment.
  • copyright verification can also be performed based on a separate audio or video extraction watermark.
  • the video file with the audio and video combined watermark added according to the embodiments described in Figs. 1 to 3 of the present application can be published directly (or after being tampered by a third party) on the video website.
  • the video files can be obtained by web visitors in the form of streaming media.
  • streaming media refers to a technology and process in which a series of multimedia data is sent in segments via the Internet (usually compressed data), and video and audio are instantly transmitted on the Internet for viewing.
  • Streaming media data can refer to multimedia data sent in segments via the Internet. This technology enables data packets to be sent and watched continuously, without the need to download the entire media file before use.
  • watermark extraction can be performed on complete video files, or watermark extraction can be performed on video streams in the form of streaming media.
  • the present application can also be implemented as a streaming media watermark extraction method, including: acquiring watermark-embedded streaming media data, the streaming media data being generated from the above-mentioned watermark-embedded video file; from the watermark-embedded stream Extracting video data and audio data from the media data; extracting the first watermark information embedded in the video data; and extracting the second watermark information embedded in the audio data.
  • streaming media data of a certain length of time can be acquired cumulatively, and audio and video data can be extracted from the streaming media data (which can be regarded as video data fragments) within a certain length of time, and Extraction of the first and second watermark information.
  • audio and video data can be extracted from the streaming media data (which can be regarded as video data fragments) within a certain length of time, and Extraction of the first and second watermark information.
  • the embedding of the watermark can also be performed on video files in the form of streaming media.
  • this application can also be implemented as a streaming media data processing method, including: acquiring video data and audio data of the streaming media data; embedding first watermark information into the video data; Correlating and embedding the second watermark information; and obtaining the watermark-embedded streaming media data.
  • the foregoing operations can be performed on accumulated streaming media data (which can be regarded as video data fragments) of a certain length of time, or streaming media data obtained in real time can be embedded, which is not limited in this application.
  • this application can also be implemented as a video file processing device, including: a video parsing unit for obtaining video data and audio data in the video file; a video watermark embedding unit for embedding in the video data The first watermark information; an audio watermark embedding unit for embedding the second watermark information into the audio data; and a video mixing unit for mixing the video data embedded in the first watermark information and the video data embedded in the second watermark information
  • the audio data is used to obtain a video file embedded with a watermark.
  • the video analysis unit may analyze the video file in the form of streaming media, for example, real-time analysis.
  • the present application can also be implemented as a video watermark extraction device, including: a video analysis unit for acquiring the video data and audio data in the video file embedded with the watermark as described above; and a video watermark extraction unit for obtaining Extracting the first watermark information from the video data; and an audio watermark extraction unit for extracting and embedding the second watermark information from the audio data.
  • the device may further include: a watermark information generating unit, configured to perform a weighted summation of the watermark data included in each of the first watermark information and the second watermark information to generate the extracted watermark of the video file.
  • the video parsing unit may analyze video files in the form of streaming media, for example, real-time parsing.
  • Fig. 7 shows a schematic structural diagram of a computing device that can be used to implement the above video processing and watermark extraction method according to an embodiment of the present application.
  • the computing device 700 includes a memory 710 and a processor 720.
  • the processor 720 may be a multi-core processor, or may include multiple processors.
  • the processor 720 may include a general-purpose main processor and one or more special co-processors, such as a graphics processing unit (GPU), a digital signal processor (DSP), and so on.
  • the processor 720 may be implemented using a customized circuit, such as an Application Specific Integrated Circuit (ASIC) or a Field Programmable Gate Array (FPGA, Field Programmable Gate Arrays).
  • ASIC Application Specific Integrated Circuit
  • FPGA Field Programmable Gate Arrays
  • the memory 710 may include various types of storage units, such as a system memory, a read only memory (ROM), and a permanent storage device.
  • the ROM may store static data or instructions required by the processor 720 or other modules of the computer.
  • the permanent storage device may be a readable and writable storage device.
  • the permanent storage device may be a non-volatile storage device that does not lose stored instructions and data even after the computer is powered off.
  • the permanent storage device adopts a large-capacity storage device (such as a magnetic or optical disk, flash memory) as the permanent storage device.
  • the permanent storage device may be a removable storage device (for example, a floppy disk, an optical drive).
  • the system memory can be a readable and writable storage device or a volatile readable and writable storage device, such as dynamic random access memory.
  • the system memory can store some or all of the instructions and data needed by the processor at runtime.
  • the memory 710 may include any combination of computer-readable storage media, including various types of semiconductor memory chips (DRAM, SRAM, SDRAM, flash memory, programmable read-only memory), and magnetic disks and/or optical disks may also be used.
  • the memory 710 may include a removable storage device that can be read and/or written, such as a compact disc (CD), a read-only digital versatile disc (for example, DVD-ROM, double-layer DVD-ROM), Read-only Blu-ray discs, ultra-density discs, flash memory cards (such as SD cards, min SD cards, Micro-SD cards, etc.), magnetic floppy disks, etc.
  • a removable storage device such as a compact disc (CD), a read-only digital versatile disc (for example, DVD-ROM, double-layer DVD-ROM), Read-only Blu-ray discs, ultra-density discs, flash memory cards (such as SD cards, min SD cards, Micro-SD cards, etc.), magnetic floppy disks, etc.
  • the computer-readable storage medium does not include carrier waves and instantaneous electronic signals transmitted wirelessly or wiredly.
  • the memory 710 stores executable codes.
  • the processor 720 can be made to execute the video processing and watermark extraction methods described above.
  • This application embeds audio watermark and video watermark in the video file at the same time. The two do not interfere with each other and complement each other.
  • the audio and video watermark extraction information can be adaptively fused, thereby greatly improving the robustness of the video file watermark, especially Fight against malicious editing attacks on video content.
  • the weights of audio watermarks and video watermarks can be dynamically adjusted according to their respective reliability to ensure the reliability of the fusion watermark.
  • the segmented synchronization code method can be used to achieve dual watermark synchronization.
  • the method according to the present application can also be implemented as a computer program or computer program product, and the computer program or computer program product includes computer program code instructions for executing the above-mentioned steps defined in the above-mentioned method of the present application.
  • this application can also be implemented as a non-transitory machine-readable storage medium (or computer-readable storage medium, or machine-readable storage medium) on which executable code (or computer program, or computer instruction code) is stored ), when the executable code (or computer program, or computer instruction code) is executed by the processor of the electronic device (or computing device, server, etc.), the processor is made to execute each step of the above-mentioned method according to the present application .
  • each block in the flowchart or block diagram may represent a module, program segment, or part of the code, and the module, program segment, or part of the code contains one or more functions for realizing the specified logical function.
  • Executable instructions may also occur in a different order than marked in the drawings. For example, two consecutive blocks can actually be executed substantially in parallel, or they can sometimes be executed in the reverse order, depending on the functions involved.
  • each block in the block diagram and/or flowchart, and the combination of the blocks in the block diagram and/or flowchart can be implemented by a dedicated hardware-based system that performs the specified functions or operations Or it can be realized by a combination of dedicated hardware and computer instructions.

Abstract

Disclosed are a video file processing method and device and a watermark extraction method and device. The video file processing method comprises: obtaining video data and audio data of a video file; embedding first watermark information in the video data; embedding second watermark information in the audio data in an associated manner; and obtaining a watermark embedded video file. In the present application, mutually related audio and video watermarks are add into the video file by using the temporal and spatial relevance of a video sequence and an audio data contained in the video, thereby increasing the watermark coverage of the video file. In addition, the mutual verification between audio and video watermarks improves the robustness of embedded watermarks against malicious tampering.

Description

视频文件处理与水印提取的方法和装置Method and device for video file processing and watermark extraction
本申请要求2020年03月24日递交的申请号为202010215301.1、发明名称为“视频文件处理与水印提取的方法和装置”的中国专利申请的优先权,其全部内容通过引用结合在本申请中。This application claims the priority of the Chinese patent application with the application number 202010215301.1 and the invention title of "Video file processing and watermark extraction method and device" filed on March 24, 2020, the entire content of which is incorporated into this application by reference.
技术领域Technical field
本公开涉及数字媒体处理技术领域,尤其涉及一种视频文件处理与水印提取的方法和装置。The present disclosure relates to the technical field of digital media processing, and in particular to a method and device for video file processing and watermark extraction.
背景技术Background technique
随着互联网的飞速发展,视频的制作和观看变得日益便捷与流行。在利益驱使下,视频的盗版问题也逐渐凸显。某些第三方会通过一些技术手段窃取视频生成方的展示视频。例如,短视频平台上很多UP主通过简单编辑他人视频生成伪原创视频来获取利益。上述问题扰乱了视频及其相关产业的发展。With the rapid development of the Internet, the production and viewing of videos has become increasingly convenient and popular. Driven by interests, the problem of video piracy has gradually become prominent. Some third parties will steal the video displayed by the video producer through some technical means. For example, many UP owners on short video platforms obtain benefits by simply editing other people's videos to generate pseudo-original videos. The above-mentioned problems have disrupted the development of video and related industries.
为此,需要一种能够更好地确认视频版权的方法。For this reason, a method that can better confirm the copyright of the video is needed.
发明内容Summary of the invention
为了解决如上至少一个问题,本公开提供了一种视频文件处理以及相应的水印提取方法。该方法利用视频中包含视频序列和音频数据在时间和空间上的关联性,向视频文件添加彼此关联的音视频水印,从而提升了视频文件的水印覆盖面。另外,通过音视频双水印之间的彼此验证,提升了嵌入水印对抗恶意篡改的鲁棒性。In order to solve at least one of the above problems, the present disclosure provides a video file processing and corresponding watermark extraction method. This method utilizes the temporal and spatial relevance of the video sequence and audio data contained in the video, and adds mutually related audio and video watermarks to the video file, thereby improving the watermark coverage of the video file. In addition, through mutual verification between audio and video double watermarks, the robustness of embedded watermarks against malicious tampering is improved.
根据本公开的第一个方面,提供了一种视频文件处理方法,包括:获取所述视频文件的视频数据和音频数据;向所述视频数据中嵌入第一水印信息;向所述音频数据中关联嵌入第二水印信息;以及获取嵌入水印的视频文件。According to a first aspect of the present disclosure, there is provided a video file processing method, including: acquiring video data and audio data of the video file; embedding first watermark information in the video data; Associating and embedding the second watermark information; and obtaining the watermark-embedded video file.
根据本公开的第二个方面,提供了一种视频水印提取方法,包括:获取根据第一个方面所述的嵌入水印的视频文件;从所述嵌入水印的视频文件中抽取视频数据和音频数据;提取所述视频数据中嵌入的第一水印信息;以及提取所述音频数据中嵌入的第二水印信息。According to a second aspect of the present disclosure, there is provided a method for extracting a video watermark, including: obtaining the watermark-embedded video file according to the first aspect; and extracting video data and audio data from the watermark-embedded video file ; Extracting the first watermark information embedded in the video data; and extracting the second watermark information embedded in the audio data.
根据本公开的第三个方面,提供了一种流媒体水印提取方法,包括:获取嵌入水印的流媒体数据,所述流媒体数据由第一方面所述的嵌入水印的视频文件生成;从所述嵌 入水印的流媒体数据中抽取视频数据和音频数据;提取所述视频数据中嵌入的第一水印信息;以及提取所述音频数据中嵌入的第二水印信息。According to a third aspect of the present disclosure, there is provided a method for extracting a streaming media watermark, including: acquiring streaming media data embedded with a watermark, the streaming media data being generated from the watermark-embedded video file described in the first aspect; Extracting video data and audio data from the watermark-embedded streaming media data; extracting the first watermark information embedded in the video data; and extracting the second watermark information embedded in the audio data.
根据本公开的第四个方面,提供了一种流媒体数据处理方法,包括:获取所述流媒体数据的视频数据和音频数据;向所述视频数据中嵌入第一水印信息;向所述音频数据中关联嵌入第二水印信息;以及获取嵌入水印的流媒体数据。According to a fourth aspect of the present disclosure, there is provided a streaming media data processing method, including: acquiring video data and audio data of the streaming media data; embedding first watermark information into the video data; The second watermark information is associated and embedded in the data; and the watermark-embedded streaming media data is obtained.
根据本公开的第五个方面,提供了一种视频文件处理装置,包括:视频解析单元,用于获取所述视频文件中的视频数据和音频数据;视频水印嵌入单元,用于向所述视频数据中嵌入第一水印信息;音频水印嵌入单元,用于向所述音频数据中关联嵌入第二水印信息;以及视频混合单元,用于混合嵌入第一水印信息的所述视频数据和嵌入第二水印信息的所述音频数据,以获取嵌入水印的视频文件。According to a fifth aspect of the present disclosure, there is provided a video file processing device, including: a video parsing unit for acquiring video data and audio data in the video file; a video watermark embedding unit for sending the video to the video The first watermark information is embedded in the data; the audio watermark embedding unit is used to associate and embed the second watermark information into the audio data; and the video mixing unit is used to mix the video data embedded in the first watermark information and the second watermark information. The audio data of the watermark information is used to obtain a watermark embedded video file.
根据本公开的第六个方面,提供了一种视频水印提取装置,包括:视频解析单元,用于获取根据第一个方面所述的嵌入水印的视频文件中的视频数据和音频数据;视频水印提取单元,用于从所述视频数据中提取第一水印信息;以及音频水印提取单元,用于从所述音频数据中提取嵌入第二水印信息。According to a sixth aspect of the present disclosure, there is provided a video watermark extraction device, including: a video parsing unit for obtaining video data and audio data in a watermark-embedded video file according to the first aspect; video watermark The extraction unit is used to extract the first watermark information from the video data; and the audio watermark extraction unit is used to extract and embed the second watermark information from the audio data.
根据本公开的第七个方面,提供了一种计算设备,包括:处理器;以及存储器,其上存储有可执行代码,当可执行代码被处理器执行时,使处理器执行如上述第一至第四方面所述的方法。According to a seventh aspect of the present disclosure, there is provided a computing device, including: a processor; and a memory on which executable code is stored. When the executable code is executed by the processor, the processor is caused to execute the first To the method described in the fourth aspect.
根据本公开的第八个方面,提供了一种非暂时性机器可读存储介质,其上存储有可执行代码,当可执行代码被电子设备的处理器执行时,使处理器执行如上述第一至第四方面所述的方法。According to an eighth aspect of the present disclosure, there is provided a non-transitory machine-readable storage medium having executable code stored thereon. When the executable code is executed by a processor of an electronic device, the processor executes the above-mentioned One to the method described in the fourth aspect.
本申请对视频文件同时嵌入音频水印和视频水印,两者互不干扰,相互补充,提取水印时可将音视频水印提取信息进行自适应融合,从而大幅提高视频文件水印的鲁棒性,特别是对抗针对视频内容的恶意编辑攻击。This application embeds audio watermark and video watermark in the video file at the same time. The two do not interfere with each other and complement each other. When extracting the watermark, the audio and video watermark extraction information can be adaptively fused, thereby greatly improving the robustness of the video file watermark, especially Fight against malicious editing attacks on video content.
附图说明Description of the drawings
通过结合附图对本公开示例性实施方式进行更详细的描述,本公开的上述以及其它目的、特征和优势将变得更加明显,其中,在本公开示例性实施方式中,相同的参考标号通常代表相同部件。Through a more detailed description of the exemplary embodiments of the present disclosure in conjunction with the accompanying drawings, the above and other objectives, features and advantages of the present disclosure will become more apparent. Among them, in the exemplary embodiments of the present disclosure, the same reference numerals generally represent The same parts.
图1示出了根据本申请的一种视频文件处理方法的示意性流程图;Fig. 1 shows a schematic flowchart of a video file processing method according to the present application;
图2示出了为视频序列和音频码流分别嵌入水印信息的一个例子;Figure 2 shows an example of embedding watermark information for video sequence and audio code stream respectively;
图3示出了根据本申请的联合水印嵌入流程的一个例子;Figure 3 shows an example of the joint watermark embedding process according to the present application;
图4示出了根据本申请一个实施例的视频水印提取方法的示意性流程图;Figure 4 shows a schematic flowchart of a video watermark extraction method according to an embodiment of the present application;
图5示出了根据本申请的联合水印提取流程的一个例子;Figure 5 shows an example of a joint watermark extraction process according to the present application;
图6示出了本申请水印嵌入和提取操作的简要方案示意图;Figure 6 shows a schematic diagram of a brief scheme of the watermark embedding and extraction operations of this application;
图7示出了根据本申请一实施例可用于实现上述视频处理和水印提取方法的计算设备的结构示意图。Fig. 7 shows a schematic structural diagram of a computing device that can be used to implement the above video processing and watermark extraction method according to an embodiment of the present application.
具体实施方式Detailed ways
下面将参照附图更详细地描述本公开的优选实施方式。虽然附图中显示了本公开的优选实施方式,然而应该理解,可以以各种形式实现本公开而不应被这里阐述的实施方式所限制。相反,提供这些实施方式是为了使本公开更加透彻和完整,并且能够将本公开的范围完整地传达给本领域的技术人员。Hereinafter, preferred embodiments of the present disclosure will be described in more detail with reference to the accompanying drawings. Although the drawings show preferred embodiments of the present disclosure, it should be understood that the present disclosure can be implemented in various forms and should not be limited by the embodiments set forth herein. On the contrary, these embodiments are provided to make the present disclosure more thorough and complete, and to fully convey the scope of the present disclosure to those skilled in the art.
随着互联网的飞速发展,视频的制作和观看变得日益便捷与流行。在利益驱使下,视频的盗版问题也逐渐凸显。某些第三方会通过一些技术手段窃取视频生成方的展示视频。例如,短视频平台上很多UP主通过简单编辑他人视频生成伪原创视频来获取利益。上述问题扰乱了视频及其相关产业的发展。With the rapid development of the Internet, the production and viewing of videos has become increasingly convenient and popular. Driven by interests, the problem of video piracy has gradually become prominent. Some third parties will steal the video displayed by the video producer through some technical means. For example, many UP owners on short video platforms obtain benefits by simply editing other people's videos to generate pseudo-original videos. The above-mentioned problems have disrupted the development of video and related industries.
近年来,数字水印技术在视频版权保护领域取得一定的成果。该技术能够将水印信息嵌入载体信息(例如,冗余信息)中,从而对视频版权进行保护。目前常见平台上的视频通常都包含视频序列和音频数据两部分,很多数字水印的应用都是将音视频作为一个整体进行处理。第三方会在盗版视频制作过程中针对其中局部视频序列或局部音频进行编辑伪造(例如去除logo水印、增加新logo水印、插入广告、修改分辨率等),这些编辑操作的组合对原视频信号引入严重失真,导致该局部对水印信息无法提取,因此仅针对视频序列或音频数据嵌入单一水印的方案并不能保证视频更高级别的版权保护要求。In recent years, digital watermarking technology has achieved certain results in the field of video copyright protection. This technology can embed the watermark information in the carrier information (for example, redundant information), thereby protecting the video copyright. At present, videos on common platforms usually contain two parts: video sequence and audio data. Many digital watermark applications process audio and video as a whole. The third party will edit and falsify the partial video sequence or partial audio in the pirated video production process (such as removing the logo watermark, adding a new logo watermark, inserting advertisements, modifying the resolution, etc.). The combination of these editing operations introduces the original video signal Severe distortion results in that the watermark information cannot be extracted in this part. Therefore, the solution of embedding a single watermark for video sequences or audio data does not guarantee the higher-level copyright protection requirements of the video.
为此,本申请提出了一种音视频联合水印方案。视频中包含的不同数据内容(视频序列和音频数据)在时间和空间上有很强的关联性,彼此制约。现有的视频水印算法多数都是只对视频码流做嵌入提取处理,没有利用音视频之间的关联性,或是根本就没有考虑到音频数据的存在。本申请通过将音视频数据之间的关联性引入视频水印处理,使得视频水印更为精确和鲁棒,由此版权保护和数字内容鉴定能够取得更好的效果。For this reason, this application proposes an audio-video joint watermarking scheme. The different data content (video sequence and audio data) contained in the video has a strong correlation in time and space and restricts each other. Most of the existing video watermarking algorithms only do embedding and extraction processing on the video code stream, without using the correlation between audio and video, or without considering the existence of audio data at all. This application introduces the correlation between audio and video data into the video watermarking process to make the video watermark more accurate and robust, so that copyright protection and digital content authentication can achieve better results.
本申请首先可以实现为一种向音视频添加水印的方案。上述方案首先可以实现为一 种视频文件处理方法。图1示出了根据本申请的一种视频文件处理方法的示意性流程图。在此,“视频文件”指代通常包括视频和音频两种信息的文件。换句话说,视频文件既包括图像内容,也包括声音内容。This application can first be implemented as a solution for adding a watermark to audio and video. The above solution can first be implemented as a video file processing method. Fig. 1 shows a schematic flowchart of a video file processing method according to the present application. Here, "video file" refers to a file that usually includes both video and audio information. In other words, video files include both image content and sound content.
在步骤S110,获取所述视频文件的视频数据和音频数据。对于待嵌入水印的视频文件,可以利用现有工具进行音视频分离。例如,可以利用视频文件解析工具对待嵌入水印的视频文件进行操作,以分别抽取视频数据(通常为视频序列)和音频数据(通常为音频码流)。In step S110, the video data and audio data of the video file are acquired. For video files to be embedded with watermarks, existing tools can be used to separate audio and video. For example, a video file parsing tool can be used to operate on the video file to be embedded with the watermark to extract video data (usually a video sequence) and audio data (usually an audio stream).
在步骤S120,向所述视频数据中嵌入第一水印信息。In step S120, the first watermark information is embedded in the video data.
在步骤S130,向所述音频数据中关联嵌入第二水印信息。应该了解的是,上述视频水印和音频水印的插入步骤,也可以是同时,或是音频水印插入在先。只要视频和音箱的水印嵌入存在关联性即可。后续在水印提取时,可以利用上述关联性,进行音频和视频水印的关联验证,由此提升水印的抗篡改能力。In step S130, the second watermark information is associated and embedded into the audio data. It should be understood that the steps of inserting the video watermark and the audio watermark can also be performed at the same time, or the audio watermark can be inserted first. As long as there is a correlation between the watermark embedding of the video and the speaker. In the subsequent extraction of watermarks, the above-mentioned correlation can be used to perform correlation verification of audio and video watermarks, thereby improving the anti-tampering ability of the watermarks.
在某些实施例中,上述关联性可以是时间上的关联,即,向所述音频数据嵌入第二水印信息的时间可以与向所述视频数据嵌入第一水印信息的时间相关联。作为替换或者补充,上述关联性也可以是内容上的关联性,即,向所述音频数据嵌入第二水印信息的内容可以与向所述视频数据嵌入第一水印信息的内容相关联。在一个优选实施例中,音频和视频水印添加的时间和内容可以都有预先确定的关联关系,以方便在后续针对提取的音视频水印信息的相互验证。In some embodiments, the above-mentioned association may be a time association, that is, the time when the second watermark information is embedded in the audio data may be associated with the time when the first watermark information is embedded in the video data. As an alternative or supplement, the above-mentioned association may also be a content association, that is, the content of the second watermark information embedded in the audio data may be associated with the content of the first watermark information embedded in the video data. In a preferred embodiment, the time and content of the audio and video watermarking may have a predetermined association relationship, so as to facilitate subsequent mutual verification of the extracted audio and video watermark information.
上述时间和内容上的关联性可以是彼此相同。例如,可以在视频和音频时间轴的相同位置上分别添加音频和视频水印。考虑到水印数据的长度往往需要涉及一个时间段内的持续添加,并且同样时长的视频序列和音频码流存在嵌入容量差异(即可能只需要7秒时长视频序列可完成嵌入64比特,而可能需要10秒时长音频码流才能完成64比特的嵌入),因此上述时间上的相同可以是嵌入的初始时间相同。于是,在一个实施例中,向所述音频数据嵌入第二水印信息的起始时间可以与向所述视频数据嵌入第一水印信息的起始时间相同。由此,在双水印提取时,方便音视频水印的彼此对齐和相互验证。The above-mentioned relevance in time and content may be the same as each other. For example, you can add audio and video watermarks at the same position on the video and audio timelines, respectively. Considering that the length of watermark data often needs to involve continuous addition within a period of time, and there is a difference in embedding capacity between video sequences and audio streams of the same duration (that is, it may only take 7 seconds for a video sequence to complete the embedding of 64 bits, but may require The 10-second audio code stream can only complete the 64-bit embedding), so the same time may be the same initial time of embedding. Therefore, in one embodiment, the start time of embedding the second watermark information into the audio data may be the same as the start time of embedding the first watermark information into the video data. Therefore, when the double watermark is extracted, it is convenient to align and verify each other of the audio and video watermarks.
虽然在某些实施例中,音视频的水印内容可以基于一定的映射关系彼此关联,而非完全相同。但是为了提升视频发布后抵御篡改的鲁棒性,优选音视频具有相同的水印内容。换句话说,向所述音频数据嵌入第二水印信息的内容可以与向所述视频数据嵌入第一水印信息的内容相同。由此,进一步提升水印提取时音视频水印相互验证的鲁棒性。Although in some embodiments, the watermark content of audio and video may be related to each other based on a certain mapping relationship, rather than being completely the same. However, in order to improve the robustness of the video against tampering after publishing, it is preferable that the audio and video have the same watermark content. In other words, the content of the second watermark information embedded in the audio data may be the same as the content of the first watermark information embedded in the video data. As a result, the robustness of mutual verification of audio and video watermarks during watermark extraction is further improved.
由于视频文件通常都具有一定的时长,为了防止通过简单的截取文件部分内容进行 盗取(例如,选取一个3分钟视频中精华的30秒进行发布),可以向一个视频文件中的视频序列和音频码流多次添加水印信息以增强保护。为此,向所述视频数据中嵌入第一水印信息可以包括:以第一预定时间间隔,向所述视频数据中嵌入多个第一水印信息。相应地,向所述音频数据中嵌入关联的第二水印信息包括:以第二预定时间间隔,向所述音频数据中嵌入多个第二水印信息。如前所述,为了提升鲁棒性并方便对齐,优选以相同的时间间隔进行音频和视频的水印嵌入。另外,虽然预定时间间隔可以是非均匀的时间间隔,例如,前一分钟每10秒嵌入一个,后一分钟每15秒嵌入一个,但是仍然优选以均匀的预定时间间隔进行水印插入,以提升水印的抗篡改能力。Since video files usually have a certain length of time, in order to prevent theft by simply intercepting part of the content of the file (for example, select 30 seconds of the essence of a 3-minute video to publish), you can send the video sequence and audio in a video file Watermark information is added to the code stream multiple times to enhance protection. To this end, embedding the first watermark information into the video data may include: embedding a plurality of first watermark information into the video data at a first predetermined time interval. Correspondingly, embedding the associated second watermark information into the audio data includes: embedding a plurality of second watermark information into the audio data at a second predetermined time interval. As mentioned above, in order to improve robustness and facilitate alignment, it is preferable to embed audio and video watermarks at the same time interval. In addition, although the predetermined time interval may be a non-uniform time interval, for example, embedding one every 10 seconds in the first minute and embedding one every 15 seconds in the next minute, it is still preferable to perform the watermark insertion at a uniform predetermined time interval to improve the watermarking performance. Anti-tampering ability.
另外,如前所述,由于水印数据的长度往往需要涉及一个时间段内的持续添加,并且同样时长的视频序列和音频码流存在嵌入容量差异(即可能只需要7秒时长视频序列可完成嵌入64比特,而可能需要10秒时长音频码流才能完成64比特的嵌入),因此在重复设置多个水印的情况下还需要考虑嵌入提取时音视频水印设置同步。为此,所述多个第一水印信息中的每个第一水印信息可以包括:第一排序数据和第一水印数据。相应地,所述多个第二水印信息中的每个第二水印信息可以包括:第二排序数据和第二水印数据。换句话说,实际添加的水印信息可以包括排序码以及水印信息本身,并且按照一定的时间间隔重复。In addition, as mentioned earlier, because the length of watermark data often involves continuous addition within a period of time, and there is a difference in embedding capacity between video sequences and audio streams of the same length (that is, it may only take 7 seconds for a video sequence to complete the embedding. 64 bits, and it may take 10 seconds to complete the 64-bit embedding of the audio stream. Therefore, when multiple watermarks are repeatedly set, it is also necessary to consider the synchronization of audio and video watermark settings during embedding and extraction. To this end, each of the plurality of first watermark information may include: first ranking data and first watermark data. Correspondingly, each second watermark information in the plurality of second watermark information may include: second ranking data and second watermark data. In other words, the actually added watermark information may include the sort code and the watermark information itself, and it is repeated at a certain time interval.
图2示出了为视频序列和音频码流分别嵌入水印信息的一个例子。如图2所示,可以在时间轴对齐的视频序列和音频码流中,以10秒的时间间隔重复插入相同的水印信息。每个时间间隔内插入的水印信息包括同步码以及单周期水印信息。Figure 2 shows an example of embedding watermark information for a video sequence and an audio code stream respectively. As shown in Figure 2, the same watermark information can be repeatedly inserted in the video sequence and audio code stream aligned on the time axis at a time interval of 10 seconds. The watermark information inserted in each time interval includes synchronization code and single-period watermark information.
在此,“同步码”可以是在音频或是视频内部用来为添加的单周期水印信息进行排序计数的码。例如,在第一个时间间隔的起始时刻(即,第0秒)添加的单周期水印信息是“000……0000”,在第二个时间间隔的起始时刻(即,第10秒)添加的单周期水印信息可以加1,变为“000……0001”,在第三个时间间隔的起始时刻(即,第20秒)添加的单周期水印信息可以加1,变为“000……0010”,并以此类推。进一步地,“同步码”的“同步”可以指代视频序列和音频码流之间的同步。即,在后续的水印提取阶段,视频序列和音频码流可以借助恢复出来的同步码进行时间上的对齐。Here, the "synchronization code" can be a code used in audio or video to sort and count the added single-period watermark information. For example, the single-period watermark information added at the beginning of the first time interval (ie, the 0th second) is "000...0000", at the beginning of the second time interval (ie, the 10th second) The added single-period watermark information can be increased by 1 to become "000……0001", and the single-period watermark information added at the beginning of the third time interval (ie, the 20th second) can be increased by 1 and become "000. ……0010", and so on. Further, the "synchronization" of the "synchronization code" may refer to the synchronization between the video sequence and the audio code stream. That is, in the subsequent watermark extraction stage, the video sequence and the audio code stream can be aligned in time with the help of the recovered synchronization code.
在此,“单周期水印信息”可以指代用于唯一表示视频身份或是视频制作者身份(或是视频发布方身份)的识别码,例如图2中示出的“48位的一个周期的完整水印信息”,由此与整个音频或视频中插入的所有水印相关信息(包括不断增大的同步码和其后一直重复的水印识别码)相区别。Here, "single-period watermark information" can refer to an identification code used to uniquely represent the identity of the video or the identity of the video producer (or the identity of the video publisher), for example, the "48-bit complete cycle of one period" shown in Figure 2 "Watermark information" is thus distinguished from all watermark-related information inserted in the entire audio or video (including the ever-increasing synchronization code and the watermark identification code that has been repeated thereafter).
如图2所示,同步码和单周期水印信息都是由0和1表示的二进制数据,与实际音视频处理中的数据进制相同。As shown in Figure 2, both the synchronization code and the single-period watermark information are binary data represented by 0 and 1, which is the same as the data system in the actual audio and video processing.
由于同样时长的视频序列和音频码流存在嵌入容量差异,只需要7秒时长视频序列就可完成嵌入64比特,而需要10秒时长音频码流才能完成64比特的嵌入,因此重复插入水印的间隔应该不小于10秒时长。虽然在本例中示出了以10秒为预定间隔持续进行水印插入,但应该理解的是,在其他实施例中,也可以以更长的时间间隔进行水印插入,例如,每20秒插入一次。类似地,每一次嵌入的同步码和单周期水印信息也可以具有不同的位数,例如,同步码32位,水印信息32位等,并且同步码与水印信息的整体长度也可以是64位之外的其他长度。Due to the difference in embedding capacity between video sequence and audio code stream of the same duration, it only takes 7 seconds to complete the 64-bit embedding of the video sequence, while it takes 10 seconds to complete the 64-bit embedding of the audio code stream, so the watermarking interval is repeated. It should be no less than 10 seconds long. Although the watermark insertion is continuously performed at a predetermined interval of 10 seconds in this example, it should be understood that in other embodiments, the watermark insertion may also be performed at a longer time interval, for example, once every 20 seconds. . Similarly, each embedded synchronization code and single-period watermark information can also have a different number of bits, for example, synchronization code 32 bits, watermark information 32 bits, etc., and the overall length of synchronization code and watermark information can also be less than 64 bits. Other lengths.
在各自完成了水印插入之后,在步骤S140,可以获取嵌入水印的视频文件。After the watermark insertion is completed respectively, in step S140, the watermark-embedded video file can be obtained.
具体地,可以将含有相同水印信息的视频序列与音频数据进行码流混合,得到含双水印的视频文件。上述嵌入双水印的视频文件随后可以发布,并在需要时,基于预定方法进行水印提取和恢复。Specifically, a video sequence and audio data containing the same watermark information can be stream-mixed to obtain a video file containing a double watermark. The video file with the double watermark embedded above can be released later, and when necessary, the watermark is extracted and restored based on a predetermined method.
在不同的实施例中,可以选择不同的音频和视频水印嵌入方法进行分别针对音频和视频的水印数据插入,只要其嵌入存在关联性即可。In different embodiments, different audio and video watermark embedding methods can be selected to insert watermark data for audio and video respectively, as long as the embedding has relevance.
在一个实施例中,为了防止水印嵌入对视听效果的影响,可以基于音频和视频数据各自的属性,选择相应的区域进行嵌入操作。为此,向所述视频数据中嵌入第一水印信息可以包括:向所述视频数据中视频帧的非显著区域添加所述第一水印信息。优选地,可以提取视频关键帧,并进行内容分析,以选取非显著区域用于水印信息的添加。作为替换或者补充,向所述音频数据中嵌入第二水印信息包括:向所述音频数据中音频帧的听觉不敏感区域添加所述第二水印信息。优选地,对音频进行分帧处理,以选取每帧音频中频带听觉不敏感的区域用于水印信息的添加。In one embodiment, in order to prevent the influence of watermark embedding on the audiovisual effect, a corresponding area may be selected for embedding operation based on the respective attributes of the audio and video data. To this end, embedding the first watermark information into the video data may include: adding the first watermark information to a non-significant area of a video frame in the video data. Preferably, video key frames can be extracted, and content analysis can be performed to select non-significant areas for adding watermark information. As an alternative or supplement, embedding the second watermark information into the audio data includes: adding the second watermark information to the auditory insensitive area of the audio frame in the audio data. Preferably, the audio is divided into frames to select areas that are not auditorily sensitive to the frequency band in each frame of audio for adding watermark information.
在具体嵌入时,可以选择进行能量水印的嵌入。在此,“能量水印”指代通过调整媒体内容某个变换域相邻区域的能量关系而实现嵌入的水印算法。在水印信息包括较多比特数的情况下,每一个完整水印的插入通常需要针对一系列的视频帧和音频帧持续进行。为此,向所述视频数据中嵌入第一水印信息可以包括:通过调整所述视频数据中视频帧变换域相邻区域的能量关系,嵌入所述第一水印信息。进一步地,通过调整所述视频数据中视频帧变换域相邻区域的能量关系,嵌入所述第一水印信息可以包括:选取视频序列中的一系列特定视频帧,所述视频数据是视频序列;以及调整所述一系列特定视频帧的变换域相邻区域的能量关系,并逐一嵌入所述第一水印信息的组成比特信息。在 此,选取视频序列中的一系列特定视频帧可以包括:选取所述视频序列中的视频关键帧。例如,每个关键帧内嵌入64比特信息中的1个比特,最终在这64个连续的关键帧内完成同步码和单周期水印信息的嵌入。In the specific embedding, you can choose to embed the energy watermark. Here, "energy watermark" refers to an embedded watermark algorithm by adjusting the energy relationship between adjacent regions of a certain transformation domain of media content. When the watermark information includes a large number of bits, the insertion of each complete watermark usually needs to be performed continuously for a series of video frames and audio frames. To this end, embedding the first watermark information into the video data may include: embedding the first watermark information by adjusting the energy relationship between adjacent regions of the video frame transform domain in the video data. Further, by adjusting the energy relationship between adjacent regions of the video frame transform domain in the video data, embedding the first watermark information may include: selecting a series of specific video frames in a video sequence, and the video data is a video sequence; And adjust the energy relationship of the adjacent regions of the transform domain of the series of specific video frames, and embed the constituent bit information of the first watermark information one by one. Here, selecting a series of specific video frames in the video sequence may include: selecting video key frames in the video sequence. For example, one bit of the 64-bit information is embedded in each key frame, and finally the synchronization code and the single-period watermark information are embedded in the 64 consecutive key frames.
相应地,向所述音频数据中嵌入关联的第二水印信息可以包括:调整相邻音频帧的能量关系,嵌入所述第二水印信息。进一步地,调整相邻音频帧的能量关系,嵌入所述第二水印信息可以包括:调整所述一系列相邻音频帧的相邻频带的能量关系,并逐一嵌入所述第二水印信息的组成比特信息。例如,可以调整相邻两个频带之间的能量关系,并嵌入64比特信息中的1个比特,最终在65个连续的相邻帧之间完成同步码和单周期水印信息的嵌入。Correspondingly, embedding the associated second watermark information into the audio data may include: adjusting the energy relationship of adjacent audio frames, and embedding the second watermark information. Further, adjusting the energy relationship of adjacent audio frames, and embedding the second watermark information may include: adjusting the energy relationship of adjacent frequency bands of the series of adjacent audio frames, and embedding the composition of the second watermark information one by one Bit information. For example, the energy relationship between two adjacent frequency bands can be adjusted, and one bit of the 64-bit information can be embedded. Finally, the synchronization code and the single-period watermark information can be embedded between 65 consecutive adjacent frames.
本申请的上述视频处理方法尤其适于实现为一种音视频联合水印添加的方案。图3示出了根据本申请的联合水印嵌入流程的一个例子。对于要进行水印添加的视频文件,可以首先利用视频文件解析工具分别抽取视频序列和音频数据。随后,如图3所示,流程分为两个并列的分支。The above-mentioned video processing method of the present application is particularly suitable for being implemented as an audio-video joint watermarking solution. Figure 3 shows an example of the joint watermark embedding process according to the present application. For the video file to be watermarked, you can first use the video file analysis tool to extract the video sequence and audio data separately. Subsequently, as shown in Figure 3, the process is divided into two parallel branches.
针对抽取的视频序列,提取视频关键帧,并进行内容分析选取非显著区域。针对选取的非显著区域,可以基于相邻块的能量关系嵌入同步码和水印比特,得到含水印信息的视频序列。结合图2的例子说明,可以提取10秒至17秒时段内的视频关键帧,进行内容分析选取非显著区域,随后进行诸如DCT(离散余弦变换)的变换得到其变换域,在通过变换域相邻区域的能量关系的调整,进行水印信息的嵌入,例如,一个关键帧的选定非显著区域,嵌入一个比特,从而使得64个连续关键帧包含64位的同步码和单周期水印信息,以方便后续提取流程中通过关键帧提取、非显著区域选取和相邻区域能量关系查找来定位水印信息。For the extracted video sequence, the video key frames are extracted, and content analysis is performed to select non-salient areas. For the selected insignificant areas, synchronization codes and watermark bits can be embedded based on the energy relationship of adjacent blocks to obtain a video sequence with watermarked information. In combination with the example in Figure 2, it is possible to extract the key frames of the video from 10 seconds to 17 seconds, perform content analysis to select insignificant areas, and then perform transformations such as DCT (Discrete Cosine Transform) to obtain its transform domain. Adjust the energy relationship of neighboring regions to embed watermark information. For example, a selected non-salient area of a key frame is embedded with one bit, so that 64 consecutive key frames contain 64-bit synchronization code and single-period watermark information. It is convenient for the subsequent extraction process to locate the watermark information through key frame extraction, non-significant region selection and adjacent region energy relationship search.
另一方面,可以对音频进行分帧处理,并选取每帧音频中频带听觉不敏感的区域。对选取的听觉不敏感的区域,基于相邻频带的能量关系嵌入同步码和水印比特,得到含水印信息的音频。结合图2的例子说明,可以提取10秒至20秒时段内的音频帧,选取听觉不敏感的频带,随后进行诸如DCT(离散余弦变换)的变换得到其变换域,在通过变换域相邻频带的能量关系的调整,进行水印信息的嵌入,例如,两个相邻帧之间选定听觉不敏感的区域,嵌入一个比特,从而使得65个连续帧之间包含64位的同步码和单周期水印信息,以方便后续提取流程中通过音频分帧、听觉不敏感区域选取和能量关系查找来定位水印信息。On the other hand, you can divide the audio into frames and select areas that are insensitive to hearing in the mid-band of each frame of audio. For the selected areas that are not sensitive to hearing, synchronization codes and watermark bits are embedded based on the energy relationship of adjacent frequency bands to obtain audio with watermarked information. In combination with the example in Figure 2, you can extract audio frames within a period of 10 to 20 seconds, select a frequency band that is not sensitive to hearing, and then perform a transformation such as DCT (Discrete Cosine Transform) to obtain its transform domain. The adjustment of the energy relationship, the embedding of watermark information, for example, select a hearing-insensitive area between two adjacent frames and embed a bit, so that 65 consecutive frames contain a 64-bit synchronization code and a single cycle Watermark information to facilitate the subsequent extraction process to locate the watermark information through audio framing, auditory insensitive area selection, and energy relationship search.
在对视频和视频都进行了水印添加执行之后,可以将含有相同水印信息的视频序列 与音频数据进行码流混合,得到含双水印的视频文件。After watermarking is performed on both the video and the video, the video sequence and audio data containing the same watermark information can be stream-mixed to obtain a video file with double watermarks.
随后,添加了上述音视频双水印的视频文件可以被发布,以供观看和使用。由于在视频文件被发布后,可能会遭受篡改并作为伪原创视频进行二次发布,此时,可以对这些视频进行水印提取操作,以明确该视频的初始发布者身份。Subsequently, the video file with the above-mentioned audio and video double watermark can be released for viewing and use. After the video files are released, they may be tampered with and released as pseudo-original videos for a second time. At this time, the watermark extraction operation can be performed on these videos to clarify the identity of the original publisher of the video.
为此,本申请还可以实现位一种视频水印提取方法。图4示出了根据本申请一个实施例的视频水印提取方法的示意性流程图。For this reason, this application can also implement a video watermark extraction method. Fig. 4 shows a schematic flowchart of a video watermark extraction method according to an embodiment of the present application.
在步骤S410,获取嵌入水印的视频文件。在此,嵌入水印的视频文件可以是如上结合图1至图3描述的经处理的视频文件,所述视频文件可以是嵌入了音视频双水印的视频文件。水印的嵌入除了能在视频无篡改的时候验证视频的身份,在很多情况下还需要在视频被攻击和篡改时仍能保证被提取。为此,步骤S410获取的可以是经篡改的嵌入水印的视频文件。In step S410, a video file embedded with a watermark is obtained. Here, the video file with embedded watermark may be the processed video file described above in conjunction with FIG. 1 to FIG. 3, and the video file may be a video file with embedded audio and video double watermarks. In addition to verifying the identity of the video when the video is not tampered with, the embedding of the watermark also needs to be able to be extracted even when the video is attacked and tampered with in many cases. For this reason, what is obtained in step S410 may be a tampered video file with embedded watermark.
随后在步骤S420,从所述嵌入水印的视频文件中抽取视频数据和音频数据。类似地,对于待提取水印的视频文件,可以利用现有工具进行音视频分离。例如,可以利用视频文件解析工具对待提取水印的视频文件进行操作,以分别抽取视频数据(通常为视频序列)和音频数据(通常为音频码流)。Subsequently, in step S420, video data and audio data are extracted from the watermark-embedded video file. Similarly, for the video file whose watermark is to be extracted, existing tools can be used to separate audio and video. For example, a video file parsing tool can be used to operate on the video file to be watermarked to extract video data (usually a video sequence) and audio data (usually an audio stream).
在步骤430,提取所述视频数据中嵌入的第一水印信息。In step 430, the first watermark information embedded in the video data is extracted.
在步骤S440,提取所述音频数据中嵌入的第二水印信息。应该了解的是,上述视频水印和音频水印的提取步骤,也可以是同时,或是音频水印提取在先。提取后的水印可以利用在前水印添加时的关联性,进行音频和视频水印的相互验证,以提升水印的抗篡改能力。In step S440, the second watermark information embedded in the audio data is extracted. It should be understood that the above extraction steps of the video watermark and the audio watermark may also be performed at the same time, or the audio watermark may be extracted first. The extracted watermark can use the relevance when the previous watermark is added to perform mutual verification of audio and video watermarks, so as to improve the anti-tampering ability of the watermark.
由上可知,上述关联性可以是时间上的关联,也可以是内容上的关联性。优选地,上述时间和内容上的关联性可以是彼此相同。为了提升鲁棒性,可以对待嵌入的视频文件进行按时间轴分段处理,相同时间点的视频序列和音频数据嵌入相同水印信息,以方便提取时相同时间点的视频水印信息和音频水印信息的相互验证。It can be seen from the above that the above-mentioned relevance can be a time relevance or a content relevance. Preferably, the above-mentioned relevance in time and content may be the same as each other. In order to improve the robustness, the video file to be embedded can be segmented according to the time axis. The video sequence and audio data at the same time point are embedded with the same watermark information to facilitate the extraction of the video watermark information and audio watermark information at the same time point. Mutual authentication.
为此,本申请的水印提取方法还可以包括根据提取的所述第一水印信息和所述第二水印信息,生成所述视频文件的提取水印。最终提取水印的生成,可以基于预先获取的音视频水印之间的相关性来确定。具体地,在确定第一水印信息和所述第二水印信息包括相同的水印数据的情况下,可以对所述第一水印信息和所述第二水印信息中各自包括的水印数据进行加权求和,生成所述视频文件的提取水印。具体地,可以根据置信度,调整所述第一水印信息和所述第二水印信息中各自包括的水印数据的权值。To this end, the watermark extraction method of the present application may further include generating the extracted watermark of the video file according to the extracted first watermark information and the second watermark information. The generation of the final extracted watermark can be determined based on the correlation between the audio and video watermarks obtained in advance. Specifically, in the case where it is determined that the first watermark information and the second watermark information include the same watermark data, a weighted summation may be performed on the watermark data included in each of the first watermark information and the second watermark information To generate the extracted watermark of the video file. Specifically, the weight of the watermark data included in each of the first watermark information and the second watermark information may be adjusted according to the degree of confidence.
如前所述,为了对视频的整个时长都加以保护,添加的第一水印信息可以包括含有第一排序数据和第一水印数据的多组水印数据,并且添加的第二水印信息则可包括含有第二排序数据和第二水印数据的多组水印数据。为此,提取所述视频数据中嵌入的第一水印信息可以包括:基于提取的第一排序数据,确定后续的第一水印数据,并且提取所述视频数据中嵌入的第二水印信息可以包括:基于提取的第二排序数据,确定后续的第二水印数据。换句话说,可以通过对更容易分辨出的排序数据的定位,来进一步定位水印信息的存在。例如,通过找出如图2所示的同步码“000……0001”,来定位其后续紧跟的单周期同步码。As mentioned above, in order to protect the entire duration of the video, the added first watermark information may include multiple sets of watermark data containing the first ranking data and the first watermark data, and the added second watermark information may include Multiple sets of watermark data of the second sorted data and the second watermark data. To this end, extracting the first watermark information embedded in the video data may include: determining subsequent first watermark data based on the extracted first ranking data, and extracting the second watermark information embedded in the video data may include: Based on the extracted second ranking data, the subsequent second watermark data is determined. In other words, the existence of watermark information can be further located by locating the sorted data that is easier to distinguish. For example, by finding the synchronization code "000...0001" as shown in FIG. 2 to locate the single-cycle synchronization code immediately following it.
具体在进行水印提取时,可以针对音视频数据的特定部分进行。为此,提取所述视频数据中嵌入的第一水印信息包括:确定所述视频数据中包含所述第一水印信息的视频帧和/或视频区域,并且提取所述音频数据中嵌入的第二水印信息包括:确定所述音频数据中包含所述第二水印信息的音频帧和/或音频区域。上述区域的确定可以是基于水印嵌入时的区域而反推。例如,选取关键视频帧的非显著区域,和/或选择相邻音频帧中的频带听觉不敏感区域。Specifically, when the watermark is extracted, it can be performed on a specific part of the audio and video data. To this end, extracting the first watermark information embedded in the video data includes: determining the video frame and/or video area that contains the first watermark information in the video data, and extracting the second watermark information embedded in the audio data. The watermark information includes: determining an audio frame and/or audio area in the audio data that contains the second watermark information. The determination of the above-mentioned area can be reversed based on the area when the watermark is embedded. For example, selecting a non-salient area of a key video frame, and/or selecting a frequency band auditory insensitive area in an adjacent audio frame.
水印比特的提取,则同样可以基于嵌入算法进行反推。在嵌入水印为“能量水印”时,可以从确定的一系列视频帧和/或视频区域中提取符合预定能量关系的所述第一水印信息的组成比特信息;并将提取的所述组成比特信息组合成所述第一水印信息。另外,对于音频水印,则可以从确定的一系列音频帧和/或音频区域中提取符合预定能量关系的所述第二水印信息的组成比特信息;以及将提取的所述组成比特信息组合成所述第二水印信息。The extraction of watermark bits can also be reversed based on the embedding algorithm. When the embedded watermark is an "energy watermark", the constituent bit information of the first watermark information that conforms to a predetermined energy relationship can be extracted from a series of determined video frames and/or video regions; and the extracted constituent bit information Combined into the first watermark information. In addition, for audio watermarks, the constituent bit information of the second watermark information that conforms to a predetermined energy relationship can be extracted from a series of determined audio frames and/or audio regions; and the extracted constituent bit information can be combined into all The second watermark information.
图5示出了根据本申请的联合水印提取流程的一个例子。图5的水印提取流程可以看作是图3的水印嵌入流程的对应操作。Figure 5 shows an example of a joint watermark extraction process according to the present application. The watermark extraction process in FIG. 5 can be regarded as the corresponding operation of the watermark embedding process in FIG. 3.
对于要进行水印提取的视频文件,可以首先利用视频文件解析工具分别抽取视频序列和音频数据。随后,如图5所示,流程分为两个并列的分支。For the video file to be watermarked, you can first use the video file analysis tool to extract the video sequence and audio data separately. Subsequently, as shown in Figure 5, the process is divided into two parallel branches.
对于视频分支,可以提取视频关键帧,并进行内容分析选取非显著区域。随后,对视频序列基于相邻块的能量关系提取同步码和水印比特,得到水印信息wm1。在此,水印信息wm1可以看作是图2示例中向视频序列添加的单周期水印信息。For video branches, video key frames can be extracted, and content analysis can be performed to select non-salient areas. Subsequently, the synchronization code and watermark bits are extracted from the video sequence based on the energy relationship of the neighboring blocks, and the watermark information wm1 is obtained. Here, the watermark information wm1 can be regarded as the single-period watermark information added to the video sequence in the example in FIG. 2.
对于音频分支,可以对音频进行分帧处理,并选取每帧音频中频带听觉不敏感的区域。随后,对音频基于相邻频带的能量关系提取同步码和水印比特,得到水印信息wm2。在此,水印信息wm1可以看作是图2示例中向音频码流添加的单周期水印信息。For the audio branch, the audio can be divided into frames, and an area that is insensitive to hearing in the frequency band of each frame of audio can be selected. Subsequently, the synchronization code and watermark bits are extracted from the audio based on the energy relationship of the adjacent frequency bands, and the watermark information wm2 is obtained. Here, the watermark information wm1 can be regarded as the single-period watermark information added to the audio code stream in the example in FIG. 2.
可以利用同步码将音频水印信息和视频水印进行对齐操作,将对齐后两个来源的水印信息进行加权相加,权重a1、a2根据当前时间轴位置能否成功提取水印进行自适应调整,当视频水印攻击相对强时a1变小,当音频水印攻击相对强时a2变小,除了音视频水印同时提取失败外,保持a1+a2=1,wm=wm1*a1+wm2*a2,最终得到水印信息wm。The synchronization code can be used to align the audio watermark information and the video watermark, and the watermark information of the two sources after the alignment is weighted and added. The weights a1 and a2 are adaptively adjusted according to whether the current time axis position can successfully extract the watermark. When the video When the watermark attack is relatively strong, a1 becomes smaller, and when the audio watermark attack is relatively strong, a2 becomes smaller. Except for the failure to extract audio and video watermarks at the same time, keep a1+a2=1, wm=wm1*a1+wm2*a2, and finally get the watermark information wm.
图6示出了本申请水印嵌入和提取操作的简要方案示意图。如方案示意图所示,对待保护的视频文件进行按时间轴分段处理,相同时间点的视频序列和音频数据嵌入相同水印信息。为简化原理描述,图中将水印信息示出为在每个分段开始时刻嵌入的0或1。而在更为实际的操作中,考虑到同样时长的视频序列和音频码流,存在嵌入容量差异,即可能只需要7秒时长视频序列可完成嵌入64比特,而可能需要10秒时长音频码流才完成嵌入64比特,需要考虑嵌入提取时音视频水印设置同步。为此,可以如图2所示,在初始时间相同的持续时间段内,根据水印算法,嵌入更长位数的同步码和水印信息。Figure 6 shows a schematic diagram of a brief scheme of the watermark embedding and extraction operations of this application. As shown in the schematic diagram of the scheme, the video file to be protected is segmented according to the time axis, and the video sequence and audio data at the same time point are embedded with the same watermark information. To simplify the principle description, the figure shows the watermark information as 0 or 1 embedded at the beginning of each segment. In a more practical operation, considering the video sequence and audio code stream of the same duration, there is a difference in embedding capacity, that is, it may only take 7 seconds for a video sequence to complete the embedding of 64 bits, while it may take 10 seconds for an audio code stream. To complete the 64-bit embedding, it is necessary to consider the synchronization of audio and video watermark settings during embedding and extraction. To this end, as shown in Figure 2, within the same duration of the initial time, according to the watermark algorithm, a longer-digit synchronization code and watermark information can be embedded.
在视频文件遭受恶意编辑攻击后,需要证明该视频的版权信息时,可以进行水印提取操作。如图6右侧所示,可以基于相同的时间轴分段提取出在前嵌入的视频水印wm1和音频水印wm2,并通过加权求和wm=wm1*a1+wm2*a2来获取最终提取的水印信息。After a video file has been attacked by malicious editing, the watermark extraction operation can be performed when the copyright information of the video needs to be proved. As shown on the right side of Figure 6, the previously embedded video watermark wm1 and audio watermark wm2 can be extracted based on the same time axis segment, and the final extracted watermark can be obtained by weighted summation wm=wm1*a1+wm2*a2 information.
如下将结合一个具体实施例来描述水印嵌入和关联提取方案。A是一个待嵌入水印的视频文件(例如影视作品)。为了保护A的版权,可以为其嵌入联合水印。首先,利用视频文件解析工具对待嵌入水印的视频文件分别抽取视频序列和音频数据。随后,在视频分支,可以提取视频关键帧,并基于密钥及内容分析选取嵌入区域。选取一种视频水印算法对视频序列(即,上一步骤中选取的嵌入区域)嵌入同步码和水印比特,得到含水印信息的视频序列。在音频分支,则可对音频进行分帧处理,并基于密钥选取每帧音频中嵌入区域。选取一种音频水印算法对音频嵌入同步码和水印比特,得到含水印信息的音频。最后,将含有相同水印信息的视频序列与音频数据进行码流混合,得到含双水印的视频文件。The watermark embedding and association extraction scheme will be described below in conjunction with a specific embodiment. A is a video file to be embedded with a watermark (such as a film and television work). In order to protect the copyright of A, a joint watermark can be embedded for it. First, the video file analysis tool is used to extract the video sequence and audio data separately from the video file to be embedded with the watermark. Subsequently, in the video branch, video key frames can be extracted, and the embedding area can be selected based on the key and content analysis. A video watermarking algorithm is selected to embed the synchronization code and watermark bits in the video sequence (that is, the embedding area selected in the previous step) to obtain a video sequence with watermarked information. In the audio branch, the audio can be divided into frames, and the embedded area in each frame of audio can be selected based on the key. Select an audio watermarking algorithm to embed the synchronization code and watermark bits in the audio to obtain the audio with watermarked information. Finally, the video sequence containing the same watermark information is mixed with the audio data to obtain a video file containing double watermarks.
在后续需要验证A的版权时,进行的联合水印的提取过程则可包括与如上嵌入过程相对应的操作。首先,同样可以利用视频文件解析工具对含水印的视频文件分别抽取视频序列和音频数据。随后,在视频分支,可以提取视频关键帧,并基于密钥及内容分析选取提取区域。选取对应视频水印算法对视频序列提取同步码和水印比特w1。在音频分支,则可对音频进行分帧处理,并基于密钥选取每帧音频的提取区域。选取对应音频水印算法对音频提取同步码和水印比特wm2。最后,可以利用同步码将音频水印信息和视频水印进行对齐操作,将对齐后两个来源的水印信息进行加权相加,权重a1、a2根据当 前时间轴位置包含音视频水印的置信度进行自适应调整。When the copyright of A needs to be verified later, the extraction process of the joint watermark may include operations corresponding to the above embedding process. First, the video file analysis tool can also be used to extract video sequences and audio data from watermarked video files. Subsequently, in the video branch, video key frames can be extracted, and the extraction area can be selected based on the key and content analysis. Select the corresponding video watermark algorithm to extract the synchronization code and watermark bit w1 from the video sequence. In the audio branch, the audio can be divided into frames, and the extraction area of each frame of audio can be selected based on the key. Select the corresponding audio watermark algorithm to extract the synchronization code and watermark bit wm2 from the audio. Finally, the synchronization code can be used to align the audio watermark information and the video watermark, and the watermark information of the two sources after the alignment is weighted and added, and the weights a1 and a2 are adaptive according to the confidence that the current time axis position contains the audio and video watermark. Adjustment.
另外,在无法同时提取音视频水印的情况下(例如,替换了音频或视频),也可以根据单独的音频或视频提取水印进行版权验证。In addition, in the case where the audio and video watermark cannot be extracted at the same time (for example, audio or video is replaced), copyright verification can also be performed based on a separate audio or video extraction watermark.
根据本申请图1至图3所述实施例生成的添加了音视频联合水印的视频文件可以被直接(或是经过第三方篡改后)发布至视频网站上。视频文件被发布至视频网站上之后,可以以流媒体的方式被网页访问者获取。在此,流媒体是指将一连串的多媒体数据经过互联网分段发送(通常经压缩的数据),在互联网上即时传输影音以供观赏的一种技术与过程。流媒体数据则可以指经互联网分段发送的多媒体数据。此技术使得数据数据包得以像流水一样发送并被持续观看,而无需在使用前下载整个媒体文件。The video file with the audio and video combined watermark added according to the embodiments described in Figs. 1 to 3 of the present application can be published directly (or after being tampered by a third party) on the video website. After the video files are published on the video website, they can be obtained by web visitors in the form of streaming media. Here, streaming media refers to a technology and process in which a series of multimedia data is sent in segments via the Internet (usually compressed data), and video and audio are instantly transmitted on the Internet for viewing. Streaming media data can refer to multimedia data sent in segments via the Internet. This technology enables data packets to be sent and watched continuously, without the need to download the entire media file before use.
在进行水印提取时,可以针对完整的视频文件进行水印提取,也可以对流媒体形式的视频流进行水印提取。为此,本申请还可以实现为一种流媒体水印提取方法,包括:获取嵌入水印的流媒体数据,所述流媒体数据由如上述的嵌入水印的视频文件生成;从所述嵌入水印的流媒体数据中抽取视频数据和音频数据;提取所述视频数据中嵌入的第一水印信息;以及提取所述音频数据中嵌入的第二水印信息。When performing watermark extraction, watermark extraction can be performed on complete video files, or watermark extraction can be performed on video streams in the form of streaming media. To this end, the present application can also be implemented as a streaming media watermark extraction method, including: acquiring watermark-embedded streaming media data, the streaming media data being generated from the above-mentioned watermark-embedded video file; from the watermark-embedded stream Extracting video data and audio data from the media data; extracting the first watermark information embedded in the video data; and extracting the second watermark information embedded in the audio data.
在一些实施例中,可以累积获取一定时长的流媒体数据(例如,30秒或1分钟),并对着一定时长内的流媒体数据(可以看作是视频数据片段)进行音视频数据抽取以及第一和第二水印信息的提取。在其他实施例中,也可以对实时获取的流媒体数据进行音视频数据的流式抽取以及第一和第二水印信息的流式提取。本申请对此不做限制。In some embodiments, streaming media data of a certain length of time (for example, 30 seconds or 1 minute) can be acquired cumulatively, and audio and video data can be extracted from the streaming media data (which can be regarded as video data fragments) within a certain length of time, and Extraction of the first and second watermark information. In other embodiments, it is also possible to perform streaming extraction of audio and video data and streaming extraction of the first and second watermark information on streaming media data acquired in real time. This application does not impose restrictions on this.
进一步地,在更为广义的实施例中,水印的嵌入也可以针对流媒体形式的视频文件进行。为此,本申请还可以实现为一种流媒体数据处理方法,包括:获取所述流媒体数据的视频数据和音频数据;向所述视频数据中嵌入第一水印信息;向所述音频数据中关联嵌入第二水印信息;以及获取嵌入水印的流媒体数据。类似地,上述操作可以是针对累积获取的一定时长的流媒体数据(可以看作是视频数据片段)进行,也可以针对实时获取的流媒体数据进行流式嵌入,本申请对此不做限制。Further, in a more generalized embodiment, the embedding of the watermark can also be performed on video files in the form of streaming media. To this end, this application can also be implemented as a streaming media data processing method, including: acquiring video data and audio data of the streaming media data; embedding first watermark information into the video data; Correlating and embedding the second watermark information; and obtaining the watermark-embedded streaming media data. Similarly, the foregoing operations can be performed on accumulated streaming media data (which can be regarded as video data fragments) of a certain length of time, or streaming media data obtained in real time can be embedded, which is not limited in this application.
传统视频水印方案在下采样重压缩的鲁棒性非常不理想,因为下采样重压缩导致视频关键帧、分块、局部内容等发生剧烈变化,导致对应位置的视频水印信息无法提取。本方案通过在音频相同时间轴位置嵌入相同水印信息,形成互补,再提取水印阶段进行视频水印与音频水印加权相加,这样成功避免传统视频水印在下采样重压缩后容易失效的缺点。The robustness of traditional video watermarking schemes in down-sampling and recompression is very unsatisfactory, because down-sampling and re-compression leads to drastic changes in video key frames, blocks, local content, etc., resulting in the inability to extract the video watermark information at the corresponding position. This solution embeds the same watermark information in the same time axis position of the audio to form a complement, and then performs the weighted addition of the video watermark and the audio watermark in the watermark stage, so as to successfully avoid the disadvantage that traditional video watermarks are easy to fail after downsampling and recompression.
此外,本申请还可以实现为一种视频文件处理装置,包括:视频解析单元,用于获 取所述视频文件中的视频数据和音频数据;视频水印嵌入单元,用于向所述视频数据中嵌入第一水印信息;音频水印嵌入单元,用于向所述音频数据中关联嵌入第二水印信息;以及视频混合单元,用于混合嵌入第一水印信息的所述视频数据和嵌入第二水印信息的所述音频数据,以获取嵌入水印的视频文件。在一个实施例中,视频解析单元可以针对流媒体形式的视频文件进行解析,例如,实时解析。In addition, this application can also be implemented as a video file processing device, including: a video parsing unit for obtaining video data and audio data in the video file; a video watermark embedding unit for embedding in the video data The first watermark information; an audio watermark embedding unit for embedding the second watermark information into the audio data; and a video mixing unit for mixing the video data embedded in the first watermark information and the video data embedded in the second watermark information The audio data is used to obtain a video file embedded with a watermark. In one embodiment, the video analysis unit may analyze the video file in the form of streaming media, for example, real-time analysis.
相应地,本申请还可以实现为一种视频水印提取装置,包括:视频解析单元,用于获取如上所述的嵌入水印的视频文件中的视频数据和音频数据;视频水印提取单元,用于从所述视频数据中提取第一水印信息;以及音频水印提取单元,用于从所述音频数据中提取嵌入第二水印信息。该装置还可以包括:水印信息生成单元,用于对所述第一水印信息和所述第二水印信息中各自包括的水印数据进行加权求和,生成所述视频文件的提取水印。类似地,在一个实施例中,视频解析单元可以针对流媒体形式的视频文件进行解析,例如,实时解析。Correspondingly, the present application can also be implemented as a video watermark extraction device, including: a video analysis unit for acquiring the video data and audio data in the video file embedded with the watermark as described above; and a video watermark extraction unit for obtaining Extracting the first watermark information from the video data; and an audio watermark extraction unit for extracting and embedding the second watermark information from the audio data. The device may further include: a watermark information generating unit, configured to perform a weighted summation of the watermark data included in each of the first watermark information and the second watermark information to generate the extracted watermark of the video file. Similarly, in one embodiment, the video parsing unit may analyze video files in the form of streaming media, for example, real-time parsing.
图7示出了根据本申请一实施例可用于实现上述视频处理和水印提取方法的计算设备的结构示意图。Fig. 7 shows a schematic structural diagram of a computing device that can be used to implement the above video processing and watermark extraction method according to an embodiment of the present application.
参见图7,计算设备700包括存储器710和处理器720。Referring to FIG. 7, the computing device 700 includes a memory 710 and a processor 720.
处理器720可以是一个多核的处理器,也可以包含多个处理器。在一些实施例中,处理器720可以包含一个通用的主处理器以及一个或多个特殊的协处理器,例如图形处理器(GPU)、数字信号处理器(DSP)等等。在一些实施例中,处理器720可以使用定制的电路实现,例如特定用途集成电路(ASIC,Application Specific Integrated Circuit)或者现场可编程逻辑门阵列(FPGA,Field Programmable Gate Arrays)。The processor 720 may be a multi-core processor, or may include multiple processors. In some embodiments, the processor 720 may include a general-purpose main processor and one or more special co-processors, such as a graphics processing unit (GPU), a digital signal processor (DSP), and so on. In some embodiments, the processor 720 may be implemented using a customized circuit, such as an Application Specific Integrated Circuit (ASIC) or a Field Programmable Gate Array (FPGA, Field Programmable Gate Arrays).
存储器710可以包括各种类型的存储单元,例如系统内存、只读存储器(ROM),和永久存储装置。其中,ROM可以存储处理器720或者计算机的其他模块需要的静态数据或者指令。永久存储装置可以是可读写的存储装置。永久存储装置可以是即使计算机断电后也不会失去存储的指令和数据的非易失性存储设备。在一些实施方式中,永久性存储装置采用大容量存储装置(例如磁或光盘、闪存)作为永久存储装置。另外一些实施方式中,永久性存储装置可以是可移除的存储设备(例如软盘、光驱)。系统内存可以是可读写存储设备或者易失性可读写存储设备,例如动态随机访问内存。系统内存可以存储一些或者所有处理器在运行时需要的指令和数据。此外,存储器710可以包括任意计算机可读存储媒介的组合,包括各种类型的半导体存储芯片(DRAM,SRAM,SDRAM,闪存,可编程只读存储器),磁盘和/或光盘也可以采用。在一些实施方式中, 存储器710可以包括可读和/或写的可移除的存储设备,例如激光唱片(CD)、只读数字多功能光盘(例如DVD-ROM,双层DVD-ROM)、只读蓝光光盘、超密度光盘、闪存卡(例如SD卡、min SD卡、Micro-SD卡等等)、磁性软盘等等。计算机可读存储媒介不包含载波和通过无线或有线传输的瞬间电子信号。The memory 710 may include various types of storage units, such as a system memory, a read only memory (ROM), and a permanent storage device. Among them, the ROM may store static data or instructions required by the processor 720 or other modules of the computer. The permanent storage device may be a readable and writable storage device. The permanent storage device may be a non-volatile storage device that does not lose stored instructions and data even after the computer is powered off. In some embodiments, the permanent storage device adopts a large-capacity storage device (such as a magnetic or optical disk, flash memory) as the permanent storage device. In other embodiments, the permanent storage device may be a removable storage device (for example, a floppy disk, an optical drive). The system memory can be a readable and writable storage device or a volatile readable and writable storage device, such as dynamic random access memory. The system memory can store some or all of the instructions and data needed by the processor at runtime. In addition, the memory 710 may include any combination of computer-readable storage media, including various types of semiconductor memory chips (DRAM, SRAM, SDRAM, flash memory, programmable read-only memory), and magnetic disks and/or optical disks may also be used. In some embodiments, the memory 710 may include a removable storage device that can be read and/or written, such as a compact disc (CD), a read-only digital versatile disc (for example, DVD-ROM, double-layer DVD-ROM), Read-only Blu-ray discs, ultra-density discs, flash memory cards (such as SD cards, min SD cards, Micro-SD cards, etc.), magnetic floppy disks, etc. The computer-readable storage medium does not include carrier waves and instantaneous electronic signals transmitted wirelessly or wiredly.
存储器710上存储有可执行代码,当可执行代码被处理器720处理时,可以使处理器720执行上文述及的视频处理以及水印提取方法。The memory 710 stores executable codes. When the executable codes are processed by the processor 720, the processor 720 can be made to execute the video processing and watermark extraction methods described above.
上文中已经参考附图详细描述了根据本申请的视频处理以及水印频提方法和装置。本申请对视频文件同时嵌入音频水印和视频水印,两者互不干扰,相互补充,提取水印时可将音视频水印提取信息进行自适应融合,从而大幅提高视频文件水印的鲁棒性,特别是对抗针对视频内容的恶意编辑攻击。具体地,基于音视频双水印融合的自适应思想,音频水印和视频水印的权重可以根据各自可靠性动态调整,保证融合后水印的可靠性。另外,针对相同内容的音频水印和视频水印存在不同步情况,可以利用分段同步码方式实现双水印同步。The method and device for video processing and watermark extraction according to the present application have been described in detail above with reference to the accompanying drawings. This application embeds audio watermark and video watermark in the video file at the same time. The two do not interfere with each other and complement each other. When extracting the watermark, the audio and video watermark extraction information can be adaptively fused, thereby greatly improving the robustness of the video file watermark, especially Fight against malicious editing attacks on video content. Specifically, based on the adaptive idea of fusion of audio and video dual watermarks, the weights of audio watermarks and video watermarks can be dynamically adjusted according to their respective reliability to ensure the reliability of the fusion watermark. In addition, for the audio watermark and video watermark of the same content are out of synchronization, the segmented synchronization code method can be used to achieve dual watermark synchronization.
此外,根据本申请的方法还可以实现为一种计算机程序或计算机程序产品,该计算机程序或计算机程序产品包括用于执行本申请的上述方法中限定的上述各步骤的计算机程序代码指令。In addition, the method according to the present application can also be implemented as a computer program or computer program product, and the computer program or computer program product includes computer program code instructions for executing the above-mentioned steps defined in the above-mentioned method of the present application.
或者,本申请还可以实施为一种非暂时性机器可读存储介质(或计算机可读存储介质、或机器可读存储介质),其上存储有可执行代码(或计算机程序、或计算机指令代码),当所述可执行代码(或计算机程序、或计算机指令代码)被电子设备(或计算设备、服务器等)的处理器执行时,使所述处理器执行根据本申请的上述方法的各个步骤。Alternatively, this application can also be implemented as a non-transitory machine-readable storage medium (or computer-readable storage medium, or machine-readable storage medium) on which executable code (or computer program, or computer instruction code) is stored ), when the executable code (or computer program, or computer instruction code) is executed by the processor of the electronic device (or computing device, server, etc.), the processor is made to execute each step of the above-mentioned method according to the present application .
本领域技术人员还将明白的是,结合这里的公开所描述的各种示例性逻辑块、模块、电路和算法步骤可以被实现为电子硬件、计算机软件或两者的组合。Those skilled in the art will also understand that the various exemplary logic blocks, modules, circuits, and algorithm steps described in conjunction with the disclosure herein can be implemented as electronic hardware, computer software, or a combination of both.
附图中的流程图和框图显示了根据本申请的多个实施例的系统和方法的可能实现的体系架构、功能和操作。在这点上,流程图或框图中的每个方框可以代表一个模块、程序段或代码的一部分,所述模块、程序段或代码的一部分包含一个或多个用于实现规定的逻辑功能的可执行指令。也应当注意,在有些作为替换的实现中,方框中所标记的功能也可以以不同于附图中所标记的顺序发生。例如,两个连续的方框实际上可以基本并行地执行,它们有时也可以按相反的顺序执行,这依所涉及的功能而定。也要注意的是,框图和/或流程图中的每个方框、以及框图和/或流程图中的方框的组合,可以用执行规定的功能或操作的专用的基于硬件的系统来实现,或者可以用专用硬件与计算机指令的组 合来实现。The flowcharts and block diagrams in the accompanying drawings show the possible implementation of the system architecture, functions, and operations of the system and method according to multiple embodiments of the present application. In this regard, each block in the flowchart or block diagram may represent a module, program segment, or part of the code, and the module, program segment, or part of the code contains one or more functions for realizing the specified logical function. Executable instructions. It should also be noted that in some alternative implementations, the functions marked in the block may also occur in a different order than marked in the drawings. For example, two consecutive blocks can actually be executed substantially in parallel, or they can sometimes be executed in the reverse order, depending on the functions involved. It should also be noted that each block in the block diagram and/or flowchart, and the combination of the blocks in the block diagram and/or flowchart, can be implemented by a dedicated hardware-based system that performs the specified functions or operations Or it can be realized by a combination of dedicated hardware and computer instructions.
以上已经描述了本申请的各实施例,上述说明是示例性的,并非穷尽性的,并且也不限于所披露的各实施例。在不偏离所说明的各实施例的范围和精神的情况下,对于本技术领域的普通技术人员来说许多修改和变更都是显而易见的。本文中所用术语的选择,旨在最好地解释各实施例的原理、实际应用或对市场中的技术的改进,或者使本技术领域的其它普通技术人员能理解本文披露的各实施例。The embodiments of the present application have been described above, and the above description is exemplary, not exhaustive, and is not limited to the disclosed embodiments. Without departing from the scope and spirit of the illustrated embodiments, many modifications and changes are obvious to those of ordinary skill in the art. The choice of terms used herein is intended to best explain the principles, practical applications, or improvements to technologies in the market of the embodiments, or to enable other ordinary skilled in the art to understand the embodiments disclosed herein.

Claims (28)

  1. 一种视频文件处理方法,包括:A video file processing method, including:
    获取所述视频文件的视频数据和音频数据;Acquiring video data and audio data of the video file;
    向所述视频数据中嵌入第一水印信息;Embedding the first watermark information into the video data;
    向所述音频数据中关联嵌入第二水印信息;以及Associatively embedding the second watermark information into the audio data; and
    获取嵌入水印的视频文件。Get the video file with embedded watermark.
  2. 如权利要求1所述的方法,其中,向所述音频数据中关联嵌入第二水印信息包括如下至少一项:The method according to claim 1, wherein associating and embedding the second watermark information into the audio data includes at least one of the following:
    向所述音频数据嵌入第二水印信息的时间与向所述视频数据嵌入第一水印信息的时间相关联;以及The time when the second watermark information is embedded in the audio data is associated with the time when the first watermark information is embedded in the video data; and
    向所述音频数据嵌入第二水印信息的内容与向所述视频数据嵌入第一水印信息的内容相关联。The content of the second watermark information embedded in the audio data is associated with the content of the first watermark information embedded in the video data.
  3. 如权利要求2所述的方法,其中,向所述音频数据中嵌入关联的第二水印信息包括如下至少一项:The method according to claim 2, wherein embedding the associated second watermark information into the audio data includes at least one of the following:
    向所述音频数据嵌入第二水印信息的起始时间与向所述视频数据嵌入第一水印信息的起始时间相同;以及The start time of embedding the second watermark information into the audio data is the same as the start time of embedding the first watermark information into the video data; and
    向所述音频数据嵌入第二水印信息的内容与向所述视频数据嵌入第一水印信息的内容相同。The content of the second watermark information embedded in the audio data is the same as the content of the first watermark information embedded in the video data.
  4. 如权利要求1所述的方法,其中,向所述视频数据中嵌入第一水印信息包括:The method of claim 1, wherein embedding the first watermark information into the video data comprises:
    以第一预定时间间隔,向所述视频数据中嵌入多个第一水印信息,并且Embed a plurality of first watermark information into the video data at a first predetermined time interval, and
    向所述音频数据中嵌入关联的第二水印信息包括:Embedding the associated second watermark information into the audio data includes:
    以第二预定时间间隔,向所述音频数据中嵌入多个第二水印信息。Embed a plurality of second watermark information into the audio data at a second predetermined time interval.
  5. 如权利要求4所述的方法,其中,所述多个第一水印信息中的每个第一水印信息包括:The method according to claim 4, wherein each first watermark information in the plurality of first watermark information comprises:
    第一排序数据和第一水印数据,并且The first sorted data and the first watermark data, and
    所述多个第二水印信息中的每个第二水印信息包括:Each second watermark information in the plurality of second watermark information includes:
    第二排序数据和第二水印数据。The second ranking data and the second watermark data.
  6. 如权利要求1所述的方法,其中,向所述视频数据中嵌入第一水印信息包括:The method of claim 1, wherein embedding the first watermark information into the video data comprises:
    向所述视频数据中视频帧的非显著区域添加所述第一水印信息。Adding the first watermark information to an insignificant area of a video frame in the video data.
  7. 如权利要求1所述的方法,其中,向所述音频数据中嵌入第二水印信息包括:The method of claim 1, wherein embedding second watermark information into the audio data comprises:
    向所述音频数据中音频帧的听觉不敏感区域添加所述第二水印信息。Adding the second watermark information to the auditory insensitive area of the audio frame in the audio data.
  8. 如权利要求1所述的方法,其中,向所述视频数据中嵌入第一水印信息包括:The method of claim 1, wherein embedding the first watermark information into the video data comprises:
    通过调整所述视频数据中视频帧变换域相邻区域的能量关系,嵌入所述第一水印信息。The first watermark information is embedded by adjusting the energy relationship between adjacent regions of the video frame transform domain in the video data.
  9. 如权利要求8所述的方法,其中,通过调整所述视频数据中视频帧变换域相邻区域的能量关系,嵌入所述第一水印信息包括:8. The method of claim 8, wherein by adjusting the energy relationship between adjacent regions of the video frame transform domain in the video data, embedding the first watermark information comprises:
    选取视频序列中的一系列特定视频帧,所述视频数据是视频序列;Selecting a series of specific video frames in a video sequence, where the video data is a video sequence;
    调整所述一系列特定视频帧的变换域相邻区域的能量关系,并逐一嵌入所述第一水印信息的组成比特信息。Adjusting the energy relationship between adjacent regions in the transform domain of the series of specific video frames, and embedding the constituent bit information of the first watermark information one by one.
  10. 如权利要求9所述的方法,其中,选取视频序列中的一系列特定视频帧包括:9. The method of claim 9, wherein selecting a series of specific video frames in the video sequence comprises:
    选取所述视频序列中的视频关键帧。Select video key frames in the video sequence.
  11. 如权利要求1所述的方法,其中,向所述音频数据中嵌入关联的第二水印信息包括:The method of claim 1, wherein embedding the associated second watermark information into the audio data comprises:
    调整相邻音频帧的能量关系,嵌入所述第二水印信息。Adjust the energy relationship of adjacent audio frames, and embed the second watermark information.
  12. 如权利要求11所述的方法,其中,调整相邻音频帧的能量关系,嵌入所述第二水印信息包括:The method of claim 11, wherein adjusting the energy relationship of adjacent audio frames and embedding the second watermark information comprises:
    调整一系列相邻音频帧的相邻频带的能量关系,并逐一嵌入所述第二水印信息的组成比特信息。Adjust the energy relationship of adjacent frequency bands of a series of adjacent audio frames, and embed the constituent bit information of the second watermark information one by one.
  13. 如权利要求1所述的方法,其中,向所述视频数据中嵌入第一水印信息包括:The method of claim 1, wherein embedding the first watermark information into the video data comprises:
    基于第一密码选取所述视频数据的视频帧中用于添加所述第一水印信息的第一添加区域,和/或Selecting the first adding area for adding the first watermark information in the video frame of the video data based on the first password, and/or
    向所述音频数据中嵌入关联的第二水印信息包括:Embedding the associated second watermark information into the audio data includes:
    基于第二密码选取所述音频数据的音频帧中用于添加所述第二水印信息的第二添加区域。The second adding area used for adding the second watermark information in the audio frame of the audio data is selected based on the second password.
  14. 一种视频水印提取方法,包括:A video watermark extraction method, including:
    获取如权利要求1-13中任一项所述的嵌入水印的视频文件;Obtain the watermark-embedded video file according to any one of claims 1-13;
    从所述嵌入水印的视频文件中抽取视频数据和音频数据;Extracting video data and audio data from the watermark-embedded video file;
    提取所述视频数据中嵌入的第一水印信息;以及Extracting the first watermark information embedded in the video data; and
    提取所述音频数据中嵌入的第二水印信息。Extract the second watermark information embedded in the audio data.
  15. 如权利要求14所述的方法,还包括:The method of claim 14, further comprising:
    根据提取的所述第一水印信息和所述第二水印信息,生成所述视频文件的提取水印。According to the extracted first watermark information and the second watermark information, an extracted watermark of the video file is generated.
  16. 如权利要求15所述的方法,其中,根据提取的所述第一水印信息和所述第二水印信息,生成所述视频文件的提取水印包括:The method according to claim 15, wherein generating the extracted watermark of the video file according to the extracted first watermark information and the second watermark information comprises:
    确定第一水印信息和所述第二水印信息包括相同的水印数据;以及Determining that the first watermark information and the second watermark information include the same watermark data; and
    对所述第一水印信息和所述第二水印信息中各自包括的水印数据进行加权求和,生成所述视频文件的提取水印。A weighted summation is performed on the watermark data included in each of the first watermark information and the second watermark information to generate an extracted watermark of the video file.
  17. 如权利要求16所述的方法,其中,根据提取的所述第一水印信息和所述第二水印信息,生成所述视频文件的提取水印还包括:The method of claim 16, wherein, according to the extracted first watermark information and the second watermark information, generating the extracted watermark of the video file further comprises:
    根据置信度,调整所述第一水印信息和所述第二水印信息中各自包括的水印数据的权值。According to the degree of confidence, the weight of the watermark data included in the first watermark information and the second watermark information is adjusted.
  18. 如权利要求16所述的方法,其中,所述第一水印信息包括含有第一排序数据和第一水印数据的多组水印数据,所述第二水印信息包括含有第二排序数据和第二水印数据的多组水印数据,The method according to claim 16, wherein the first watermark information includes a plurality of sets of watermark data including a first ranking data and a first watermark data, and the second watermark information includes a plurality of sets of watermark data including a second ranking data and a second watermark. Multiple sets of watermark data of the data,
    其中,提取所述视频数据中嵌入的第一水印信息包括:Wherein, extracting the first watermark information embedded in the video data includes:
    基于提取的第一排序数据,确定后续的第一水印数据,并且Based on the extracted first ranking data, determine the subsequent first watermark data, and
    提取所述视频数据中嵌入的第二水印信息包括:Extracting the second watermark information embedded in the video data includes:
    基于提取的第二排序数据,确定后续的第二水印数据。Based on the extracted second ranking data, the subsequent second watermark data is determined.
  19. 如权利要求14所述的方法,其中,提取所述视频数据中嵌入的第一水印信息包括:The method of claim 14, wherein extracting the first watermark information embedded in the video data comprises:
    确定所述视频数据中包含所述第一水印信息的视频帧和/或视频区域,并且Determine the video frame and/or video area in the video data containing the first watermark information, and
    提取所述音频数据中嵌入的第二水印信息包括:Extracting the second watermark information embedded in the audio data includes:
    确定所述音频数据中包含所述第二水印信息的音频帧和/或音频区域。Determine the audio frame and/or audio area in the audio data that contains the second watermark information.
  20. 如权利要求19所述的方法,其中,提取所述视频数据中嵌入的第一水印信息还包括:The method of claim 19, wherein extracting the first watermark information embedded in the video data further comprises:
    从确定的一系列视频帧和/或视频区域中提取符合预定能量关系的所述第一水印信息的组成比特信息;Extracting constituent bit information of the first watermark information that conforms to a predetermined energy relationship from a determined series of video frames and/or video regions;
    将提取的所述组成比特信息组合成所述第一水印信息,以及Combining the extracted constituent bit information into the first watermark information, and
    提取所述音频数据中嵌入的第二水印信息还包括:Extracting the second watermark information embedded in the audio data further includes:
    从确定的一系列音频帧和/或音频区域中提取符合预定能量关系的所述第二水印信 息的组成比特信息;以及Extracting constituent bit information of the second watermark information that conforms to a predetermined energy relationship from a determined series of audio frames and/or audio regions; and
    将提取的所述组成比特信息组合成所述第二水印信息。Combining the extracted constituent bit information into the second watermark information.
  21. 如权利要求20所述的方法,其中,所述视频帧和/或视频区域基于如下至少一项确定:The method according to claim 20, wherein the video frame and/or video area are determined based on at least one of the following:
    所述视频帧和/或视频区域的内容;The content of the video frame and/or video area;
    所述视频帧和/或视频区域的选取密码,The selected password of the video frame and/or video area,
    并且,所述音频帧和/或音频区域基于如下至少一项确定:And, the audio frame and/or audio area are determined based on at least one of the following:
    所述音频帧和/或音频区域的频谱内容;The frequency spectrum content of the audio frame and/or audio region;
    所述音频帧和/或音频区域的选取密码。The selected password of the audio frame and/or audio area.
  22. 一种流媒体水印提取方法,包括:A streaming media watermark extraction method, including:
    获取嵌入水印的流媒体数据,所述流媒体数据由如权利要求1-13任一项所述的嵌入水印的视频文件生成;Acquiring watermark-embedded streaming media data, the streaming media data being generated from the watermark-embedded video file according to any one of claims 1-13;
    从所述嵌入水印的流媒体数据中抽取视频数据和音频数据;Extracting video data and audio data from the watermark-embedded streaming media data;
    提取所述视频数据中嵌入的第一水印信息;以及Extracting the first watermark information embedded in the video data; and
    提取所述音频数据中嵌入的第二水印信息。Extract the second watermark information embedded in the audio data.
  23. 一种流媒体数据处理方法,包括:A streaming media data processing method, including:
    获取所述流媒体数据的视频数据和音频数据;Acquiring video data and audio data of the streaming media data;
    向所述视频数据中嵌入第一水印信息;Embedding the first watermark information into the video data;
    向所述音频数据中关联嵌入第二水印信息;以及Associatively embedding the second watermark information into the audio data; and
    获取嵌入水印的流媒体数据。Get the streaming media data embedded with the watermark.
  24. 一种视频文件处理装置,包括:A video file processing device, including:
    视频解析单元,用于获取所述视频文件中的视频数据和音频数据;A video parsing unit for obtaining video data and audio data in the video file;
    视频水印嵌入单元,用于向所述视频数据中嵌入第一水印信息;A video watermark embedding unit, configured to embed the first watermark information into the video data;
    音频水印嵌入单元,用于向所述音频数据中关联嵌入第二水印信息;以及The audio watermark embedding unit is used to associate and embed the second watermark information into the audio data; and
    视频混合单元,用于混合嵌入第一水印信息的所述视频数据和嵌入第二水印信息的所述音频数据,以获取嵌入水印的视频文件。The video mixing unit is configured to mix the video data embedded with the first watermark information and the audio data embedded with the second watermark information to obtain a watermark-embedded video file.
  25. 一种视频水印提取装置,包括:A video watermark extraction device, including:
    视频解析单元,用于获取如权利要求1-13中任一项所述的嵌入水印的视频文件中的视频数据和音频数据;A video parsing unit, configured to obtain the video data and audio data in the watermarked video file according to any one of claims 1-13;
    视频水印提取单元,用于从所述视频数据中提取第一水印信息;以及The video watermark extraction unit is used to extract the first watermark information from the video data; and
    音频水印提取单元,用于从所述音频数据中提取嵌入第二水印信息。The audio watermark extraction unit is used to extract and embed the second watermark information from the audio data.
  26. 如权利要求25所述的装置,还包括:The device of claim 25, further comprising:
    水印信息生成单元,用于对所述第一水印信息和所述第二水印信息中各自包括的水印数据进行加权求和,生成所述视频文件的提取水印。The watermark information generating unit is configured to perform a weighted summation of the watermark data included in each of the first watermark information and the second watermark information to generate the extracted watermark of the video file.
  27. 一种计算设备,包括:A computing device including:
    处理器;以及Processor; and
    存储器,其上存储有可执行代码,当所述可执行代码被所述处理器执行时,使所述处理器执行如权利要求1-23中任一项所述的方法。The memory has executable code stored thereon, and when the executable code is executed by the processor, the processor is caused to execute the method according to any one of claims 1-23.
  28. 一种非暂时性机器可读存储介质,其上存储有可执行代码,当所述可执行代码被电子设备的处理器执行时,使所述处理器执行如权利要求1-23中任一项所述的方法。A non-transitory machine-readable storage medium with executable code stored thereon, and when the executable code is executed by a processor of an electronic device, the processor is caused to execute any one of claims 1-23 The method described.
PCT/CN2021/081259 2020-03-24 2021-03-17 Video file processing method and device, and watermark extraction method and device WO2021190372A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202010215301.1 2020-03-24
CN202010215301.1A CN113453039B (en) 2020-03-24 2020-03-24 Method and device for processing video file and extracting watermark

Publications (1)

Publication Number Publication Date
WO2021190372A1 true WO2021190372A1 (en) 2021-09-30

Family

ID=77807451

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2021/081259 WO2021190372A1 (en) 2020-03-24 2021-03-17 Video file processing method and device, and watermark extraction method and device

Country Status (2)

Country Link
CN (1) CN113453039B (en)
WO (1) WO2021190372A1 (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114826744A (en) * 2022-04-28 2022-07-29 中国银行股份有限公司 Information processing method, device, equipment and storage medium
TWI814427B (en) * 2022-06-07 2023-09-01 宏正自動科技股份有限公司 Method for synchronizing audio and video

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101261865A (en) * 2007-04-20 2008-09-10 炬力集成电路设计有限公司 Making method, device, playing device and method for media electronic file
CN104581202A (en) * 2013-10-25 2015-04-29 腾讯科技(北京)有限公司 Audio and video synchronization method and system, encoding device and decoding device
WO2017026714A1 (en) * 2015-08-07 2017-02-16 엘지전자 주식회사 Broadcast signal transmission device, broadcast signal reception device, broadcast signal transmission method, and broadcast signal reception method
CN106878827A (en) * 2017-03-22 2017-06-20 河海大学 A kind of high robust audio frequency and video intersect watermarking algorithm
CN109151157A (en) * 2017-06-28 2019-01-04 成都宇飞信息工程有限责任公司 A kind of multimedia digital watermark evidence obtaining mobile phone

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP5246804B2 (en) * 2006-07-18 2013-07-24 トムソン ライセンシング Method and system for time synchronization
WO2014014252A1 (en) * 2012-07-16 2014-01-23 Lg Electronics Inc. Method and apparatus for processing digital service signals
WO2015138798A1 (en) * 2014-03-13 2015-09-17 Verance Corporation Interactive content acquisition using embedded codes
CN107749990B (en) * 2017-09-27 2021-02-19 深圳大学 Video copyright protection method and device based on digital watermark

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101261865A (en) * 2007-04-20 2008-09-10 炬力集成电路设计有限公司 Making method, device, playing device and method for media electronic file
CN104581202A (en) * 2013-10-25 2015-04-29 腾讯科技(北京)有限公司 Audio and video synchronization method and system, encoding device and decoding device
WO2017026714A1 (en) * 2015-08-07 2017-02-16 엘지전자 주식회사 Broadcast signal transmission device, broadcast signal reception device, broadcast signal transmission method, and broadcast signal reception method
CN106878827A (en) * 2017-03-22 2017-06-20 河海大学 A kind of high robust audio frequency and video intersect watermarking algorithm
CN109151157A (en) * 2017-06-28 2019-01-04 成都宇飞信息工程有限责任公司 A kind of multimedia digital watermark evidence obtaining mobile phone

Also Published As

Publication number Publication date
CN113453039B (en) 2023-04-18
CN113453039A (en) 2021-09-28

Similar Documents

Publication Publication Date Title
US7224819B2 (en) Integrating digital watermarks in multimedia content
US8259938B2 (en) Efficient and secure forensic marking in compressed
US9323902B2 (en) Conditional access using embedded watermarks
EP1256086B1 (en) Methods and apparatus for multi-layer data hiding
US9547753B2 (en) Coordinated watermarking
US7643649B2 (en) Integrating digital watermarks in multimedia content
TWI462592B (en) Method and system for utilizing gps information to secure digital media
US20130151855A1 (en) Watermark embedding workflow improvements
US20170316189A1 (en) Object-based watermarking
WO2005011279A1 (en) Content identification for broadcast media
WO2021190372A1 (en) Video file processing method and device, and watermark extraction method and device
JP2010158024A (en) Fingerprinting of data object
US20100017614A1 (en) Encoding and detecting apparatus
US20100226425A1 (en) Encoding and detecting apparatus
US8315877B2 (en) Encoding and detecting apparatus
US8332954B2 (en) Methods for embedding data in digital audio data
Petrovic Audio watermarking in compressed domain
US20090185683A1 (en) Encoding and detecting apparatus
CN114067811A (en) Method, device and system for audio data processing and watermark extraction
Mani et al. Intelligent embedding and decoding scheme for real time audio processing

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 21774961

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 21774961

Country of ref document: EP

Kind code of ref document: A1