WO2005122586A1 - Compensating watermark irregularities caused by moved objects - Google Patents
Compensating watermark irregularities caused by moved objects Download PDFInfo
- Publication number
- WO2005122586A1 WO2005122586A1 PCT/IB2005/051767 IB2005051767W WO2005122586A1 WO 2005122586 A1 WO2005122586 A1 WO 2005122586A1 IB 2005051767 W IB2005051767 W IB 2005051767W WO 2005122586 A1 WO2005122586 A1 WO 2005122586A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- watermark
- coefficients
- additional data
- signal
- embedded
- Prior art date
Links
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/134—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
- H04N19/136—Incoming video signal characteristics or properties
- H04N19/137—Motion inside a coding unit, e.g. average field, frame or block difference
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/46—Embedding additional information in the video signal during the compression process
- H04N19/467—Embedding additional information in the video signal during the compression process characterised by the embedded information being invisible, e.g. watermarking
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/48—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using compressed domain processing techniques other than decoding, e.g. modification of transform coefficients, variable length coding [VLC] data or run-length data
Definitions
- the present invention generally relates to the field of watermarking of media signals, preferably video signals for instance coded according to the MPEG coding scheme. More particularly the present invention is directed towards a method, device and computer program product for determining additional data to be embedded in a media signal as well as a media signal processing device having such a device for determining additional data.
- a watermark is here normally a pseudo-random noise code that is inserted in the media signal. In the watermarking process it is necessary that the watermark is not perceptible. A watermark that is embedded in for mstance a video signal should then not be visible for an end user. It should however be possible to detect the watermark safely using a watermark detector, therefore the watermark should furthermore retain its structure throughout the signal.
- One known watermarking scheme for a video signal is described in WO- 02/060182. Here a watermark is embedded in an MPEG video signal.
- An MPEG signal is received and comprises VLC (Variable-Length Coding) coded quantised DCT (Discrete Cosine Transform) samples of a video stream divided into frames, where each frame includes a number of blocks of pixel information.
- VLC Very-Length Coding
- quantised DCT Discrete Cosine Transform
- a watermark is here embedded in the quantised DCT components of a block of size 8x8 under the use of a bit-rate controller, such that only the small DCT levels with ⁇ 1 are modified into a zero value. These values are furthermore only modified if the bit rate of the stream is not increased.
- this object is achieved by a method of determining additional data to be embedded in a media signal and comprising the steps of: obtaining, from a media signal divided into frames having blocks of a number of signal sample values, at least one motion vector of a current frame that is associated with a first block of signal samples, retrieving additional data embedded in a previous frame of said signal in dependence of the motion vector, determining additional data coefficients to be embedded in said signal based on the retrieved additional data and additional reference data, and embedding said additional data coefficients into said first block.
- a device for determining additional data to be embedded in a media signal comprising an embedding unit having: a motion compensating unit arranged to: obtain, from a media signal divided into frames having blocks of a number of signal sample values, at least one motion vector of a current frame that is associated with a first block of signal samples, retrieve additional data embedded in a previous frame of said signal in dependence of the motion vector, a determining unit arranged to determine additional data coefficients to be embedded in said signal based on said retrieved additional data and additional reference data, and a data embedding unit arranged to embed the said additional data coefficients into said first block.
- this object is also achieved by a media signal processing device comprising a device for determining additional data according to the second aspect.
- this object is also achieved by a computer program product for determining additional data to be embedded in a media signal, comprising computer program code, to make a computer do, when said program is loaded in the computer: obtain, from a media signal divided into frames having blocks of a number of signal sample values, at least one motion vector of a current frame that is associated with a first block of signal samples, retrieve additional data embedded in a previous frame of said signal in dependence of the motion vector, and determine additional data coefficients to be embedded in said signal based on said retrieved additional data and additional reference data, and embed said additional data coefficients into said first block.
- additional data retrieved using one motion vector is provided for a second block of said previous frame that the motion vector is pointing to
- additional reference data is data identifying what the additional data to be embedded should resemble.
- the additional data is a watermark and the direction of change of the coefficients of a retrieved part of a previous frame watermark is compared with the direction of change of the coefficients of a corresponding part of the reference watermark, and those direction of changes of the retrieved watermark coefficients that differ from the direction of changes of the reference watermark coefficients are changed into the direction of changes of the reference watermark coefficients by means of adding corresponding correcting coefficients.
- the correcting coefficients are then embedded in the signal.
- the correcting coefficients are added to the part of the retrieved watermark and the result is stored as a part of a previous frame watermark for correction of following frames, which ensures that the watermark can be restored also in other frames.
- the retrieving is performed in the spatial domain, and the correction and embedding are performed in the DCT domain.
- the motion vector is associated with the spatial domain, which means that the retrieving then has to be performed there, while the watermark embedding has to be made in the DCT domain.
- the current frame is a frame that is predicted only based on a frame to be presented before the current frame.
- the present invention has the advantage of restoring the embedded additional data to what it should be in case an object coded in a media signal is moved. This allows the retaining of a high correlation between the embedded additional data and the additional data intended to be embedded.
- additional data which is to be embedded in a signal where a coded object is moved, is motion compensated with motion vectors associated with the object. The motion compensated additional data and additional reference data are then used for determining additional data to be embedded, in order to restore the intended information of the additional data.
- Fig. 1 schematically shows a number of frames of video information in a media signal
- Fig. 2 schematically shows one such frame of video information where a watermark has been provided, where the frame is divided into a number blocks
- Fig. 3 shows an example of a number of luminance levels in the spatial domain for one intraframe coded block
- Fig. 4 shows DCT levels corresponding to the luminance levels in Fig. 3 for the block
- Fig. 5 shows the default intra quantizer matrix for the block in Fig. 3 and 4
- Fig. 6 shows the scanning of quantised DCT coefficients for obtaining a VLC coded video signal
- Fig. 1 schematically shows a number of frames of video information in a media signal
- Fig. 2 schematically shows one such frame of video information where a watermark has been provided, where the frame is divided into a number blocks
- Fig. 3 shows an example of a number of luminance levels in the spatial domain for one intraframe coded block
- Fig. 4 shows DCT levels corresponding to
- FIG. 7 shows the default inter quantizer matrix for an intercoded block
- Fig. 8 shows a device for embedding additional data according to the present invention
- Fig. 9 shows a block schematic of an embedding unit in more detail according to the present invention
- Fig. 10 schematically shows a computer program product comprising computer program code for performing the method according to the invention.
- the invention is directed towards the embedding of additional data in a media signal.
- additional data is preferably a watermark.
- the media signal will in the following be described in relation to a video signal and then an MPEG coded video signal. It should be realised that the invention is not limited to MPEG coding, but other types of coding can just as well be contemplated.
- a video signal or stream X according to the MPEG standard is schematically shown in Fig. 1.
- An MPEG stream X comprises a number of transmitted frames or pictures denoted I, B and P.
- Fig. 1 shows a number of such frames shown one after the other.
- first line of numbers is shown, where these numbers indicate the display order, i.e. the order in which the information relating to the frames is to be displayed.
- second line of numbers indicating the transmission and decoding order, i.e. the order in which the frames are received and decoded in order to display a video sequence.
- arrows that indicate how the frames refer to each other. It should be realised that the stream also includes other information such as overhead information.
- the different types of frames are divided into I-, B- and P-pictures, where one such picture that is a P-picture is indicated with reference numeral 10. An I-picture is denoted with reference numeral 11.
- 1-pictures are so-called intraframe coded pictures. These pictures are coded independently of other pictures and thus contains all the information necessary for displaying an image.
- P- and B-pictures are so called interframe coded pictures that exploit the temporal redundancy between consecutive pictures and they use motion compensation to minimize the prediction error.
- P-pictures refer to one picture in the past, which previous picture can be an I-picture or a P-picture.
- B-pictures refer to two pictures one in the past and one in the future, where the picture referred to can be an I- or a P-picture. Because of this the B-picture has to be transmitted after the pictures it refers to, which leads to the transmission order being different than the display order.
- the frame contains a number of pixels, where the luminance and chrominance are provided for each pixel.
- focus will be made on the luminance, since watermarks are embedded into this property of a pixel.
- Each such frame is further divided into 8x8 pixel blocks of luminance values.
- One such frame 11 is shown in Fig. 2, which shows an object 12 provided in the stream.
- Fig. 3 shows an example of some luminance values y for the block indicated in Fig. 2.
- a DCT Discrete Cosine Transform
- Fig. 4 shows such a DCT coefficient block for the block in Fig. 3.
- the coefficients contain information on the horizontal and vertical spatial frequencies of the input block.
- the coefficient corresponding to zero horizontal and vertical frequency is called a DC component, which is the coefficient in the upper left corner of Fig. 4.
- DC component which is the coefficient in the upper left corner of Fig. 4.
- these coefficients are not evenly distributed, but the transformation tends to concentrate the energy to the low frequency coefficients, which are in the upper left corner of Fig. 4.
- the AC coefficients in the intracoded block are quantised by applying a quantisation step q * Qi n tra(m, n)/16.
- Fig. 5 shows the default quantisation values Qi ntra used here.
- the quantisation step q can be set differently from block to block and can vary between 1 and 112. After this quantisation the coefficients in the blocks are serialized into a one dimensional array of 64 coefficients.
- This serialisation scheme is here a zigzag scheme as shown in Fig. 6, where the first coefficient is the DC component and the last entry represents the highest spatial frequencies in the lower corner on the right side. From the DC component to this latest component the coefficients are connected to each other in a zigzag pattern.
- the one dimensional array is then compressed or entropy coded using a VLC (variable length code. This is done through providing a limited number of code words based on the array.
- Each code word denotes a run of zero values, i.e. the number of zero valued coefficients preceding a quantised DCT coefficient followed by a non zero coefficient of a particular level. This leads to the creation of the following line of code words for the values in Fig. 6:
- an I-frame only comprises intracoded blocks.
- P- and B- frames include mtercoded blocks where the coefficients represent prediction errors instead.
- motion vectors related to the intercoded blocks In the overhead information of such a frame there is also provided motion vectors related to the intercoded blocks.
- P- and B-frames might also contain intracoded blocks.
- An intercoded block is, as was mentioned above, handled in a similar manner as an intracoded block when being coded. The difference here is that the DCT coefficients do not represent luminance values but rather prediction errors, which are however treated in the same way as the intracoded coefficients.
- a quantisation step is applied according to q * Q n0n -i n tra(m, n)/16.
- Fig. 7 shows the default quantisation values Q n0n -intra used here.
- the quantisation step q can be set differently from block to block and can also here vary between 1 and 112.
- additional information in the form of a watermark is embedded in the different blocks.
- a typical algorithm is the so-called run-merge algorithm described in WO-02/060182, which is herein incorporated by reference.
- a watermark w in the form of a pseudo-random noise sequence, is embedded in the blocks of a frame.
- a watermark is here provided as a number of identical tiles provided over the whole image and where one tile can have the size of 128x128 pixels.
- the watermark tile is divided into blocks corresponding to the size of the DCT blocks and transformed into the DCT domain and these DCT blocks are then stored in a watermark buffer.
- the watermark is embedded in the quantised DCT coefficients under the control of a bit-rate controller.
- the watermark is embedded by adding ⁇ 1 to the smallest quantised DCT level.
- ⁇ 1 since many of the signal coefficients are zero an addition of ⁇ 1 may lead to an increased bit rate, which is disadvantageous. There is furthermore a risk that the watermark will be visible.
- the media processing device includes a parsing unit 18, a device for determining additional data 20 and an output stage 22.
- the parsing unit is connected to the device 20 as well as to the output stage 22, also the device 20 is connected to the output stage 22.
- the device 20 includes a first processing unit 26, connected to an embedding unit 28 and a second processing unit 30.
- a watermark buffer 24 is connected to the embedding unit 28. This watermark buffer 24 will later be called a reference watermark buffer for reasons that will become clear by the description.
- the parsing unit 18 receives a media signal X in the form of a number of video images or frames including blocks with VLC coded code words.
- the parsing unit separates the VLC coded code words from other types of information and sends the VLC coded code words to the first processing unit 26 of device 20, which processes the stream X in order to recreate the run-level pairs of each block.
- the parsing unit 18 also separates motion vectors V associated with intercoded blocks and provided in the overhead information of B- and P-frames and provides these motion vectors V to the embedding unit 28, which obtains them in this way.
- the run-level pairs received by the first processing unit 26, i.e. the quantised DCT coefficient matrix, are then sent to the embedding unit 28.
- the embedding unit 28 embeds a watermark stored in the watermark buffer, provides the watermarked DCT matrix to the second processing unit 30, that VLC codes it and provides it to the combining unit 22 for combination with the other MPEG codes. From the combining unit 22 the watermarked signal X' is then provided.
- Watermarking is according to the present invention normally handled as outlined in WO-02/060182, but possibly allowing higher or lower levels than ⁇ 1 of the watermark coefficients. During normal watermarking of blocks other watermarking levels than ⁇ 1 are allowed.
- the watermark coefficient for the signal coefficient is taken from the watermark buffer 24, where it is stored in the DCT domain.
- the watermark coefficient here has a value that defines the amount and direction (i.e. the sign) that the corresponding dequantized signal coefficient is to change.
- the embedding unit 28 for solving the above mentioned problem is shown in a block schematic in Fig. 9.
- the embedding unit 28 comprises a motion compensating unit 32 connected to a preceding frame watermark buffer 25.
- the motion compensating unit 32 is furthermore connected to a DCT transforming unit 34.
- the DCT transforming unit 34 is connected to a determining unit 36, which in turn is connected to a data embedding unit 38.
- the determining unit 36 is furthermore connected to the reference watermark buffer 24 and to an inverse DCT transforming unit 40, which is also connected to the preceding frame watermark buffer 25.
- the preceding frame watermark buffer has here been divided into a first buffer 25A and a second buffer 25B.
- the first buffer 25A comprises the watermark embedded in a previous frame
- the second buffer 25B comprises the watermark embedded in the present or current frame, which will be used as a reference watermark for the following frame.
- the functioning of the device in Fig. 9 will now be described under the assumption that the object 12 in Fig. 2 is moved in a P-frame.
- a preceding watermark Wpo in the spatial domain related to a previous frame has been stored in the first buffer 25A.
- the motion compensating unit 32 obtains the vectors V of all blocks of the P-frame in a consecutive fashion by counting rows and columns of the frame using a block counter and getting the vectors of the positions one by one. Each vector is associated with a first block position of the current frame and also points out a second position of a previous frame from where the corresponding block has been moved. If no motion vector is associated with a block, the vector in question has zero length. For each vector, the motion compensating unit then retrieves a previous frame watermark W PO block corresponding to the second position the vector is pointing to. In case the vector is zero the first and second positions are the same. The retrieved previous frame watermark W P0 blocks are then moved to the first positions of the current blocks, i.e. the positions associated with the vectors.
- the previous frame watermark block being motion compensated using the vector V such that now it has moved from the second to the first position.
- the retrieved and reordered previous frame watermark blocks W PO are then provided to the DCT transforming unit 34, which transforms the previous frame blocks from the spatial domain into the DCT domain and provides them to the determining unit 36.
- the watermark to be embedded is determined based on the retrieved and reordered previous frame watermark and a reference watermark. This is done through the reference watermark W R , which comprises data supposed to be embedded, being compared block by block with the reordered previous frame watermark Wp 0 .
- a first block of the reference watermark is compared with a second block of the previous frame watermark.
- the determination which is here done by correcting the previous frame watermark Wpo, is done in the following way.
- the directions of changes or signs of the motion compensated previous frame watermark coefficients are compared with the signs of the corresponding reference watermark coefficients. For a given first and second block combination those coefficients of the motion compensated second block that are the same as the signs of the first block, nothing is done. If the coefficients of the second block were all zero, i.e. no watermark was provided in that block of the previous frame, nothing is done also in this case.
- the signs of the second block coefficients are changed to the opposite sign, i.e. + is turned into - and vice versa. This is done by adding correcting coefficients.
- the correcting coefficients are added to the second block coefficients such that they resemble the first block coefficients.
- the motion compensated watermark coefficients receive the same sign as the reference watermark coefficients. This is what always happens. If the bit-rate is not increased, the levels of the correcting coefficients are furthermore raised and ideally they receive double the value of the second block coefficients in order to completely restore the watermark.
- the reason for this is that the prediction error is added to the motion compensated frame, and therefore to obtain a watermark with an opposite sign the watermark has to be embedded with twice the energy. It might here be necessary to have a lower level than twice the original energy level, to only limit the change to a sign change or to provide a zero level, i.e. to skip the level, for other reasons than increased bit-rate and that is when the quantisation step associated with the block to be corrected is too large. Such a level correction could then lead to the watermark becoming visible.
- the watermark coefficients can here also be quantised instead of dequantised.
- the correcting coefficients are supplied to the data embedding unit 38, which embeds them in the signal X.
- the data embedding unit then first quantises them before embedding and in case they were already quantised, they are directly embedded in the signal.
- the determining unit 36 furthermore, adds the correcting coefficients to the retrieved and reordered watermark. For coefficients where no correction takes place the sum only consists of the retrieved coefficient.
- the result of the addition is then provided to the inverse DCT transforming unit 40, which performs an inverse DCT transformation in order to obtain a previous frame watermark Wpi in the spatial domain.
- This previous frame watermark is then provided to the second watermark buffer 25B for storing as a new previous frame watermark for a following frame.
- the watermark retains the structure that it should have, which is important when detecting the watermark.
- the watermark furthermore remains invisible.
- By changing the signs of the coefficients a high correlation is retained when only the signs have been used to embed the watermark in the DCT domain.
- a P-frame may also comprise intracoded blocks, where the correction according to the invention is not used. However the watermark coefficients for this block will then be stored in the second buffer of the preceding frame watermark buffer. It is possible to restrict the correction to only the above-described P- pictures, since these pictures are used as reference for other P- and B-pictures. This means that only for I- and P-frames the embedded watermark is stored in buffers for future use, and the watermark is motion compensated in the P-pictures, which reduces the amount of processing needed. It should however be realised that it can also be implemented for B- pictures. In the case of B-frames there would be needed an extra previous frame buffer because the motion compensation depends on at most two buffers.
- the coefficients of the two previous frames are furthermore added to each other and divided by two.
- the correction process thus becomes more complex for a B-frame.
- the motion compensation might be possible to perform in the DCT domain, in which case the reference watermark might be stored also in this domain and in which case there would be no need for the DCT transforming unit and the inverse DCT transforming unit.
- the present invention has been described in relation to a watermark embedding unit.
- This embedding unit is preferably provided in the form of one of more processors containing program code for performing the method according to the present invention.
- This program code can also be provided on a computer program medium, like a CD ROM 42, which is generally shown in Fig. 10.
- the method according to the invention is then performed when the CD ROM is loaded in a computer.
- the program code can furthermore be downloaded from a server, for example via the Internet.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Compression Or Coding Systems Of Tv Signals (AREA)
- Editing Of Facsimile Originals (AREA)
Abstract
Description
Claims
Priority Applications (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR1020067025592A KR20070032674A (en) | 2004-06-08 | 2005-05-31 | Compensation for Watermark Irregularities Caused by Moved Objects |
EP05746696A EP1757104A1 (en) | 2004-06-08 | 2005-05-31 | Compensating watermark irregularities caused by moved objects |
US11/569,976 US20070223693A1 (en) | 2004-06-08 | 2005-05-31 | Compensating Watermark Irregularities Caused By Moved Objects |
JP2007526625A JP2008502256A (en) | 2004-06-08 | 2005-05-31 | Compensation for watermark irregularities caused by moved objects |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
EP04102595.8 | 2004-06-08 | ||
EP04102595 | 2004-06-08 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2005122586A1 true WO2005122586A1 (en) | 2005-12-22 |
Family
ID=34970346
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/IB2005/051767 WO2005122586A1 (en) | 2004-06-08 | 2005-05-31 | Compensating watermark irregularities caused by moved objects |
Country Status (6)
Country | Link |
---|---|
US (1) | US20070223693A1 (en) |
EP (1) | EP1757104A1 (en) |
JP (1) | JP2008502256A (en) |
KR (1) | KR20070032674A (en) |
CN (1) | CN1965584A (en) |
WO (1) | WO2005122586A1 (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9076220B2 (en) | 2010-04-29 | 2015-07-07 | Thomson Licensing | Method of processing an image based on the determination of blockiness level |
Families Citing this family (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP5103479B2 (en) * | 2006-10-18 | 2012-12-19 | デスティニー ソフトウェア プロダクションズ インコーポレイテッド | Method of adding digital watermark to media data |
US8228993B2 (en) * | 2007-04-06 | 2012-07-24 | Shalini Priti | System and method for encoding and decoding information in digital signal content |
US8798133B2 (en) | 2007-11-29 | 2014-08-05 | Koplar Interactive Systems International L.L.C. | Dual channel encoding and detection |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2000056058A1 (en) * | 1999-03-18 | 2000-09-21 | British Broadcasting Corporation | Watermarking |
JP2002084510A (en) * | 2000-09-08 | 2002-03-22 | Jisedai Joho Hoso System Kenkyusho:Kk | Method and apparatus for embedding electronic watermark |
WO2002060182A1 (en) * | 2001-01-23 | 2002-08-01 | Koninklijke Philips Electronics N.V. | Watermarking a compressed information signal |
US20020181706A1 (en) * | 2001-06-05 | 2002-12-05 | Yuuki Matsumura | Digital watermark embedding device and digital watermark embedding method |
Family Cites Families (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7809138B2 (en) * | 1999-03-16 | 2010-10-05 | Intertrust Technologies Corporation | Methods and apparatus for persistent control and protection of content |
JP4035257B2 (en) * | 1998-04-10 | 2008-01-16 | キヤノン株式会社 | Image processing apparatus, image processing method, and computer-readable storage medium |
US20030118181A1 (en) * | 1999-11-12 | 2003-06-26 | Kunihiko Miwa | Method and Apparatus for Controlling Digital Data |
CN1279532C (en) * | 2000-10-31 | 2006-10-11 | 索尼公司 | Apparatus and method for recording/reproducing audio data embedded with additive information |
JP3861624B2 (en) * | 2001-06-05 | 2006-12-20 | ソニー株式会社 | Digital watermark embedding processing apparatus, digital watermark embedding processing method, and program |
JP2005514717A (en) * | 2002-01-11 | 2005-05-19 | コーニンクレッカ フィリップス エレクトロニクス エヌ ヴィ | Generation of unique watermarks for receivers of multimedia multicast transmissions |
JP4564753B2 (en) * | 2002-02-06 | 2010-10-20 | ソニー ヨーロッパ リミテッド | Bitstream changing method and apparatus |
-
2005
- 2005-05-31 CN CNA200580018840XA patent/CN1965584A/en active Pending
- 2005-05-31 US US11/569,976 patent/US20070223693A1/en not_active Abandoned
- 2005-05-31 JP JP2007526625A patent/JP2008502256A/en active Pending
- 2005-05-31 WO PCT/IB2005/051767 patent/WO2005122586A1/en active Application Filing
- 2005-05-31 EP EP05746696A patent/EP1757104A1/en not_active Withdrawn
- 2005-05-31 KR KR1020067025592A patent/KR20070032674A/en not_active Application Discontinuation
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2000056058A1 (en) * | 1999-03-18 | 2000-09-21 | British Broadcasting Corporation | Watermarking |
JP2002084510A (en) * | 2000-09-08 | 2002-03-22 | Jisedai Joho Hoso System Kenkyusho:Kk | Method and apparatus for embedding electronic watermark |
WO2002060182A1 (en) * | 2001-01-23 | 2002-08-01 | Koninklijke Philips Electronics N.V. | Watermarking a compressed information signal |
US20020181706A1 (en) * | 2001-06-05 | 2002-12-05 | Yuuki Matsumura | Digital watermark embedding device and digital watermark embedding method |
Non-Patent Citations (2)
Title |
---|
PATENT ABSTRACTS OF JAPAN vol. 2002, no. 07 3 July 2002 (2002-07-03) * |
SCHIMMEL S: "Motion Sensitive Video Watermarking", NAT. LAB. UNCLASSIFIED REPORT 2001/825, PHILIPS ELECTRONICS, August 2001 (2001-08-01), XP002340828, Retrieved from the Internet <URL:http://www.extra.research.philips.com/publ/rep/nl-ur/NL-UR2001-825.pdf> [retrieved on 20050816] * |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9076220B2 (en) | 2010-04-29 | 2015-07-07 | Thomson Licensing | Method of processing an image based on the determination of blockiness level |
Also Published As
Publication number | Publication date |
---|---|
CN1965584A (en) | 2007-05-16 |
US20070223693A1 (en) | 2007-09-27 |
JP2008502256A (en) | 2008-01-24 |
EP1757104A1 (en) | 2007-02-28 |
KR20070032674A (en) | 2007-03-22 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP4248241B2 (en) | Watermarking of compressed information signals | |
US20040202350A1 (en) | Digital watermarking technique | |
EP0928110A2 (en) | Image signal processing for electronic watermarking | |
KR20110061551A (en) | Context-based adaptive binary arithmetic coding (cabac) video stream compliance | |
JP2006157910A (en) | Filtering method for removing blocking artifacts and apparatus therefor | |
EP1413143B1 (en) | Processing a compressed media signal | |
JP2003517796A (en) | How to reduce the "uneven picture" effect | |
EP2199970B1 (en) | Watermarking compressed video data by changing blocks' prediction mode | |
JP2004336529A (en) | Motion picture data processing apparatus, method and program | |
JP2006505173A (en) | Digital watermarking method for variable bit rate signal | |
US20070223693A1 (en) | Compensating Watermark Irregularities Caused By Moved Objects | |
JP2004241869A (en) | Watermark embedding and image compressing section | |
US20050089189A1 (en) | Embedding a watermark in an image signal | |
US8848791B2 (en) | Compressed domain video watermarking | |
WO2005122081A1 (en) | Watermarking based on motion vectors | |
EP1395036A1 (en) | Method for burying data in image, and method for extracting the data | |
JP4642509B2 (en) | Apparatus for embedding confidential information in compressed image data and apparatus for encoding confidential data | |
WO2005122080A1 (en) | Variance based variation of watermarking depth in a media signal | |
JP2007535262A (en) | How to watermark a compressed information signal | |
JP2011130050A (en) | Image encoding apparatus | |
JP4931077B2 (en) | Digital watermark embedding method, apparatus and program, digital watermark detection method, apparatus and program | |
JP2009278179A (en) | Image decoder, image decoding method, program and recording medium | |
JP3566924B2 (en) | Digital watermark embedding method, detection method, digital watermark embedding device, detection device, recording medium recording digital watermark embedding program, and recording medium recording detection program | |
JP4020302B2 (en) | Information embedding device and information extracting device | |
JP5174878B2 (en) | Secret information extraction device and rewrite device |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AK | Designated states |
Kind code of ref document: A1 Designated state(s): AE AG AL AM AT AU AZ BA BB BG BR BW BY BZ CA CH CN CO CR CU CZ DE DK DM DZ EC EE EG ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KM KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX MZ NA NG NI NO NZ OM PG PH PL PT RO RU SC SD SE SG SK SL SM SY TJ TM TN TR TT TZ UA UG US UZ VC VN YU ZA ZM ZW |
|
AL | Designated countries for regional patents |
Kind code of ref document: A1 Designated state(s): GM KE LS MW MZ NA SD SL SZ TZ UG ZM ZW AM AZ BY KG KZ MD RU TJ TM AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IS IT LT LU MC NL PL PT RO SE SI SK TR BF BJ CF CG CI CM GA GN GQ GW ML MR NE SN TD TG |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application | ||
WWE | Wipo information: entry into national phase |
Ref document number: 2005746696 Country of ref document: EP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2007526625 Country of ref document: JP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 11569976 Country of ref document: US Ref document number: 2007223693 Country of ref document: US |
|
WWE | Wipo information: entry into national phase |
Ref document number: 1020067025592 Country of ref document: KR |
|
WWE | Wipo information: entry into national phase |
Ref document number: 200580018840.X Country of ref document: CN |
|
WWE | Wipo information: entry into national phase |
Ref document number: 71/CHENP/2007 Country of ref document: IN |
|
WWP | Wipo information: published in national office |
Ref document number: 2005746696 Country of ref document: EP |
|
WWP | Wipo information: published in national office |
Ref document number: 1020067025592 Country of ref document: KR |
|
WWP | Wipo information: published in national office |
Ref document number: 11569976 Country of ref document: US |