WO2019082268A1 - Image processing device, and image processing method - Google Patents

Image processing device, and image processing method

Info

Publication number
WO2019082268A1
WO2019082268A1 PCT/JP2017/038339 JP2017038339W WO2019082268A1 WO 2019082268 A1 WO2019082268 A1 WO 2019082268A1 JP 2017038339 W JP2017038339 W JP 2017038339W WO 2019082268 A1 WO2019082268 A1 WO 2019082268A1
Authority
WO
WIPO (PCT)
Prior art keywords
picture
pictures
block
evaluation value
calculation unit
Prior art date
Application number
PCT/JP2017/038339
Other languages
French (fr)
Japanese (ja)
Inventor
彰 峯澤
守屋 芳美
杉本 和夫
Original Assignee
三菱電機株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 三菱電機株式会社 filed Critical 三菱電機株式会社
Priority to PCT/JP2017/038339 priority Critical patent/WO2019082268A1/en
Priority to JP2018550470A priority patent/JP6782789B2/en
Priority to PCT/JP2018/013789 priority patent/WO2019082413A1/en
Priority to TW107127425A priority patent/TWI673652B/en
Publication of WO2019082268A1 publication Critical patent/WO2019082268A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/136Incoming video signal characteristics or properties
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/142Detection of scene cut or scene change
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/17Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
    • H04N19/176Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/85Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression
    • H04N19/87Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression involving scene cut or scene change detection in combination with video compression

Definitions

  • the present invention relates to an image processing apparatus and an image processing method for determining the presence or absence of a scene change which is a change point of a scene.
  • Patent Document 1 discloses an image processing apparatus that determines the presence or absence of a scene change.
  • This image processing apparatus acquires a motion vector, which is one of the coding parameters of a plurality of pictures, from the encoded video stream, and determines the presence or absence of a scene change from the difference of the motion vectors among the plurality of pictures. I have to.
  • the presence or absence of a scene change can be determined if a motion vector is included as a coding parameter of a picture in a coded video stream.
  • the coding mode of a picture is, for example, an intra-frame coding mode
  • a motion vector may not be included as a coding parameter of the picture.
  • the present invention has been made to solve the above-described problems, and it is an object of the present invention to provide an image processing apparatus and an image processing method capable of determining the presence or absence of a scene change even for a picture containing no motion vector. To aim.
  • the image processing apparatus selects one of a plurality of pictures as a picture to be evaluated for the presence or absence of a scene change, which is a scene change point, by a picture selection unit and a picture selection unit.
  • the feature quantity calculation unit that calculates the feature quantity of each block included in the picture and the feature quantity of the block unit included in the two pictures calculated by the feature quantity calculation unit.
  • an evaluation value calculation unit that calculates an evaluation value used to determine the presence or absence of a change, and the presence / absence determination unit compares the evaluation value calculated by the evaluation value calculation unit with a threshold to obtain between two pictures. It is determined whether there is a scene change.
  • the evaluation value calculation unit is provided for calculating an evaluation value used to determine the presence or absence of a scene change from the feature amounts in block units included in two pictures calculated by the feature amount calculation unit.
  • the picture including no motion vector is also included. There is an effect that it is possible to determine the presence or absence of a scene change.
  • FIG. 1 is a block diagram showing an image processing apparatus according to Embodiment 1 of the present invention. It is a hardware block diagram which shows the image processing apparatus by Embodiment 1 of this invention.
  • FIG. 17 is a hardware configuration diagram of a computer in the case where the image processing apparatus is realized by software or firmware.
  • 7 is a flowchart showing an image processing method which is a processing procedure when the image processing apparatus is realized by software or firmware.
  • FIG. 7 is an explanatory drawing showing an example of selecting an evaluation target picture from among a plurality of pictures included in a coded video stream.
  • FIG. 7 is an explanatory drawing showing an example of selecting an evaluation target picture from among a plurality of pictures included in a coded video stream.
  • the coding method is AVC / H.
  • FIG. 1 is a block diagram showing an image processing apparatus according to Embodiment 1 of the present invention.
  • FIG. 2 is a hardware configuration diagram showing an image processing apparatus according to Embodiment 1 of the present invention.
  • the picture selection unit 1 is realized by, for example, the picture selection circuit 11 shown in FIG.
  • the picture selection unit 1 executes a process of acquiring a coded video stream including coded data of a plurality of pictures.
  • the picture selection unit 1 determines a candidate picture to be evaluated for the presence or absence of a scene change, which is a scene change point, from among a plurality of pictures having encoded data encoded in encoded block units (block units), A process of selecting two pictures is performed as a target picture for which the presence or absence of a scene change is to be evaluated from candidate pictures.
  • the method of determining candidate pictures is assumed to be previously defined.
  • the coded data of a picture is, for example, data coded in coding block units called macroblocks, or in coding block units called CTU (Coding Tree Unit).
  • the H.264 standard is disclosed in the following non-patent document 1.
  • the H.265 standard is disclosed in the following non-patent document 2.
  • Non-Patent Document 1 ISO / IEC 14496-10 / ITU-T H.2.
  • H.264 standard ISO / IEC 23008-2 / ITU-T H.264. 265 standard
  • the feature quantity calculation unit 2 is realized by, for example, the feature quantity calculation circuit 12 shown in FIG.
  • the feature amount calculation unit 2 performs, for each of the pictures selected by the picture selection unit 1, a process of calculating a feature amount in units of processing blocks from coded data in units of coding blocks in the picture.
  • FIG. 14 shows an example in which a processing block is composed of four coding blocks. Further, as in the example shown in FIG. 15, the number of coding blocks in the vertical direction and the horizontal direction constituting the processing block may be different.
  • the feature amount in processing block units is calculated from the coded data of each coding block in the processing block.
  • the coded data of each coding block is decoded, the code amount of each coding block is specified from the decoding result of the coded data, and the coding blocks belonging to each processing block are identified.
  • the sum of code amounts is taken as a feature amount in processing block units. Note that the processing block sizes of two pictures selected by the picture selection unit 1 (configurations of coding blocks in processing blocks to be defined in advance) are the same.
  • the evaluation value calculation unit 3 is realized by, for example, the evaluation value calculation circuit 13 shown in FIG.
  • the evaluation value calculation unit 3 executes a process of calculating an evaluation value to be used to determine the presence or absence of a scene change from the feature amounts of the processing block unit included in the two pictures calculated by the feature amount calculation unit 2.
  • the presence / absence determination unit 4 is realized by, for example, the presence / absence determination circuit 14 illustrated in FIG.
  • the presence / absence determination unit 4 executes a process of determining the presence / absence of a scene change between two pictures by comparing the evaluation value calculated by the evaluation value calculation unit 3 with a threshold.
  • each of the picture selection unit 1, the feature amount calculation unit 2, the evaluation value calculation unit 3, and the presence / absence determination unit 4 which are components of the image processing apparatus is realized by dedicated hardware as shown in FIG. 2.
  • the thing is assumed. That is, what is realized by the picture selection circuit 11, the feature amount calculation circuit 12, the evaluation value calculation circuit 13, and the presence / absence determination circuit 14 is assumed.
  • the picture selection circuit 11, the feature value calculation circuit 12, the evaluation value calculation circuit 13, and the presence / absence determination circuit 14 are, for example, a single circuit, a composite circuit, a processor that has been programmed, a processor that is parallel programmed, ASIC (Application Specific) An integrated circuit), an FPGA (field-programmable gate array), or a combination thereof is applicable.
  • the components of the image processing apparatus are not limited to those realized by dedicated hardware, and the image processing apparatus may be realized by software, firmware, or a combination of software and firmware.
  • the software or firmware is stored as a program in the memory of the computer.
  • a computer means hardware that executes a program, and corresponds to, for example, a central processing unit (CPU), a central processing unit, a processing unit, an arithmetic unit, a microprocessor, a microcomputer, a processor, a digital signal processor (DSP), etc. .
  • the memory of the computer may be, for example, non-volatile or volatile such as random access memory (RAM), read only memory (ROM), flash memory, erasable programmable read only memory (EPROM), electrically erasable programmable read only memory (EEPROM).
  • RAM random access memory
  • ROM read only memory
  • EPROM erasable programmable read only memory
  • EEPROM electrically erasable programmable read only memory
  • a semiconductor memory a magnetic disk, a flexible disk, an optical disk, a compact disk, a mini disk, a DVD (Digital Versatile Disc), etc. correspond.
  • FIG. 3 is a hardware configuration diagram of a computer when the image processing apparatus is realized by software or firmware.
  • the memory 21 has a program for causing a computer to execute the processing procedure of the picture selection unit 1, feature quantity calculation unit 2, evaluation value calculation unit 3, and presence / absence determination unit 4.
  • the program may be stored and the processor 22 of the computer may execute the program stored in the memory 21.
  • FIG. 4 is a flowchart showing an image processing method which is a processing procedure when the image processing apparatus is realized by software or firmware.
  • FIG. 2 shows an example in which each of the components of the image processing apparatus is realized by dedicated hardware
  • FIG. 3 shows an example in which the image processing apparatus is realized by software or firmware. Some components of the image processing apparatus may be realized by dedicated hardware, and the remaining components may be realized by software, firmware or the like.
  • the picture selection unit 1 obtains a coded video stream including coded data of a plurality of pictures.
  • the picture selecting unit 1 determines a candidate picture from the obtained encoded video stream according to a method for determining a candidate picture defined in advance, and evaluates the presence or absence of a scene change which is a scene change point among the candidate pictures.
  • Two pictures are selected as the picture of (1) (step ST1 in FIG. 4).
  • the later picture in the reproduction order is referred to as the “evaluation target picture”
  • the previous picture in the reproduction order is referred to as the “reference picture”.
  • “Evaluate the presence or absence of a scene change between two pictures” means that there is a scene change from the picture one after the reference picture in the playback order to the evaluation target picture (including the evaluation target picture itself) It shows that it evaluates whether it is not.
  • all of the plurality of pictures in which the encoded data is included in the encoded video stream are set as candidate pictures, and the evaluation target pictures are sequentially selected in the reproduction order from the candidate pictures.
  • An example is shown in which all pictures included in the encoded video stream are selected as pictures to be evaluated.
  • a plurality of pictures in which the encoded data is included in the encoded video stream has a GOP (Group Of Picture) structure, and the intra picture of the random access point in each GOP is taken as a candidate picture.
  • GOP Group Of Picture
  • the reference picture is a candidate picture one before the evaluation target picture in reproduction order in any of FIGS. 5 and 6. Therefore, in the example of FIG. 5, it is evaluated whether there is a scene change between two adjacent pictures, that is, whether the evaluation target picture itself is a scene change point. In the example of FIG. 6, the presence or absence of a scene change between the intra pictures of random access points in adjacent GOPs is evaluated.
  • FIG. 7 shows that the coding method is AVC / H. H.264 or HEVC / H. It is explanatory drawing which shows the structural example of the encoding video stream in the case of being H.265.
  • a plurality of access units are arranged in coding order, and one access unit indicates data of one picture.
  • Each access unit is composed of a plurality of NAL (Network Abst Ruction Layer) units.
  • Each NAL unit is classified into header information (non-VCL) and picture data (VCL) which is coded data of a picture.
  • the header information (non-VCL) includes an access unit delimiter, a sequence level header and a picture level header.
  • one or more NAL units constituting one access unit are continuous, and a method of identifying the head position of the access unit is defined so that the division position between adjacent access units on the stream can be known.
  • an access unit delimiter which is an NAL unit indicating the beginning of an access unit is defined.
  • the NAL unit illustrated in FIG. 7 is an example, and the configuration pattern of the NAL unit conforms to the specifications of the respective standards.
  • the head position of the access unit should be identified according to the identification method of the head position of the access unit defined in the standard. It is possible.
  • the picture selection unit 1 can identify the break position of the access unit which is a picture by identifying the start position of the access unit according to the identification method of the start position of the access unit defined in the standard. Since the picture selection unit 1 can identify the break position of the access unit, it can select two pictures from among a plurality of access units, and calculates the feature amount of the encoded data of the selected two pictures Output to part 2.
  • the feature quantity calculation unit 2 calculates feature quantities in processing block units from the encoded data in coding block units for each of the pictures selected by the picture selection unit 1 (step ST2 in FIG. 4). The process of calculating the feature amount in units of processing blocks by the feature amount calculation unit 2 will be specifically described below.
  • n is a block number for specifying a processing block to be subjected to feature amount calculation in a picture, and is numbered 1, 2,..., NUM_BL t in raster scan order from the upper left processing block.
  • the feature amount calculation unit 2 decodes each of all the encoding parameters included in the encoded data of each encoding block belonging to the processing block for which the feature amount is to be calculated, The sum of code amounts is specified as a code amount Sn, t .
  • the coding parameter is a parameter necessary for generating a decoded image of a coding block, and corresponds to, for example, a parameter indicating a coding mode, an intra prediction parameter, an inter prediction parameter, a prediction difference coding parameter, a motion vector, and the like.
  • the feature amount calculating unit 2 the code amount S ALL feature quantity calculation target picture P t, identifies t, as shown in the following equation (1), the feature quantity calculation target picture P t code amount S ALL, in t, the code amount S n of the processing blocks, normalized by dividing t, feature amount C n of the processing block, as t, the code amount S n of normalized processing block units, the t M Double.
  • M is a preset constant.
  • the feature amount calculating section 2 the code amount S ALL feature quantity calculation target picture P t
  • the code amount S ALL feature quantity calculation target picture P t is shown an example of specifying a t, the feature quantity calculation target picture P t code amount S ALL, as t, code amount S n of all of the processing blocks included in the feature quantity calculation target picture P t, example of calculating the sum of t is considered.
  • the code amount S ALL feature quantity calculation target picture P t, as t, the size of the picture data (VCL) of the feature quantity calculation target picture P t may be specified.
  • the size of the picture data (VCL) without performing the decoding process of syntax in the feature quantity calculation target picture P t, can be identified.
  • FIG. 8 is an explanatory diagram of an example of the code amount in processing block units.
  • FIG. 8 shows an example in which the processing block is composed of 2 ⁇ 2 coded blocks and the coded block is a macroblock or CTU, and the code amount of each coded block is indicated by a numeral. Furthermore, at the center of each processing block, the sum of the code amounts of the coding blocks belonging to the processing block is represented as the code amount of the processing block.
  • the unit of the code amount of the coding block is represented by bytes in the example of FIG. 8, this is merely an example, and may be represented by bit units.
  • Evaluation value calculation unit 3 the feature amount C n of the processing block in the reference picture P tp calculated by the feature amount calculation unit 2, and tp, the feature amount C n of the processing block in the evaluation object picture P tc, and tc Then, the evaluation value J used to determine the presence or absence of a scene change is calculated (step ST3 in FIG. 4). For example, as the evaluation value J used to determine the presence or absence of a scene change, the evaluation value calculation unit 3 performs processing blocks between the reference picture P tp and the evaluation target picture P tc as shown in the following equation (2). Calculate the sum of absolute differences of the feature amounts of the unit.
  • MIN_NUM_BL is the minimum value of NUM_BL tp and NUM_BL tc . That is, when the number of processing blocks of the reference picture P tp and the evaluation target picture P tc is different, only the processing block whose feature amount is calculated for both pictures is the calculation target of the evaluation value J.
  • the decoding processing time also tends to be largely different.
  • the time allocated to decoding processing of the encoded data of each picture and calculation of the feature amount in processing block units is uniformly limited, the result that the number of process blocks that can be processed is different in both pictures It becomes.
  • the evaluation value J can be obtained for only the processing block whose feature amount has been calculated for both pictures.
  • the presence / absence determination unit 4 compares the evaluation value J calculated by the evaluation value calculation unit 3 with the threshold value Th set in advance (step ST4 in FIG. 4). If the evaluation value J calculated by the evaluation value calculation unit 3 is equal to or greater than the threshold Th (if step ST4 in FIG. 4: YES), the presence / absence determination unit 4 determines between the reference picture Ptp and the evaluation target picture Ptc. It is determined that there is a scene change (step ST5 in FIG. 4). If the evaluation value J calculated by the evaluation value calculation unit 3 is less than the threshold Th (in the case of NO in step ST4 of FIG. 4), the presence / absence determination unit 4 selects between the reference picture Ptp and the evaluation target picture Ptc.
  • the presence determination unit 4 outputs the determination result of the presence or absence of a scene change.
  • the threshold Th may be a fixed value set in advance, or may be switched or changed according to a specific condition.
  • a method of preparing and switching the threshold Th for each type of content for each type of content (drama, news, sports, etc.), a method of preparing and switching the threshold Th for each broadcasting station, and for each encoder generating a stream
  • a method of preparing and switching the threshold Th a method of adaptively calculating the threshold Th according to the setting of the encoder that generates the stream (a method of calculating the threshold Th with the setting value of the encoder as a variable)
  • Method of calculating threshold Th adaptively according to code amount transition Prepare a formula for calculating threshold Th with time change of code amount (1st to nth differential in the time direction, n is an integer of 1 or more) as a variable
  • step ST7 in FIG. 4 If there is still a picture not selected as an evaluation target picture among the candidate pictures determined by the picture selection unit 1 (step ST7 in FIG. 4: YES), the process returns to step ST1 and step ST1. The processing of ⁇ ST7 is repeated. In the process of step ST1 returned, the process of calculating the feature amount by the feature amount calculating unit 2 is omitted for the picture of which the feature amount is already calculated among the two pictures selected by the picture selecting unit 1 You may do so. By doing this, it is not necessary to recalculate feature amounts, and processing time can be shortened. For example, as described in FIG. 5 and FIG.
  • the evaluation target picture when the evaluation target picture is selected from among the candidate pictures in the reproduction order and the reference picture is made one candidate picture in reproduction order from the evaluation target picture, the evaluation target The feature amount C n, tc in the picture is the feature amount C n, tp in the reference picture when the candidate picture one ahead in the reproduction order is selected as the evaluation target picture. Therefore, the feature amount calculation process can be omitted by sequentially storing the calculated feature amounts C n and tc and reading out the feature amounts C n and tp from the stored feature amounts.
  • the evaluation target picture is selected in the decoding order, and the presence or absence of a scene change is evaluated as a candidate picture closest to the evaluation target picture in decoding order before the reference object You may Also in this case, for the picture whose feature amount has already been calculated, the process may be omitted by diverting the calculation result of the feature amount. If no candidate picture not selected as the evaluation target picture remains in the candidate pictures determined by the picture selection unit 1 (in the case of NO at step ST7 in FIG. 4), a series of processing ends.
  • the evaluation value used to determine the presence or absence of a scene change is calculated from the feature amount in units of blocks included in the two pictures calculated by the feature amount calculation unit 2
  • the evaluation value calculation unit 3 is provided, and the presence / absence determination unit 4 determines the presence / absence of a scene change between two pictures by comparing the evaluation value calculated by the evaluation value calculation unit 3 with a threshold. Since the configuration is made, it is possible to determine the presence or absence of a scene change even for a picture that does not include a motion vector.
  • FIG. 9 is a block diagram showing another image processing apparatus according to Embodiment 1 of the present invention.
  • the media transmission stream may be, for example, a media transmission format such as MPEG-2 TS (Transport Stream of ISO / IEC 13818-1 / ITU-T H.222.0) or MMT (ISO / IEC 23008-1).
  • MPEG-2 TS Transport Stream of ISO / IEC 13818-1 / ITU-T H.222.0
  • MMT ISO / IEC 23008-1).
  • a packetized media transport stream is assumed.
  • the demultiplexer 5 can know the division between each access unit and the type and size of the NAL unit from the header information of the given media transport stream. If the demultiplexer 5 gives information indicating the division between each access unit to the picture selection unit 1, the picture selection unit 1 can encode the encoded video stream based on the information provided from the demultiplexer 5. A desired picture can be selected from In addition, if the demultiplexer 5 supplies information indicating the type and size of the NAL unit to the feature amount calculation unit 2, the feature amount calculation unit 2 does not perform the decoding process of the syntax in the evaluation target picture. The code amount of the evaluation target picture can be specified.
  • the feature amount calculation unit 2 targets the processing block of the whole picture in the calculation processing of the code amount Sn, t of the processing block.
  • FIG. 10 is an explanatory diagram of a processing block for specifying a code amount among processing blocks constituting a picture.
  • the processing block in which “1” is described is a processing block that specifies a code amount
  • the processing block in which “0” is described is a processing block that does not specify a code amount.
  • FIG. 11 is an explanatory view showing a processing block for specifying a code amount among processing blocks constituting a picture.
  • the processing block in which “1” is described is a processing block that specifies the code amount
  • the processing block in which “0” is described is a processing block that does not specify the code amount.
  • FIG. 12 is an explanatory view showing a processing block for specifying a code amount among processing blocks constituting a picture.
  • the processing block in which “1” is described is a processing block that specifies a code amount
  • the processing block in which “0” is described is a processing block that does not specify a code amount.
  • the processing block in which “1” is described and the processing block in which “0” is described may be reversed.
  • the feature quantity calculation unit 2 decodes the encoded data of the encoded block, specifies the code amount of the processing block from the decoding result of the encoded data, and processes from the code amount of the processing block The example which calculates the feature-value of a block is shown.
  • the feature quantity calculation unit 2 sets the processing block as the feature quantity C n, t of the processing block. The number of divided blocks included in may be used.
  • the number of divided blocks included in the processing block is calculated by adding the number of divided blocks included in each coding block belonging to the processing block.
  • the number of divided blocks included in the coding block is, for example, the coding method of AVC / H. H.264 corresponds to the number of blocks in a macroblock determined by the macroblock type (mb_type).
  • the number of blocks in a macroblock is obtained by performing syntax decoding processing on a picture to be evaluated.
  • the coding method is HEVC / H. If it is 265, it corresponds to the number of CUs (Coding Units) in the CTU which is a coding block.
  • the number of CUs in a CTU can be obtained by performing syntax decoding processing on a picture to be evaluated.
  • the feature amount calculating unit 2 When specifying the number of divided blocks included in the processing block as the feature amount of the processing block, the feature amount calculating unit 2 does not need to perform processing for normalizing the code amount when calculating the feature amount of the processing block As a result, the processing load of the feature amount calculation unit 2 is reduced.
  • the feature quantity calculation unit 2 decodes the encoded data of the encoded block, specifies the code amount of the processing block from the decoding result of the encoded data, and processes from the code amount of the processing block The example which calculates the feature-value of a block is shown.
  • the feature quantity calculation unit 2 sets the processing block as the feature quantity C n, t of the processing block. The highest hierarchical number of divided blocks in the above may be used.
  • the deepest hierarchy number of divided blocks in the processing block is calculated by obtaining the maximum value of the deepest hierarchy number of divided blocks in each coding block belonging to the processing block.
  • the number of layers in the deepest layer in the coding block is the coding scheme HEVC / H. If it is 265, as shown in FIG. 13, the hierarchy of division corresponds to the deepest number among the plurality of CUs included in the CTU which is a coding block.
  • FIG. 13 is an explanatory drawing showing the number of layers of divided blocks in a coding block. In the example of FIG. 13, since the maximum value of CU depth is 3, the number of deepest layers in the divided block in the coding block is three.
  • the deepest hierarchy number of the divided blocks can be obtained by performing syntax decoding processing on the evaluation target picture.
  • the feature amount calculation unit 2 does not need to perform processing for normalizing the code amount when calculating the feature amount of the processing block. As a result, the processing load of the feature amount calculation unit 2 is reduced.
  • the evaluation value calculation unit 3 calculates the feature amount C n, tp of the processing block unit in the evaluation target picture P tp calculated by the feature amount calculation unit 2 and the processing block unit in the evaluation target picture P tc
  • processing is performed according to the flowchart of FIG. 4 as in the first embodiment, but the processing content is changed to the processing described below.
  • the picture selection unit 1 obtains a coded video stream including coded data of a plurality of pictures.
  • the picture selection unit 1 determines a candidate picture from among a plurality of pictures in which the encoded data is included in the encoded video stream according to a method for determining a candidate picture defined in advance, and performs the above implementation from among the candidate pictures. Similar to mode 1, the reference picture P tp and the evaluation target picture P tc are selected. Also, the picture selection unit 1 selects a secondary reference picture Ptp ′ from the candidate pictures. That is, in step ST1 of FIG. 4, in addition to the evaluation target picture and the reference picture, the secondary reference picture is also selected.
  • the feature quantity calculation unit 2 calculates, for each of the pictures selected by the picture selection unit 1, the feature of processing block unit from the encoded data of coding block unit included in the evaluation target picture.
  • the quantities are calculated respectively (step ST2 in FIG. 4). That is, the feature amount calculation unit 2 calculates the feature amount C n of the processing blocks included in the reference picture P tp, and tp, the feature amount C n of the processing blocks included in the evaluation object picture P tc, and tc . Also, the feature quantity calculation unit 2 calculates feature quantities C n, tp ′ in units of processing blocks included in the secondary reference picture Ptp ′ .
  • the evaluation value calculation unit 3 calculates the feature amount C n, tp of the processing block unit included in the reference picture P tp calculated by the feature amount calculation unit 2 and the feature amount C of the processing block unit included in the evaluation target picture P tc
  • An evaluation value J is calculated from n, tc and the feature amount C n, tp 'in units of processing blocks included in the secondary reference picture P tp' (step ST3 in FIG. 4).
  • the evaluation value calculation unit 3 calculates the sum of absolute differences of the feature amounts in processing block units between the reference picture P tp and the evaluation target picture P tc .
  • the evaluation value calculation unit 3 calculates the sum of absolute differences of feature quantities in processing block units between the reference picture P tp and the secondary reference picture P tp ′ .
  • the evaluation value calculation unit 3 calculates, as the evaluation value J, the difference between the calculated difference absolute value sums as shown in the following equation (3).
  • MIN_NUM_BL ' is the minimum value of three values of NUM_BL tp' , NUM_BL tp , and NUM_BL tc .
  • the magnitude of the change in the pattern between the reference picture P tp and the evaluation target picture P t c is evaluated.
  • the pattern in the reference picture P tp Since the magnitude of the difference between the degree of change of the pattern and the degree of change of the pattern in the evaluation target picture P tc is evaluated, the pattern of the whole screen is large in gradual change of the pattern between the pictures that are not scene changes. It is possible to reduce the possibility of erroneous detection as a changing scene change, and to improve the scene change detection accuracy more than the first to third embodiments.
  • the definition method of the processing block to be the target of calculation of the feature amount in processing block units is the same as that of the first embodiment (all processing blocks in a picture, partial processing blocks as shown in FIGS. It can be defined arbitrarily.
  • the presence / absence determination unit 4 compares the evaluation value J calculated by the evaluation value calculation unit 3 with the threshold value Th set in advance (step ST4 in FIG. 4). As in the first embodiment, if the evaluation value J calculated by the evaluation value calculation unit 3 is equal to or larger than the threshold value Th, the presence / absence determination unit 4 sets the scene between the reference picture P tp and the evaluation target picture P tc. It is determined that there is a change (step ST5 in FIG. 4). As in the first embodiment, if the evaluation value J calculated by the evaluation value calculation unit 3 is smaller than the threshold value Th, the presence / absence determination unit 4 sets the scene between the reference picture P tp and the evaluation target picture P tc. It is determined that there is no change (step ST6 in FIG. 4). The presence determination unit 4 outputs the determination result of the presence or absence of a scene change. Further, the method of setting the threshold value Th is the same as that of the first embodiment.
  • step ST7 in FIG. 4 If there is still a picture not selected as an evaluation target picture among the candidate pictures determined by the picture selection unit 1 (step ST7 in FIG. 4: YES), the process returns to step ST1 and step ST1. The processing of ⁇ ST7 is repeated. In the process of step ST1 returned, among the three pictures selected by the picture selection unit 1, for the picture for which the feature amount has already been calculated, the process of calculating the feature amount by the feature amount calculation unit 2 is omitted. You may do so. By doing this, it is not necessary to recalculate feature amounts, and processing time can be shortened. If there is no picture which has not been selected as a target picture for evaluation yet in the candidate pictures determined by the picture selection unit 1 (in the case of NO at step ST7 in FIG. 4), a series of processing ends.
  • the present invention allows free combination of each embodiment, or modification of any component of each embodiment, or omission of any component in each embodiment. .
  • the present invention is suitable for an image processing apparatus and an image processing method for determining the presence or absence of a scene change which is a change point of a scene.
  • Reference Signs List 1 picture selection unit, 2 feature amount calculation unit, 3 evaluation value calculation unit, 4 presence / absence determination unit, 5 demultiplexers, 11 picture selection circuit, 12 feature amount calculation circuit, 13 evaluation value calculation circuit, 14 presence / absence determination circuit, 21 memory , 22 processors.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)
  • Image Analysis (AREA)

Abstract

There is provided an evaluation value calculation unit (3) for calculating an evaluation value used in determining the presence of a scene change from a feature quantity in units of blocks included in two pictures calculated by a feature quantity calculation unit (2), and a presence determination unit (4) compares the evaluation value calculated by the evaluation value calculation unit (3) and a threshold value and thereby determines the presence of a scene change between the two pictures. This makes it possible to determine the presence of a scene change even for a picture in which no motion vectors are included.

Description

画像処理装置及び画像処理方法Image processing apparatus and image processing method
 この発明は、シーンの変化点であるシーンチェンジの有無を判定する画像処理装置及び画像処理方法に関するものである。 The present invention relates to an image processing apparatus and an image processing method for determining the presence or absence of a scene change which is a change point of a scene.
 以下の特許文献1には、シーンチェンジの有無を判定する画像処理装置が開示されている。
 この画像処理装置は、符号化映像ストリームから、複数のピクチャの符号化パラメータの一つである動きベクトルを取得し、複数のピクチャの間での動きベクトルの差分からシーンチェンジの有無を判定するようにしている。
Patent Document 1 below discloses an image processing apparatus that determines the presence or absence of a scene change.
This image processing apparatus acquires a motion vector, which is one of the coding parameters of a plurality of pictures, from the encoded video stream, and determines the presence or absence of a scene change from the difference of the motion vectors among the plurality of pictures. I have to.
国際公開第2010/125757号International Publication No. 2010/125757
 従来の画像処理装置は以上のように構成されているので、符号化映像ストリームの中に、ピクチャの符号化パラメータとして、動きベクトルが含まれていれば、シーンチェンジの有無を判定することができる。しかし、ピクチャの符号化モードが例えば画面内符号化モードであるために、当該ピクチャの符号化パラメータとして、動きベクトルが含まれていないことがある。符号化パラメータとして、動きベクトルが含まれていないピクチャについては、シーンチェンジの有無を判定することができないという課題があった。 Since the conventional image processing apparatus is configured as described above, the presence or absence of a scene change can be determined if a motion vector is included as a coding parameter of a picture in a coded video stream. . However, since the coding mode of a picture is, for example, an intra-frame coding mode, a motion vector may not be included as a coding parameter of the picture. There is a problem that it is not possible to determine the presence or absence of a scene change for a picture that does not include a motion vector as a coding parameter.
 この発明は上記のような課題を解決するためになされたもので、動きベクトルが含まれていないピクチャについても、シーンチェンジの有無を判定することができる画像処理装置及び画像処理方法を得ることを目的とする。 The present invention has been made to solve the above-described problems, and it is an object of the present invention to provide an image processing apparatus and an image processing method capable of determining the presence or absence of a scene change even for a picture containing no motion vector. To aim.
 この発明に係る画像処理装置は、複数のピクチャの中から、シーンの変化点であるシーンチェンジの有無を評価する対象のピクチャとして、2つのピクチャを選択するピクチャ選択部と、ピクチャ選択部により選択された2つのピクチャのそれぞれについて、ピクチャに含まれるブロック単位の特徴量をそれぞれ算出する特徴量算出部と、特徴量算出部により算出された2つのピクチャに含まれるブロック単位の特徴量から、シーンチェンジの有無の判定に用いる評価値を算出する評価値算出部とを設け、有無判定部が、評価値算出部により算出された評価値と閾値を比較することで、2つのピクチャの間でのシーンチェンジの有無を判定するようにしたものである。 The image processing apparatus according to the present invention selects one of a plurality of pictures as a picture to be evaluated for the presence or absence of a scene change, which is a scene change point, by a picture selection unit and a picture selection unit. The feature quantity calculation unit that calculates the feature quantity of each block included in the picture and the feature quantity of the block unit included in the two pictures calculated by the feature quantity calculation unit There is provided an evaluation value calculation unit that calculates an evaluation value used to determine the presence or absence of a change, and the presence / absence determination unit compares the evaluation value calculated by the evaluation value calculation unit with a threshold to obtain between two pictures. It is determined whether there is a scene change.
 この発明によれば、特徴量算出部により算出された2つのピクチャに含まれるブロック単位の特徴量から、シーンチェンジの有無の判定に用いる評価値を算出する評価値算出部を設け、有無判定部が、評価値算出部により算出された評価値と閾値を比較することで、2つのピクチャの間でのシーンチェンジの有無を判定するように構成したので、動きベクトルが含まれていないピクチャについても、シーンチェンジの有無を判定することができる効果がある。 According to the present invention, the evaluation value calculation unit is provided for calculating an evaluation value used to determine the presence or absence of a scene change from the feature amounts in block units included in two pictures calculated by the feature amount calculation unit. However, since it is configured to determine the presence or absence of a scene change between two pictures by comparing the threshold with the evaluation value calculated by the evaluation value calculation unit, the picture including no motion vector is also included. There is an effect that it is possible to determine the presence or absence of a scene change.
この発明の実施の形態1による画像処理装置を示す構成図である。FIG. 1 is a block diagram showing an image processing apparatus according to Embodiment 1 of the present invention. この発明の実施の形態1による画像処理装置を示すハードウェア構成図である。It is a hardware block diagram which shows the image processing apparatus by Embodiment 1 of this invention. 画像処理装置がソフトウェア又はファームウェアなどで実現される場合のコンピュータのハードウェア構成図である。FIG. 17 is a hardware configuration diagram of a computer in the case where the image processing apparatus is realized by software or firmware. 画像処理装置がソフトウェア又はファームウェアなどで実現される場合の処理手順である画像処理方法を示すフローチャートである。7 is a flowchart showing an image processing method which is a processing procedure when the image processing apparatus is realized by software or firmware. 符号化映像ストリームに含まれている複数のピクチャの中から、評価対象ピクチャを選択する例を示す説明図である。FIG. 7 is an explanatory drawing showing an example of selecting an evaluation target picture from among a plurality of pictures included in a coded video stream. 符号化映像ストリームに含まれている複数のピクチャの中から、評価対象ピクチャを選択する例を示す説明図である。FIG. 7 is an explanatory drawing showing an example of selecting an evaluation target picture from among a plurality of pictures included in a coded video stream. 符号化方式がAVC/H.264又はHEVC/H.265である場合の符号化映像ストリームの構成例を示す説明図である。The coding method is AVC / H. H.264 or HEVC / H. It is explanatory drawing which shows the structural example of the encoding video stream in the case of being H.265. 処理ブロック単位の符号量の一例を示す説明図である。It is explanatory drawing which shows an example of the code amount of processing block unit. この発明の実施の形態1による他の画像処理装置を示す構成図である。It is a block diagram which shows the other image processing apparatus by Embodiment 1 of this invention. 評価対象ピクチャに含まれている処理ブロックのうち、符号量を特定する処理ブロックを示す説明図である。It is an explanatory view showing a processing block which specifies code amount among processing blocks contained in a picture for evaluation. 評価対象ピクチャに含まれている処理ブロックのうち、符号量を特定する処理ブロックを示す説明図である。It is an explanatory view showing a processing block which specifies code amount among processing blocks contained in a picture for evaluation. 評価対象ピクチャに含まれている処理ブロックのうち、符号量を特定する処理ブロックを示す説明図である。It is an explanatory view showing a processing block which specifies code amount among processing blocks contained in a picture for evaluation. 符号化ブロックにおける分割ブロックの階層数を示す説明図である。It is explanatory drawing which shows the hierarchy number of the division block in a coding block. 処理ブロックと符号化ブロックの例を示す説明図である。It is an explanatory view showing an example of a processing block and a coding block. 処理ブロックと符号化ブロックの例を示す説明図である。It is an explanatory view showing an example of a processing block and a coding block.
 以下、この発明をより詳細に説明するために、この発明を実施するための形態について、添付の図面に従って説明する。 Hereinafter, in order to explain the present invention in more detail, a mode for carrying out the present invention will be described according to the attached drawings.
実施の形態1.
 図1は、この発明の実施の形態1による画像処理装置を示す構成図である。図2は、この発明の実施の形態1による画像処理装置を示すハードウェア構成図である。
 図1において、ピクチャ選択部1は、例えば図2に示すピクチャ選択回路11で実現される。
 ピクチャ選択部1は、複数のピクチャの符号化データが含まれている符号化映像ストリームを取得する処理を実施する。
 ピクチャ選択部1は、符号化ブロック単位(ブロック単位)に符号化されている符号化データを有する複数のピクチャの中から、シーンの変化点であるシーンチェンジの有無を評価する候補ピクチャを定め、候補ピクチャの中からシーンチェンジの有無を評価する対象ピクチャとして、2つのピクチャを選択する処理を実施する。ただし、候補ピクチャの決定方法は、予め定義されているものとする。
Embodiment 1
FIG. 1 is a block diagram showing an image processing apparatus according to Embodiment 1 of the present invention. FIG. 2 is a hardware configuration diagram showing an image processing apparatus according to Embodiment 1 of the present invention.
In FIG. 1, the picture selection unit 1 is realized by, for example, the picture selection circuit 11 shown in FIG.
The picture selection unit 1 executes a process of acquiring a coded video stream including coded data of a plurality of pictures.
The picture selection unit 1 determines a candidate picture to be evaluated for the presence or absence of a scene change, which is a scene change point, from among a plurality of pictures having encoded data encoded in encoded block units (block units), A process of selecting two pictures is performed as a target picture for which the presence or absence of a scene change is to be evaluated from candidate pictures. However, the method of determining candidate pictures is assumed to be previously defined.
 ピクチャの符号化データは、例えば、マクロブロックと呼ばれる符号化ブロック単位、あるいは、CTU(Coding Tree Unit)と呼ばれる符号化ブロック単位に符号化されているデータである。
 マクロブロック単位に符号化する符号化方式の規格として、例えば、AVC/H.264規格が以下の非特許文献1に開示されている。
 また、CTU単位に符号化する符号化方式の規格として、例えば、HEVC/H.265規格が以下の非特許文献2に開示されている。
[非特許文献1]ISO/IEC 14496-10/ITU-T H.264規格
[非特許文献2]ISO/IEC 23008-2/ITU-T H.265規格
The coded data of a picture is, for example, data coded in coding block units called macroblocks, or in coding block units called CTU (Coding Tree Unit).
As a standard of a coding scheme for coding in macro block units, for example, AVC / H. The H.264 standard is disclosed in the following non-patent document 1.
In addition, as a standard of a coding scheme for coding in CTU units, for example, HEVC / H. The H.265 standard is disclosed in the following non-patent document 2.
[Non-Patent Document 1] ISO / IEC 14496-10 / ITU-T H.2. H.264 standard [non-patent document 2] ISO / IEC 23008-2 / ITU-T H.264. 265 standard
 特徴量算出部2は、例えば図2に示す特徴量算出回路12で実現される。
 特徴量算出部2は、ピクチャ選択部1により選択されたピクチャ毎に、当該ピクチャにおける符号化ブロック単位の符号化データから、処理ブロック単位の特徴量をそれぞれ算出する処理を実施する。処理ブロックは、1以上の符号化ブロックから構成されており、その構成は予め定義されている。処理ブロックは、1以上の符号化ブロックから構成されるため、処理ブロックの定義可能な最小単位は符号化ブロックとなる(処理ブロック=符号化ブロック)。
 図14は、4つの符号化ブロックから処理ブロックを構成する例を示している。また、図15に示す例のように、処理ブロックを構成する縦方向と横方向の符号化ブロック数が異なるように構成してもよい。
 処理ブロック単位の特徴量は、処理ブロック内の各々の符号化ブロックの符号化データから算出する。この実施の形態1では、各々の符号化ブロックの符号化データを復号して、符号化データの復号結果から各々の符号化ブロックの符号量を特定し、各々の処理ブロックに属する符号化ブロックの符号量の総和を処理ブロック単位の特徴量とする。
 なお、ピクチャ選択部1により選択された2つのピクチャの処理ブロックサイズ(予め定義する処理ブロックにおける符号化ブロックの構成)は同一とする。
The feature quantity calculation unit 2 is realized by, for example, the feature quantity calculation circuit 12 shown in FIG.
The feature amount calculation unit 2 performs, for each of the pictures selected by the picture selection unit 1, a process of calculating a feature amount in units of processing blocks from coded data in units of coding blocks in the picture. A processing block is composed of one or more coding blocks, and the configuration is predefined. Since the processing block is composed of one or more coding blocks, the smallest definable unit of the processing block is the coding block (processing block = coding block).
FIG. 14 shows an example in which a processing block is composed of four coding blocks. Further, as in the example shown in FIG. 15, the number of coding blocks in the vertical direction and the horizontal direction constituting the processing block may be different.
The feature amount in processing block units is calculated from the coded data of each coding block in the processing block. In the first embodiment, the coded data of each coding block is decoded, the code amount of each coding block is specified from the decoding result of the coded data, and the coding blocks belonging to each processing block are identified. The sum of code amounts is taken as a feature amount in processing block units.
Note that the processing block sizes of two pictures selected by the picture selection unit 1 (configurations of coding blocks in processing blocks to be defined in advance) are the same.
 評価値算出部3は、例えば図2に示す評価値算出回路13で実現される。
 評価値算出部3は、特徴量算出部2により算出された2つのピクチャに含まれる処理ブロック単位の特徴量から、シーンチェンジの有無の判定に用いる評価値を算出する処理を実施する。
The evaluation value calculation unit 3 is realized by, for example, the evaluation value calculation circuit 13 shown in FIG.
The evaluation value calculation unit 3 executes a process of calculating an evaluation value to be used to determine the presence or absence of a scene change from the feature amounts of the processing block unit included in the two pictures calculated by the feature amount calculation unit 2.
 有無判定部4は、例えば図2に示す有無判定回路14で実現される。
 有無判定部4は、評価値算出部3により算出された評価値と閾値を比較することで、2つのピクチャの間でのシーンチェンジの有無を判定する処理を実施する。
The presence / absence determination unit 4 is realized by, for example, the presence / absence determination circuit 14 illustrated in FIG.
The presence / absence determination unit 4 executes a process of determining the presence / absence of a scene change between two pictures by comparing the evaluation value calculated by the evaluation value calculation unit 3 with a threshold.
 図1では、画像処理装置の構成要素であるピクチャ選択部1、特徴量算出部2、評価値算出部3及び有無判定部4のそれぞれが、図2に示すような専用のハードウェアで実現されるものを想定している。即ち、ピクチャ選択回路11、特徴量算出回路12、評価値算出回路13及び有無判定回路14で実現されるものを想定している。
 ここで、ピクチャ選択回路11、特徴量算出回路12、評価値算出回路13及び有無判定回路14は、例えば、単一回路、複合回路、プログラム化したプロセッサ、並列プログラム化したプロセッサ、ASIC(Application Specific Integrated Circuit)、FPGA(Field-Programmable Gate Array)、または、これらを組み合わせたものが該当する。
In FIG. 1, each of the picture selection unit 1, the feature amount calculation unit 2, the evaluation value calculation unit 3, and the presence / absence determination unit 4 which are components of the image processing apparatus is realized by dedicated hardware as shown in FIG. 2. The thing is assumed. That is, what is realized by the picture selection circuit 11, the feature amount calculation circuit 12, the evaluation value calculation circuit 13, and the presence / absence determination circuit 14 is assumed.
Here, the picture selection circuit 11, the feature value calculation circuit 12, the evaluation value calculation circuit 13, and the presence / absence determination circuit 14 are, for example, a single circuit, a composite circuit, a processor that has been programmed, a processor that is parallel programmed, ASIC (Application Specific) An integrated circuit), an FPGA (field-programmable gate array), or a combination thereof is applicable.
 画像処理装置の構成要素は、専用のハードウェアで実現されるものに限るものではなく、画像処理装置がソフトウェア、ファームウェア、または、ソフトウェアとファームウェアとの組み合わせで実現されるものであってもよい。
 ソフトウェア又はファームウェアはプログラムとして、コンピュータのメモリに格納される。コンピュータは、プログラムを実行するハードウェアを意味し、例えば、CPU(Central Processing Unit)、中央処理装置、処理装置、演算装置、マイクロプロセッサ、マイクロコンピュータ、プロセッサ、DSP(Digital Signal Processor)などが該当する。
 コンピュータのメモリは、例えば、RAM(Random Access Memory)、ROM(Read Only Memory)、フラッシュメモリ、EPROM(Erasable Programmable Read Only Memory)、EEPROM(Electrically Erasable Programmable Read Only Memory)などの不揮発性又は揮発性の半導体メモリや、磁気ディスク、フレキシブルディスク、光ディスク、コンパクトディスク、ミニディスク、DVD(Digital Versatile Disc)などが該当する。
The components of the image processing apparatus are not limited to those realized by dedicated hardware, and the image processing apparatus may be realized by software, firmware, or a combination of software and firmware.
The software or firmware is stored as a program in the memory of the computer. A computer means hardware that executes a program, and corresponds to, for example, a central processing unit (CPU), a central processing unit, a processing unit, an arithmetic unit, a microprocessor, a microcomputer, a processor, a digital signal processor (DSP), etc. .
The memory of the computer may be, for example, non-volatile or volatile such as random access memory (RAM), read only memory (ROM), flash memory, erasable programmable read only memory (EPROM), electrically erasable programmable read only memory (EEPROM). A semiconductor memory, a magnetic disk, a flexible disk, an optical disk, a compact disk, a mini disk, a DVD (Digital Versatile Disc), etc. correspond.
 図3は、画像処理装置がソフトウェア又はファームウェアなどで実現される場合のコンピュータのハードウェア構成図である。
 画像処理装置がソフトウェア又はファームウェアなどで実現される場合、ピクチャ選択部1、特徴量算出部2、評価値算出部3及び有無判定部4の処理手順をコンピュータに実行させるためのプログラムをメモリ21に格納し、コンピュータのプロセッサ22がメモリ21に格納されているプログラムを実行するようにすればよい。
 図4は、画像処理装置がソフトウェア又はファームウェアなどで実現される場合の処理手順である画像処理方法を示すフローチャートである。
FIG. 3 is a hardware configuration diagram of a computer when the image processing apparatus is realized by software or firmware.
When the image processing apparatus is realized by software or firmware, the memory 21 has a program for causing a computer to execute the processing procedure of the picture selection unit 1, feature quantity calculation unit 2, evaluation value calculation unit 3, and presence / absence determination unit 4. The program may be stored and the processor 22 of the computer may execute the program stored in the memory 21.
FIG. 4 is a flowchart showing an image processing method which is a processing procedure when the image processing apparatus is realized by software or firmware.
 また、図2では、画像処理装置の構成要素のそれぞれが専用のハードウェアで実現される例を示し、図3では、画像処理装置がソフトウェアやファームウェアなどで実現される例を示しているが、画像処理装置における一部の構成要素が専用のハードウェアで実現され、残りの構成要素がソフトウェアやファームウェアなどで実現されるものであってもよい。 Also, FIG. 2 shows an example in which each of the components of the image processing apparatus is realized by dedicated hardware, and FIG. 3 shows an example in which the image processing apparatus is realized by software or firmware. Some components of the image processing apparatus may be realized by dedicated hardware, and the remaining components may be realized by software, firmware or the like.
 次に動作について説明する。
 ピクチャ選択部1は、複数のピクチャの符号化データが含まれている符号化映像ストリームを取得する。
 ピクチャ選択部1は、取得した符号化映像ストリームから、予め定義された候補ピクチャの決定方法に従って候補ピクチャを決定し、候補ピクチャの中から、シーンの変化点であるシーンチェンジの有無を評価する対象のピクチャとして、2つのピクチャを選択する(図4のステップST1)。
 ここで、選択した2つのピクチャのうち、再生順で後のピクチャを「評価対象ピクチャ」、再生順で先のピクチャを「参照ピクチャ」と称する。「2つのピクチャの間のシーンチェンジの有無を評価する」とは、参照ピクチャより再生順で一つ後のピクチャから評価対象ピクチャまで(評価対象ピクチャ自身を含む)にシーンチェンジが発生しているか否かを評価することを示すものである。
Next, the operation will be described.
The picture selection unit 1 obtains a coded video stream including coded data of a plurality of pictures.
The picture selecting unit 1 determines a candidate picture from the obtained encoded video stream according to a method for determining a candidate picture defined in advance, and evaluates the presence or absence of a scene change which is a scene change point among the candidate pictures. Two pictures are selected as the picture of (1) (step ST1 in FIG. 4).
Here, of the two selected pictures, the later picture in the reproduction order is referred to as the “evaluation target picture”, and the previous picture in the reproduction order is referred to as the “reference picture”. “Evaluate the presence or absence of a scene change between two pictures” means that there is a scene change from the picture one after the reference picture in the playback order to the evaluation target picture (including the evaluation target picture itself) It shows that it evaluates whether it is not.
 図5及び図6は、符号化データが符号化映像ストリームに含まれているピクチャの中から、評価対象ピクチャとなる候補ピクチャを決定し、候補ピクチャの中から、評価対象ピクチャと参照ピクチャを選択する例を示す説明図である。
 図5では、符号化データが符号化映像ストリームに含まれている複数のピクチャの全てを候補ピクチャとし、候補ピクチャの中から、再生順で順番に評価対象ピクチャを選択することで、最終的に、符号化映像ストリームに含まれている全てのピクチャが評価対象ピクチャとして選択される例を示している。
 図6では、符号化データが符号化映像ストリームに含まれている複数のピクチャがGOP(Group Of Picture)構造を有しており、各々のGOPにおけるランダムアクセスポイントのイントラピクチャを候補ピクチャとし、候補ピクチャの中から、再生順で順番に評価対象ピクチャとして選択される例を示している。
 AVC/H.264では、IDR(Instantaneous Decoding Refresh)ピクチャが、ランダムアクセスポイントのイントラピクチャである。
 HEVC/H.265では、IRAP(Intra Random Access Point)ピクチャが、ランダムアクセスポイントのイントラピクチャである。
 参照ピクチャは、図5及び図6のいずれでも、に評価対象ピクチャより再生順で一つ前の候補ピクチャとしている。したがって、図5の例では、隣接している2つのピクチャの間でのシーンチェンジの有無、即ち、評価対象ピクチャ自身がシーンチェンジ点であるか否かが評価される。図6の例では、隣接しているGOPにおけるランダムアクセスポイントのイントラピクチャの間でのシーンチェンジの有無が評価される。
 なお、上記以外にもN個(Nは0以上の整数)のピクチャ間隔で評価対象ピクチャを選択する(N=0の場合、図5と同じとなる)など、候補ピクチャの定義を任意に設定することができる。
5 and 6 determine a candidate picture to be an evaluation target picture from among pictures in which encoded data is included in a coded video stream, and select an evaluation target picture and a reference picture from candidate pictures. It is an explanatory view showing an example to do.
In FIG. 5, all of the plurality of pictures in which the encoded data is included in the encoded video stream are set as candidate pictures, and the evaluation target pictures are sequentially selected in the reproduction order from the candidate pictures. An example is shown in which all pictures included in the encoded video stream are selected as pictures to be evaluated.
In FIG. 6, a plurality of pictures in which the encoded data is included in the encoded video stream has a GOP (Group Of Picture) structure, and the intra picture of the random access point in each GOP is taken as a candidate picture. An example is shown in which pictures are selected as evaluation target pictures in order of reproduction from pictures.
AVC / H. In H.264, Instantaneous Decoding Refresh (IDR) pictures are intra pictures of random access points.
HEVC / H. At 265, an Intra Random Access Point (IRAP) picture is an intra picture of a random access point.
The reference picture is a candidate picture one before the evaluation target picture in reproduction order in any of FIGS. 5 and 6. Therefore, in the example of FIG. 5, it is evaluated whether there is a scene change between two adjacent pictures, that is, whether the evaluation target picture itself is a scene change point. In the example of FIG. 6, the presence or absence of a scene change between the intra pictures of random access points in adjacent GOPs is evaluated.
In addition to the above, candidate picture definitions are optionally set, such as selecting an evaluation target picture at N (N is an integer of 0 or more) picture intervals (in the case of N = 0, the same as in FIG. 5). can do.
 図7は、符号化方式がAVC/H.264又はHEVC/H.265である場合の符号化映像ストリームの構成例を示す説明図である。
 図7では、複数のアクセスユニットが符号化順に並んでおり、1つのアクセスユニットが1つのピクチャのデータを示している。
 各々のアクセスユニットは、複数のNAL(Network Abst ruction Layer)ユニットから構成される。
 各々のNALユニットは、ヘッダ情報(non-VCL)と、ピクチャの符号化データであるピクチャデータ(VCL)とに分類される。
 ヘッダ情報(non-VCL)は、アクセスユニットデリミタ、シーケンスレベルヘッダ及びピクチャレベルヘッダを有している。
 AVC/H.264又はHEVC/H.265では、1つのアクセスユニットを構成する1つ以上のNALユニットが連続しており、ストリーム上で隣接するアクセスユニット間の区切り位置が分かるように、アクセスユニットの先頭位置の識別方法が定義されている。その一つの例として、アクセスユニットの先頭を示すNALユニットであるアクセスユニットデリミタが定義されている。
 なお、図7に示すNALユニットは一例であり、NALユニットの構成パターンは、それぞれの規格の仕様に準ずる。このとき、アクセスユニットを構成するNALユニットにアクセスユニットデリミタが存在しない場合もあり、その場合においても、規格で定義されたアクセスユニットの先頭位置の識別方法に従ってアクセスユニットの先頭位置を識別することが可能である。
FIG. 7 shows that the coding method is AVC / H. H.264 or HEVC / H. It is explanatory drawing which shows the structural example of the encoding video stream in the case of being H.265.
In FIG. 7, a plurality of access units are arranged in coding order, and one access unit indicates data of one picture.
Each access unit is composed of a plurality of NAL (Network Abst Ruction Layer) units.
Each NAL unit is classified into header information (non-VCL) and picture data (VCL) which is coded data of a picture.
The header information (non-VCL) includes an access unit delimiter, a sequence level header and a picture level header.
AVC / H. H.264 or HEVC / H. At 265, one or more NAL units constituting one access unit are continuous, and a method of identifying the head position of the access unit is defined so that the division position between adjacent access units on the stream can be known. There is. As one example, an access unit delimiter which is an NAL unit indicating the beginning of an access unit is defined.
The NAL unit illustrated in FIG. 7 is an example, and the configuration pattern of the NAL unit conforms to the specifications of the respective standards. At this time, there are cases where no access unit delimiter exists in the NAL unit that constitutes the access unit, and even in that case, the head position of the access unit should be identified according to the identification method of the head position of the access unit defined in the standard. It is possible.
 したがって、ピクチャ選択部1は、規格で定義されているアクセスユニットの先頭位置の識別方法に従ってアクセスユニットの先頭位置を識別することで、ピクチャであるアクセスユニットの区切り位置を識別することができる。
 ピクチャ選択部1は、アクセスユニットの区切り位置を識別することができるため、複数のアクセスユニットの中から、2つのピクチャを選択することができ、選択した2つのピクチャの符号化データを特徴量算出部2に出力する。
Therefore, the picture selection unit 1 can identify the break position of the access unit which is a picture by identifying the start position of the access unit according to the identification method of the start position of the access unit defined in the standard.
Since the picture selection unit 1 can identify the break position of the access unit, it can select two pictures from among a plurality of access units, and calculates the feature amount of the encoded data of the selected two pictures Output to part 2.
 特徴量算出部2は、ピクチャ選択部1により選択されたピクチャ毎に、符号化ブロック単位の符号化データから、処理ブロック単位の特徴量をそれぞれ算出する(図4のステップST2)。
 以下、特徴量算出部2による処理ブロック単位の特徴量の算出処理を具体的に説明する。
 ここで、ピクチャ選択部1により選択されたピクチャ番号tのピクチャPに含まれる処理ブロックの特徴量をCn,tとし、参照ピクチャのピクチャ番号をt=tp、評価対象ピクチャのピクチャ番号をt=tcとする。
 nは、ピクチャ内の特徴量算出対象の処理ブロックを特定するブロック番号であり、左上の処理ブロックからラスタスキャン順に1,2,・・・,NUM_BLと番号付けされている。ここで、NUM_BLは、ピクチャ番号tのピクチャPにおける特徴量算出対象の処理ブロックの総数を示している。処理ブロックサイズが64×64画素、ピクチャの解像度が3840×2160画素であった場合、1画面の全てを対象としたとすると、NUM_BL=60×34=2040となる。ただし、最下端の列のブロックはブロックサイズが64×48画素となる。
 よって、t=tp及びt=tcにおける特徴量Cn,tp、Cn,tcをそれぞれ下記に従って算出する。
The feature quantity calculation unit 2 calculates feature quantities in processing block units from the encoded data in coding block units for each of the pictures selected by the picture selection unit 1 (step ST2 in FIG. 4).
The process of calculating the feature amount in units of processing blocks by the feature amount calculation unit 2 will be specifically described below.
Here, the feature amount of the processing block included in the picture P t of the picture number t selected by the picture selection unit 1 is C n, t , the picture number of the reference picture is t = tp, and the picture number of the evaluation target picture is Let t = tc.
n is a block number for specifying a processing block to be subjected to feature amount calculation in a picture, and is numbered 1, 2,..., NUM_BL t in raster scan order from the upper left processing block. Here, NUM_BL t indicates the total number of processing blocks for feature amount calculation in the picture P t of the picture number t. If the processing block size is 64 × 64 pixels and the resolution of the picture is 3840 × 2160 pixels, NUM_BL t = 60 × 34 = 2040, assuming that the whole of one screen is targeted. However, the block in the lowermost row has a block size of 64 × 48 pixels.
Therefore, feature quantities C n, tp and C n, tc at t = tp and t = tc are respectively calculated according to the following.
 特徴量算出部2は、特徴量算出対象ピクチャPにおけるシンタックスの復号処理を実施し、シンタックスの復号処理結果から、処理ブロック単位の符号量Sn,t(n=1,2,・・・,NUM_BL)をそれぞれ特定する。
 例えば、特徴量算出部2は、特徴量算出対象の処理ブロックに属する各々の符号化ブロックの符号化データに含まれている全ての符号化パラメータをそれぞれ復号し、復号した全ての符号化パラメータの符号量の総和を符号量Sn,tとして特定する。
 符号化パラメータは、符号化ブロックの復号画像の生成に必要なパラメータであり、例えば、符号化モードを示すパラメータ、イントラ予測パラメータ、インター予測パラメータ、予測差分符号化パラメータ、動きベクトルなどが該当する。
Feature amount calculation unit 2, implemented decoding syntax in the feature quantity calculation target picture P t, from the decoding processing result of the syntax, the code amount S n of the processing block, t (n = 1,2, ·・ ・, NUM_BL t ) is specified respectively.
For example, the feature amount calculation unit 2 decodes each of all the encoding parameters included in the encoded data of each encoding block belonging to the processing block for which the feature amount is to be calculated, The sum of code amounts is specified as a code amount Sn, t .
The coding parameter is a parameter necessary for generating a decoded image of a coding block, and corresponds to, for example, a parameter indicating a coding mode, an intra prediction parameter, an inter prediction parameter, a prediction difference coding parameter, a motion vector, and the like.
 次に、特徴量算出部2は、特徴量算出対象ピクチャPの符号量SALL,tを特定し、以下の式(1)に示すように、特徴量算出対象ピクチャPの符号量SALL,tで、処理ブロック単位の符号量Sn,tを除算することで正規化し、処理ブロック単位の特徴量Cn,tとして、正規化した処理ブロック単位の符号量Sn,tをM倍する。Mは、予め設定した定数である。
Figure JPOXMLDOC01-appb-I000001
Next, the feature amount calculating unit 2, the code amount S ALL feature quantity calculation target picture P t, identifies t, as shown in the following equation (1), the feature quantity calculation target picture P t code amount S ALL, in t, the code amount S n of the processing blocks, normalized by dividing t, feature amount C n of the processing block, as t, the code amount S n of normalized processing block units, the t M Double. M is a preset constant.
Figure JPOXMLDOC01-appb-I000001
 ここでは、特徴量算出部2が、特徴量算出対象ピクチャPの符号量SALL,tを特定する例を示しているが、特徴量算出対象ピクチャPの符号量SALL,tとして、特徴量算出対象ピクチャPに含まれている全ての処理ブロックの符号量Sn,tの総和を算出する例が考えられる。
 ただし、これは一例に過ぎず、特徴量算出対象ピクチャPの符号量SALL,tとして、特徴量算出対象ピクチャPのピクチャデータ(VCL)のサイズを特定するようにしてもよい。ピクチャデータ(VCL)のサイズは、特徴量算出対象ピクチャPにおけるシンタックスの復号処理を実施することなく、特定することができる。
Here, the feature amount calculating section 2, the code amount S ALL feature quantity calculation target picture P t, is shown an example of specifying a t, the feature quantity calculation target picture P t code amount S ALL, as t, code amount S n of all of the processing blocks included in the feature quantity calculation target picture P t, example of calculating the sum of t is considered.
However, this is only an example, the code amount S ALL feature quantity calculation target picture P t, as t, the size of the picture data (VCL) of the feature quantity calculation target picture P t may be specified. The size of the picture data (VCL), without performing the decoding process of syntax in the feature quantity calculation target picture P t, can be identified.
 式(1)で算出される処理ブロック単位の特徴量Cn,tは、各々の処理ブロックの符号量の比を表しており、ピクチャ毎の総符号量の大きさのばらつきの影響を抑えた特徴量になっている。
 図8は、処理ブロック単位の符号量の一例を示す説明図である。
 図8は、処理ブロックが2×2個の符号化ブロックから構成され、符号化ブロックがマクロブロック又はCTUである例を示しており、各々の符号化ブロックの符号量を数字で表している。さらに、各処理ブロックの中央に、当該処理ブロックに属する符号化ブロックの符号量の総和を当該処理ブロックの符号量として表している。
 図8の例では、符号化ブロックの符号量の単位をbyteで表しているが、一例に過ぎず、bit単位で表されていてもよい。
The feature amount C n, t of the processing block unit calculated by the equation (1) represents the ratio of the code amount of each processing block, and the influence of the variation of the size of the total code amount for each picture is suppressed It is a feature amount.
FIG. 8 is an explanatory diagram of an example of the code amount in processing block units.
FIG. 8 shows an example in which the processing block is composed of 2 × 2 coded blocks and the coded block is a macroblock or CTU, and the code amount of each coded block is indicated by a numeral. Furthermore, at the center of each processing block, the sum of the code amounts of the coding blocks belonging to the processing block is represented as the code amount of the processing block.
Although the unit of the code amount of the coding block is represented by bytes in the example of FIG. 8, this is merely an example, and may be represented by bit units.
 評価値算出部3は、特徴量算出部2により算出された参照ピクチャPtpにおける処理ブロック単位の特徴量Cn,tpと、評価対象ピクチャPtcにおける処理ブロック単位の特徴量Cn,tcとから、シーンチェンジの有無の判定に用いる評価値Jを算出する(図4のステップST3)。
 例えば、評価値算出部3は、シーンチェンジの有無の判定に用いる評価値Jとして、以下の式(2)に示すように、参照ピクチャPtpと評価対象ピクチャPtcとの間での処理ブロック単位の特徴量の差分絶対値和を算出する。
Figure JPOXMLDOC01-appb-I000002
 評価値算出部3により算出された評価値Jは、参照ピクチャPtpと評価対象ピクチャPtcとの間での絵柄の変化が大きい程、参照ピクチャPtpと評価対象ピクチャPtcの処理ブロック単位の符号量の分布の差異が大きくなり、値が大きくなる。
 また、MIN_NUM_BLは、NUM_BLtpとNUM_BLtcの最小値である。即ち、参照ピクチャPtpと評価対象ピクチャPtcの処理ブロック数が異なる場合、両ピクチャ共に特徴量を算出している処理ブロックのみが評価値Jの算出対象となる。
例えば、参照ピクチャPtpと評価対象ピクチャPtcの符号化データのサイズが大きく異なる場合、復号処理時間も大きく異なる傾向にある。そのようなケースにて、各ピクチャの符号化データの復号処理や処理ブロック単位の特徴量の算出に割かれる時間が一律に制限されている場合、処理可能な処理ブロック数が両ピクチャで異なる結果となる。このような場合、評価値Jは、両ピクチャ共に特徴量を算出している処理ブロックのみを対象として求められる。なお、上記のように各々のピクチャで消費できる処理時間Tを決定することで、所望の処理速度(フレームレート1/T)での特徴量算出処理が可能となる。
Evaluation value calculation unit 3, the feature amount C n of the processing block in the reference picture P tp calculated by the feature amount calculation unit 2, and tp, the feature amount C n of the processing block in the evaluation object picture P tc, and tc Then, the evaluation value J used to determine the presence or absence of a scene change is calculated (step ST3 in FIG. 4).
For example, as the evaluation value J used to determine the presence or absence of a scene change, the evaluation value calculation unit 3 performs processing blocks between the reference picture P tp and the evaluation target picture P tc as shown in the following equation (2). Calculate the sum of absolute differences of the feature amounts of the unit.
Figure JPOXMLDOC01-appb-I000002
Evaluation value J calculated by the evaluation value calculating section 3, as the pattern of change between the reference picture P tp evaluated picture P tc is large, the reference process block units of picture P tp evaluated picture P tc The difference in the distribution of the code amount of increases, and the value increases.
Also, MIN_NUM_BL is the minimum value of NUM_BL tp and NUM_BL tc . That is, when the number of processing blocks of the reference picture P tp and the evaluation target picture P tc is different, only the processing block whose feature amount is calculated for both pictures is the calculation target of the evaluation value J.
For example, when the sizes of the encoded data of the reference picture P tp and the evaluation target picture P tc are largely different, the decoding processing time also tends to be largely different. In such a case, when the time allocated to decoding processing of the encoded data of each picture and calculation of the feature amount in processing block units is uniformly limited, the result that the number of process blocks that can be processed is different in both pictures It becomes. In such a case, the evaluation value J can be obtained for only the processing block whose feature amount has been calculated for both pictures. By determining the processing time T that can be consumed in each picture as described above, feature amount calculation processing at a desired processing speed (frame rate 1 / T) becomes possible.
 有無判定部4は、評価値算出部3により算出された評価値Jと、事前に設定された閾値Thとを比較する(図4のステップST4)。
 有無判定部4は、評価値算出部3により算出された評価値Jが閾値Th以上であれば(図4のステップST4:YESの場合)、参照ピクチャPtpと評価対象ピクチャPtcとの間にシーンチェンジが有ると判定する(図4のステップST5)。
 有無判定部4は、評価値算出部3により算出された評価値Jが閾値Th未満であれば(図4のステップST4:NOの場合)、参照ピクチャPtpと評価対象ピクチャPtcとの間にシーンチェンジが無いと判定する(図4のステップST6)。
 有無判定部4は、シーンチェンジの有無の判定結果を出力する。
 なお、閾値Thは、予め設定する固定値であってもよいし、特定の条件に従って切り替えたり、変化したりするものであってもよい。例えば、コンテンツの種類(ドラマ、ニュース、スポーツ等)毎に閾値Thを用意して切り替えるようにする方法、放送局毎に閾値Thを用意して切り替えるようにする方法、ストリームを生成するエンコーダ毎に閾値Thを用意して切り替えるようにする方法、ストリームを生成するエンコーダの設定に従って適応的に閾値Thを算出する(エンコーダの設定値を変数とした閾値Thの算出式を用意する)方法、ストリームの符号量推移に従って適応的に閾値Thを算出する(符号量の時間変化(時間方向の1~n次微分、nは1以上の整数)を変数とした閾値Thの算出式を用意する)方法などが考えられる。
The presence / absence determination unit 4 compares the evaluation value J calculated by the evaluation value calculation unit 3 with the threshold value Th set in advance (step ST4 in FIG. 4).
If the evaluation value J calculated by the evaluation value calculation unit 3 is equal to or greater than the threshold Th (if step ST4 in FIG. 4: YES), the presence / absence determination unit 4 determines between the reference picture Ptp and the evaluation target picture Ptc. It is determined that there is a scene change (step ST5 in FIG. 4).
If the evaluation value J calculated by the evaluation value calculation unit 3 is less than the threshold Th (in the case of NO in step ST4 of FIG. 4), the presence / absence determination unit 4 selects between the reference picture Ptp and the evaluation target picture Ptc. It is determined that there is no scene change (step ST6 in FIG. 4).
The presence determination unit 4 outputs the determination result of the presence or absence of a scene change.
Note that the threshold Th may be a fixed value set in advance, or may be switched or changed according to a specific condition. For example, a method of preparing and switching the threshold Th for each type of content (drama, news, sports, etc.), a method of preparing and switching the threshold Th for each broadcasting station, and for each encoder generating a stream A method of preparing and switching the threshold Th, a method of adaptively calculating the threshold Th according to the setting of the encoder that generates the stream (a method of calculating the threshold Th with the setting value of the encoder as a variable), Method of calculating threshold Th adaptively according to code amount transition (Prepare a formula for calculating threshold Th with time change of code amount (1st to nth differential in the time direction, n is an integer of 1 or more) as a variable) Is considered.
 ピクチャ選択部1により決定された候補ピクチャの中に、未だ評価対象ピクチャとして選択していないピクチャが残っていれば(図4のステップST7:YESの場合)、ステップST1の処理に戻り、ステップST1~ST7の処理が繰り返される。
 なお、戻ったステップST1の処理で、ピクチャ選択部1により選択された2つのピクチャのうち、特徴量が既に算出されているピクチャについては、特徴量算出部2による特徴量の算出処理を省略するようにしてもよい。このようにすることで、特徴量の再計算を実施せずに済み、処理時間を短縮することができる。例えば、図5、図6で説明したように、候補ピクチャの中から、再生順に評価対象ピクチャを選択し、参照ピクチャを評価対象ピクチャより再生順で一つ前の候補ピクチャとする場合、評価対象ピクチャにおける特徴量Cn,tcは、再生順で一つ先の候補ピクチャを評価対象ピクチャに選択した際の参照ピクチャにおける特徴量Cn,tpとなる。したがって、算出した特徴量Cn,tcを順次保存しておき、特徴量Cn,tpは、保存された特徴量から読み出すようにすることで特徴量の算出処理を省略することができる。
 また、復号順と再生順が異なる場合、復号順に評価対象ピクチャを選択し、参照ピクチャを評価対象ピクチャより復号順で前かつ再生順で最も評価対象ピクチャに近い候補ピクチャとしてシーンチェンジの有無を評価してもよい。この場合においても、特徴量が既に算出されているピクチャは、特徴量の算出結果を流用して処理を省略するようにしてもよい。
 ピクチャ選択部1により決定された候補ピクチャの中に、未だ評価対象ピクチャとして選択していない候補ピクチャが残っていなければ(図4のステップST7:NOの場合)、一連の処理が終了する。
If there is still a picture not selected as an evaluation target picture among the candidate pictures determined by the picture selection unit 1 (step ST7 in FIG. 4: YES), the process returns to step ST1 and step ST1. The processing of ̃ST7 is repeated.
In the process of step ST1 returned, the process of calculating the feature amount by the feature amount calculating unit 2 is omitted for the picture of which the feature amount is already calculated among the two pictures selected by the picture selecting unit 1 You may do so. By doing this, it is not necessary to recalculate feature amounts, and processing time can be shortened. For example, as described in FIG. 5 and FIG. 6, when the evaluation target picture is selected from among the candidate pictures in the reproduction order and the reference picture is made one candidate picture in reproduction order from the evaluation target picture, the evaluation target The feature amount C n, tc in the picture is the feature amount C n, tp in the reference picture when the candidate picture one ahead in the reproduction order is selected as the evaluation target picture. Therefore, the feature amount calculation process can be omitted by sequentially storing the calculated feature amounts C n and tc and reading out the feature amounts C n and tp from the stored feature amounts.
In addition, when the decoding order and the reproduction order are different, the evaluation target picture is selected in the decoding order, and the presence or absence of a scene change is evaluated as a candidate picture closest to the evaluation target picture in decoding order before the reference object You may Also in this case, for the picture whose feature amount has already been calculated, the process may be omitted by diverting the calculation result of the feature amount.
If no candidate picture not selected as the evaluation target picture remains in the candidate pictures determined by the picture selection unit 1 (in the case of NO at step ST7 in FIG. 4), a series of processing ends.
 以上で明らかなように、この実施の形態1によれば、特徴量算出部2により算出された2つのピクチャに含まれるブロック単位の特徴量から、シーンチェンジの有無の判定に用いる評価値を算出する評価値算出部3を設け、有無判定部4が、評価値算出部3により算出された評価値と閾値を比較することで、2つのピクチャの間でのシーンチェンジの有無を判定するように構成したので、動きベクトルが含まれていないピクチャについても、シーンチェンジの有無を判定することができる効果を奏する。 As apparent from the above, according to the first embodiment, the evaluation value used to determine the presence or absence of a scene change is calculated from the feature amount in units of blocks included in the two pictures calculated by the feature amount calculation unit 2 The evaluation value calculation unit 3 is provided, and the presence / absence determination unit 4 determines the presence / absence of a scene change between two pictures by comparing the evaluation value calculated by the evaluation value calculation unit 3 with a threshold. Since the configuration is made, it is possible to determine the presence or absence of a scene change even for a picture that does not include a motion vector.
 この実施の形態1では、符号化映像ストリームが画像処理装置のピクチャ選択部1に与えられる例を示しているが、符号化映像ストリームを格納するメディア伝送ストリームが画像処理装置に与えられることがある。
 この場合、図9に示すように、メディア伝送ストリームから符号化映像ストリームを取り出して、符号化映像ストリームをピクチャ選択部1に出力するデマルチプレクサ5を備えるようにしてもよい。
 図9は、この発明の実施の形態1による他の画像処理装置を示す構成図である。
 メディア伝送ストリームとしては、例えば、MPEG-2 TS(ISO/IEC 13818-1/ITU-T H.222.0のTransport Stream)、あるいは、MMT(ISO/IEC 23008-1)などのメディア伝送フォーマットでパケット化されているメディア伝送ストリームなどが想定される。
Although the first embodiment shows an example in which the coded video stream is given to the picture selection unit 1 of the image processing apparatus, a media transport stream storing the coded video stream may be given to the image processing apparatus. .
In this case, as shown in FIG. 9, the demultiplexer 5 may be provided to extract the encoded video stream from the media transmission stream and output the encoded video stream to the picture selection unit 1.
FIG. 9 is a block diagram showing another image processing apparatus according to Embodiment 1 of the present invention.
The media transmission stream may be, for example, a media transmission format such as MPEG-2 TS (Transport Stream of ISO / IEC 13818-1 / ITU-T H.222.0) or MMT (ISO / IEC 23008-1). A packetized media transport stream is assumed.
 デマルチプレクサ5は、与えられたメディア伝送ストリームのヘッダ情報から、各々のアクセスユニットの間の区切り、NALユニットの種類及びサイズを知ることができる。
 デマルチプレクサ5が、各々のアクセスユニットの間の区切りを示す情報をピクチャ選択部1に与えるようにすれば、ピクチャ選択部1では、デマルチプレクサ5から与えられた情報に基づいて、符号化映像ストリームの中から所望のピクチャを選択することができる。
 また、デマルチプレクサ5が、NALユニットの種類及びサイズを示す情報を特徴量算出部2に与えるようにすれば、特徴量算出部2では、評価対象ピクチャにおけるシンタックスの復号処理を実施することなく、上記評価対象ピクチャの符号量を特定することができる。
The demultiplexer 5 can know the division between each access unit and the type and size of the NAL unit from the header information of the given media transport stream.
If the demultiplexer 5 gives information indicating the division between each access unit to the picture selection unit 1, the picture selection unit 1 can encode the encoded video stream based on the information provided from the demultiplexer 5. A desired picture can be selected from
In addition, if the demultiplexer 5 supplies information indicating the type and size of the NAL unit to the feature amount calculation unit 2, the feature amount calculation unit 2 does not perform the decoding process of the syntax in the evaluation target picture. The code amount of the evaluation target picture can be specified.
 この実施の形態1では、特徴量算出部2が、処理ブロックの符号量Sn,tの算出処理において、ピクチャ全体の処理ブロックを対象とする例を示している。
 これは一例に過ぎず、特徴量算出部2の処理負荷を軽減するために、処理ブロックの符号量Sn,tの算出処理において、ピクチャ内の一部の処理ブロックだけを対象とするようにしてもよい。
In the first embodiment, an example is shown in which the feature amount calculation unit 2 targets the processing block of the whole picture in the calculation processing of the code amount Sn, t of the processing block.
This is merely an example, and in order to reduce the processing load of the feature amount calculation unit 2, only a part of processing blocks in a picture are targeted in the calculation process of the code amount Sn, t of the processing block. May be
 具体的には、図10に示すように、行方向及び列方向のそれぞれで、1つおきの処理ブロックだけを対象とするようにしてもよい。
 図10は、ピクチャを構成する処理ブロックのうち、符号量を特定する処理ブロックを示す説明図である。
 図10において、“1”が記述されている処理ブロックは、符号量を特定する処理ブロックであり、“0”が記述されている処理ブロックは、符号量を特定しない処理ブロックである。
Specifically, as shown in FIG. 10, only every other processing block may be targeted in each of the row direction and the column direction.
FIG. 10 is an explanatory diagram of a processing block for specifying a code amount among processing blocks constituting a picture.
In FIG. 10, the processing block in which “1” is described is a processing block that specifies a code amount, and the processing block in which “0” is described is a processing block that does not specify a code amount.
 また、図11に示すように、1行おきの処理ブロック、あるいは、1列おきの処理ブロックだけを対象とするようにしてもよい。
 図11は、ピクチャを構成する処理ブロックのうち、符号量を特定する処理ブロックを示す説明図である。
 図11において、“1”が記述されている処理ブロックは、符号量を特定する処理ブロックであり、“0”が記述されている処理ブロックは、符号量を特定しない処理ブロックである。
Further, as shown in FIG. 11, processing blocks every other row or processing blocks every other column may be targeted.
FIG. 11 is an explanatory view showing a processing block for specifying a code amount among processing blocks constituting a picture.
In FIG. 11, the processing block in which “1” is described is a processing block that specifies the code amount, and the processing block in which “0” is described is a processing block that does not specify the code amount.
 図12に示すように、処理ブロックの符号量Sn,tの算出処理において、ピクチャの左上の処理ブロックからラスタスキャン順に特定の処理ブロックまでを対象とし、残りの処理ブロックは対象としないようにしてもよい。
 図12は、ピクチャを構成する処理ブロックのうち、符号量を特定する処理ブロックを示す説明図である。
 図12において、“1”が記述されている処理ブロックは、符号量を特定する処理ブロックであり、“0”が記述されている処理ブロックは、符号量を特定しない処理ブロックである。
 なお、図10から図12において、“1”が記述されている処理ブロックと、“0”が記述されている処理ブロックとが逆であってもよい。
As shown in FIG. 12, in the process of calculating the code amount S n, t of the processing block, the processing block from the upper left of the picture to the specific processing block in raster scan order is targeted, and the remaining processing blocks are not targeted. May be
FIG. 12 is an explanatory view showing a processing block for specifying a code amount among processing blocks constituting a picture.
In FIG. 12, the processing block in which “1” is described is a processing block that specifies a code amount, and the processing block in which “0” is described is a processing block that does not specify a code amount.
In FIGS. 10 to 12, the processing block in which “1” is described and the processing block in which “0” is described may be reversed.
実施の形態2.
 上記実施の形態1では、特徴量算出部2が、符号化ブロックの符号化データを復号して、符号化データの復号結果から処理ブロックの符号量を特定し、処理ブロックの符号量から、処理ブロックの特徴量を算出する例を示している。
 この実施の形態2では、特徴量算出部2が、ピクチャ選択部1により選択されたピクチャ番号tのピクチャPの処理ブロック毎に、当該処理ブロックの特徴量Cn,tとして、当該処理ブロックに含まれている分割ブロックの個数を用いるようにしてもよい。
Second Embodiment
In the first embodiment, the feature quantity calculation unit 2 decodes the encoded data of the encoded block, specifies the code amount of the processing block from the decoding result of the encoded data, and processes from the code amount of the processing block The example which calculates the feature-value of a block is shown.
In the second embodiment, for each processing block of the picture P t of the picture number t selected by the picture selection unit 1, the feature quantity calculation unit 2 sets the processing block as the feature quantity C n, t of the processing block. The number of divided blocks included in may be used.
 当該処理ブロックに含まれている分割ブロックの個数は、当該処理ブロックに属する各符号化ブロックに含まれている分割ブロックの個数を加算することで算出される。
 符号化ブロックに含まれている分割ブロックの個数は、例えば、符号化方式がAVC/H.264であれば、マクロブロックタイプ(mb_type)で決まるマクロブロック内のブロック数に相当する。マクロブロック内のブロック数は、評価対象ピクチャにおけるシンタックスの復号処理を実施することで得られる。
 符号化方式がHEVC/H.265であれば、符号化ブロックであるCTU内のCU(Coding Unit)の数に相当する。CTU内のCUの数は、評価対象ピクチャにおけるシンタックスの復号処理を実施することで得られる。
 処理ブロックの特徴量として、当該処理ブロックに含まれている分割ブロックの個数を特定する場合、特徴量算出部2が、処理ブロックの特徴量を算出する際に符号量を正規化する処理が不要になり、特徴量算出部2の処理負荷が軽減される。
The number of divided blocks included in the processing block is calculated by adding the number of divided blocks included in each coding block belonging to the processing block.
The number of divided blocks included in the coding block is, for example, the coding method of AVC / H. H.264 corresponds to the number of blocks in a macroblock determined by the macroblock type (mb_type). The number of blocks in a macroblock is obtained by performing syntax decoding processing on a picture to be evaluated.
The coding method is HEVC / H. If it is 265, it corresponds to the number of CUs (Coding Units) in the CTU which is a coding block. The number of CUs in a CTU can be obtained by performing syntax decoding processing on a picture to be evaluated.
When specifying the number of divided blocks included in the processing block as the feature amount of the processing block, the feature amount calculating unit 2 does not need to perform processing for normalizing the code amount when calculating the feature amount of the processing block As a result, the processing load of the feature amount calculation unit 2 is reduced.
実施の形態3.
 上記実施の形態1では、特徴量算出部2が、符号化ブロックの符号化データを復号して、符号化データの復号結果から処理ブロックの符号量を特定し、処理ブロックの符号量から、処理ブロックの特徴量を算出する例を示している。
 この実施の形態3では、特徴量算出部2が、ピクチャ選択部1により選択されたピクチャ番号tのピクチャPの処理ブロック毎に、当該処理ブロックの特徴量Cn,tとして、当該処理ブロックにおける分割ブロックの最深の階層数を用いるようにしてもよい。
Third Embodiment
In the first embodiment, the feature quantity calculation unit 2 decodes the encoded data of the encoded block, specifies the code amount of the processing block from the decoding result of the encoded data, and processes from the code amount of the processing block The example which calculates the feature-value of a block is shown.
In the third embodiment, for each processing block of the picture P t of the picture number t selected by the picture selection unit 1, the feature quantity calculation unit 2 sets the processing block as the feature quantity C n, t of the processing block. The highest hierarchical number of divided blocks in the above may be used.
 当該処理ブロックにおける分割ブロックの最深の階層数は、当該処理ブロックに属する各符号化ブロックにおける分割ブロックの最深の階層数の最大値を求めることで算出される。
 符号化ブロックにおける分割ブロックの最深の階層数は、例えば、符号化方式がHEVC/H.265であれば、図13に示すように、符号化ブロックであるCTUに含まれている複数のCUのうち、分割の階層が最も深い数に相当する。
 図13は、符号化ブロックにおける分割ブロックの階層数を示す説明図である。
 図13の例では、CU depthの最大値が3であるため、符号化ブロックにおける分割ブロックの最深の階層数は、3である。
 分割ブロックの最深の階層数は、評価対象ピクチャにおけるシンタックスの復号処理を実施することで得られる。
 処理ブロックの特徴量として、当該処理ブロックにおける分割ブロックの最深の階層数を特定する場合、特徴量算出部2が、処理ブロックの特徴量を算出する際に符号量を正規化する処理が不要になり、特徴量算出部2の処理負荷が軽減される。
The deepest hierarchy number of divided blocks in the processing block is calculated by obtaining the maximum value of the deepest hierarchy number of divided blocks in each coding block belonging to the processing block.
For example, the number of layers in the deepest layer in the coding block is the coding scheme HEVC / H. If it is 265, as shown in FIG. 13, the hierarchy of division corresponds to the deepest number among the plurality of CUs included in the CTU which is a coding block.
FIG. 13 is an explanatory drawing showing the number of layers of divided blocks in a coding block.
In the example of FIG. 13, since the maximum value of CU depth is 3, the number of deepest layers in the divided block in the coding block is three.
The deepest hierarchy number of the divided blocks can be obtained by performing syntax decoding processing on the evaluation target picture.
When specifying the number of deepest layers of the divided block in the processing block as the feature amount of the processing block, the feature amount calculation unit 2 does not need to perform processing for normalizing the code amount when calculating the feature amount of the processing block. As a result, the processing load of the feature amount calculation unit 2 is reduced.
実施の形態4.
 上記実施の形態では、評価値算出部3が、特徴量算出部2により算出された評価対象ピクチャPtpにおける処理ブロック単位の特徴量Cn,tpと、評価対象ピクチャPtcにおける処理ブロック単位の特徴量Cn,tcとから、評価値Jを算出する例を示している。
 この実施の形態4では、評価値算出部3が、ピクチャ番号t=tpの参照ピクチャPtpにおける処理ブロック単位の特徴量Cn,tpと、ピクチャ番号t=tcの評価対象ピクチャPtcにおける処理ブロック単位の特徴量Cn,tcと、候補ピクチャの中で参照ピクチャPtpよりも再生時刻的に1つ前のピクチャ番号t=tp’の二次参照ピクチャPtp’における処理ブロック単位の特徴量Cn,tp’とから、評価値Jを算出する例を説明する。
 なお、二次参照ピクチャPtp’は参照ピクチャPtpの参照ピクチャ、即ち、参照ピクチャPtpが評価対象ピクチャPtcであったときの参照ピクチャPtpである。したがって、3つのピクチャの再生時刻的位置関係はtp’<tp<tcであり、tc-tp=tp-tp’である。
Fourth Embodiment
In the above embodiment, the evaluation value calculation unit 3 calculates the feature amount C n, tp of the processing block unit in the evaluation target picture P tp calculated by the feature amount calculation unit 2 and the processing block unit in the evaluation target picture P tc An example in which the evaluation value J is calculated from the feature amounts C n and t c is shown.
In the fourth embodiment, the evaluation value calculation unit 3 processes the feature amount C n, tp of the processing block unit in the reference picture P tp of the picture number t = tp and the evaluation target picture P tc of the picture number t = tc. Features of processing block unit in the feature quantity C n, tc of block unit and secondary reference picture P tp ' of picture number t = tp' one time before the reference picture P tp among the candidate pictures in reproduction time An example of calculating the evaluation value J from the amounts C n and tp ′ will be described.
Incidentally, the secondary reference picture P tp 'is the reference picture of the reference picture P tp, i.e., a reference picture P tp when the reference picture P tp was evaluated picture P tc. Therefore, the reproduction time positional relationship of the three pictures is tp '<tp <tc, and tc-tp = tp-tp'.
 次に動作について説明する。
 この実施の形態4では、上記実施の形態1と同様に、図4のフローチャートに従って処理されるが、処理内容は、下記に説明する処理に変更となる。
Next, the operation will be described.
In the fourth embodiment, processing is performed according to the flowchart of FIG. 4 as in the first embodiment, but the processing content is changed to the processing described below.
 ピクチャ選択部1は、上記実施の形態1と同様に、複数のピクチャの符号化データが含まれている符号化映像ストリームを取得する。
 ピクチャ選択部1は、符号化データが符号化映像ストリームに含まれている複数のピクチャの中から、予め定義された候補ピクチャの決定方法に従って候補ピクチャを決定し、候補ピクチャの中から、上記実施の形態1と同様に、参照ピクチャPtp及び評価対象ピクチャPtcを選択する。
 また、ピクチャ選択部1は、候補ピクチャの中から、二次参照ピクチャPtp’を選択する。即ち、図4のステップST1では、評価対象ピクチャと参照ピクチャに加えて、二次参照ピクチャも選択する。
As in the first embodiment, the picture selection unit 1 obtains a coded video stream including coded data of a plurality of pictures.
The picture selection unit 1 determines a candidate picture from among a plurality of pictures in which the encoded data is included in the encoded video stream according to a method for determining a candidate picture defined in advance, and performs the above implementation from among the candidate pictures. Similar to mode 1, the reference picture P tp and the evaluation target picture P tc are selected.
Also, the picture selection unit 1 selects a secondary reference picture Ptp ′ from the candidate pictures. That is, in step ST1 of FIG. 4, in addition to the evaluation target picture and the reference picture, the secondary reference picture is also selected.
 特徴量算出部2は、上記実施の形態1と同様に、ピクチャ選択部1により選択されたピクチャ毎に、当該評価対象ピクチャに含まれる符号化ブロック単位の符号化データから、処理ブロック単位の特徴量をそれぞれ算出する(図4のステップST2)。
 即ち、特徴量算出部2は、参照ピクチャPtpに含まれる処理ブロック単位の特徴量Cn,tpと、評価対象ピクチャPtcに含まれる処理ブロック単位の特徴量Cn,tcとを算出する。
 また、特徴量算出部2は、二次参照ピクチャPtp’に含まれる処理ブロック単位の特徴量Cn,tp’を算出する。
As in the first embodiment, the feature quantity calculation unit 2 calculates, for each of the pictures selected by the picture selection unit 1, the feature of processing block unit from the encoded data of coding block unit included in the evaluation target picture. The quantities are calculated respectively (step ST2 in FIG. 4).
That is, the feature amount calculation unit 2 calculates the feature amount C n of the processing blocks included in the reference picture P tp, and tp, the feature amount C n of the processing blocks included in the evaluation object picture P tc, and tc .
Also, the feature quantity calculation unit 2 calculates feature quantities C n, tp ′ in units of processing blocks included in the secondary reference picture Ptp ′ .
 評価値算出部3は、特徴量算出部2により算出された参照ピクチャPtpに含まれる処理ブロック単位の特徴量Cn,tpと、評価対象ピクチャPtcに含まれる処理ブロック単位の特徴量Cn,tcと、二次参照ピクチャPtp’に含まれる処理ブロック単位の特徴量Cn,tp’とから、評価値Jを算出する(図4のステップST3)。
 具体的には、評価値算出部3は、参照ピクチャPtpと評価対象ピクチャPtcとの間での処理ブロック単位の特徴量の差分絶対値和を算出する。
 また、評価値算出部3は、参照ピクチャPtpと二次参照ピクチャPtp’との間での処理ブロック単位の特徴量の差分絶対値和を算出する。
 評価値算出部3は、評価値Jとして、以下の式(3)に示すように、算出した双方の差分絶対値和の差分を算出する。
Figure JPOXMLDOC01-appb-I000003
 ここで、MIN_NUM_BL’は、NUM_BLtp’、NUM_BLtp、NUM_BLtcの3つの値の最小値である。即ち、二次参照ピクチャPtp’と参照ピクチャPtpと評価対象ピクチャPtcの処理ブロック数が全て同じでない場合、3つのピクチャ共通で特徴量を算出している処理ブロックのみが評価値Jの算出対象となる。
 評価値算出部3により算出された評価値Jは、二次参照ピクチャPtp’と参照ピクチャPtpの間で絵柄があまり変化せずに処理ブロック単位の特徴量の変化が小さくなり、さらに、参照ピクチャPtpと評価対象ピクチャPtcの間で絵柄が大きく変化して処理ブロック単位の特徴量の変化が大きくなる場合、値が大きくなる。
 したがって、上記実施の形態1~3では、参照ピクチャPtpと評価対象ピクチャPtcの間の絵柄の変化の大きさを評価していたが、この実施の形態4では、参照ピクチャPtpにおける絵柄の変化の程度と評価対象ピクチャPtcにおける絵柄の変化の程度の差異の大きさを評価するようにしたので、シーンチェンジではないピクチャ間での緩やかな絵柄の変化を、画面全体の絵柄が大きく変化するシーンチェンジと誤検出する可能性を低減し、上記実施の形態1~3よりも、シーンチェンジの検出精度を高めることができる。
 また、処理ブロック単位の特徴量の算出対象となる処理ブロックの定義方法は、上記実施の形態1と同様(ピクチャ内全ての処理ブロック、図10~12のような一部の処理ブロック等)に任意に定義できる。
The evaluation value calculation unit 3 calculates the feature amount C n, tp of the processing block unit included in the reference picture P tp calculated by the feature amount calculation unit 2 and the feature amount C of the processing block unit included in the evaluation target picture P tc An evaluation value J is calculated from n, tc and the feature amount C n, tp 'in units of processing blocks included in the secondary reference picture P tp' (step ST3 in FIG. 4).
Specifically, the evaluation value calculation unit 3 calculates the sum of absolute differences of the feature amounts in processing block units between the reference picture P tp and the evaluation target picture P tc .
In addition, the evaluation value calculation unit 3 calculates the sum of absolute differences of feature quantities in processing block units between the reference picture P tp and the secondary reference picture P tp ′ .
The evaluation value calculation unit 3 calculates, as the evaluation value J, the difference between the calculated difference absolute value sums as shown in the following equation (3).
Figure JPOXMLDOC01-appb-I000003
Here, MIN_NUM_BL 'is the minimum value of three values of NUM_BL tp' , NUM_BL tp , and NUM_BL tc . That is, when the numbers of processing blocks of the secondary reference picture Ptp ' , the reference picture Ptp, and the evaluation target picture Ptc are not all the same, only the processing block whose feature amount is calculated for all three pictures is the evaluation value J It becomes the object of calculation.
In the evaluation value J calculated by the evaluation value calculation unit 3, the change in the feature amount in processing block units becomes small without much change in pattern between the secondary reference picture Ptp ' and the reference picture Ptp , and further When the design largely changes between the reference picture P tp and the evaluation target picture P tc and the change of the feature amount in processing block units becomes large, the value becomes large.
Therefore, in the first to third embodiments, the magnitude of the change in the pattern between the reference picture P tp and the evaluation target picture P t c is evaluated. However, in the fourth embodiment, the pattern in the reference picture P tp Since the magnitude of the difference between the degree of change of the pattern and the degree of change of the pattern in the evaluation target picture P tc is evaluated, the pattern of the whole screen is large in gradual change of the pattern between the pictures that are not scene changes. It is possible to reduce the possibility of erroneous detection as a changing scene change, and to improve the scene change detection accuracy more than the first to third embodiments.
Further, the definition method of the processing block to be the target of calculation of the feature amount in processing block units is the same as that of the first embodiment (all processing blocks in a picture, partial processing blocks as shown in FIGS. It can be defined arbitrarily.
 有無判定部4は、上記実施の形態1と同様に、評価値算出部3により算出された評価値Jと、事前に設定された閾値Thとを比較する(図4のステップST4)。
 有無判定部4は、上記実施の形態1と同様に、評価値算出部3により算出された評価値Jが閾値Th以上であれば、参照ピクチャPtpと評価対象ピクチャPtcとの間にシーンチェンジが有ると判定する(図4のステップST5)。
 有無判定部4は、上記実施の形態1と同様に、評価値算出部3により算出された評価値Jが閾値Th未満であれば、参照ピクチャPtpと評価対象ピクチャPtcとの間にシーンチェンジが無いと判定する(図4のステップST6)。
 有無判定部4は、シーンチェンジの有無の判定結果を出力する。
 また、閾値Thの設定方法は、上記実施の形態1と同様である。
Similar to the first embodiment, the presence / absence determination unit 4 compares the evaluation value J calculated by the evaluation value calculation unit 3 with the threshold value Th set in advance (step ST4 in FIG. 4).
As in the first embodiment, if the evaluation value J calculated by the evaluation value calculation unit 3 is equal to or larger than the threshold value Th, the presence / absence determination unit 4 sets the scene between the reference picture P tp and the evaluation target picture P tc. It is determined that there is a change (step ST5 in FIG. 4).
As in the first embodiment, if the evaluation value J calculated by the evaluation value calculation unit 3 is smaller than the threshold value Th, the presence / absence determination unit 4 sets the scene between the reference picture P tp and the evaluation target picture P tc. It is determined that there is no change (step ST6 in FIG. 4).
The presence determination unit 4 outputs the determination result of the presence or absence of a scene change.
Further, the method of setting the threshold value Th is the same as that of the first embodiment.
 ピクチャ選択部1により決定された候補ピクチャの中に、未だ評価対象ピクチャとして選択していないピクチャが残っていれば(図4のステップST7:YESの場合)、ステップST1の処理に戻り、ステップST1~ST7の処理が繰り返される。
 なお、戻ったステップST1の処理で、ピクチャ選択部1により選択された3つのピクチャのうち、特徴量が既に算出されているピクチャについては、特徴量算出部2による特徴量の算出処理を省略するようにしてもよい。このようにすることで、特徴量の再計算を実施せずに済み、処理時間を短縮することができる。
 ピクチャ選択部1により決定された候補ピクチャの中に、未だ評価対象ピクチャとして選択していないピクチャが残っていなければ(図4のステップST7:NOの場合)、一連の処理が終了する。
If there is still a picture not selected as an evaluation target picture among the candidate pictures determined by the picture selection unit 1 (step ST7 in FIG. 4: YES), the process returns to step ST1 and step ST1. The processing of ̃ST7 is repeated.
In the process of step ST1 returned, among the three pictures selected by the picture selection unit 1, for the picture for which the feature amount has already been calculated, the process of calculating the feature amount by the feature amount calculation unit 2 is omitted. You may do so. By doing this, it is not necessary to recalculate feature amounts, and processing time can be shortened.
If there is no picture which has not been selected as a target picture for evaluation yet in the candidate pictures determined by the picture selection unit 1 (in the case of NO at step ST7 in FIG. 4), a series of processing ends.
 なお、本願発明はその発明の範囲内において、各実施の形態の自由な組み合わせ、あるいは各実施の形態の任意の構成要素の変形、もしくは各実施の形態において任意の構成要素の省略が可能である。 In the scope of the invention, the present invention allows free combination of each embodiment, or modification of any component of each embodiment, or omission of any component in each embodiment. .
 この発明は、シーンの変化点であるシーンチェンジの有無を判定する画像処理装置及び画像処理方法に適している。 The present invention is suitable for an image processing apparatus and an image processing method for determining the presence or absence of a scene change which is a change point of a scene.
 1 ピクチャ選択部、2 特徴量算出部、3 評価値算出部、4 有無判定部、5 デマルチプレクサ、11 ピクチャ選択回路、12 特徴量算出回路、13 評価値算出回路、14 有無判定回路、21 メモリ、22 プロセッサ。 Reference Signs List 1 picture selection unit, 2 feature amount calculation unit, 3 evaluation value calculation unit, 4 presence / absence determination unit, 5 demultiplexers, 11 picture selection circuit, 12 feature amount calculation circuit, 13 evaluation value calculation circuit, 14 presence / absence determination circuit, 21 memory , 22 processors.

Claims (7)

  1.  複数のピクチャの中から、シーンの変化点であるシーンチェンジの有無を評価する対象のピクチャとして、2つのピクチャを選択するピクチャ選択部と、
     前記ピクチャ選択部により選択された2つのピクチャのそれぞれについて、ピクチャに含まれるブロック単位の特徴量をそれぞれ算出する特徴量算出部と、
     前記特徴量算出部により算出された2つのピクチャに含まれるブロック単位の特徴量から、シーンチェンジの有無の判定に用いる評価値を算出する評価値算出部と、
     前記評価値算出部により算出された評価値と閾値を比較することで、前記2つのピクチャの間でのシーンチェンジの有無を判定する有無判定部と
     を備えた画像処理装置。
    A picture selection unit for selecting two pictures as pictures to be evaluated for the presence or absence of a scene change which is a scene change point among a plurality of pictures;
    A feature amount calculation unit that calculates feature amounts in units of blocks included in each of the two pictures selected by the picture selection unit;
    An evaluation value calculation unit that calculates an evaluation value used to determine the presence or absence of a scene change from the feature amount in block units included in the two pictures calculated by the feature amount calculation unit;
    An image processing apparatus comprising: a presence / absence determination unit that determines presence / absence of a scene change between the two pictures by comparing the threshold with the evaluation value calculated by the evaluation value calculation unit.
  2.  前記特徴量算出部は、前記ピクチャ選択部により選択されたピクチャのブロック毎に、当該ブロックの符号量を特定し、当該ブロックの符号量から、当該ブロックの特徴量を算出することを特徴とする請求項1記載の画像処理装置。 The feature quantity calculation unit specifies, for each block of the picture selected by the picture selection unit, a code quantity of the block, and calculates a feature quantity of the block from the code quantity of the block. An image processing apparatus according to claim 1.
  3.  前記特徴量算出部は、前記ピクチャ選択部により選択されたピクチャのブロック毎に、当該ブロックの特徴量として、当該ブロックに含まれている分割ブロックの個数を特定することを特徴とする請求項1記載の画像処理装置。 The feature quantity calculation unit specifies, for each block of a picture selected by the picture selection unit, the number of divided blocks included in the block as the feature quantity of the block. Image processing apparatus as described.
  4.  前記特徴量算出部は、前記ピクチャ選択部により選択されたピクチャのブロック毎に、当該ブロックの特徴量として、当該ブロックに含まれている分割ブロックの最深の階層数を特定することを特徴とする請求項1記載の画像処理装置。 The feature quantity calculation unit is characterized by specifying, for each block of the picture selected by the picture selection unit, as the feature quantity of the block, the deepest hierarchy number of divided blocks included in the block. An image processing apparatus according to claim 1.
  5.  前記評価値算出部は、シーンチェンジの有無の判定に用いる評価値として、前記特徴量算出部により算出された2つのピクチャの間でのブロック単位の特徴量の差分絶対値和を算出することを特徴とする請求項1記載の画像処理装置。 The evaluation value calculating unit may calculate a sum of absolute differences of feature amounts in units of blocks between two pictures calculated by the feature amount calculating unit as an evaluation value used to determine the presence or absence of a scene change. The image processing apparatus according to claim 1, characterized in that
  6.  前記ピクチャ選択部は、シーンチェンジの有無を評価する対象のピクチャとして、3つのピクチャを選択し、
     前記特徴量算出部は、前記ピクチャ選択部により選択された3つのピクチャのそれぞれについて、ピクチャに含まれるブロック単位の特徴量をそれぞれ算出し、
     前記評価値算出部は、前記特徴量算出部によりブロック単位の特徴量が算出された3つのピクチャのうち、再生順で、2番目のピクチャと3番目のピクチャとの間でのブロック単位の特徴量の差分絶対値和を算出するとともに、再生順で、1番目のピクチャと2番目のピクチャとの間でのブロック単位の特徴量の差分絶対値和を算出し、シーンチェンジの有無の判定に用いる評価値として、算出した双方の差分絶対値和の差分を算出し、
     前記有無判定部は、前記評価値算出部により算出された評価値と閾値を比較することで、再生順で、2番目のピクチャと3番目のピクチャとの間でのシーンチェンジの有無を判定することを特徴とする請求項1記載の画像処理装置。
    The picture selection unit selects three pictures as pictures to be evaluated for the presence or absence of a scene change,
    The feature quantity calculation unit calculates feature quantities in units of blocks included in each of the three pictures selected by the picture selection unit,
    The evaluation value calculation unit is a block unit feature between the second picture and the third picture in reproduction order among the three pictures for which the feature amount calculation unit calculates the feature amount in block units. In addition to calculating the sum of absolute differences of quantities, the sum of absolute differences of feature quantities in units of blocks between the first picture and the second picture in reproduction order is calculated, and the presence or absence of a scene change is determined. As the evaluation value to be used, calculate the difference between the calculated sum of absolute differences of both,
    The presence / absence determination unit compares the evaluation value calculated by the evaluation value calculation unit with a threshold to determine the presence or absence of a scene change between the second picture and the third picture in the reproduction order. An image processing apparatus according to claim 1, characterized in that.
  7.  ピクチャ選択部が、複数のピクチャの中から、シーンチェンジの有無を評価する対象のピクチャとして、2つのピクチャを選択し、
     特徴量算出部が、前記ピクチャ選択部により選択された2つのピクチャのそれぞれについて、ピクチャに含まれるブロック単位の特徴量をそれぞれ算出し、
     評価値算出部が、前記特徴量算出部により算出された2つのピクチャに含まれるブロック単位の特徴量から、シーンチェンジの有無の判定に用いる評価値を算出し、
     有無判定部が、前記評価値算出部により算出された評価値と閾値を比較することで、前記2つのピクチャの間でのシーンチェンジの有無を判定する
     画像処理方法。
    The picture selection unit selects two pictures from among the plurality of pictures as pictures to be evaluated for the presence or absence of a scene change,
    The feature amount calculation unit calculates feature amounts in units of blocks included in each of the two pictures selected by the picture selection unit,
    An evaluation value calculation unit calculates an evaluation value to be used to determine the presence or absence of a scene change from the feature amounts in block units included in the two pictures calculated by the feature amount calculation unit,
    An image processing method, comprising: determining a presence or absence of a scene change between the two pictures by comparing a threshold with the evaluation value calculated by the evaluation value calculator.
PCT/JP2017/038339 2017-10-24 2017-10-24 Image processing device, and image processing method WO2019082268A1 (en)

Priority Applications (4)

Application Number Priority Date Filing Date Title
PCT/JP2017/038339 WO2019082268A1 (en) 2017-10-24 2017-10-24 Image processing device, and image processing method
JP2018550470A JP6782789B2 (en) 2017-10-24 2018-03-30 Image processing device and image processing method
PCT/JP2018/013789 WO2019082413A1 (en) 2017-10-24 2018-03-30 Image processing device, and image processing method
TW107127425A TWI673652B (en) 2017-10-24 2018-08-07 Image processing device and image processing method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/JP2017/038339 WO2019082268A1 (en) 2017-10-24 2017-10-24 Image processing device, and image processing method

Publications (1)

Publication Number Publication Date
WO2019082268A1 true WO2019082268A1 (en) 2019-05-02

Family

ID=66247342

Family Applications (2)

Application Number Title Priority Date Filing Date
PCT/JP2017/038339 WO2019082268A1 (en) 2017-10-24 2017-10-24 Image processing device, and image processing method
PCT/JP2018/013789 WO2019082413A1 (en) 2017-10-24 2018-03-30 Image processing device, and image processing method

Family Applications After (1)

Application Number Title Priority Date Filing Date
PCT/JP2018/013789 WO2019082413A1 (en) 2017-10-24 2018-03-30 Image processing device, and image processing method

Country Status (3)

Country Link
JP (1) JP6782789B2 (en)
TW (1) TWI673652B (en)
WO (2) WO2019082268A1 (en)

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH0636026A (en) * 1992-07-13 1994-02-10 Toshiba Corp Moving picture processor
JPH06133305A (en) * 1992-09-07 1994-05-13 Fujitsu Ltd Picture data coding method, picture data coder using the method, picture data decoding method, picture data decoder using the method, scene change detection method, scene change detection method, scene change detector, scene change recorder, and picture data scene change recording and reproducing device
JPH07152779A (en) * 1993-11-29 1995-06-16 Toshiba Corp Processing method for detecting moving picture index and moving picture processor having moving picture index detection processing function
JPH10294923A (en) * 1997-04-18 1998-11-04 Matsushita Electric Ind Co Ltd Scene change detection method and scene change detector
JP2001251630A (en) * 2000-03-06 2001-09-14 Nec Corp Image signal coder and its method
WO2010125757A1 (en) * 2009-04-28 2010-11-04 三菱電機株式会社 Video/audio reproduction device, video/audio recording/ reproduction device, video/audio reproduction method, and video/audio recording/reproduction method
JP2017011587A (en) * 2015-06-24 2017-01-12 日本放送協会 Scene change detector, scene change detection method and program

Family Cites Families (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7382417B2 (en) * 2004-12-23 2008-06-03 Intel Corporation Method and algorithm for detection of scene cuts or similar images in video images
JP4618166B2 (en) * 2006-03-07 2011-01-26 ソニー株式会社 Image processing apparatus, image processing method, and program
TW200822739A (en) * 2006-11-10 2008-05-16 Benq Corp Method for chaptering an image datum according to scene change
JP4420056B2 (en) * 2007-04-20 2010-02-24 ソニー株式会社 Image processing apparatus, image processing method, image processing program, reproduction information generation apparatus, reproduction information generation method, and reproduction information generation program
CN101681518B (en) * 2007-05-21 2012-11-14 三菱电机株式会社 Image difference detection method and apparatus, scene change detection method and apparatus, as well as image difference value detection method and apparatus
US7859547B2 (en) * 2007-09-25 2010-12-28 Himax Technologies Limited Display parameter adjusting method and apparatus for scene change compensation
JP5191240B2 (en) * 2008-01-09 2013-05-08 オリンパス株式会社 Scene change detection apparatus and scene change detection program
JP5178616B2 (en) * 2009-04-22 2013-04-10 三菱電機株式会社 Scene change detection device and video recording device
JP2015216632A (en) * 2014-04-22 2015-12-03 ソニー株式会社 Coding device and coding method
JP6373681B2 (en) * 2014-07-30 2018-08-15 Kddi株式会社 Objective image quality evaluation apparatus, objective image quality evaluation method, and program

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH0636026A (en) * 1992-07-13 1994-02-10 Toshiba Corp Moving picture processor
JPH06133305A (en) * 1992-09-07 1994-05-13 Fujitsu Ltd Picture data coding method, picture data coder using the method, picture data decoding method, picture data decoder using the method, scene change detection method, scene change detection method, scene change detector, scene change recorder, and picture data scene change recording and reproducing device
JPH07152779A (en) * 1993-11-29 1995-06-16 Toshiba Corp Processing method for detecting moving picture index and moving picture processor having moving picture index detection processing function
JPH10294923A (en) * 1997-04-18 1998-11-04 Matsushita Electric Ind Co Ltd Scene change detection method and scene change detector
JP2001251630A (en) * 2000-03-06 2001-09-14 Nec Corp Image signal coder and its method
WO2010125757A1 (en) * 2009-04-28 2010-11-04 三菱電機株式会社 Video/audio reproduction device, video/audio recording/ reproduction device, video/audio reproduction method, and video/audio recording/reproduction method
JP2017011587A (en) * 2015-06-24 2017-01-12 日本放送協会 Scene change detector, scene change detection method and program

Also Published As

Publication number Publication date
TWI673652B (en) 2019-10-01
TW201917631A (en) 2019-05-01
JP6782789B2 (en) 2020-11-11
JPWO2019082413A1 (en) 2019-11-21
WO2019082413A1 (en) 2019-05-02

Similar Documents

Publication Publication Date Title
US10142632B2 (en) Moving image coding apparatus and moving image decoding apparatus
EP3389276B1 (en) Hash-based encoder decisions for video coding
CN107529063B (en) Encoding method, encoding device, decoding method, and decoding device
US9800896B2 (en) Method for depth lookup table signaling
JP6005157B2 (en) Depth map encoding and decoding
US9167260B2 (en) Apparatus and method for video processing
JP2018529255A (en) Image prediction method and apparatus
US10178405B2 (en) Enhanced coding and decoding using intra block copy mode
US20240107008A1 (en) Video coding and decoding
TW201720154A (en) Chroma quantization in video coding
KR20170021337A (en) Encoder decisions based on results of hash-based block matching
US9681126B2 (en) Motion vector coding apparatus, method and program for coding motion vector, motion vector decoding apparatus, and method and program for decoding motion vector
EP2960855B1 (en) Method and device for determining a set of modifiable elements in a group of pictures
WO2010052837A1 (en) Image decoding device, image decoding method, integrated circuit, and program
JP2007124408A (en) Motion vector detector and motion vector detecting method
CN113014926B (en) Video transcoding method and device, electronic equipment and storage medium
KR101552664B1 (en) Method and device for encoding images, method and device for decoding images, and programs therefor
JP2013229674A (en) Image coding device, image decoding device, image coding method, image decoding method, image coding program, and image decoding program
CN114402608A (en) Coordinated early termination of BDOF and DMVR in video coding and decoding
US9762916B2 (en) Encoding apparatus and method
US20080112631A1 (en) Method of obtaining a motion vector in block-based motion estimation
JP5178616B2 (en) Scene change detection device and video recording device
US11849138B2 (en) Video coding and decoding
WO2019082268A1 (en) Image processing device, and image processing method
JP2009260494A (en) Image coding apparatus and its control method

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 17930075

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 17930075

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: JP