CN109285178A - Image partition method, device and storage medium - Google Patents

Image partition method, device and storage medium Download PDF

Info

Publication number
CN109285178A
CN109285178A CN201811252058.XA CN201811252058A CN109285178A CN 109285178 A CN109285178 A CN 109285178A CN 201811252058 A CN201811252058 A CN 201811252058A CN 109285178 A CN109285178 A CN 109285178A
Authority
CN
China
Prior art keywords
pixel
target image
image
identifier
area
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201811252058.XA
Other languages
Chinese (zh)
Inventor
郭益林
郑文
赵松涛
宋丛礼
于永航
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Dajia Internet Information Technology Co Ltd
Original Assignee
Beijing Dajia Internet Information Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Dajia Internet Information Technology Co Ltd filed Critical Beijing Dajia Internet Information Technology Co Ltd
Priority to CN201811252058.XA priority Critical patent/CN109285178A/en
Publication of CN109285178A publication Critical patent/CN109285178A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/194Segmentation; Edge detection involving foreground-background segmentation

Landscapes

  • Engineering & Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Image Analysis (AREA)

Abstract

The disclosure is directed to a kind of image partition method, device and storage mediums, belong to field of image processing.Method includes: the pixel classifications model for obtaining and having trained, pixel classifications model is used to determine the class indication of each pixel in any image, class indication includes first identifier and second identifier, first identifier is used to indicate corresponding pixel and belongs to foreground area, and second identifier is used to indicate corresponding pixel and belongs to background area;Target image to be processed is input in pixel classifications model, pixel classifications model is based on, determines the class indication of each pixel in target image;According to the class indication of pixel each in target image, foreground area and the background area of target image are determined, so as to extract foreground area or the background area of target image.By way of training pixel classifications model, it can be directly based upon pixel classifications model, carry out image segmentation to target image reduces calculation amount, save the time without largely being calculated.

Description

Image partition method, device and storage medium
Technical field
This disclosure relates to field of image processing more particularly to a kind of image partition method, device and storage medium.
Background technique
Piece image generally includes foreground area and background area, and image Segmentation Technology can determine the prospect in image Region and background area, foreground area and background area are distinguished, independent to carry out to foreground area or background area Ground processing, is widely used at present in the multiple fields such as target detection, object tracking, video monitoring, recognition of face.
The nontransparent degree of the foreground of the display color of any pixel and the pixel, background colour and foreground has in image Following incidence relation: I=α F+ (1- α) B, wherein I indicates that the display color of pixel, F indicate the foreground of pixel, and B indicates picture The background colour of element, α indicate the nontransparent degree of pixel foreground.When to be split to target image, calculated using matting Method, can be according to the display color and above-mentioned incidence relation of pixels multiple in the target image, before calculating each pixel The nontransparent degree of scenery generates prospect probability graph (mask) according to the nontransparent degree of the foreground of each pixel, the prospect probability Figure is used to indicate the foreground area in the target image and the position where background area, according to the prospect probability in subsequent process Figure handles the target image, and foreground area or background area can be extracted from target image.
Above scheme needs to carry out a large amount of nontransparent degree for calculating and capable of just calculating the foreground of each pixel, calculation amount It is larger, it takes a long time.
Summary of the invention
Present disclose provides a kind of image partition method, device and storage mediums, can overcome present in the relevant technologies Problem.
According to the first aspect of the embodiments of the present disclosure, a kind of image partition method is provided, which comprises
The pixel classifications model trained is obtained, the pixel classifications model is for determining each pixel in any image Class indication, the class indication include first identifier and second identifier, and the first identifier is used to indicate corresponding pixel category In foreground area, the second identifier is used to indicate corresponding pixel and belongs to background area;
Target image to be processed is input in the pixel classifications model, the pixel classifications model is based on, is determined The class indication of each pixel in the target image;
According to the class indication of pixel each in the target image, the foreground area and background of the target image are determined Region.
In a kind of mode in the cards, the method also includes:
Obtain the class indication of each pixel in multiple sample images and the multiple sample image;
Model is carried out according to the class indication of each pixel in the multiple sample image and the multiple sample image Training, obtains the pixel classifications model.
In the mode of alternatively possible realization, the class indication according to pixel each in the target image, really The foreground area of the fixed target image and background area, comprising:
According to the class indication of pixel each in the target image, prospect probability graph is generated, the prospect probability graph is used Position where the foreground area and background area for indicating the target image.
In the mode of alternatively possible realization, the method also includes:
According to the prospect probability graph, enhancing processing is carried out to the foreground area of the target image;Alternatively,
According to the prospect probability graph, Fuzzy processing is carried out to the background area of the target image.
It is described according to the prospect probability graph in the mode of alternatively possible realization, to the back in the target image Scene area carries out Fuzzy processing, comprising:
According to the prospect probability graph, the foreground area of the target image is extracted;
Fuzzy processing is carried out to the target image, obtains the first image, and according to the prospect probability graph, extracts institute State the background area of the first image;
The background area of the foreground area of the target image and the first image is combined, the second figure is obtained Picture.
In the mode of alternatively possible realization, in the prospect probability graph, the pixel value of pixel is 1 in foreground area, The pixel value of pixel is 0 in background area, described according to the prospect probability graph, is carried out to the background area of the target image Fuzzy processing, comprising:
The background area of the target image is carried out at blurring using following formula according to the prospect probability graph Reason:
Target=Source*mask+Gaussian* (255-mask);
Wherein, after Source indicates that the target image, Gaussian indicate that the target image carries out Fuzzy processing Obtained image, mask indicate that the prospect probability graph, Target indicate that the background area of the target image is blurred The image obtained after processing.
According to the second aspect of an embodiment of the present disclosure, a kind of image segmentation device is provided, described device includes:
Model obtains module, is configured as obtaining the pixel classifications model trained, the pixel classifications model is for true Determine the class indication of each pixel in any image, the class indication includes first identifier and second identifier, first mark Knowledge is used to indicate corresponding pixel and belongs to foreground area, and the second identifier is used to indicate corresponding pixel and belongs to background area;
Determining module is identified, is configured as target image to be processed being input in the pixel classifications model, be based on The pixel classifications model, determines the class indication of each pixel in the target image;
Area determination module is configured as determining the mesh according to the class indication of pixel each in the target image The foreground area of logo image and background area.
In a kind of mode in the cards, described device further include:
Sample acquisition module is configured as obtaining each pixel in multiple sample images and the multiple sample image Class indication;
Training module is configured as according to each pixel in the multiple sample image and the multiple sample image Class indication carries out model training, obtains the pixel classifications model.
In the mode of alternatively possible realization, the area determination module includes:
Generation unit is configured as the class indication according to pixel each in the target image, generates prospect probability graph, The foreground area and the position where background area that the prospect probability graph is used to indicate the target image.
In the mode of alternatively possible realization, described device further include:
Processing module is configured as enhancing the foreground area of the target image according to the prospect probability graph Processing;Alternatively, the processing module, is additionally configured to according to the prospect probability graph, to the background area of the target image Carry out Fuzzy processing.
In the mode of alternatively possible realization, the processing module includes:
Extraction unit is configured as extracting the foreground area of the target image according to the prospect probability graph;
The extraction unit is additionally configured to carry out Fuzzy processing to the target image, obtains the first image, and root According to the prospect probability graph, the background area of the first image is extracted;
Assembled unit is configured as carrying out the background area of the foreground area of the target image and the first image Combination, obtains the second image.
In the mode of alternatively possible realization, in the prospect probability graph, the pixel value of pixel is 1 in foreground area, The pixel value of pixel is 0 in background area, and the processing module is additionally configured to according to the prospect probability graph, and use is following Formula carries out Fuzzy processing to the background area of the target image:
Target=Source*mask+Gaussian* (255-mask);
Wherein, after Source indicates that the target image, Gaussian indicate that the target image carries out Fuzzy processing Obtained image, mask indicate that the prospect probability graph, Target indicate that the background area of the target image is blurred The image obtained after processing.
According to the third aspect of an embodiment of the present disclosure, a kind of image segmentation device is provided, described device includes:
Processor;
Memory for storage processor executable command;
Wherein, the processor is configured to:
The pixel classifications model trained is obtained, the pixel classifications model is for determining each pixel in any image Class indication, the class indication include first identifier and second identifier, and the first identifier is used to indicate corresponding pixel category In foreground area, the second identifier is used to indicate corresponding pixel and belongs to background area;
Target image to be processed is input in the pixel classifications model, the pixel classifications model is based on, is determined The class indication of each pixel in the target image;
According to the class indication of pixel each in the target image, the foreground area and background of the target image are determined Region.
According to a fourth aspect of embodiments of the present disclosure, a kind of non-transitorycomputer readable storage medium is provided, when described When instruction in storage medium is executed by the processor of processing unit, so that processing unit is able to carry out a kind of image segmentation side Method, which comprises
The pixel classifications model trained is obtained, the pixel classifications model is for determining each pixel in any image Class indication, the class indication include first identifier and second identifier, and the first identifier is used to indicate corresponding pixel category In foreground area, the second identifier is used to indicate corresponding pixel and belongs to background area;
Target image to be processed is input in the pixel classifications model, the pixel classifications model is based on, is determined The class indication of each pixel in the target image;
According to the class indication of pixel each in the target image, the foreground area and background of the target image are determined Region.
According to a fifth aspect of the embodiments of the present disclosure, a kind of computer program product is provided, when the computer program produces When instruction in product is executed by the processor of processing unit, so that processing unit is able to carry out a kind of image partition method, it is described Method includes:
The pixel classifications model trained is obtained, the pixel classifications model is for determining each pixel in any image Class indication, the class indication include first identifier and second identifier, and the first identifier is used to indicate corresponding pixel category In foreground area, the second identifier is used to indicate corresponding pixel and belongs to background area;
Target image to be processed is input in the pixel classifications model, the pixel classifications model is based on, is determined The class indication of each pixel in the target image;
According to the class indication of pixel each in the target image, the foreground area and background of the target image are determined Region.
The technical scheme provided by this disclosed embodiment can include the following benefits:
The pixel classifications model trained is obtained, pixel classifications model is used to determine the classification of each pixel in any image Mark, class indication includes first identifier and second identifier, and first identifier is used to indicate corresponding pixel and belongs to foreground area, the Two marks are used to indicate corresponding pixel and belong to background area, and target image to be processed is input in pixel classifications model, Based on pixel classifications model, the class indication of each pixel in target image is determined, according to point of pixel each in target image Class mark, determines foreground area and the background area of target image, so as to extract the foreground area or background of target image Region.By way of training pixel classifications model, it can be directly based upon pixel classifications model, image point is carried out to target image It cuts, without largely being calculated, reduces calculation amount, save the time.
It should be understood that above general description and following detailed description be only it is exemplary and explanatory, not The disclosure can be limited.
Detailed description of the invention
The drawings herein are incorporated into the specification and forms part of this specification, and shows the implementation for meeting the disclosure Example, and together with specification for explaining the principles of this disclosure.
Fig. 1 is a kind of flow chart of image partition method shown according to an exemplary embodiment.
Fig. 2 is a kind of flow chart of image partition method shown according to an exemplary embodiment.
Fig. 3 is a kind of block diagram of processing unit shown according to an exemplary embodiment.
Fig. 4 is a kind of block diagram of terminal for image segmentation shown according to an exemplary embodiment.
Fig. 5 is a kind of structural schematic diagram of server shown according to an exemplary embodiment.
Specific embodiment
Example embodiments are described in detail here, and the example is illustrated in the accompanying drawings.Following description is related to When attached drawing, unless otherwise indicated, the same numbers in different drawings indicate the same or similar elements.Following exemplary embodiment Described in embodiment do not represent all implementations consistent with this disclosure.On the contrary, they be only with it is such as appended The example of the consistent device and method of some aspects be described in detail in claims, the disclosure.
Fig. 1 is a kind of flow chart of image partition method shown according to an exemplary embodiment, as shown in Figure 1, the figure As dividing method is in processing unit, comprising the following steps:
In a step 101, the pixel classifications model trained is obtained, pixel classifications model is every in any image for determining The class indication of a pixel, class indication include first identifier and second identifier, and first identifier is used to indicate corresponding pixel category In foreground area, second identifier is used to indicate corresponding pixel and belongs to background area.
In a step 102, target image to be processed is input in pixel classifications model, is based on pixel classifications model, Determine the class indication of each pixel in target image.
In step 103, according to the class indication of pixel each in target image, determine target image foreground area and Background area.
The method that the embodiment of the present disclosure provides obtains the pixel classifications model trained, and pixel classifications model is for determining The class indication of each pixel in any image, class indication include first identifier and second identifier, and first identifier is used to indicate Corresponding pixel belongs to foreground area, and second identifier is used to indicate corresponding pixel and belongs to background area, by target to be processed Image is input in pixel classifications model, is based on pixel classifications model, is determined the class indication of each pixel in target image, root According to the class indication of pixel each in target image, foreground area and the background area of target image are determined, so as to extract The foreground area of target image or background area.By way of training pixel classifications model, pixel classifications can be directly based upon Model, carry out image segmentation to target image reduces calculation amount, saves the time without largely being calculated.
In a kind of mode in the cards, method further include:
Obtain the class indication of each pixel in multiple sample images and multiple sample images;
Model training is carried out according to the class indication of each pixel in multiple sample images and multiple sample images, is obtained Pixel classifications model.
In the mode of alternatively possible realization, according to the class indication of pixel each in target image, target figure is determined The foreground area of picture and background area, comprising:
According to the class indication of pixel each in target image, prospect probability graph is generated, prospect probability graph is used to indicate mesh The foreground area of logo image and the position where background area.
In the mode of alternatively possible realization, method further include:
According to prospect probability graph, enhancing processing is carried out to the foreground area of target image;Alternatively,
According to prospect probability graph, Fuzzy processing is carried out to the background area of target image.
In the mode of alternatively possible realization, according to prospect probability graph, mould is carried out to the background area in target image Gelatinization processing, comprising:
According to prospect probability graph, the foreground area of target image is extracted;
Fuzzy processing is carried out to target image, obtains the first image, and according to prospect probability graph, extracts the first image Background area;
The background area of the foreground area of target image and the first image is combined, the second image is obtained.
In the mode of alternatively possible realization, in prospect probability graph, the pixel value of pixel is 1 in foreground area, background The pixel value of pixel is 0 in region, according to prospect probability graph, carries out Fuzzy processing to the background area of target image, comprising:
According to prospect probability graph, using following formula, Fuzzy processing is carried out to the background area of target image:
Target=Source*mask+Gaussian* (255-mask);
Wherein, Source indicates that target image, Gaussian indicate that target image carries out the figure obtained after Fuzzy processing Picture, mask indicate that prospect probability graph, Target indicate that the background area of target image carries out the image obtained after Fuzzy processing.
Fig. 2 is a kind of flow chart of image partition method shown according to an exemplary embodiment, as shown in Fig. 2, the figure As dividing method is in processing unit, processing unit can be the tool such as mobile phone, computer, server, camera, monitoring device There is the device of image processing function, method includes the following steps:
In step 201, processing unit obtains target image to be processed, and obtains the pixel classifications model trained.
Wherein, which can be shot to obtain by processing unit, or be mentioned from the video that processing unit takes It obtains, is perhaps downloaded from internet by processing unit and obtain or be sent to processing unit by other equipment.Alternatively, During processing unit carries out net cast, every image in available video flowing, using every picture as mesh Logo image, to handle every image in video flowing.
The target image can be a plurality of types of images, such as character image, scenery picture, cuisines picture, the target Image includes foreground area and background area, and shape of color, object that foreground area is shown with background area etc. can phase Together, or different, difference is only that: the area where object when foreground area refers to the shooting target image close to camera Domain is equivalent to " center of interest " of the target image, and background area is in addition to foreground area, for serving as a contrast or foil foreground area Region.
It further include the environment of personage behind for example, both having included personage in the portrait photographs of shooting, personage region at this time It may be considered foreground area, behind environment region may be considered background area, alternatively, both including table in cuisines picture Son further includes the cuisines on desk, and desk region may be considered background area at this time, and cuisines region can consider It is foreground area.
It, can be by the foreground area of the target image and background area area based on pixel classifications model in the embodiment of the present disclosure It branches away.Wherein, pixel classifications model is used to determine that the class indication of each pixel in any image, class indication to include first Mark and second identifier, first identifier are used to indicate corresponding pixel and belong to foreground area, and second identifier is used to indicate corresponding Pixel belongs to background area, that is to say, when the class indication of any pixel in image is first identifier, indicates the pixel category In foreground area, when the class indication of any pixel in image is second identifier, indicate that the pixel belongs to background area.
First identifier is two different marks from second identifier, for example, first identifier is 1, and second identifier is 0, or Person, first identifier 0, and second identifier is 1.
The pixel classifications model can be obtained by processing unit training, and be stored by the processing unit, alternatively, the pixel point Class model is sent to processing unit after being trained by other equipment, and is stored by the processing unit.
In training pixel classifications model, each pixel in multiple sample images and multiple sample images can be first obtained Class indication, according in multiple sample images and multiple sample images each pixel class indication carry out model training, Obtain pixel classifications model.
In a kind of mode in the cards, in training pixel classifications model, initial pixel classifications mould can be constructed Type, obtains training dataset and test data set, training dataset and test data concentrate include multiple sample images and The class indication of each pixel in each sample image.
In the training process, multiple sample images training data concentrated are as the input of pixel classifications model, by this Output of the class indication of each pixel as pixel classifications model in multiple sample images, instructs pixel classifications model Practice, learn pixel classifications model to the difference between foreground area and background area, has pixel classifications model and draw Divide the ability of foreground area and background area.Later, each sample image that test data is concentrated is input to pixel classifications mould In type, determine that the testing classification of each pixel in each sample image identifies respectively based on pixel classifications model, by testing classification It identifies and is compared with the actual classification of mark mark, pixel classifications model is modified according to comparing result.
, can be using default training algorithm in training pixel classifications model in a kind of possible implementation, this is default Training algorithm can be deep learning network algorithm, decision Tree algorithms, artificial neural network algorithm etc..
In subsequent process, the class indication of each pixel in new sample image and the sample image can also be obtained, Continue to train the pixel classifications model, so as to improve the classification accuracy of the pixel classifications model.
In step 202, target image is input in pixel classifications model by processing unit, is based on pixel classifications model, Determine the class indication of each pixel in target image.
Processing unit is handled the target image by the way that target image to be input in pixel classifications model, can be with Classified according to the class indication of each pixel in target image to the pixel in target image.
When the class indication of pixel is first identifier, indicate that the pixel belongs to foreground area, when the class indication of pixel When for second identifier, indicate that the pixel belongs to background area, the class indication of each pixel and each in combining target image The position of pixel in the target image can determine foreground area and the background area of target image, therefore the pixel classifications mould Type is actually a kind of semantic segmentation model, target image can be divided into foreground area and background area.
In step 203, processing unit generates prospect probability graph according to the class indication of pixel each in target image.
Prospect probability graph is used to indicate the position in target image where foreground area and background area, general according to the prospect Rate figure can determine that the pixel in target image on each position belongs to foreground area and still falls within background area, so as to incite somebody to action Target image is divided into foreground area and background area, realizes to the foreground area of target image or taking for background area.
In a kind of possible implementation, using the class indication of pixel each in target image as right in prospect probability graph The pixel value of pixel on position is answered, to generate prospect probability graph.Then in prospect probability graph, the pixel value of any pixel is the One mark indicates that the pixel corresponding pixel in the target image belongs to foreground area, and the pixel value of any pixel is the second mark Knowing indicates that the pixel corresponding pixel in the target image belongs to background area.
For example, first identifier is 1, second identifier 0, the white area in prospect probability graph indicates phase in the target image The position answered is foreground area, and the black region in prospect probability graph indicates that corresponding position is background area in the target image Domain.
In step 204, according to prospect probability graph, foreground area or background area to target image carry out processing unit Processing.
In order to improve the clarity of foreground area, both enhancing processing can be carried out to foreground area, it can also be to background area Domain carries out virtualization processing.Correspondingly, step 204 includes the following steps 2041 and 2042 at least one:
In step 2041, according to prospect probability graph, enhancing processing is carried out to the foreground area of target image.
Processing unit determines that pixel value is the pixel group of first identifier according to the pixel value of pixel each in prospect probability graph At image-region, by the image-region, corresponding image-region is extracted in the target image, before target image Scene area, and determine the image-region that the pixel that pixel value is second identifier forms, the image-region is right in the target image The image-region answered extracts, the background area as target image.
Later, using algorithm for image enhancement, enhancing processing is carried out to foreground area, the foreground area that obtains that treated will Foreground area that treated and background area are combined, the image that obtains that treated, to improve the quality of foreground area, before making Scene area is more clear, thus more prominent foreground area.
For example, algorithm for image enhancement can for algorithm of histogram equalization, contrast enhancement algorithms or other can increase The algorithm of strong image, this is no longer going to repeat them.
In step 2042, according to prospect probability graph, Fuzzy processing is carried out to the background area of target image.
In a kind of possible implementation, processing unit carries out Fuzzy processing to target image, obtains the first image, and According to the pixel value of pixel each in prospect probability graph, the image-region that the pixel that pixel value is second identifier forms is determined, it will The image-region corresponding image-region in the first image extracts, which is that the background area of target image passes through Region after crossing Fuzzy processing, the as background area after Fuzzy processing, and determine that pixel value is the pixel of first identifier The image-region of composition, by the image-region, corresponding image-region is extracted in the target image, as target image Background area after foreground area and Fuzzy processing is combined by foreground area later, the image that obtains that treated, real The Fuzzy processing to the background area of target image is showed, thus more prominent foreground area.
In alternatively possible implementation, in prospect probability graph, the pixel value of pixel is 1 in foreground area, background The pixel value of pixel is 0 in region, according to prospect probability graph, using following formula, carries out mould to the background area of target image Gelatinization processing:
Target=Source*mask+Gaussian* (255-mask).
Wherein, Source indicates that target image, Gaussian indicate that target image carries out the figure obtained after Fuzzy processing Picture, mask indicate that prospect probability graph, Target indicate that the background area of target image carries out the image obtained after Fuzzy processing. Source*mask indicates the foreground area of target image, and Gaussian* (255-mask) indicates target image by blurring The background area of the image obtained after reason obtains the foreground area of target image and target image after Fuzzy processing It combines the background area of image, so that it may which the background area for obtaining target image carries out the image obtained after Fuzzy processing.
It should be noted that when processing unit handles target image in the embodiment of the present disclosure, it can be first to target Image carries out enhancing processing, then carries out Fuzzy processing to target image.Alternatively, processing unit handles target image When, enhancing processing only can be carried out to target image, alternatively, only carrying out Fuzzy processing to target image.
The method that the embodiment of the present disclosure provides obtains the pixel classifications model trained, and pixel classifications model is for determining The class indication of each pixel in any image, class indication include first identifier and second identifier, and first identifier is used to indicate Corresponding pixel belongs to foreground area, and second identifier is used to indicate corresponding pixel and belongs to background area, by target to be processed Image is input in pixel classifications model, is based on pixel classifications model, is determined the class indication of each pixel in target image, root According to the class indication of pixel each in target image, foreground area and the background area of target image are determined, so as to extract The foreground area of target image or background area.By way of training pixel classifications model, pixel classifications can be directly based upon Model, carry out image segmentation to target image reduces calculation amount, saves the time, improve without largely being calculated Background blurring process speed.
Also, by the way that foreground area is carried out enhancing processing, foreground area can be made to be more clear, by background area Fuzzy processing is carried out, background area can be weakened, enhance the difference between foreground area and background area, so as to More prominent foreground area, keeps foreground area more obvious.
The embodiment of the present disclosure can be applied to recognition of face, under the scenes such as video monitoring, net cast and beautification picture, For example, human face region can be extracted from target image under recognition of face scene, ignore the area in addition to human face region Domain, to carry out recognition of face to human face region.Under video monitoring scene, it can be extracted from every image in video flowing The region in addition to target object region is ignored in target object region, to obtain the dynamic change of target object in video flowing Information realizes the tracking to target object.It, can be real in every image in the video flowing of live streaming under net cast scene When distinguish foreground area and background area, and virtualization processing is carried out to background area, so that prominent foreground area, makes foreground area It is more clear, enhancing live streaming effect.
Fig. 3 is a kind of block diagram of processing unit shown according to an exemplary embodiment.Referring to Fig. 3, which includes mould Type obtains module 301, mark determining module 302 and area determination module 303.
Model obtains module 301, is configured as obtaining the pixel classifications model trained, pixel classifications model is for determining The class indication of each pixel in any image, class indication include first identifier and second identifier, and first identifier is used to indicate Corresponding pixel belongs to foreground area, and second identifier is used to indicate corresponding pixel and belongs to background area;
Determining module 302 is identified, is configured as target image to be processed being input in pixel classifications model, is based on picture Plain disaggregated model determines the class indication of each pixel in target image;
Area determination module 303 is configured as determining target image according to the class indication of pixel each in target image Foreground area and background area.
In a kind of mode in the cards, device further include:
Sample acquisition module is configured as obtaining the classification of each pixel in multiple sample images and multiple sample images Mark;
Training module is configured as the class indication according to each pixel in multiple sample images and multiple sample images Model training is carried out, pixel classifications model is obtained.
In the mode of alternatively possible realization, area determination module 303 includes:
Generation unit is configured as the class indication according to pixel each in target image, generates prospect probability graph, prospect The foreground area and the position where background area that probability graph is used to indicate target image.
In the mode of alternatively possible realization, device further include:
Processing module, is configured as according to prospect probability graph, carries out enhancing processing to the foreground area of target image;Or Person, processing module are additionally configured to according to prospect probability graph, carry out Fuzzy processing to the background area of target image.
In the mode of alternatively possible realization, processing module includes:
Extraction unit is configured as extracting the foreground area of target image according to prospect probability graph;
Extraction unit is additionally configured to carry out Fuzzy processing to target image, obtains the first image, and general according to prospect Rate figure extracts the background area of the first image;
Assembled unit is configured as being combined the background area of the foreground area of target image and the first image, obtain To the second image.
In the mode of alternatively possible realization, in prospect probability graph, the pixel value of pixel is 1 in foreground area, background The pixel value of pixel is 0 in region, and processing module is additionally configured to according to prospect probability graph, using following formula, to target figure The background area of picture carries out Fuzzy processing:
Target=Source*mask+Gaussian* (255-mask);
Wherein, Source indicates that target image, Gaussian indicate that target image carries out the figure obtained after Fuzzy processing Picture, mask indicate that prospect probability graph, Target indicate that the background area of target image carries out the image obtained after Fuzzy processing.
About the device in above-described embodiment, wherein modules execute the concrete mode of operation in related this method Embodiment in be described in detail, no detailed explanation will be given here.
Fig. 4 is a kind of block diagram of terminal 400 for image segmentation shown according to an exemplary embodiment.The terminal 400 are used to execute step performed by processing unit in above-mentioned image partition method, can be portable mobile termianl, such as: Smart phone, tablet computer, MP3 player (Moving Picture Experts Group Audio Layer III, dynamic Image expert's compression standard audio level 3), MP4 (Moving Picture Experts Group Audio Layer IV, move State image expert's compression standard audio level 4) player, laptop or desktop computer.Terminal 400 is also possible to referred to as use Other titles such as family equipment, portable terminal, laptop terminal, terminal console.
In general, terminal 400 includes: processor 401 and memory 402.
Processor 401 may include one or more processing cores, such as 4 core processors, 8 core processors etc..Place Reason device 401 can use DSP (Digital Signal Processing, Digital Signal Processing), FPGA (Field- Programmable Gate Array, field programmable gate array), PLA (Programmable Logic Array, may be programmed Logic array) at least one of example, in hardware realize.Processor 401 also may include primary processor and coprocessor, master Processor is the processor for being handled data in the awake state, also referred to as CPU (Central Processing Unit, central processing unit);Coprocessor is the low power processor for being handled data in the standby state.? In some embodiments, processor 401 can be integrated with GPU (Graphics Processing Unit, image processor), GPU is used to be responsible for the rendering and drafting of content to be shown needed for display screen.In some embodiments, processor 401 can also be wrapped AI (Artificial Intelligence, artificial intelligence) processor is included, the AI processor is for handling related machine learning Calculating operation.
Memory 402 may include one or more computer readable storage mediums, which can To be non-transient.Memory 402 may also include high-speed random access memory and nonvolatile memory, such as one Or multiple disk storage equipments, flash memory device.In some embodiments, the non-transient computer in memory 402 can Storage medium is read for storing at least one instruction, at least one instruction by processor 401 for being had to realize this Shen Please in embodiment of the method provide image partition method.
In some embodiments, terminal 400 is also optional includes: peripheral device interface 403 and at least one peripheral equipment. It can be connected by bus or signal wire between processor 401, memory 402 and peripheral device interface 403.Each peripheral equipment It can be connected by bus, signal wire or circuit board with peripheral device interface 403.Specifically, peripheral equipment includes: radio circuit 404, at least one of touch display screen 405, camera 406, voicefrequency circuit 407, positioning component 408 and power supply 409.
Peripheral device interface 403 can be used for I/O (Input/Output, input/output) is relevant outside at least one Peripheral equipment is connected to processor 401 and memory 402.In some embodiments, processor 401, memory 402 and peripheral equipment Interface 403 is integrated on same chip or circuit board;In some other embodiments, processor 401, memory 402 and outer Any one or two in peripheral equipment interface 403 can realize on individual chip or circuit board, the present embodiment to this not It is limited.
Radio circuit 404 is for receiving and emitting RF (Radio Frequency, radio frequency) signal, also referred to as electromagnetic signal.It penetrates Frequency circuit 404 is communicated by electromagnetic signal with communication network and other communication equipments.Radio circuit 404 turns electric signal It is changed to electromagnetic signal to be sent, alternatively, the electromagnetic signal received is converted to electric signal.Optionally, radio circuit 404 wraps It includes: antenna system, RF transceiver, one or more amplifiers, tuner, oscillator, digital signal processor, codec chip Group, user identity module card etc..Radio circuit 404 can be carried out by least one wireless communication protocol with other terminals Communication.The wireless communication protocol includes but is not limited to: Metropolitan Area Network (MAN), each third generation mobile communication network (2G, 3G, 4G and 13G), wireless office Domain net and/or WiFi (Wireless Fidelity, Wireless Fidelity) network.In some embodiments, radio circuit 404 may be used also To include the related circuit of NFC (Near Field Communication, wireless near field communication), the application is not subject to this It limits.
Display screen 405 is for showing UI (User Interface, user interface).The UI may include figure, text, figure Mark, video and its their any combination.When display screen 405 is touch display screen, display screen 405 also there is acquisition to show The ability of the touch signal on the surface or surface of screen 405.The touch signal can be used as control signal and be input to processor 401 are handled.At this point, display screen 405 can be also used for providing virtual push button and/or dummy keyboard, also referred to as soft button and/or Soft keyboard.In some embodiments, display screen 405 can be one, and the front panel of terminal 400 is arranged;In other embodiments In, display screen 405 can be at least two, be separately positioned on the different surfaces of terminal 400 or in foldover design;In still other reality It applies in example, display screen 405 can be flexible display screen, be arranged on the curved surface of terminal 400 or on fold plane.Even, it shows Display screen 405 can also be arranged to non-rectangle irregular figure, namely abnormity screen.Display screen 405 can use LCD (Liquid Crystal Display, liquid crystal display), OLED (Organic Light-EmittingDiode, Organic Light Emitting Diode) Etc. materials preparation.
CCD camera assembly 406 is for acquiring image or video.Optionally, CCD camera assembly 406 include front camera and Rear camera.In general, the front panel of terminal is arranged in front camera, the back side of terminal is arranged in rear camera.One In a little embodiments, rear camera at least two is main camera, depth of field camera, wide-angle camera, focal length camera shooting respectively Any one in head, to realize that main camera and the fusion of depth of field camera realize background blurring function, main camera and wide-angle Camera fusion realizes that pan-shot and VR (Virtual Reality, virtual reality) shooting function or other fusions are clapped Camera shooting function.In some embodiments, CCD camera assembly 406 can also include flash lamp.Flash lamp can be monochromatic warm flash lamp, It is also possible to double-colored temperature flash lamp.Double-colored temperature flash lamp refers to the combination of warm light flash lamp and cold light flash lamp, can be used for not With the light compensation under colour temperature.
Voicefrequency circuit 407 may include microphone and loudspeaker.Microphone is used to acquire the sound wave of user and environment, and will Sound wave, which is converted to electric signal and is input to processor 401, to be handled, or is input to radio circuit 404 to realize voice communication. For stereo acquisition or the purpose of noise reduction, microphone can be separately positioned on the different parts of terminal 400 to be multiple.Mike Wind can also be array microphone or omnidirectional's acquisition type microphone.Loudspeaker is then used to that processor 401 or radio circuit will to be come from 404 electric signal is converted to sound wave.Loudspeaker can be traditional wafer speaker, be also possible to piezoelectric ceramic loudspeaker.When When loudspeaker is piezoelectric ceramic loudspeaker, the audible sound wave of the mankind can be not only converted electrical signals to, it can also be by telecommunications Number the sound wave that the mankind do not hear is converted to carry out the purposes such as ranging.In some embodiments, voicefrequency circuit 407 can also include Earphone jack.
Positioning component 408 is used for the current geographic position of positioning terminal 400, to realize navigation or LBS (Location Based Service, location based service).Positioning component 408 can be the GPS (Global based on the U.S. Positioning System, global positioning system), the dipper system of China, Russia Gray receive this system or European Union The positioning component of Galileo system.
Power supply 409 is used to be powered for the various components in terminal 400.Power supply 409 can be alternating current, direct current, Disposable battery or rechargeable battery.When power supply 409 includes rechargeable battery, which can support wired charging Or wireless charging.The rechargeable battery can be also used for supporting fast charge technology.
In some embodiments, terminal 400 further includes having one or more sensors 410.The one or more sensors 410 include but is not limited to: acceleration transducer 411, gyro sensor 412, pressure sensor 413, fingerprint sensor 414, Optical sensor 415 and proximity sensor 416.
The acceleration that acceleration transducer 411 can detecte in three reference axis of the coordinate system established with terminal 400 is big It is small.For example, acceleration transducer 411 can be used for detecting component of the acceleration of gravity in three reference axis.Processor 401 can With the acceleration of gravity signal acquired according to acceleration transducer 411, touch display screen 405 is controlled with transverse views or longitudinal view Figure carries out the display of user interface.Acceleration transducer 411 can be also used for the acquisition of game or the exercise data of user.
Gyro sensor 412 can detecte body direction and the rotational angle of terminal 400, and gyro sensor 412 can To cooperate with acquisition user to act the 3D of terminal 400 with acceleration transducer 411.Processor 401 is according to gyro sensor 412 Following function may be implemented in the data of acquisition: when action induction (for example changing UI according to the tilt operation of user), shooting Image stabilization, game control and inertial navigation.
The lower layer of side frame and/or touch display screen 405 in terminal 400 can be set in pressure sensor 413.Work as pressure When the side frame of terminal 400 is arranged in sensor 413, user can detecte to the gripping signal of terminal 400, by processor 401 Right-hand man's identification or prompt operation are carried out according to the gripping signal that pressure sensor 413 acquires.When the setting of pressure sensor 413 exists When the lower layer of touch display screen 405, the pressure operation of touch display screen 405 is realized to UI circle according to user by processor 401 Operability control on face is controlled.Operability control includes button control, scroll bar control, icon control, menu At least one of control.
Fingerprint sensor 414 is used to acquire the fingerprint of user, collected according to fingerprint sensor 414 by processor 401 The identity of fingerprint recognition user, alternatively, by fingerprint sensor 414 according to the identity of collected fingerprint recognition user.It is identifying When the identity of user is trusted identity out, authorize the user that there is relevant sensitive operation, the sensitive operation packet by processor 401 Include solution lock screen, check encryption information, downloading software, payment and change setting etc..Terminal can be set in fingerprint sensor 414 400 front, the back side or side.When being provided with physical button or manufacturer Logo in terminal 400, fingerprint sensor 414 can be with It is integrated with physical button or manufacturer's mark.
Optical sensor 415 is for acquiring ambient light intensity.In one embodiment, processor 401 can be according to optics The ambient light intensity that sensor 415 acquires controls the display brightness of touch display screen 405.Specifically, when ambient light intensity is higher When, the display brightness of touch display screen 405 is turned up;When ambient light intensity is lower, the display for turning down touch display screen 405 is bright Degree.In another embodiment, the ambient light intensity that processor 401 can also be acquired according to optical sensor 415, dynamic adjust The acquisition parameters of CCD camera assembly 406.
Proximity sensor 416, also referred to as range sensor are generally arranged at the front panel of terminal 400.Proximity sensor 416 For acquiring the distance between the front of user Yu terminal 400.In one embodiment, when proximity sensor 416 detects use When family and the distance between the front of terminal 400 gradually become smaller, touch display screen 405 is controlled from bright screen state by processor 401 It is switched to breath screen state;When proximity sensor 416 detects user and the distance between the front of terminal 400 becomes larger, Touch display screen 405 is controlled by processor 401 and is switched to bright screen state from breath screen state.
It will be understood by those skilled in the art that the restriction of the not structure paired terminal 400 of structure shown in Fig. 4, can wrap It includes than illustrating more or fewer components, perhaps combine certain components or is arranged using different components.
Fig. 5 is a kind of structural schematic diagram of server shown according to an exemplary embodiment, which can be because matching It sets or performance is different and generate bigger difference, may include one or more processors (central Processing units, CPU) 501 and one or more memory 502, wherein it is stored in the memory 502 There is at least one instruction, at least one instruction is loaded by the processor 501 and executed to realize that above-mentioned each method is real The method that example offer is provided.Certainly, which can also have wired or wireless network interface, keyboard and input/output interface Equal components, to carry out input and output, which can also include other for realizing the component of functions of the equipments, not do herein It repeats.
Server 500 can be used for executing step performed by processing unit in above-mentioned image partition method.
In the exemplary embodiment, a kind of non-transitorycomputer readable storage medium is additionally provided, when in storage medium Instruction by processing unit processor execute when so that processing unit is able to carry out a kind of image partition method, method includes:
The pixel classifications model trained is obtained, pixel classifications model is used to determine the classification of each pixel in any image Mark, class indication includes first identifier and second identifier, and first identifier is used to indicate corresponding pixel and belongs to foreground area, the Two marks are used to indicate corresponding pixel and belong to background area;
Target image to be processed is input in pixel classifications model, pixel classifications model is based on, determines target image In each pixel class indication;
According to the class indication of pixel each in target image, foreground area and the background area of target image are determined.
In the exemplary embodiment, a kind of computer program product is additionally provided, when the instruction in computer program product When being executed by the processor of processing unit, so that processing unit is able to carry out a kind of image partition method, method includes:
The pixel classifications model trained is obtained, pixel classifications model is used to determine the classification of each pixel in any image Mark, class indication includes first identifier and second identifier, and first identifier is used to indicate corresponding pixel and belongs to foreground area, the Two marks are used to indicate corresponding pixel and belong to background area;
Target image to be processed is input in pixel classifications model, pixel classifications model is based on, determines target image In each pixel class indication;
According to the class indication of pixel each in target image, foreground area and the background area of target image are determined.
Those skilled in the art will readily occur to other realities of the disclosure after considering specification and practicing disclosure herein Apply scheme.This application is intended to cover any variations, uses, or adaptations of the disclosure, these modifications, purposes or suitable The variation of answering property follows the general principles of this disclosure and including the undocumented common knowledge in the art of the disclosure or used Use technological means.The description and examples are only to be considered as illustrative, and the true scope and spirit of the disclosure are by following right It is required that pointing out.
It should be understood that the present disclosure is not limited to the precise structures that have been described above and shown in the drawings, and And various modifications and changes may be made without departing from the scope thereof.The scope of the present disclosure is only limited by the accompanying claims.

Claims (10)

1. a kind of image partition method, which is characterized in that the described method includes:
The pixel classifications model trained is obtained, the pixel classifications model is used to determine the classification of each pixel in any image Mark, the class indication include first identifier and second identifier, and the first identifier is used to indicate before corresponding pixel belongs to Scene area, the second identifier are used to indicate corresponding pixel and belong to background area;
Target image to be processed is input in the pixel classifications model, be based on the pixel classifications model, determine described in The class indication of each pixel in target image;
According to the class indication of pixel each in the target image, the foreground area and background area of the target image are determined Domain.
2. the method according to claim 1, wherein the method also includes:
Obtain the class indication of each pixel in multiple sample images and the multiple sample image;
Model training is carried out according to the class indication of each pixel in the multiple sample image and the multiple sample image, Obtain the pixel classifications model.
3. the method according to claim 1, wherein the classification according to pixel each in the target image Mark, determines foreground area and the background area of the target image, comprising:
According to the class indication of pixel each in the target image, prospect probability graph is generated, the prospect probability graph is for referring to The foreground area for showing the target image and the position where background area.
4. according to the method described in claim 3, it is characterized in that, the method also includes:
According to the prospect probability graph, enhancing processing is carried out to the foreground area of the target image;Alternatively,
According to the prospect probability graph, Fuzzy processing is carried out to the background area of the target image.
5. according to the method described in claim 4, it is characterized in that, described according to the prospect probability graph, to the target figure Background area as in carries out Fuzzy processing, comprising:
According to the prospect probability graph, the foreground area of the target image is extracted;
Fuzzy processing is carried out to the target image, obtains the first image, and according to the prospect probability graph, extracts described the The background area of one image;
The background area of the foreground area of the target image and the first image is combined, the second image is obtained.
6. according to the method described in claim 4, it is characterized in that, in the prospect probability graph, the picture of pixel in foreground area Element value is 1, and the pixel value of pixel is 0 in background area, described according to the prospect probability graph, to the background of the target image Region carries out Fuzzy processing, comprising:
According to the prospect probability graph, using following formula, Fuzzy processing is carried out to the background area of the target image:
Target=Source*mask+Gaussian* (255-mask);
Wherein, Source indicates that the target image, Gaussian indicate to obtain after the target image carries out Fuzzy processing Image, mask indicates that the prospect probability graph, Target indicate that the background area of the target image carries out Fuzzy processing The image obtained afterwards.
7. a kind of image segmentation device, which is characterized in that described device includes:
Model obtains module, is configured as obtaining the pixel classifications model trained, the pixel classifications model is appointed for determining The class indication of each pixel in one image, the class indication include first identifier and second identifier, and the first identifier is used Belong to foreground area in the corresponding pixel of instruction, the second identifier is used to indicate corresponding pixel and belongs to background area;
Determining module is identified, is configured as target image to be processed being input in the pixel classifications model, based on described Pixel classifications model determines the class indication of each pixel in the target image;
Area determination module is configured as the class indication according to pixel each in the target image, determines the target figure The foreground area of picture and background area.
8. device according to claim 7, which is characterized in that described device further include:
Sample acquisition module is configured as obtaining the classification of each pixel in multiple sample images and the multiple sample image Mark;
Training module is configured as the classification according to each pixel in the multiple sample image and the multiple sample image Mark carries out model training, obtains the pixel classifications model.
9. a kind of image segmentation device, which is characterized in that described device includes:
Processor;
Memory for storage processor executable command;
Wherein, the processor is configured to:
The pixel classifications model trained is obtained, the pixel classifications model is used to determine the classification of each pixel in any image Mark, the class indication include first identifier and second identifier, and the first identifier is used to indicate before corresponding pixel belongs to Scene area, the second identifier are used to indicate corresponding pixel and belong to background area;
Target image to be processed is input in the pixel classifications model, be based on the pixel classifications model, determine described in The class indication of each pixel in target image;
According to the class indication of pixel each in the target image, the foreground area and background area of the target image are determined Domain.
10. a kind of non-transitorycomputer readable storage medium, when the instruction in the storage medium is by the processing of processing unit When device executes, so that processing unit is able to carry out a kind of image partition method, which comprises
The pixel classifications model trained is obtained, the pixel classifications model is used to determine the classification of each pixel in any image Mark, the class indication include first identifier and second identifier, and the first identifier is used to indicate before corresponding pixel belongs to Scene area, the second identifier are used to indicate corresponding pixel and belong to background area;
Target image to be processed is input in the pixel classifications model, be based on the pixel classifications model, determine described in The class indication of each pixel in target image;
According to the class indication of pixel each in the target image, the foreground area and background area of the target image are determined Domain.
CN201811252058.XA 2018-10-25 2018-10-25 Image partition method, device and storage medium Pending CN109285178A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201811252058.XA CN109285178A (en) 2018-10-25 2018-10-25 Image partition method, device and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201811252058.XA CN109285178A (en) 2018-10-25 2018-10-25 Image partition method, device and storage medium

Publications (1)

Publication Number Publication Date
CN109285178A true CN109285178A (en) 2019-01-29

Family

ID=65178440

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201811252058.XA Pending CN109285178A (en) 2018-10-25 2018-10-25 Image partition method, device and storage medium

Country Status (1)

Country Link
CN (1) CN109285178A (en)

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109887045A (en) * 2019-03-29 2019-06-14 南京维沃软件技术有限公司 A kind of image display method and terminal device
CN110321865A (en) * 2019-07-09 2019-10-11 北京字节跳动网络技术有限公司 Head effect processing method and device, storage medium
CN110728632A (en) * 2019-09-04 2020-01-24 北京奇艺世纪科技有限公司 Image blurring processing method, image blurring processing device, computer device and storage medium
CN111539960A (en) * 2019-03-25 2020-08-14 华为技术有限公司 Image processing method and related device
CN111758118A (en) * 2020-05-26 2020-10-09 蜂图科技有限公司 Visual positioning method, device and equipment and readable storage medium
CN112270686A (en) * 2020-12-24 2021-01-26 北京达佳互联信息技术有限公司 Image segmentation model training method, image segmentation device and electronic equipment
CN112489053A (en) * 2020-11-26 2021-03-12 深圳市艾合芯科智慧医疗有限公司 Tongue image segmentation method and device and storage medium
CN113240611A (en) * 2021-05-28 2021-08-10 中建材信息技术股份有限公司 Foreign matter detection method based on picture sequence
CN113784207A (en) * 2021-07-30 2021-12-10 北京达佳互联信息技术有限公司 Video picture display method and device, electronic equipment and storage medium

Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7783075B2 (en) * 2006-06-07 2010-08-24 Microsoft Corp. Background blurring for video conferencing
CN102436637A (en) * 2010-09-29 2012-05-02 中国科学院计算技术研究所 Method and system for automatically segmenting hairs in head images
CN103262119A (en) * 2010-12-08 2013-08-21 阿尔卡特朗讯公司 Method and system for segmenting an image
CN104063865A (en) * 2014-06-27 2014-09-24 小米科技有限责任公司 Classification model creation method, image segmentation method and related device
CN105513105A (en) * 2015-12-07 2016-04-20 天津大学 Image background blurring method based on saliency map
CN106683147A (en) * 2017-01-23 2017-05-17 浙江大学 Method of image background blur
CN106920243A (en) * 2017-03-09 2017-07-04 桂林电子科技大学 The ceramic material part method for sequence image segmentation of improved full convolutional neural networks
CN107341805A (en) * 2016-08-19 2017-11-10 北京市商汤科技开发有限公司 Background segment and network model training, image processing method and device before image
CN107347166A (en) * 2016-08-19 2017-11-14 北京市商汤科技开发有限公司 Processing method, device and the terminal device of video image
CN107944499A (en) * 2017-12-10 2018-04-20 上海童慧科技股份有限公司 A kind of background detection method modeled at the same time for prospect background

Patent Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7783075B2 (en) * 2006-06-07 2010-08-24 Microsoft Corp. Background blurring for video conferencing
CN102436637A (en) * 2010-09-29 2012-05-02 中国科学院计算技术研究所 Method and system for automatically segmenting hairs in head images
CN103262119A (en) * 2010-12-08 2013-08-21 阿尔卡特朗讯公司 Method and system for segmenting an image
CN104063865A (en) * 2014-06-27 2014-09-24 小米科技有限责任公司 Classification model creation method, image segmentation method and related device
CN105513105A (en) * 2015-12-07 2016-04-20 天津大学 Image background blurring method based on saliency map
CN107341805A (en) * 2016-08-19 2017-11-10 北京市商汤科技开发有限公司 Background segment and network model training, image processing method and device before image
CN107347166A (en) * 2016-08-19 2017-11-14 北京市商汤科技开发有限公司 Processing method, device and the terminal device of video image
CN106683147A (en) * 2017-01-23 2017-05-17 浙江大学 Method of image background blur
CN106920243A (en) * 2017-03-09 2017-07-04 桂林电子科技大学 The ceramic material part method for sequence image segmentation of improved full convolutional neural networks
CN107944499A (en) * 2017-12-10 2018-04-20 上海童慧科技股份有限公司 A kind of background detection method modeled at the same time for prospect background

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
周修庄 等,: "《视觉跟踪中的马氏链蒙特卡洛方法》", 30 June 2018, 北京邮电大学出版社, *

Cited By (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111539960B (en) * 2019-03-25 2023-10-24 华为技术有限公司 Image processing method and related device
CN111539960A (en) * 2019-03-25 2020-08-14 华为技术有限公司 Image processing method and related device
WO2020192692A1 (en) * 2019-03-25 2020-10-01 华为技术有限公司 Image processing method and related apparatus
CN109887045A (en) * 2019-03-29 2019-06-14 南京维沃软件技术有限公司 A kind of image display method and terminal device
CN110321865A (en) * 2019-07-09 2019-10-11 北京字节跳动网络技术有限公司 Head effect processing method and device, storage medium
CN110728632B (en) * 2019-09-04 2022-07-12 北京奇艺世纪科技有限公司 Image blurring processing method, image blurring processing device, computer device and storage medium
CN110728632A (en) * 2019-09-04 2020-01-24 北京奇艺世纪科技有限公司 Image blurring processing method, image blurring processing device, computer device and storage medium
CN111758118A (en) * 2020-05-26 2020-10-09 蜂图科技有限公司 Visual positioning method, device and equipment and readable storage medium
CN111758118B (en) * 2020-05-26 2024-04-16 蜂图志科技控股有限公司 Visual positioning method, device, equipment and readable storage medium
CN112489053A (en) * 2020-11-26 2021-03-12 深圳市艾合芯科智慧医疗有限公司 Tongue image segmentation method and device and storage medium
CN112489053B (en) * 2020-11-26 2021-07-13 深圳市艾合芯科智慧医疗有限公司 Tongue image segmentation method and device and storage medium
CN112270686A (en) * 2020-12-24 2021-01-26 北京达佳互联信息技术有限公司 Image segmentation model training method, image segmentation device and electronic equipment
CN112270686B (en) * 2020-12-24 2021-03-16 北京达佳互联信息技术有限公司 Image segmentation model training method, image segmentation device and electronic equipment
CN113240611A (en) * 2021-05-28 2021-08-10 中建材信息技术股份有限公司 Foreign matter detection method based on picture sequence
CN113240611B (en) * 2021-05-28 2024-05-07 中建材信息技术股份有限公司 Foreign matter detection method based on picture sequence
CN113784207A (en) * 2021-07-30 2021-12-10 北京达佳互联信息技术有限公司 Video picture display method and device, electronic equipment and storage medium

Similar Documents

Publication Publication Date Title
CN110189340B (en) Image segmentation method and device, electronic equipment and storage medium
CN109285178A (en) Image partition method, device and storage medium
CN108594997B (en) Gesture skeleton construction method, device, equipment and storage medium
EP3828769B1 (en) Image processing method and apparatus, terminal and computer-readable storage medium
CN110992493B (en) Image processing method, device, electronic equipment and storage medium
CN109829456A (en) Image-recognizing method, device and terminal
CN110139142A (en) Virtual objects display methods, device, terminal and storage medium
WO2019024717A1 (en) Anti-counterfeiting processing method and related product
CN110650379B (en) Video abstract generation method and device, electronic equipment and storage medium
CN108401124A (en) The method and apparatus of video record
CN109360222A (en) Image partition method, device and storage medium
CN112907725B (en) Image generation, training of image processing model and image processing method and device
CN110290421A (en) Frame per second method of adjustment, device, computer equipment and storage medium
CN110222789A (en) Image-recognizing method and storage medium
CN112287852B (en) Face image processing method, face image display method, face image processing device and face image display equipment
CN113763228B (en) Image processing method, device, electronic equipment and storage medium
CN110827195B (en) Virtual article adding method and device, electronic equipment and storage medium
CN111144365A (en) Living body detection method, living body detection device, computer equipment and storage medium
CN111723803B (en) Image processing method, device, equipment and storage medium
CN109558837A (en) Face critical point detection method, apparatus and storage medium
CN109978996B (en) Method, device, terminal and storage medium for generating expression three-dimensional model
CN110149517A (en) Method, apparatus, electronic equipment and the computer storage medium of video processing
CN110290426A (en) Method, apparatus, equipment and the storage medium of showing resource
CN110139143A (en) Virtual objects display methods, device, computer equipment and storage medium
CN110991457A (en) Two-dimensional code processing method and device, electronic equipment and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20190129

RJ01 Rejection of invention patent application after publication