WO2020125495A1 - Panoramic segmentation method, apparatus and device - Google Patents

Panoramic segmentation method, apparatus and device Download PDF

Info

Publication number
WO2020125495A1
WO2020125495A1 PCT/CN2019/124334 CN2019124334W WO2020125495A1 WO 2020125495 A1 WO2020125495 A1 WO 2020125495A1 CN 2019124334 W CN2019124334 W CN 2019124334W WO 2020125495 A1 WO2020125495 A1 WO 2020125495A1
Authority
WO
WIPO (PCT)
Prior art keywords
instance
segmentation
panoramic
space
original image
Prior art date
Application number
PCT/CN2019/124334
Other languages
French (fr)
Chinese (zh)
Inventor
张维桐
张锲石
程俊
Original Assignee
中国科学院深圳先进技术研究院
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 中国科学院深圳先进技术研究院 filed Critical 中国科学院深圳先进技术研究院
Publication of WO2020125495A1 publication Critical patent/WO2020125495A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/13Edge detection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/136Segmentation; Edge detection involving thresholding
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/194Segmentation; Edge detection involving foreground-background segmentation

Definitions

  • the present application belongs to the field of image processing, and particularly relates to a panoramic segmentation method, device and equipment.
  • Panoramic segmentation as an emerging field, has important application value and development prospects in many fields, such as security control, industrial robot applications, and automobile assisted driving.
  • individual instances with different shapes, complex and diverse background environments, dynamically changing scenes between pedestrians and perspectives, strict requirements for real-time and stability of the system, etc.
  • Semantic segmentation and instance segmentation can no longer meet the current needs, panoramic segmentation should be applied It is born to supplement and solve the current problems, and poses great challenges to the panoramic segmentation problem.
  • the current deep-learning-based panoramic segmentation methods mostly rely on the selection of candidate frame regions for measurement. There is no way to identify and segment all pixels or shared pixels, and the current panoramic segmentation method is usually a combination of multiple sub-networks. Frame.
  • embodiments of the present application provide a panoramic segmentation method, device, and equipment to solve the panoramic segmentation method in the prior art, which cannot identify and segment all pixels or shared pixels, and cannot achieve an end-to-end frame problem.
  • a first aspect of an embodiment of the present application provides a panoramic segmentation method.
  • the panoramic segmentation method includes:
  • the clustering loss function is used to further distinguish different instances, and the panoramic segmentation result is obtained.
  • the step of semantically segmenting the original image includes:
  • the fully convolutional structure of the fully connected layer based on the VGG model is used as the skeleton framework, and the conditional random recursive neural network containing closely connected pairs is randomly selected as the final layer of the model to semantically segment the original image.
  • the target and background obtained by segmenting an instance are used as examples, and guided by a semantic segmentation output map, To make the centers of the instances embedded in the space mutually exclusive, the pixels within the range of the instance are attracted to the center of the instance, and the steps of segmenting the image include:
  • the centers of the instances in the embedded space are mutually repelled, and the pixels of the instance are clustered according to the preset radius of attractive force.
  • the clustering loss function is:
  • S represents the number of calibrated clusters in the standard data
  • E s represents all the elements contained in the cluster S
  • x i represents the embedding space
  • represents all the clustering centers of S
  • represents the depth space Distance
  • ⁇ pull and ⁇ push respectively represent the edge threshold of gravity and repulsion in the embedded space
  • N s represents the number of pixels included in the S clustering instance
  • ⁇ , ⁇ , ⁇ , ⁇ are adjustment parameters.
  • the target and background obtained from the instance segmentation are taken as examples, and the output graph of the semantic segmentation is used to guide, so that the centers between the instances embedded in the space are mutually exclusive ,
  • the pixels within the scope of the instance are attracted to the center of the instance, and the steps of segmenting the image include:
  • the semantic calibration and mask in the original picture are generated according to the semantic segmentation, the instance of the multi-dimensional pixel embedding is generated by the instance segmentation of the embedding space, and the cluster fusion is performed by the depth metric space, and the aggregated segmented image is output.
  • a second aspect of an embodiment of the present application provides a panoramic segmentation device, the panoramic segmentation device includes:
  • the original image acquisition unit is used to acquire the original image to be segmented
  • a segmentation unit used for semantic segmentation of the original image, and for instance segmentation of the original image through a distance learning method of embedded space;
  • the fusion unit is used to take the target and background obtained by instance segmentation as an instance, and guide through the semantic segmentation output map, so that the centers between the instances embedded in the space are mutually exclusive, and the pixels within the instance range are attracted to the center of the instance to segment the image ;
  • the loss training unit is used to further distinguish different instances by using a clustering loss function to obtain a panoramic segmentation result.
  • the fusion unit includes:
  • the instance determination subunit is used to determine the center point of the instance by using the target and background obtained from the instance segmentation as the instance;
  • the clustering unit is used to repel each instance center of the embedded space according to a preset radius of repulsive force of the instance, and to cluster pixels of the instance according to a preset radius of attractive force.
  • the clustering loss function is:
  • S represents the number of calibrated clusters in the standard data
  • E s represents all the elements contained in the cluster S
  • x i represents the embedding space
  • represents all the clustering centers of S
  • represents the depth space Distance
  • ⁇ pull and ⁇ push respectively represent the edge threshold of gravity and repulsion in the embedded space
  • N s represents the number of pixels included in the S clustering instance
  • ⁇ , ⁇ , ⁇ , ⁇ are adjustment parameters.
  • a third aspect of the embodiments of the present application provides a panoramic segmentation device, including a memory, a processor, and a computer program stored in the memory and executable on the processor, and the processor executes the computer program To realize the steps of the panoramic segmentation method according to any one of the first aspect.
  • a fourth aspect of the embodiments of the present application provides a computer-readable storage medium that stores a computer program, and when the computer program is executed by a processor, the panoramic view described in any one of the first aspect is implemented The steps of the segmentation method.
  • the embodiments of the present application have the following beneficial effects: after obtaining the original image to be segmented, the original image is semantically segmented based on the semantic segmented image, and the original image is measured by the metric distance learning method of the embedded space Semantic segmentation, embedding spatial clustering operations based on semantic segmentation images, can process all pixels of the image, and further distinguish different instances through the loss function, thus achieving an end-to-end panoramic segmentation network framework.
  • FIG. 1 is a schematic diagram of an implementation process of a panoramic segmentation method provided by an embodiment of the present application
  • FIG. 2 is a schematic structural diagram of a panoramic segmentation system provided by an embodiment of the present application.
  • FIG. 3 is a schematic structural diagram of an example of an embedded space provided by an embodiment of the present application.
  • FIG. 4 is a schematic diagram of a panoramic segmentation device provided by an embodiment of the present application.
  • FIG. 5 is a schematic diagram of a panoramic segmentation device provided by an embodiment of the present application.
  • FIG. 1 is a schematic diagram of an implementation process of a panoramic segmentation method provided by an embodiment of the present application, and details are as follows:
  • step S101 obtain the original image to be divided
  • the original image to be divided may be a single picture or an image sequence in a video.
  • step S102 perform semantic segmentation on the original image, and perform instance segmentation on the original image by using the metric distance learning method of the embedded space;
  • Semantic segmentation is used to group or categorize all pixels in the picture according to the semantic meaning expressed.
  • FCN based on VGG model can be used as a skeleton framework to include conditional random fields of recurrent neural networks with closely connected pairs As the final layer of the model.
  • the fully convolutional structure and design of the fully connected layer based on the VGG model can further improve the segmentation quality of the semantic segmentation pixel level, and calibrate the pixel standard between detection and output. It is worth noting that the whole process is a field process that can be deduced differentially. This application adopts the output of this process as an identifier for instance detection in the next stage.
  • step S103 the target and background obtained by instance segmentation are used as an instance, and the output of the semantic segmentation is used to guide the center between the instances in the embedded space.
  • the pixels within the instance range are attracted to the center of the instance to segment the image.
  • the semantic calibration and mask in the original image are generated through semantic segmentation, and the N-dimensional pixel embedding is generated through the instance segmentation processing in the embedding space, so that the instance segmentation can match the output of the semantic segmentation well.
  • Highly clustered segmented images are generated through semantic segmentation, and the N-dimensional pixel embedding is generated through the instance segmentation processing in the embedding space, so that the instance segmentation can match the output of the semantic segmentation well.
  • Attraction within an instance attract the relevant pixel embedding point to the center of the instance within the scope of an instance
  • the threshold of the action distance for the repulsive force and attractive force that is, the repulsive force radius and attractive force radius of the instance.
  • the two The instance centers are mutually exclusive, making the instance segmentation more accurate; when the distance between the pixels embedded in the space and the center point of the instance is less than the attractive radius, the pixels will be clustered by the attractive force.
  • the central point of the instance in the embedded space does not cause too much attraction to pixels in other central domains in the range of action, and sufficient repulsion is ensured between multiple central points, so that there will not be too much or too little. Negative effect. In addition, this can ensure that the embedded pixels are as close as possible to the center without independent points. Appropriate restraint and relaxation achieve the clustering effect in the embedded space as shown in Figure 2. Through multiple iterations, a clustering algorithm is used to obtain pixel-level segmentation, and a panoramic segmentation method based on semantic overlay examples is achieved.
  • step S104 a clustering loss function is used to further distinguish different instances to obtain a panoramic segmentation result.
  • the loss function may be:
  • S represents the number of calibrated clusters in the standard data
  • E s represents all the elements contained in the cluster S
  • x i represents the embedding space
  • represents all the clustering centers of S
  • represents the depth space Distance
  • ⁇ pull and ⁇ push respectively represent the edge threshold of gravity and repulsion in the embedded space
  • N s represents the number of pixels included in the S clustering instance
  • ⁇ , ⁇ , ⁇ , ⁇ are adjustment parameters.
  • ⁇ and ⁇ are 1, ⁇ is 0.001, and ⁇ is 0.7.
  • FIG. 3 is a schematic diagram of a panoramic segmentation framework provided by an embodiment of the present application.
  • the end-to-end network framework After inputting the original image to be segmented into an end-to-end network framework, the end-to-end network framework The original image is shared and decoded.
  • the decoded data is divided into two branches for processing.
  • the lower semantic segmentation branch is used to train the semantic calibration and mask in the generated picture.
  • the upper branch is divided by the instance of the embedded space to generate N( N is a natural number)-dimensional pixel embedding, so that the instance segmentation branch can match the output of semantic segmentation well.
  • each instance of the embedded space is subjected to repulsive iteration, and the iterative calculation of attraction within the instance is obtained to obtain pixel-level segmentation and achieve a panoramic view based on semantic overlay instances segmentation.
  • FIG. 4 is a schematic structural diagram of a panoramic segmentation device according to an embodiment of the present application. Details are as follows:
  • the panoramic segmentation device includes:
  • the original image obtaining unit 401 is used to obtain an original image to be divided
  • a segmentation unit 402 is used to perform semantic segmentation on the original image, and perform instance segmentation on the original image by using the metric distance learning method of the embedded space;
  • the fusion unit 403 is used to take the target and background obtained by instance segmentation as examples, and guide through the semantic segmentation output map, so that the centers between the instances embedded in the space are mutually exclusive, and the pixels within the range of the instance are attracted to the center of the instance to perform the image segmentation;
  • the loss training unit 404 is used to further distinguish different instances by using a clustering loss function to obtain a panoramic segmentation result.
  • the fusion unit includes:
  • the instance determination subunit is used to determine the center point of the instance by using the target and background obtained from the instance segmentation as the instance;
  • the clustering unit is used to repel each instance center of the embedded space according to a preset radius of repulsive force of the instance, and to cluster pixels of the instance according to a preset radius of attractive force.
  • the clustering loss function is:
  • S represents the number of calibrated clusters in the standard data
  • E s represents all the elements contained in the cluster S
  • x i represents the embedding space
  • represents all the clustering centers of S
  • represents the depth space Distance
  • ⁇ pull and ⁇ push respectively represent the edge threshold of gravity and repulsion in the embedded space
  • N s represents the number of pixels included in the S clustering instance
  • ⁇ , ⁇ , ⁇ , ⁇ are adjustment parameters.
  • the panoramic segmentation device shown in FIG. 4 corresponds to the panoramic segmentation method shown in FIG. 1.
  • FIG. 5 is a schematic diagram of a panoramic segmentation device provided by an embodiment of the present application.
  • the panoramic segmentation device 5 of this embodiment includes: a processor 50, a memory 51, and a computer program 52 stored in the memory 51 and executable on the processor 50, for example, a panoramic segmentation program.
  • the processor 50 executes the computer program 52, the steps in the foregoing embodiments of the panoramic segmentation method are implemented.
  • the processor 50 executes the computer program 52, the functions of the modules/units in the foregoing device embodiments are realized.
  • the computer program 52 may be divided into one or more modules/units, and the one or more modules/units are stored in the memory 51 and executed by the processor 50 to complete This application.
  • the one or more modules/units may be a series of computer program instruction segments capable of performing specific functions, and the instruction segments are used to describe the execution process of the computer program 52 in the panoramic segmentation device 5.
  • the computer program 52 may be divided into:
  • the original image acquisition unit is used to acquire the original image to be segmented
  • a segmentation unit used for semantic segmentation of the original image, and for instance segmentation of the original image through a distance learning method of embedded space;
  • the fusion unit is used to take the target and background obtained by instance segmentation as an instance, and guide through the semantic segmentation output map, so that the centers between the instances embedded in the space are mutually exclusive, and the pixels within the instance range are attracted to the center of the instance to segment the image ;
  • the loss training unit is used to further distinguish different instances by using a clustering loss function to obtain a panoramic segmentation result.
  • the panoramic segmentation device 5 may be a computing device such as a desktop computer, a notebook, a palmtop computer and a cloud server.
  • the panoramic segmentation device may include, but is not limited to, the processor 50 and the memory 51.
  • FIG. 5 is only an example of the panoramic segmentation device 5 and does not constitute a limitation on the panoramic segmentation device 5, and may include more or fewer components than the illustration, or a combination of certain components, or different Components, for example, the panoramic segmentation device may also include an input and output device, a network access device, a bus, and the like.
  • the so-called processor 50 may be a central processing unit (Central Processing Unit, CPU), or other general-purpose processors, digital signal processors (Digital Signal Processor, DSP), application specific integrated circuits (Application Specific Integrated Circuit, ASIC), Ready-made programmable gate array (Field-Programmable Gate Array, FPGA) or other programmable logic devices, discrete gate or transistor logic devices, discrete hardware components, etc.
  • the general-purpose processor may be a microprocessor or the processor may be any conventional processor or the like.
  • the memory 51 may be an internal storage unit of the panoramic segmentation device 5, such as a hard disk or a memory of the panoramic segmentation device 5.
  • the memory 51 may also be an external storage device of the panoramic segmentation device 5, for example, a plug-in hard disk equipped on the panoramic segmentation device 5, a smart memory card (Smart, Media, Card, SMC), and a secure digital (Secure Digital, SD) card, flash memory card (Flash Card), etc.
  • the memory 51 may also include both an internal storage unit of the panoramic segmentation device 5 and an external storage device.
  • the memory 51 is used to store the computer program and other programs and data required by the panoramic segmentation device.
  • the memory 51 can also be used to temporarily store data that has been or will be output.
  • each functional unit and module is used as an example for illustration.
  • the above-mentioned functions may be allocated by different functional units
  • Module completion means that the internal structure of the device is divided into different functional units or modules to complete all or part of the functions described above.
  • the functional units and modules in the embodiment may be integrated into one processing unit, or each unit may exist alone physically, or two or more units may be integrated into one unit.
  • the above integrated unit may use hardware It can also be implemented in the form of software functional units.
  • the specific names of each functional unit and module are only for the purpose of distinguishing each other, and are not used to limit the protection scope of the present application.
  • the disclosed device/terminal device and method may be implemented in other ways.
  • the device/terminal device embodiments described above are only schematic.
  • the division of the module or unit is only a logical function division, and in actual implementation, there may be another division manner, such as multiple units Or components can be combined or integrated into another system, or some features can be ignored or not implemented.
  • the displayed or discussed mutual coupling or direct coupling or communication connection may be indirect coupling or communication connection through some interfaces, devices or units, and may be in electrical, mechanical or other forms.
  • the units described as separate components may or may not be physically separated, and the components displayed as units may or may not be physical units, that is, they may be located in one place, or may be distributed on multiple network units. Some or all of the units may be selected according to actual needs to achieve the purpose of the solution of this embodiment.
  • each functional unit in each embodiment of the present application may be integrated into one processing unit, or each unit may exist alone physically, or two or more units may be integrated into one unit.
  • the above integrated unit can be implemented in the form of hardware or software function unit.
  • the integrated module/unit is implemented in the form of a software functional unit and sold or used as an independent product, it may be stored in a computer-readable storage medium.
  • the present application can implement all or part of the processes in the methods of the above embodiments, or it can be completed by a computer program instructing relevant hardware.
  • the computer program can be stored in a computer-readable storage medium. When the program is executed by the processor, the steps of the foregoing method embodiments may be implemented.
  • the computer program includes computer program code, and the computer program code may be in a source code form, an object code form, an executable file, or some intermediate form, etc.
  • the computer-readable medium may include: any entity or device capable of carrying the computer program code, a recording medium, a USB flash drive, a mobile hard disk, a magnetic disk, an optical disc, a computer memory, and a read-only memory (ROM, Read-Only Memory) , Random Access Memory (RAM, Random Access Memory), electrical carrier signals, telecommunications signals and software distribution media, etc.
  • ROM Read-Only Memory
  • RAM Random Access Memory
  • electrical carrier signals telecommunications signals and software distribution media, etc.

Landscapes

  • Engineering & Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Image Analysis (AREA)

Abstract

Disclosed is a panoramic segmentation method, comprising: acquiring an original image to be segmented; carrying out semantic segmentation on the original image, and by means of a measured distance learning method for an embedded space, carrying out example segmentation on the original image; taking a target and a background obtained by means of example segmentation as examples, and carrying out guidance by means of a semantic segmentation output image, such that the centers of examples of the embedded space repel each other, and pixels within an example range are attracted to the centers of the examples to segment the image; and using a clustering loss function to further distinguish between different examples in order to obtain a panoramic segmentation result. An embedded space clustering operation is performed based on semantic segmentation of the image to process all the pixels of the image, and the different examples are further distinguished by means of the loss function, thereby realizing an end-to-end panoramic segmentation network framework.

Description

一种全景分割方法、装置及设备Panorama segmentation method, device and equipment 技术领域Technical field
本申请属于图像处理领域,尤其涉及一种全景分割方法、装置及设备。The present application belongs to the field of image processing, and particularly relates to a panoramic segmentation method, device and equipment.
背景技术Background technique
全景分割作为一个新兴领域,在很多领域有重要的应用价值和发展前景,如安保布控、工业机器人应用、汽车辅助驾驶等。但形态各异的实例个体、复杂多样的背景环境、行人与视角之间动态变化的场景、系统实时性与稳定性的严格要求等,语义分割和实例分割早已不能满足当下的需求,全景分割应运而生来补充和解决当下的问题,并对全景分割问题提出了很大挑战。Panoramic segmentation, as an emerging field, has important application value and development prospects in many fields, such as security control, industrial robot applications, and automobile assisted driving. However, individual instances with different shapes, complex and diverse background environments, dynamically changing scenes between pedestrians and perspectives, strict requirements for real-time and stability of the system, etc. Semantic segmentation and instance segmentation can no longer meet the current needs, panoramic segmentation should be applied It is born to supplement and solve the current problems, and poses great challenges to the panoramic segmentation problem.
目前基于深度学习的全景分割方法,大多依赖于测定候选框架区域的选取,存在不能对所有像素或共享像素进行识别和分割,并且目前的全景分割方法通常是多个子网络的联合,不能达成端到端的框架。The current deep-learning-based panoramic segmentation methods mostly rely on the selection of candidate frame regions for measurement. There is no way to identify and segment all pixels or shared pixels, and the current panoramic segmentation method is usually a combination of multiple sub-networks. Frame.
技术问题technical problem
有鉴于此,本申请实施例提供了一种全景分割方法、装置及设备,以解决现有技术中的全景分割方法,不能对所有像素或共享像素进行识别和分割,不能达成端到端的框架的问题。In view of this, embodiments of the present application provide a panoramic segmentation method, device, and equipment to solve the panoramic segmentation method in the prior art, which cannot identify and segment all pixels or shared pixels, and cannot achieve an end-to-end frame problem.
技术解决方案Technical solution
本申请实施例的第一方面提供了一种全景分割方法,所述全景分割方法包括:A first aspect of an embodiment of the present application provides a panoramic segmentation method. The panoramic segmentation method includes:
获取待分割的原始图像;Obtain the original image to be segmented;
对所述原始图像进行语义分割,以及通过嵌入空间的度量距离学习方法,对所述原始图像进行实例分割;Performing semantic segmentation on the original image, and performing instance segmentation on the original image through a distance learning method of embedded space;
将实例分割得到的目标和背景作为实例,通过语义分割输出图进行引导,使嵌入空间的实例之间的中心互相排斥,实例范围内的像素吸引至实例中心,对图像进行分割;Take the target and background obtained by instance segmentation as examples, and guide through the output graph of semantic segmentation, so that the centers between the instances embedded in the space are mutually exclusive, and the pixels within the instance range are attracted to the center of the instance to segment the image;
采用聚类损失函数进一步区分不同实例,得到全景分割结果。The clustering loss function is used to further distinguish different instances, and the panoramic segmentation result is obtained.
结合第一方面,在第一方面的第一种可能实现方式中,所述对所述原始图像进行语义分割的步骤包括:With reference to the first aspect, in a first possible implementation manner of the first aspect, the step of semantically segmenting the original image includes:
采用基于VGG模型的全连接层的完全卷积结构作为骨骼框架,以包含有紧密连接对的递归神经网络的条件随机作为模型的最终层,对所述原始图像进行语义分割。The fully convolutional structure of the fully connected layer based on the VGG model is used as the skeleton framework, and the conditional random recursive neural network containing closely connected pairs is randomly selected as the final layer of the model to semantically segment the original image.
结合第一方面或第一方面的第一种可能实现方式,在第一方面的第二种可能实现方式中,所述将实例分割得到的目标和背景作为实例,通过语义分割输出图进行引导,使嵌入空间的实例之间的中心互相排斥,实例范围内的像素吸引至实例中心,对图像进行分割的步骤包括:With reference to the first aspect or the first possible implementation manner of the first aspect, in a second possible implementation manner of the first aspect, the target and background obtained by segmenting an instance are used as examples, and guided by a semantic segmentation output map, To make the centers of the instances embedded in the space mutually exclusive, the pixels within the range of the instance are attracted to the center of the instance, and the steps of segmenting the image include:
将实例分割得到的目标和背景作为实例,确定实例的中心点;Use the target and background obtained by segmenting the instance as an example to determine the center point of the instance;
根据预先设定的实例排斥力半径,使嵌入空间的实例中心互相排斥,以及根据预先设定的吸引力半径,对实例的像素点进行聚类。According to the preset radius of repulsive force of the instance, the centers of the instances in the embedded space are mutually repelled, and the pixels of the instance are clustered according to the preset radius of attractive force.
结合第一方面,在第一方面的第三种可能实现方式中,所述聚类损失函数为:With reference to the first aspect, in a third possible implementation manner of the first aspect, the clustering loss function is:
L=α·L pull+β·L push+γ·L nor+θ·L seg L=α·L pull +β·L push +γ·L nor +θ·L seg
其中:among them:
Figure PCTCN2019124334-appb-000001
Figure PCTCN2019124334-appb-000001
Figure PCTCN2019124334-appb-000002
Figure PCTCN2019124334-appb-000002
Figure PCTCN2019124334-appb-000003
Figure PCTCN2019124334-appb-000003
S表示标准数据当中标定的聚类个数,E s代表了聚类S当中包含的所有元素,x i代表嵌入空间,μ代表S的所有聚类中心,||||代表着深度空间当中的距离,η pull和η push分别表示引力与斥力在嵌入空间当中的作用边缘阈值,N s表示S聚类实例当中包含的像素个数,α、β、γ、θ为调节参数。 S represents the number of calibrated clusters in the standard data, E s represents all the elements contained in the cluster S, x i represents the embedding space, μ represents all the clustering centers of S, |||| represents the depth space Distance, η pull and η push respectively represent the edge threshold of gravity and repulsion in the embedded space, N s represents the number of pixels included in the S clustering instance, and α, β, γ, θ are adjustment parameters.
结合第一方面,在第一方面的第四种可能实现方式中,所述将实例分割得到的目标和背景作为实例,通过语义分割输出图进行引导,使嵌入空间的实例之间的中心互相排斥,实例范围内的像素吸引至实例中心,对图像进行分割的步骤包括:With reference to the first aspect, in a fourth possible implementation manner of the first aspect, the target and background obtained from the instance segmentation are taken as examples, and the output graph of the semantic segmentation is used to guide, so that the centers between the instances embedded in the space are mutually exclusive , The pixels within the scope of the instance are attracted to the center of the instance, and the steps of segmenting the image include:
根据语义分割生成原始图片中的语义标定和掩码,通过嵌入空间的实例分割生成多维像素嵌入的实例,通过深度度量空间进行聚类融合,输出聚集的分割图像。The semantic calibration and mask in the original picture are generated according to the semantic segmentation, the instance of the multi-dimensional pixel embedding is generated by the instance segmentation of the embedding space, and the cluster fusion is performed by the depth metric space, and the aggregated segmented image is output.
本申请实施例的第二方面提供了一种全景分割装置,所述全景分割装置包括:A second aspect of an embodiment of the present application provides a panoramic segmentation device, the panoramic segmentation device includes:
原始图像获取单元,用于获取待分割的原始图像;The original image acquisition unit is used to acquire the original image to be segmented;
分割单元,用于对所述原始图像进行语义分割,以及通过嵌入空间的度量距离学习方法,对所述原始图像进行实例分割;A segmentation unit, used for semantic segmentation of the original image, and for instance segmentation of the original image through a distance learning method of embedded space;
融合单元,用于将实例分割得到的目标和背景作为实例,通过语义分割输出图进行引导,使嵌入空间的实例之间的中心互相排斥,实例范围内的像素吸引至实例中心,对图像进行分割;The fusion unit is used to take the target and background obtained by instance segmentation as an instance, and guide through the semantic segmentation output map, so that the centers between the instances embedded in the space are mutually exclusive, and the pixels within the instance range are attracted to the center of the instance to segment the image ;
损失训练单元,用于采用聚类损失函数进一步区分不同实例,得到全景分割结果。The loss training unit is used to further distinguish different instances by using a clustering loss function to obtain a panoramic segmentation result.
结合第二方面,在第二方面的第一种可能实现方式中,所述融合单元包括:With reference to the second aspect, in a first possible implementation manner of the second aspect, the fusion unit includes:
实例确定子单元,用于将实例分割得到的目标和背景作为实例,确定实例 的中心点;The instance determination subunit is used to determine the center point of the instance by using the target and background obtained from the instance segmentation as the instance;
聚类单元,用于根据预先设定的实例排斥力半径,使嵌入空间的实例中心互相排斥,以及根据预先设定的吸引力半径,对实例的像素点进行聚类。The clustering unit is used to repel each instance center of the embedded space according to a preset radius of repulsive force of the instance, and to cluster pixels of the instance according to a preset radius of attractive force.
结合第二方面,在第二方面的第二种可能实现方式中,所述聚类损失函数为:With reference to the second aspect, in a second possible implementation manner of the second aspect, the clustering loss function is:
L=α·L pull+β·L push+γ·L nor+θ·L seg L=α·L pull +β·L push +γ·L nor +θ·L seg
其中:among them:
Figure PCTCN2019124334-appb-000004
Figure PCTCN2019124334-appb-000004
Figure PCTCN2019124334-appb-000005
Figure PCTCN2019124334-appb-000005
Figure PCTCN2019124334-appb-000006
Figure PCTCN2019124334-appb-000006
S表示标准数据当中标定的聚类个数,E s代表了聚类S当中包含的所有元素,x i代表嵌入空间,μ代表S的所有聚类中心,||||代表着深度空间当中的距离,η pull和η push分别表示引力与斥力在嵌入空间当中的作用边缘阈值,N s表示S聚类实例当中包含的像素个数,α、β、γ、θ为调节参数。 S represents the number of calibrated clusters in the standard data, E s represents all the elements contained in the cluster S, x i represents the embedding space, μ represents all the clustering centers of S, |||| represents the depth space Distance, η pull and η push respectively represent the edge threshold of gravity and repulsion in the embedded space, N s represents the number of pixels included in the S clustering instance, and α, β, γ, θ are adjustment parameters.
本申请实施例的第三方面提供了一种全景分割设备,包括存储器、处理器以及存储在所述存储器中并可在所述处理器上运行的计算机程序,所述处理器执行所述计算机程序时实现如第一方面任一项所述全景分割方法的步骤。A third aspect of the embodiments of the present application provides a panoramic segmentation device, including a memory, a processor, and a computer program stored in the memory and executable on the processor, and the processor executes the computer program To realize the steps of the panoramic segmentation method according to any one of the first aspect.
本申请实施例的第四方面提供了一种计算机可读存储介质,所述计算机可读存储介质存储有计算机程序,所述计算机程序被处理器执行时实现如第一方面任一项所述全景分割方法的步骤。A fourth aspect of the embodiments of the present application provides a computer-readable storage medium that stores a computer program, and when the computer program is executed by a processor, the panoramic view described in any one of the first aspect is implemented The steps of the segmentation method.
有益效果Beneficial effect
本申请实施例与现有技术相比存在的有益效果是:在获取到待分割的原始图像后,通过基于语义分割图像对原始图像进行语义分割,通过嵌入空间的度量距离学习方法对原始图像进行语义分割,基于语义分割图像进行嵌入空间聚类操作,得以对图像所有像素进行处理,并通过损失函数进一步区分不同实例,从而实现了端到端的全景分割网络框架。Compared with the prior art, the embodiments of the present application have the following beneficial effects: after obtaining the original image to be segmented, the original image is semantically segmented based on the semantic segmented image, and the original image is measured by the metric distance learning method of the embedded space Semantic segmentation, embedding spatial clustering operations based on semantic segmentation images, can process all pixels of the image, and further distinguish different instances through the loss function, thus achieving an end-to-end panoramic segmentation network framework.
附图说明BRIEF DESCRIPTION
为了更清楚地说明本申请实施例中的技术方案,下面将对实施例或现有技术描述中所需要使用的附图作简单地介绍,显而易见地,下面描述中的附图仅仅是本申请的一些实施例,对于本领域普通技术人员来讲,在不付出创造性劳动性的前提下,还可以根据这些附图获得其他的附图。In order to more clearly explain the technical solutions in the embodiments of the present application, the following will briefly introduce the drawings used in the embodiments or the description of the prior art. Obviously, the drawings in the following description are only for the application In some embodiments, for those of ordinary skill in the art, without paying creative labor, other drawings may be obtained based on these drawings.
图1是本申请实施例提供的一种全景分割方法的实现流程示意图;1 is a schematic diagram of an implementation process of a panoramic segmentation method provided by an embodiment of the present application;
图2是本申请实施例提供的一种全景分割的系统结构示意图;2 is a schematic structural diagram of a panoramic segmentation system provided by an embodiment of the present application;
图3是本申请实施例提供的一种嵌入空间的实例结构示意图;3 is a schematic structural diagram of an example of an embedded space provided by an embodiment of the present application;
图4是本申请实施例提供的一种全景分割装置的示意图;4 is a schematic diagram of a panoramic segmentation device provided by an embodiment of the present application;
图5是本申请实施例提供的全景分割设备的示意图。FIG. 5 is a schematic diagram of a panoramic segmentation device provided by an embodiment of the present application.
本发明的实施方式Embodiments of the invention
以下描述中,为了说明而不是为了限定,提出了诸如特定系统结构、技术之类的具体细节,以便透彻理解本申请实施例。然而,本领域的技术人员应当清楚,在没有这些具体细节的其它实施例中也可以实现本申请。在其它情况中,省略对众所周知的系统、装置、电路以及方法的详细说明,以免不必要的细节妨碍本申请的描述。In the following description, for the purpose of illustration rather than limitation, specific details such as specific system structures and technologies are proposed to thoroughly understand the embodiments of the present application. However, those skilled in the art should understand that the present application can also be implemented in other embodiments without these specific details. In other cases, detailed descriptions of well-known systems, devices, circuits, and methods are omitted to avoid unnecessary details hindering the description of the present application.
为了说明本申请所述的技术方案,下面通过具体实施例来进行说明。In order to explain the technical solutions described in the present application, the following will be described with specific embodiments.
图1为本申请实施例提供的一种全景分割方法的实现流程示意图,详述如 下:FIG. 1 is a schematic diagram of an implementation process of a panoramic segmentation method provided by an embodiment of the present application, and details are as follows:
在步骤S101中,获取待分割的原始图像;In step S101, obtain the original image to be divided;
所述待分割的原始图像,可以为单个的图片,也可以视频中的图像序列。The original image to be divided may be a single picture or an image sequence in a video.
在步骤S102中,对所述原始图像进行语义分割,以及通过嵌入空间的度量距离学习方法,对所述原始图像进行实例分割;In step S102, perform semantic segmentation on the original image, and perform instance segmentation on the original image by using the metric distance learning method of the embedded space;
语义分割用于对图片中所有像素,根据所表达的语义含义进行不同的分组或标定类别标签。Semantic segmentation is used to group or categorize all pixels in the picture according to the semantic meaning expressed.
在本申请中,可以采用基于VGG模型的FCN(英文全称为:fully connected layer,中文全称为全连接层)完全卷积结构作为骨骼框架,以包含有紧密连接对的递归神经网络的条件随机场作为模型的最终层。In this application, FCN based on VGG model (full name in English: fully connected layer, full name in Chinese is fully connected layer) can be used as a skeleton framework to include conditional random fields of recurrent neural networks with closely connected pairs As the final layer of the model.
借此基于VGG模型的全连接层的完全卷积结构框架和设计,可以进一步提高语义分割像素级别的分割质量,并校准检测和输出之间的像素标准。值得注意的是,整个过程是一个可微分推导的场过程。本申请通过采用这一过程的输出作为下一阶段实例检测的标识。Based on this, the fully convolutional structure and design of the fully connected layer based on the VGG model can further improve the segmentation quality of the semantic segmentation pixel level, and calibrate the pixel standard between detection and output. It is worth noting that the whole process is a field process that can be deduced differentially. This application adopts the output of this process as an identifier for instance detection in the next stage.
为了进一步解析语义所分割得到的目标和背景的语义,我们设计了第二个支路用于引入嵌入空间的实例分割。目前,主流方法都是基于候选框检测后进行分割和识别。然后这种基于候选框检测后进行分割和识别的方式并不适用于本申请的全景分割任务,正是本申请进行改进的初衷和所关注的缺陷。因此,我们采用了基于嵌入空间的度量距离学习方法,这样不仅易于嵌入标准的前馈网络,而且能够达成端到端的框架应用。In order to further analyze the semantics of the target and background obtained by semantic segmentation, we designed a second branch for instance segmentation that introduces embedded space. At present, mainstream methods are based on segmentation and recognition after candidate frame detection. Then this method of segmentation and recognition based on candidate frame detection is not suitable for the panoramic segmentation task of this application, which is the original intention of this application for improvement and the defects concerned. Therefore, we adopt the metric distance learning method based on embedding space, so that it is not only easy to embed standard feed-forward network, but also can achieve end-to-end frame application.
在步骤S103中,将实例分割得到的目标和背景作为实例,通过语义分割输出图进行引导,使嵌入空间的实例之间的中心互相排斥,实例范围内的像素吸引至实例中心,对图像进行分割;In step S103, the target and background obtained by instance segmentation are used as an instance, and the output of the semantic segmentation is used to guide the center between the instances in the embedded space. The pixels within the instance range are attracted to the center of the instance to segment the image. ;
通过语义分割生成原始图像中的语义标定和掩码,通过嵌入空间的实例分割处理生成N维像素嵌入,使得实例分割可以很好的匹配语义分割的输出,经过深度度量空间进行聚类融合,输出高度聚集的分割图像。The semantic calibration and mask in the original image are generated through semantic segmentation, and the N-dimensional pixel embedding is generated through the instance segmentation processing in the embedding space, so that the instance segmentation can match the output of the semantic segmentation well. Highly clustered segmented images.
在本申请中,我们可以将语义分割所得到的目标和背景都看作一个实例,通过在语义分割输出图的引导下,需要在嵌入空间的每一个实例间和实例内达成两个目的:In this application, we can regard the target and background obtained by semantic segmentation as an instance. Under the guidance of the output graph of semantic segmentation, we need to achieve two goals between and within each instance of the embedded space:
(1)实例间排斥力:将嵌入空间当中的实例之间的中心互相排斥。(1) Inter-instance repulsive force: The centers between the instances embedded in the space are mutually repulsive.
(2)实例内吸引力:在一个实例所属的范围内将相关的像素嵌入点吸引至实例中心(2) Attraction within an instance: attract the relevant pixel embedding point to the center of the instance within the scope of an instance
相应的,我们为排斥力和吸引力的作用范围各自设定了作用距离阈值,即分别设定实例排斥力半径和吸引力半径。如图2所示,在确定实例的中心点后,通过设定的实例排斥力半径,当嵌入空间的实例的中心点之间的距离小于所述两倍所述排斥力半径时,这两个实例中心互相排斥,使得实例分割更加准确;当嵌入空间的像素与实例的中心点的距离小于所述吸引力半径时,则会受到吸引力对像素点进行聚类。Correspondingly, we set the threshold of the action distance for the repulsive force and attractive force, that is, the repulsive force radius and attractive force radius of the instance. As shown in FIG. 2, after determining the center point of the instance, by setting the radius of the repulsive force of the instance, when the distance between the center points of the instances embedded in the space is less than twice the repulsive radius, the two The instance centers are mutually exclusive, making the instance segmentation more accurate; when the distance between the pixels embedded in the space and the center point of the instance is less than the attractive radius, the pixels will be clustered by the attractive force.
这样,在嵌入空间的实例的中心点,在作用范围内不对其他中心域的像素产生过多的吸引力,在多个中心点之间保证了足够的排斥性,从而不会产生过多过少的负效果。除此以外,这样可以保证嵌入像素尽可能的靠拢中心而不会存在独立点的情况。适当的约束和放松达成了在嵌入空间的聚类效果如图2所示。通过多次迭代,采用聚类算法,得到像素级别的分割,达成了基于语义叠加实例的全景分割方法。In this way, the central point of the instance in the embedded space does not cause too much attraction to pixels in other central domains in the range of action, and sufficient repulsion is ensured between multiple central points, so that there will not be too much or too little. Negative effect. In addition, this can ensure that the embedded pixels are as close as possible to the center without independent points. Appropriate restraint and relaxation achieve the clustering effect in the embedded space as shown in Figure 2. Through multiple iterations, a clustering algorithm is used to obtain pixel-level segmentation, and a panoramic segmentation method based on semantic overlay examples is achieved.
在步骤S104中,采用聚类损失函数进一步区分不同实例,得到全景分割结果。In step S104, a clustering loss function is used to further distinguish different instances to obtain a panoramic segmentation result.
将多任务和模块做到了集成进入一个端到端的框架当中,实现了全景分割这一任务时,需要计算联合损失。通过采用聚类损失函数,可以对实例分割这一支路得到更好的训练。因此,我们的损失函数可以集中在实例分割和嵌入空间部分。所述损失函数可以为:Integrating multi-tasks and modules into an end-to-end framework, when implementing the task of panoramic segmentation, the joint loss needs to be calculated. By using the clustering loss function, the branch of the instance can be better trained. Therefore, our loss function can focus on the instance segmentation and embedding space. The loss function may be:
L=α·L pull+β·L push+γ·L nor+θ·L seg L=α·L pull +β·L push +γ·L nor +θ·L seg
其中:among them:
Figure PCTCN2019124334-appb-000007
Figure PCTCN2019124334-appb-000007
Figure PCTCN2019124334-appb-000008
Figure PCTCN2019124334-appb-000008
Figure PCTCN2019124334-appb-000009
Figure PCTCN2019124334-appb-000009
S表示标准数据当中标定的聚类个数,E s代表了聚类S当中包含的所有元素,x i代表嵌入空间,μ代表S的所有聚类中心,||||代表着深度空间当中的距离,η pull和η push分别表示引力与斥力在嵌入空间当中的作用边缘阈值,N s表示S聚类实例当中包含的像素个数,α、β、γ、θ为调节参数。 S represents the number of calibrated clusters in the standard data, E s represents all the elements contained in the cluster S, x i represents the embedding space, μ represents all the clustering centers of S, |||| represents the depth space Distance, η pull and η push respectively represent the edge threshold of gravity and repulsion in the embedded space, N s represents the number of pixels included in the S clustering instance, and α, β, γ, θ are adjustment parameters.
另外,我们可以设置一个正则化过程,以能够保证迭代计算不会过于超出空间,在予以分割部分采用公共的交叉熵计算损失,可以取得很好的效果。如果斥力半径阈值大于或等于5倍于引力半径阈值时,可以结束迭代过程,可以提高系统的迭代效率。作为本申请优选的一种实施方式中,通过随机梯度下降,α和β为1,γ为0.001,θ为0.7。In addition, we can set up a regularization process to ensure that the iterative calculation will not exceed the space too much, and use the common cross entropy calculation loss in the divided part, which can achieve good results. If the repulsion radius threshold is greater than or equal to 5 times the gravitational radius threshold, the iterative process can be ended, and the iterative efficiency of the system can be improved. As a preferred embodiment of the present application, through random gradient descent, α and β are 1, γ is 0.001, and θ is 0.7.
应理解,上述实施例中各步骤的序号的大小并不意味着执行顺序的先后,各过程的执行顺序应以其功能和内在逻辑确定,而不应对本申请实施例的实施 过程构成任何限定。It should be understood that the size of the sequence numbers of the steps in the above embodiments does not mean the order of execution, and the execution order of each process should be determined by its function and internal logic, and should not constitute any limitation on the implementation process of the embodiments of the present application.
图3为本申请实施例提供的一种全景分割框架示意图,如图3所示,输入待分割的原始图像至端到端的网络框架后,在所述端到端的网络框架中,可以先对所述原始图像进行共享解码,解码后的数据分为两个分支进行处理,下方的语义分割分支用于训练生成图片当中的语义标定和掩码,上方的分支通过嵌入空间的实例分割,生成N(N为自然数)维的像素嵌入,这样,实例分割分支部分可以很好的匹配语义分割的输出。然后两个分支经过融合,在语义分割输出图的引导下,对嵌入空间的每一个实例间进行排斥迭代,以及对实例内进行吸引迭代计算,得到像素级别的分割,达成基于语义叠加实例的全景分割。FIG. 3 is a schematic diagram of a panoramic segmentation framework provided by an embodiment of the present application. As shown in FIG. 3, after inputting the original image to be segmented into an end-to-end network framework, the end-to-end network framework The original image is shared and decoded. The decoded data is divided into two branches for processing. The lower semantic segmentation branch is used to train the semantic calibration and mask in the generated picture. The upper branch is divided by the instance of the embedded space to generate N( N is a natural number)-dimensional pixel embedding, so that the instance segmentation branch can match the output of semantic segmentation well. Then the two branches are merged, and under the guidance of the semantic segmentation output graph, each instance of the embedded space is subjected to repulsive iteration, and the iterative calculation of attraction within the instance is obtained to obtain pixel-level segmentation and achieve a panoramic view based on semantic overlay instances segmentation.
然后通过设定的损失函数,进一步区分不同实例,实现了端到端的全景分类网络框架,输出全景分割图像。Then, through the set loss function, further distinguish different examples, realize the end-to-end panoramic classification network framework, and output panoramic segmented images.
图4为本申请实施例提供的一种全景分割装置的结构示意图,详述如下:FIG. 4 is a schematic structural diagram of a panoramic segmentation device according to an embodiment of the present application. Details are as follows:
所述全景分割装置包括:The panoramic segmentation device includes:
原始图像获取单元401,用于获取待分割的原始图像;The original image obtaining unit 401 is used to obtain an original image to be divided;
分割单元402,用于对所述原始图像进行语义分割,以及通过嵌入空间的度量距离学习方法,对所述原始图像进行实例分割;A segmentation unit 402 is used to perform semantic segmentation on the original image, and perform instance segmentation on the original image by using the metric distance learning method of the embedded space;
融合单元403,用于将实例分割得到的目标和背景作为实例,通过语义分割输出图进行引导,使嵌入空间的实例之间的中心互相排斥,实例范围内的像素吸引至实例中心,对图像进行分割;The fusion unit 403 is used to take the target and background obtained by instance segmentation as examples, and guide through the semantic segmentation output map, so that the centers between the instances embedded in the space are mutually exclusive, and the pixels within the range of the instance are attracted to the center of the instance to perform the image segmentation;
损失训练单元404,用于采用聚类损失函数进一步区分不同实例,得到全景分割结果。The loss training unit 404 is used to further distinguish different instances by using a clustering loss function to obtain a panoramic segmentation result.
优选的,所述融合单元包括:Preferably, the fusion unit includes:
实例确定子单元,用于将实例分割得到的目标和背景作为实例,确定实例 的中心点;The instance determination subunit is used to determine the center point of the instance by using the target and background obtained from the instance segmentation as the instance;
聚类单元,用于根据预先设定的实例排斥力半径,使嵌入空间的实例中心互相排斥,以及根据预先设定的吸引力半径,对实例的像素点进行聚类。The clustering unit is used to repel each instance center of the embedded space according to a preset radius of repulsive force of the instance, and to cluster pixels of the instance according to a preset radius of attractive force.
优选的,所述聚类损失函数为:Preferably, the clustering loss function is:
L=α·L pull+β·L push+γ·L nor+θ·L seg L=α·L pull +β·L push +γ·L nor +θ·L seg
其中:among them:
Figure PCTCN2019124334-appb-000010
Figure PCTCN2019124334-appb-000010
Figure PCTCN2019124334-appb-000011
Figure PCTCN2019124334-appb-000011
Figure PCTCN2019124334-appb-000012
Figure PCTCN2019124334-appb-000012
S表示标准数据当中标定的聚类个数,E s代表了聚类S当中包含的所有元素,x i代表嵌入空间,μ代表S的所有聚类中心,||||代表着深度空间当中的距离,η pull和η push分别表示引力与斥力在嵌入空间当中的作用边缘阈值,N s表示S聚类实例当中包含的像素个数,α、β、γ、θ为调节参数。 S represents the number of calibrated clusters in the standard data, E s represents all the elements contained in the cluster S, x i represents the embedding space, μ represents all the clustering centers of S, |||| represents the depth space Distance, η pull and η push respectively represent the edge threshold of gravity and repulsion in the embedded space, N s represents the number of pixels included in the S clustering instance, and α, β, γ, θ are adjustment parameters.
图4所述全景分割装置,与图1所述全景分割方法对应。The panoramic segmentation device shown in FIG. 4 corresponds to the panoramic segmentation method shown in FIG. 1.
图5是本申请一实施例提供的全景分割设备的示意图。如图5所示,该实施例的全景分割设备5包括:处理器50、存储器51以及存储在所述存储器51中并可在所述处理器50上运行的计算机程序52,例如全景分割程序。所述处理器50执行所述计算机程序52时实现上述各个全景分割方法实施例中的步骤。或者,所述处理器50执行所述计算机程序52时实现上述各装置实施例中各模块/单元的功能。FIG. 5 is a schematic diagram of a panoramic segmentation device provided by an embodiment of the present application. As shown in FIG. 5, the panoramic segmentation device 5 of this embodiment includes: a processor 50, a memory 51, and a computer program 52 stored in the memory 51 and executable on the processor 50, for example, a panoramic segmentation program. When the processor 50 executes the computer program 52, the steps in the foregoing embodiments of the panoramic segmentation method are implemented. Alternatively, when the processor 50 executes the computer program 52, the functions of the modules/units in the foregoing device embodiments are realized.
示例性的,所述计算机程序52可以被分割成一个或多个模块/单元,所述一个或者多个模块/单元被存储在所述存储器51中,并由所述处理器50执行,以完成本申请。所述一个或多个模块/单元可以是能够完成特定功能的一系列计算机程序指令段,该指令段用于描述所述计算机程序52在所述全景分割设备5中的执行过程。例如,所述计算机程序52可以被分割成:Exemplarily, the computer program 52 may be divided into one or more modules/units, and the one or more modules/units are stored in the memory 51 and executed by the processor 50 to complete This application. The one or more modules/units may be a series of computer program instruction segments capable of performing specific functions, and the instruction segments are used to describe the execution process of the computer program 52 in the panoramic segmentation device 5. For example, the computer program 52 may be divided into:
原始图像获取单元,用于获取待分割的原始图像;The original image acquisition unit is used to acquire the original image to be segmented;
分割单元,用于对所述原始图像进行语义分割,以及通过嵌入空间的度量距离学习方法,对所述原始图像进行实例分割;A segmentation unit, used for semantic segmentation of the original image, and for instance segmentation of the original image through a distance learning method of embedded space;
融合单元,用于将实例分割得到的目标和背景作为实例,通过语义分割输出图进行引导,使嵌入空间的实例之间的中心互相排斥,实例范围内的像素吸引至实例中心,对图像进行分割;The fusion unit is used to take the target and background obtained by instance segmentation as an instance, and guide through the semantic segmentation output map, so that the centers between the instances embedded in the space are mutually exclusive, and the pixels within the instance range are attracted to the center of the instance to segment the image ;
损失训练单元,用于采用聚类损失函数进一步区分不同实例,得到全景分割结果。The loss training unit is used to further distinguish different instances by using a clustering loss function to obtain a panoramic segmentation result.
所述全景分割设备5可以是桌上型计算机、笔记本、掌上电脑及云端服务器等计算设备。所述全景分割设备可包括,但不仅限于,处理器50、存储器51。本领域技术人员可以理解,图5仅仅是全景分割设备5的示例,并不构成对全景分割设备5的限定,可以包括比图示更多或更少的部件,或者组合某些部件,或者不同的部件,例如所述全景分割设备还可以包括输入输出设备、网络接入设备、总线等。The panoramic segmentation device 5 may be a computing device such as a desktop computer, a notebook, a palmtop computer and a cloud server. The panoramic segmentation device may include, but is not limited to, the processor 50 and the memory 51. Those skilled in the art may understand that FIG. 5 is only an example of the panoramic segmentation device 5 and does not constitute a limitation on the panoramic segmentation device 5, and may include more or fewer components than the illustration, or a combination of certain components, or different Components, for example, the panoramic segmentation device may also include an input and output device, a network access device, a bus, and the like.
所称处理器50可以是中央处理单元(Central Processing Unit,CPU),还可以是其他通用处理器、数字信号处理器(Digital Signal Processor,DSP)、专用集成电路(Application Specific Integrated Circuit,ASIC)、现成可编程门阵列(Field-Programmable Gate Array,FPGA)或者其他可编 程逻辑器件、分立门或者晶体管逻辑器件、分立硬件组件等。通用处理器可以是微处理器或者该处理器也可以是任何常规的处理器等。The so-called processor 50 may be a central processing unit (Central Processing Unit, CPU), or other general-purpose processors, digital signal processors (Digital Signal Processor, DSP), application specific integrated circuits (Application Specific Integrated Circuit, ASIC), Ready-made programmable gate array (Field-Programmable Gate Array, FPGA) or other programmable logic devices, discrete gate or transistor logic devices, discrete hardware components, etc. The general-purpose processor may be a microprocessor or the processor may be any conventional processor or the like.
所述存储器51可以是所述全景分割设备5的内部存储单元,例如全景分割设备5的硬盘或内存。所述存储器51也可以是所述全景分割设备5的外部存储设备,例如所述全景分割设备5上配备的插接式硬盘,智能存储卡(Smart Media Card,SMC),安全数字(Secure Digital,SD)卡,闪存卡(Flash Card)等。进一步地,所述存储器51还可以既包括所述全景分割设备5的内部存储单元也包括外部存储设备。所述存储器51用于存储所述计算机程序以及所述全景分割设备所需的其他程序和数据。所述存储器51还可以用于暂时地存储已经输出或者将要输出的数据。The memory 51 may be an internal storage unit of the panoramic segmentation device 5, such as a hard disk or a memory of the panoramic segmentation device 5. The memory 51 may also be an external storage device of the panoramic segmentation device 5, for example, a plug-in hard disk equipped on the panoramic segmentation device 5, a smart memory card (Smart, Media, Card, SMC), and a secure digital (Secure Digital, SD) card, flash memory card (Flash Card), etc. Further, the memory 51 may also include both an internal storage unit of the panoramic segmentation device 5 and an external storage device. The memory 51 is used to store the computer program and other programs and data required by the panoramic segmentation device. The memory 51 can also be used to temporarily store data that has been or will be output.
所属领域的技术人员可以清楚地了解到,为了描述的方便和简洁,仅以上述各功能单元、模块的划分进行举例说明,实际应用中,可以根据需要而将上述功能分配由不同的功能单元、模块完成,即将所述装置的内部结构划分成不同的功能单元或模块,以完成以上描述的全部或者部分功能。实施例中的各功能单元、模块可以集成在一个处理单元中,也可以是各个单元单独物理存在,也可以两个或两个以上单元集成在一个单元中,上述集成的单元既可以采用硬件的形式实现,也可以采用软件功能单元的形式实现。另外,各功能单元、模块的具体名称也只是为了便于相互区分,并不用于限制本申请的保护范围。上述系统中单元、模块的具体工作过程,可以参考前述方法实施例中的对应过程,在此不再赘述。Those skilled in the art can clearly understand that, for convenience and conciseness of description, only the above-mentioned division of each functional unit and module is used as an example for illustration. In practical applications, the above-mentioned functions may be allocated by different functional units, Module completion means that the internal structure of the device is divided into different functional units or modules to complete all or part of the functions described above. The functional units and modules in the embodiment may be integrated into one processing unit, or each unit may exist alone physically, or two or more units may be integrated into one unit. The above integrated unit may use hardware It can also be implemented in the form of software functional units. In addition, the specific names of each functional unit and module are only for the purpose of distinguishing each other, and are not used to limit the protection scope of the present application. For the specific working processes of the units and modules in the above system, reference may be made to the corresponding processes in the foregoing method embodiments, which will not be repeated here.
在上述实施例中,对各个实施例的描述都各有侧重,某个实施例中没有详述或记载的部分,可以参见其它实施例的相关描述。In the above embodiments, the description of each embodiment has its own emphasis. For a part that is not detailed or recorded in an embodiment, you can refer to the related descriptions of other embodiments.
本领域普通技术人员可以意识到,结合本文中所公开的实施例描述的各示 例的单元及算法步骤,能够以电子硬件、或者计算机软件和电子硬件的结合来实现。这些功能究竟以硬件还是软件方式来执行,取决于技术方案的特定应用和设计约束条件。专业技术人员可以对每个特定的应用来使用不同方法来实现所描述的功能,但是这种实现不应认为超出本申请的范围。Those of ordinary skill in the art may realize that the units and algorithm steps of the examples described in conjunction with the embodiments disclosed herein can be implemented by electronic hardware, or a combination of computer software and electronic hardware. Whether these functions are executed in hardware or software depends on the specific application of the technical solution and design constraints. Professional technicians can use different methods to implement the described functions for each specific application, but such implementation should not be considered beyond the scope of this application.
在本申请所提供的实施例中,应该理解到,所揭露的装置/终端设备和方法,可以通过其它的方式实现。例如,以上所描述的装置/终端设备实施例仅仅是示意性的,例如,所述模块或单元的划分,仅仅为一种逻辑功能划分,实际实现时可以有另外的划分方式,例如多个单元或组件可以结合或者可以集成到另一个系统,或一些特征可以忽略,或不执行。另一点,所显示或讨论的相互之间的耦合或直接耦合或通讯连接可以是通过一些接口,装置或单元的间接耦合或通讯连接,可以是电性,机械或其它的形式。In the embodiments provided in this application, it should be understood that the disclosed device/terminal device and method may be implemented in other ways. For example, the device/terminal device embodiments described above are only schematic. For example, the division of the module or unit is only a logical function division, and in actual implementation, there may be another division manner, such as multiple units Or components can be combined or integrated into another system, or some features can be ignored or not implemented. In addition, the displayed or discussed mutual coupling or direct coupling or communication connection may be indirect coupling or communication connection through some interfaces, devices or units, and may be in electrical, mechanical or other forms.
所述作为分离部件说明的单元可以是或者也可以不是物理上分开的,作为单元显示的部件可以是或者也可以不是物理单元,即可以位于一个地方,或者也可以分布到多个网络单元上。可以根据实际的需要选择其中的部分或者全部单元来实现本实施例方案的目的。The units described as separate components may or may not be physically separated, and the components displayed as units may or may not be physical units, that is, they may be located in one place, or may be distributed on multiple network units. Some or all of the units may be selected according to actual needs to achieve the purpose of the solution of this embodiment.
另外,在本申请各个实施例中的各功能单元可以集成在一个处理单元中,也可以是各个单元单独物理存在,也可以两个或两个以上单元集成在一个单元中。上述集成的单元既可以采用硬件的形式实现,也可以采用软件功能单元的形式实现。In addition, each functional unit in each embodiment of the present application may be integrated into one processing unit, or each unit may exist alone physically, or two or more units may be integrated into one unit. The above integrated unit can be implemented in the form of hardware or software function unit.
所述集成的模块/单元如果以软件功能单元的形式实现并作为独立的产品销售或使用时,可以存储在一个计算机可读取存储介质中。基于这样的理解,本申请实现上述实施例方法中的全部或部分流程,也可以通过计算机程序来指令相关的硬件来完成,所述的计算机程序可存储于一计算机可读存储介质中, 该计算机程序在被处理器执行时,可实现上述各个方法实施例的步骤。。其中,所述计算机程序包括计算机程序代码,所述计算机程序代码可以为源代码形式、对象代码形式、可执行文件或某些中间形式等。所述计算机可读介质可以包括:能够携带所述计算机程序代码的任何实体或装置、记录介质、U盘、移动硬盘、磁碟、光盘、计算机存储器、只读存储器(ROM,Read-Only Memory)、随机存取存储器(RAM,Random Access Memory)、电载波信号、电信信号以及软件分发介质等。需要说明的是,所述计算机可读介质包含的内容可以根据司法管辖区内立法和专利实践的要求进行适当的增减,例如在某些司法管辖区,根据立法和专利实践,计算机可读介质不包括是电载波信号和电信信号。以上所述实施例仅用以说明本申请的技术方案,而非对其限制;尽管参照前述实施例对本申请进行了详细的说明,本领域的普通技术人员应当理解:其依然可以对前述各实施例所记载的技术方案进行修改,或者对其中部分技术特征进行等同替换;而这些修改或者替换,并不使相应技术方案的本质脱离本申请各实施例技术方案的精神和范围,均应包含在本申请的保护范围之内。If the integrated module/unit is implemented in the form of a software functional unit and sold or used as an independent product, it may be stored in a computer-readable storage medium. Based on this understanding, the present application can implement all or part of the processes in the methods of the above embodiments, or it can be completed by a computer program instructing relevant hardware. The computer program can be stored in a computer-readable storage medium. When the program is executed by the processor, the steps of the foregoing method embodiments may be implemented. . Wherein, the computer program includes computer program code, and the computer program code may be in a source code form, an object code form, an executable file, or some intermediate form, etc. The computer-readable medium may include: any entity or device capable of carrying the computer program code, a recording medium, a USB flash drive, a mobile hard disk, a magnetic disk, an optical disc, a computer memory, and a read-only memory (ROM, Read-Only Memory) , Random Access Memory (RAM, Random Access Memory), electrical carrier signals, telecommunications signals and software distribution media, etc. It should be noted that the content contained in the computer-readable medium can be appropriately increased or decreased according to the requirements of legislation and patent practice in jurisdictions. For example, in some jurisdictions, according to legislation and patent practice, computer-readable media Excluded are electrical carrier signals and telecommunications signals. The above-mentioned embodiments are only used to illustrate the technical solutions of the present application, not to limit them; although the present application has been described in detail with reference to the foregoing embodiments, persons of ordinary skill in the art should understand that they can still implement the foregoing The technical solutions described in the examples are modified, or some of the technical features are equivalently replaced; and these modifications or replacements do not deviate from the spirit and scope of the technical solutions of the embodiments of the present application. Within the scope of protection of this application.

Claims (10)

  1. 一种全景分割方法,其特征在于,所述全景分割方法包括:A panoramic segmentation method, characterized in that the panoramic segmentation method includes:
    获取待分割的原始图像;Obtain the original image to be segmented;
    对所述原始图像进行语义分割,以及通过嵌入空间的度量距离学习方法,对所述原始图像进行实例分割;Performing semantic segmentation on the original image, and performing instance segmentation on the original image through a distance learning method of embedded space;
    将实例分割得到的目标和背景作为实例,通过语义分割输出图进行引导,使嵌入空间的实例之间的中心互相排斥,实例范围内的像素吸引至实例中心,对图像进行分割;Take the target and background obtained by instance segmentation as examples, and guide through the output graph of semantic segmentation, so that the centers between the instances embedded in the space are mutually exclusive, and the pixels within the instance range are attracted to the center of the instance to segment the image;
    采用聚类损失函数进一步区分不同实例,得到全景分割结果。The clustering loss function is used to further distinguish different instances, and the panoramic segmentation result is obtained.
  2. 根据权利要求1所述的全景分割方法,其特征在于,所述对所述原始图像进行语义分割的步骤包括:The panoramic segmentation method according to claim 1, wherein the step of performing semantic segmentation on the original image includes:
    采用基于VGG模型的全连接层的完全卷积结构作为骨骼框架,以包含有紧密连接对的递归神经网络的条件随机作为模型的最终层,对所述原始图像进行语义分割。The fully convolutional structure of the fully connected layer based on the VGG model is used as the skeleton framework, and the conditional random recursive neural network containing closely connected pairs is randomly selected as the final layer of the model to semantically segment the original image.
  3. 根据权利要求1或2所述的全景分割方法,其特征在于,所述将实例分割得到的目标和背景作为实例,通过语义分割输出图进行引导,使嵌入空间的实例之间的中心互相排斥,实例范围内的像素吸引至实例中心,对图像进行分割的步骤包括:The panoramic segmentation method according to claim 1 or 2, wherein the target and background obtained by segmenting an instance are used as an instance, guided by a semantic segmentation output image, so that the centers between the instances embedded in the space are mutually exclusive, The pixels within the scope of the instance are attracted to the center of the instance, and the steps of segmenting the image include:
    将实例分割得到的目标和背景作为实例,确定实例的中心点;Use the target and background obtained by segmenting the instance as an example to determine the center point of the instance;
    根据预先设定的实例排斥力半径,使嵌入空间的实例中心互相排斥,以及根据预先设定的吸引力半径,对实例的像素点进行聚类。According to the preset radius of repulsive force of the instance, the centers of the instances in the embedded space are mutually repelled, and the pixels of the instance are clustered according to the preset radius of attractive force.
  4. 根据权利要求1所述的全景分割方法,其特征在于,所述聚类损失函数为:The panoramic segmentation method according to claim 1, wherein the clustering loss function is:
    L=α·L pull+β·L push+γ·L nor+θ·L seg L=α·L pull +β·L push +γ·L nor +θ·L seg
    其中:among them:
    Figure PCTCN2019124334-appb-100001
    Figure PCTCN2019124334-appb-100001
    Figure PCTCN2019124334-appb-100002
    Figure PCTCN2019124334-appb-100002
    Figure PCTCN2019124334-appb-100003
    Figure PCTCN2019124334-appb-100003
    S表示标准数据当中标定的聚类个数,E s代表了聚类S当中包含的所有元素,x i代表嵌入空间,μ代表S的所有聚类中心,||||代表着深度空间当中的距离,η pull和η push分别表示引力与斥力在嵌入空间当中的作用边缘阈值,N s表示S聚类实例当中包含的像素个数,α、β、γ、θ为调节参数。 S represents the number of calibrated clusters in the standard data, E s represents all the elements contained in the cluster S, x i represents the embedding space, μ represents all the clustering centers of S, |||| represents the depth space Distance, η pull and η push respectively represent the edge threshold of gravity and repulsion in the embedded space, N s represents the number of pixels included in the S clustering instance, and α, β, γ, θ are adjustment parameters.
  5. 根据权利要求1所述的全景分割方法,基特征在于,所述将实例分割得到的目标和背景作为实例,通过语义分割输出图进行引导,使嵌入空间的实例之间的中心互相排斥,实例范围内的像素吸引至实例中心,对图像进行分割的步骤包括:The panoramic segmentation method according to claim 1, characterized in that the target and background obtained by segmenting an instance are used as an instance, and guided by a semantic segmentation output image, so that the centers between the instances embedded in the space are mutually exclusive, and the instance range The pixels inside are attracted to the center of the instance, and the steps to segment the image include:
    根据语义分割生成原始图片中的语义标定和掩码,通过嵌入空间的实例分割生成多维像素嵌入的实例,通过深度度量空间进行聚类融合,输出聚集的分割图像。The semantic calibration and mask in the original picture are generated according to the semantic segmentation, the instance of the multi-dimensional pixel embedding is generated by the instance segmentation of the embedding space, and the cluster fusion is performed by the depth metric space, and the aggregated segmented image is output.
  6. 一种全景分割装置,其特征在于,所述全景分割装置包括:A panoramic segmentation device, characterized in that the panoramic segmentation device includes:
    原始图像获取单元,用于获取待分割的原始图像;The original image acquisition unit is used to acquire the original image to be segmented;
    分割单元,用于对所述原始图像进行语义分割,以及通过嵌入空间的度量距离学习方法,对所述原始图像进行实例分割;A segmentation unit, used for semantic segmentation of the original image, and for instance segmentation of the original image through a distance learning method of embedded space;
    融合单元,用于将实例分割得到的目标和背景作为实例,通过语义分割输出图进行引导,使嵌入空间的实例之间的中心互相排斥,实例范围内的像素吸引至实例中心,对图像进行分割;The fusion unit is used to take the target and background obtained by instance segmentation as an instance, and guide through the semantic segmentation output map, so that the centers between the instances embedded in the space are mutually exclusive, and the pixels within the instance range are attracted to the center of the instance to segment the image ;
    损失训练单元,用于采用聚类损失函数进一步区分不同实例,得到全景分割结果。The loss training unit is used to further distinguish different instances by using a clustering loss function to obtain a panoramic segmentation result.
  7. 根据权利要求6所述的全景分割装置,其特征在于,所述融合单元包括:The panoramic segmentation device according to claim 6, wherein the fusion unit comprises:
    实例确定子单元,用于将实例分割得到的目标和背景作为实例,确定实例的中心点;The instance determination subunit is used to determine the center point of the instance by using the target and background obtained by the instance segmentation as the instance;
    聚类单元,用于根据预先设定的实例排斥力半径,使嵌入空间的实例中心互相排斥,以及根据预先设定的吸引力半径,对实例的像素点进行聚类。The clustering unit is used to repel each instance center of the embedded space according to a preset radius of repulsive force of the instance, and to cluster pixels of the instance according to a preset radius of attractive force.
  8. 根据权利要求6所述的全景分割装置,其特征在于,所述聚类损失函数为:The panoramic segmentation device according to claim 6, wherein the clustering loss function is:
    L=α·L pull+β·L push+γ·L nor+θ·L seg L=α·L pull +β·L push +γ·L nor +θ·L seg
    其中:among them:
    Figure PCTCN2019124334-appb-100004
    Figure PCTCN2019124334-appb-100004
    Figure PCTCN2019124334-appb-100005
    Figure PCTCN2019124334-appb-100005
    Figure PCTCN2019124334-appb-100006
    Figure PCTCN2019124334-appb-100006
    S表示标准数据当中标定的聚类个数,E s代表了聚类S当中包含的所有元素,x i代表嵌入空间,μ代表S的所有聚类中心,||||代表着深度空间当中的距 离,η pull和η push分别表示引力与斥力在嵌入空间当中的作用边缘阈值,N s表示S聚类实例当中包含的像素个数,α、β、γ、θ为调节参数。 S represents the number of calibrated clusters in the standard data, E s represents all the elements contained in the cluster S, x i represents the embedding space, μ represents all the clustering centers of S, |||| represents the depth space Distance, η pull and η push respectively represent the edge threshold of gravity and repulsion in the embedded space, N s represents the number of pixels included in the S clustering instance, and α, β, γ, θ are adjustment parameters.
  9. 一种全景分割设备,包括存储器、处理器以及存储在所述存储器中并可在所述处理器上运行的计算机程序,其特征在于,所述处理器执行所述计算机程序时实现如权利要求1至5任一项所述全景分割方法的步骤。A panoramic segmentation device, including a memory, a processor, and a computer program stored in the memory and runable on the processor, characterized in that, when the processor executes the computer program, it is implemented as claimed in claim 1. To any one of the steps of the panoramic segmentation method described in any one of 5.
  10. 一种计算机可读存储介质,所述计算机可读存储介质存储有计算机程序,其特征在于,所述计算机程序被处理器执行时实现如权利要求1至5任一项所述全景分割方法的步骤。A computer-readable storage medium storing a computer program, characterized in that, when the computer program is executed by a processor, the steps of the panoramic segmentation method according to any one of claims 1 to 5 are implemented .
PCT/CN2019/124334 2018-12-17 2019-12-10 Panoramic segmentation method, apparatus and device WO2020125495A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201811542525.2 2018-12-17
CN201811542525.2A CN109801307A (en) 2018-12-17 2018-12-17 A kind of panorama dividing method, device and equipment

Publications (1)

Publication Number Publication Date
WO2020125495A1 true WO2020125495A1 (en) 2020-06-25

Family

ID=66556927

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2019/124334 WO2020125495A1 (en) 2018-12-17 2019-12-10 Panoramic segmentation method, apparatus and device

Country Status (2)

Country Link
CN (1) CN109801307A (en)
WO (1) WO2020125495A1 (en)

Cited By (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111754505A (en) * 2020-06-30 2020-10-09 创新奇智(成都)科技有限公司 Auxiliary material detection method and device, electronic equipment and storage medium
CN112614134A (en) * 2020-12-17 2021-04-06 北京迈格威科技有限公司 Image segmentation method and device, electronic equipment and storage medium
CN113139549A (en) * 2021-03-25 2021-07-20 北京化工大学 Parameter self-adaptive panorama segmentation method based on multitask learning
CN113160257A (en) * 2021-04-23 2021-07-23 深圳市优必选科技股份有限公司 Image data labeling method and device, electronic equipment and storage medium
CN113379762A (en) * 2021-05-28 2021-09-10 上海商汤智能科技有限公司 Image segmentation method and device, electronic equipment and storage medium
CN113569620A (en) * 2021-05-24 2021-10-29 惠州市德赛西威智能交通技术研究院有限公司 Pavement marker instantiation identification method based on monocular vision
CN113916245A (en) * 2021-10-09 2022-01-11 上海大学 Semantic map construction method based on instance segmentation and VSLAM
CN114022865A (en) * 2021-10-29 2022-02-08 北京百度网讯科技有限公司 Image processing method, apparatus, device and medium based on lane line recognition model
CN114495174A (en) * 2022-01-29 2022-05-13 建信金融科技有限责任公司 Pedestrian re-identification method and device
CN114758128A (en) * 2022-04-11 2022-07-15 西安交通大学 Scene panorama segmentation method and system based on controlled pixel embedding representation explicit interaction
CN114838729A (en) * 2022-04-27 2022-08-02 中国建设银行股份有限公司 Path planning method, device and equipment
CN115361536A (en) * 2022-07-26 2022-11-18 鹏城实验室 Panoramic image compression method and device, intelligent equipment and storage medium
CN115393892A (en) * 2022-07-20 2022-11-25 东北电力大学 Crowd scene pedestrian detection method based on improved double-candidate-frame cross replacement strategy and loss function
CN116079749A (en) * 2023-04-10 2023-05-09 南京师范大学 Robot vision obstacle avoidance method based on cluster separation conditional random field and robot

Families Citing this family (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109801307A (en) * 2018-12-17 2019-05-24 中国科学院深圳先进技术研究院 A kind of panorama dividing method, device and equipment
CN110276765B (en) * 2019-06-21 2021-04-23 北京交通大学 Image panorama segmentation method based on multitask learning deep neural network
CN110232370B (en) * 2019-06-21 2022-04-26 华北电力大学(保定) Power transmission line aerial image hardware detection method for improving SSD model
CN110378278B (en) * 2019-07-16 2021-11-02 北京地平线机器人技术研发有限公司 Neural network training method, object searching method, device and electronic equipment
CN112802028A (en) * 2019-11-13 2021-05-14 北京深睿博联科技有限责任公司 Image processing method and device for mediastinal organ segmentation
CN111681244B (en) * 2020-05-29 2022-06-21 山东大学 Blade image segmentation method, system, equipment and storage medium
CN112053358B (en) * 2020-09-28 2024-09-13 腾讯科技(深圳)有限公司 Method, device, equipment and storage medium for determining instance category of pixel in image
CN112308867B (en) * 2020-11-10 2022-07-22 上海商汤智能科技有限公司 Tooth image processing method and device, electronic equipment and storage medium
CN112489060B (en) * 2020-12-07 2022-05-10 北京医准智能科技有限公司 System and method for pneumonia focus segmentation
CN112668579A (en) * 2020-12-24 2021-04-16 西安电子科技大学 Weak supervision semantic segmentation method based on self-adaptive affinity and class distribution
CN113052858B (en) * 2021-03-23 2023-02-14 电子科技大学 Panorama segmentation method based on semantic stream
CN113096140B (en) * 2021-04-15 2022-11-22 北京市商汤科技开发有限公司 Instance partitioning method and device, electronic device and storage medium
CN116778170B (en) * 2023-08-25 2023-11-07 安徽蔚来智驾科技有限公司 Point cloud panorama segmentation method, control device, readable storage medium and vehicle

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20120075409A1 (en) * 2010-09-27 2012-03-29 Hon Hai Precision Industry Co., Ltd. Image segmentation system and method thereof
JP2013218432A (en) * 2012-04-05 2013-10-24 Dainippon Printing Co Ltd Image processing device, image processing method, program for image processing, and recording medium
CN106780536A (en) * 2017-01-13 2017-05-31 深圳市唯特视科技有限公司 A kind of shape based on object mask network perceives example dividing method
CN107564032A (en) * 2017-09-01 2018-01-09 深圳市唯特视科技有限公司 A kind of video tracking object segmentation methods based on outward appearance network
CN108170751A (en) * 2017-12-21 2018-06-15 百度在线网络技术(北京)有限公司 For handling the method and apparatus of image
CN108230329A (en) * 2017-12-18 2018-06-29 孙颖 Semantic segmentation method based on multiple dimensioned convolutional neural networks
CN109801307A (en) * 2018-12-17 2019-05-24 中国科学院深圳先进技术研究院 A kind of panorama dividing method, device and equipment

Family Cites Families (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10424064B2 (en) * 2016-10-18 2019-09-24 Adobe Inc. Instance-level semantic segmentation system
CN107451620A (en) * 2017-08-11 2017-12-08 深圳市唯特视科技有限公司 A kind of scene understanding method based on multi-task learning
CN107704862A (en) * 2017-11-06 2018-02-16 深圳市唯特视科技有限公司 A kind of video picture segmentation method based on semantic instance partitioning algorithm
CN108053420B (en) * 2018-01-05 2021-11-02 昆明理工大学 Partition method based on finite space-time resolution class-independent attribute dynamic scene
CN108062756B (en) * 2018-01-29 2020-04-14 重庆理工大学 Image semantic segmentation method based on deep full convolution network and conditional random field
CN108596184B (en) * 2018-04-25 2021-01-12 清华大学深圳研究生院 Training method of image semantic segmentation model, readable storage medium and electronic device
CN108986136B (en) * 2018-07-23 2020-07-24 南昌航空大学 Binocular scene flow determination method and system based on semantic segmentation

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20120075409A1 (en) * 2010-09-27 2012-03-29 Hon Hai Precision Industry Co., Ltd. Image segmentation system and method thereof
JP2013218432A (en) * 2012-04-05 2013-10-24 Dainippon Printing Co Ltd Image processing device, image processing method, program for image processing, and recording medium
CN106780536A (en) * 2017-01-13 2017-05-31 深圳市唯特视科技有限公司 A kind of shape based on object mask network perceives example dividing method
CN107564032A (en) * 2017-09-01 2018-01-09 深圳市唯特视科技有限公司 A kind of video tracking object segmentation methods based on outward appearance network
CN108230329A (en) * 2017-12-18 2018-06-29 孙颖 Semantic segmentation method based on multiple dimensioned convolutional neural networks
CN108170751A (en) * 2017-12-21 2018-06-15 百度在线网络技术(北京)有限公司 For handling the method and apparatus of image
CN109801307A (en) * 2018-12-17 2019-05-24 中国科学院深圳先进技术研究院 A kind of panorama dividing method, device and equipment

Cited By (21)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111754505A (en) * 2020-06-30 2020-10-09 创新奇智(成都)科技有限公司 Auxiliary material detection method and device, electronic equipment and storage medium
CN111754505B (en) * 2020-06-30 2024-03-15 创新奇智(成都)科技有限公司 Auxiliary material detection method and device, electronic equipment and storage medium
CN112614134A (en) * 2020-12-17 2021-04-06 北京迈格威科技有限公司 Image segmentation method and device, electronic equipment and storage medium
CN113139549A (en) * 2021-03-25 2021-07-20 北京化工大学 Parameter self-adaptive panorama segmentation method based on multitask learning
CN113139549B (en) * 2021-03-25 2024-03-15 北京化工大学 Parameter self-adaptive panoramic segmentation method based on multitask learning
CN113160257A (en) * 2021-04-23 2021-07-23 深圳市优必选科技股份有限公司 Image data labeling method and device, electronic equipment and storage medium
CN113160257B (en) * 2021-04-23 2024-01-16 深圳市优必选科技股份有限公司 Image data labeling method, device, electronic equipment and storage medium
CN113569620A (en) * 2021-05-24 2021-10-29 惠州市德赛西威智能交通技术研究院有限公司 Pavement marker instantiation identification method based on monocular vision
CN113569620B (en) * 2021-05-24 2024-09-13 惠州市德赛西威智能交通技术研究院有限公司 Pavement marking instantiation identification method based on monocular vision
CN113379762A (en) * 2021-05-28 2021-09-10 上海商汤智能科技有限公司 Image segmentation method and device, electronic equipment and storage medium
CN113916245A (en) * 2021-10-09 2022-01-11 上海大学 Semantic map construction method based on instance segmentation and VSLAM
CN114022865A (en) * 2021-10-29 2022-02-08 北京百度网讯科技有限公司 Image processing method, apparatus, device and medium based on lane line recognition model
CN114495174A (en) * 2022-01-29 2022-05-13 建信金融科技有限责任公司 Pedestrian re-identification method and device
CN114758128A (en) * 2022-04-11 2022-07-15 西安交通大学 Scene panorama segmentation method and system based on controlled pixel embedding representation explicit interaction
CN114758128B (en) * 2022-04-11 2024-04-16 西安交通大学 Scene panorama segmentation method and system based on controlled pixel embedding characterization explicit interaction
CN114838729A (en) * 2022-04-27 2022-08-02 中国建设银行股份有限公司 Path planning method, device and equipment
CN115393892B (en) * 2022-07-20 2023-08-04 东北电力大学 Congestion scene pedestrian detection method based on improved double-candidate-frame cross replacement strategy and loss function
CN115393892A (en) * 2022-07-20 2022-11-25 东北电力大学 Crowd scene pedestrian detection method based on improved double-candidate-frame cross replacement strategy and loss function
CN115361536B (en) * 2022-07-26 2023-06-06 鹏城实验室 Panoramic image compression method and device, intelligent device and storage medium
CN115361536A (en) * 2022-07-26 2022-11-18 鹏城实验室 Panoramic image compression method and device, intelligent equipment and storage medium
CN116079749A (en) * 2023-04-10 2023-05-09 南京师范大学 Robot vision obstacle avoidance method based on cluster separation conditional random field and robot

Also Published As

Publication number Publication date
CN109801307A (en) 2019-05-24

Similar Documents

Publication Publication Date Title
WO2020125495A1 (en) Panoramic segmentation method, apparatus and device
WO2022001623A1 (en) Image processing method and apparatus based on artificial intelligence, and device and storage medium
CN108229504B (en) Image analysis method and device
WO2020253127A1 (en) Facial feature extraction model training method and apparatus, facial feature extraction method and apparatus, device, and storage medium
CN111369427B (en) Image processing method, image processing device, readable medium and electronic equipment
TWI747120B (en) Method, device and electronic equipment for depth model training and storage medium thereof
CN111476284A (en) Image recognition model training method, image recognition model training device, image recognition method, image recognition device and electronic equipment
CN110570352B (en) Image labeling method, device and system and cell labeling method
CN110379020B (en) Laser point cloud coloring method and device based on generation countermeasure network
CN111414879B (en) Face shielding degree identification method and device, electronic equipment and readable storage medium
WO2021129466A1 (en) Watermark detection method, device, terminal and storage medium
CN113689372B (en) Image processing method, apparatus, storage medium, and program product
CN114746898A (en) Method and system for generating trisection images of image matting
WO2022227218A1 (en) Drug name recognition method and apparatus, and computer device and storage medium
WO2023082453A1 (en) Image processing method and device
CN110349161B (en) Image segmentation method, image segmentation device, electronic equipment and storage medium
WO2022166258A1 (en) Behavior recognition method and apparatus, terminal device, and computer-readable storage medium
CN110211195B (en) Method, device, electronic equipment and computer-readable storage medium for generating image set
CN113706562B (en) Image segmentation method, device and system and cell segmentation method
CN112598673A (en) Panorama segmentation method, device, electronic equipment and computer readable medium
CN114549557A (en) Portrait segmentation network training method, device, equipment and medium
WO2024099026A1 (en) Image processing method and apparatus, device, storage medium and program product
CN114333062A (en) Pedestrian re-recognition model training method based on heterogeneous dual networks and feature consistency
CN111815748B (en) Animation processing method and device, storage medium and electronic equipment
WO2023056833A1 (en) Background picture generation method and apparatus, image fusion method and apparatus, and electronic device and readable medium

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 19901301

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 11.11.2021)

122 Ep: pct application non-entry in european phase

Ref document number: 19901301

Country of ref document: EP

Kind code of ref document: A1