CN107369158A - The estimation of indoor scene layout and target area extracting method based on RGB D images - Google Patents

The estimation of indoor scene layout and target area extracting method based on RGB D images Download PDF

Info

Publication number
CN107369158A
CN107369158A CN201710442910.9A CN201710442910A CN107369158A CN 107369158 A CN107369158 A CN 107369158A CN 201710442910 A CN201710442910 A CN 201710442910A CN 107369158 A CN107369158 A CN 107369158A
Authority
CN
China
Prior art keywords
plane
mrow
segmentation
rgb
target area
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201710442910.9A
Other languages
Chinese (zh)
Other versions
CN107369158B (en
Inventor
吴晓秋
霍智勇
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nanjing Post and Telecommunication University
Nanjing University of Posts and Telecommunications
Original Assignee
Nanjing Post and Telecommunication University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nanjing Post and Telecommunication University filed Critical Nanjing Post and Telecommunication University
Priority to CN201710442910.9A priority Critical patent/CN107369158B/en
Publication of CN107369158A publication Critical patent/CN107369158A/en
Application granted granted Critical
Publication of CN107369158B publication Critical patent/CN107369158B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/11Region-based segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/13Edge detection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/50Depth or shape recovery
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/10Terrestrial scenes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10024Color image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10028Range image; Depth image; 3D point clouds
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20024Filtering details

Abstract

The present invention discloses a kind of indoor scene layout estimation based on RGB D images and target area extracting method, including:Scene layout estimates;Over-segmentation is done to pretreated depth map and RGB figures using the partitioning algorithm based on figure and constrained parameters minimal cut algorithm, obtains different size of regional ensemble;Over-segmentation level is grouped, and is carried out region merging technique using four kinds of different measuring similarity modes to complete regional level packet, is obtained the target area of all scale sizes;And object boundary frame matching.The present invention realizes the target area extraction of efficient, the high recall rate of indoor scene.

Description

The estimation of indoor scene layout and target area extracting method based on RGB-D images
Technical field
The invention belongs to artificial intelligence computing technique field, particularly a kind of indoor scene layout based on RGB-D images Estimation and target area extracting method, applied to indoor service robot technology.
Background technology
The research of indoor scene parsing is one of study hotspot of domestic and foreign scholars, for the semantic positioning of Indoor Robot and Map generation have important application value, simultaneously for solve some high level computer visual problems also have it is very important Meaning.Target Segmentation and extraction algorithm purpose are target positioning and the example segmentation result for obtaining high quality, are scene parsings One of committed step.Objective extraction result is usually object candidate area or object boundary frame, by development for many years, target Extraction algorithm can be divided into two classes at present:The first kind is the algorithm based on sliding window detection thought, and the second class is based on segmentation Algorithm, include image over-segmentation and segmentation split strategy.First kind algorithm comparison classics are DPM (Deformable Parts Model) algorithm of target detection, there is very strong robustness using HOG features and SVM classifier, the deformation to target is improved, but It is that this kind of algorithm calculation cost is larger, and more complicated character representation can not be used.
More classical in second class algorithm is image segmentations of the GBS (graph based segmentation) based on figure Algorithm, the algorithm is realized simply, speed, can find out visually consistent region, but easily cause over-segmentation;Also The Target Segmentation algorithm of constrained parameters minimal cut, the algorithm segmentation effect is good, but only includes foreground segmentation region.In recent years by In the popularization of depth transducer, there are the RGB-D image data sets for largely including depth image, researcher starts with RGB- D data sets lift effect by increasing geometric properties or depth information, but these algorithms generally all to there is supervision algorithm, it is necessary to The skeleton pattern being previously obtained is trained, computation complexity is larger, although part improves the degree of accuracy of Objective extraction, target Classification is less and recall rate is relatively low, and easily ignores plane domain object when detecting.Also there are some unsupervised RGB-D in addition Objective extraction and partitioning algorithm, calculating speed is very fast, but more sensitive to brightness of image change, noise etc., and robustness is not high.
Although object extraction algorithm is constantly developing, due to the office of the features such as RGB image texture, color, brightness It is sex-limited, in the indoor scene applied to complexity, problems with still be present:1) occlusion issue, because occlusion detection is less than one A little big targets;2) the problem of plane domain object and small-size object are easily ignored so that recall rate is relatively low;3) calculate complicated Spend larger, it is necessary to which pre-training, is unsuitable for real system application;4) it is poor that respond is influenceed on the uncertain factor in image, Robustness is low.
The content of the invention
The present invention in order to overcome the above-mentioned deficiencies of the prior art, takes into account recall rate and rapidity, solves complex indoor scene The problem of lower layout estimation and Objective extraction, it is proposed that a kind of estimation of indoor scene layout and target area based on RGB-D images Domain extracting method, realize the target area extraction of efficient, the high recall rate of indoor scene.
The estimation of indoor scene layout and target area extracting method based on RGB-D images, comprise the following steps,
Step 1, scene layout's estimation:Depth map is converted into intensive 3D point cloud, by three-dimensional European between calculating point cloud Distance carries out plane segmentation division plane domain and non-planar area, and gained plane domain is classified, and is divided into border and puts down Face and non-boundary plane;
Step 2, image over-segmentation:Using based on the partitioning algorithm of figure with constrained parameters minimal cut algorithm to pretreated Depth map does over-segmentation with RGB figures, obtains different size of regional ensemble;
Step 3, the packet of over-segmentation level:Utilize color, texture, size, four kinds of different measuring similarity modes of coincideing Region merging technique is carried out to complete regional level packet, obtains the target area of all scale sizes;
Step 4, the matching of object boundary frame:To plane domain and the target of non-planar area, it is divided into boundary plane, non-border In plane, plane domain, four kinds of non-planar area situation take the minimum rectangle bounding box that different strategy matchings includes target, Obtain target area boundaries frame.
The present invention does plane segmentation and classification using the 3D point cloud of input, make use of the geometric continuity of a cloud to reduce and blocks Influence to layout estimation, improve the effect of scene layout's estimation;It is minimum using the partitioning algorithm based on figure and constrained parameters Cut algorithm and over-segmentation is done to pretreated depth map and RGB figures, combine depth information and RGB information, improve segmentation effect Fruit;Region merging technique is carried out using four kinds of different measuring similarity modes, obtains the target area of all scale sizes, and A variety of image conditions are considered, add the robustness of algorithm;The target area of plane domain and non-planar area is taken not Same bounding box matching strategy, had both remained the object in plane domain, and it is undue caused by due to blocking to improve big object again Cut problem;Rate is overlapped using bounding box and eliminates redundancy bounding box, leaves optimum target zone boundary frame, is producing less candidate side Object boundary frame recall rate is effectively increased in the case of boundary's frame;Whole process does not need pre-training, and computation complexity is low, is easy to real Existing, calculating speed is fast.
Brief description of the drawings
Fig. 1 is the flow of the estimation of indoor scene layout and the embodiment of target area extracting method one based on RGB-D images Figure;
Fig. 2 is depth map and 3D point cloud schematic diagram in Fig. 1 embodiments;
Fig. 3 is plane segmentation and the classifying quality figure of different scenes;
Fig. 4 is homomorphic filtering process chart in Fig. 1 embodiments;
Fig. 5 is the bounding box design sketch of the target area extraction of an implement scene in Fig. 1 embodiments.
Embodiment
In order to make the purpose , technical scheme and advantage of the present invention be clearer, it is right below in conjunction with drawings and Examples The present invention is further elaborated.It should be appreciated that the specific embodiments described herein are merely illustrative of the present invention, and It is not used in the restriction present invention.
Fig. 1 is estimated by the layout of the indoor scene based on RGB-D images that the embodiment of the present invention proposes and target area carries Take the overview flow chart of method.The embodiment step is as follows
Step (1) scene layout estimates:Depth map is converted into intensive 3D point cloud first, as shown in Fig. 2 then passing through meter The three-dimensional Euclidean distance calculated between point cloud carries out plane segmentation division plane domain and non-planar area, and gained plane domain is entered Row classification, is divided into boundary plane and non-boundary plane.
Step (1.1) plane is split:Carry out uniformity on depth map to sample to obtain triangle point set, for therein each Triangulation point group uses one candidate plane of RANSAC algorithmic match;Then search point in plane in 3d space, it is each in point can be by A pixel and its corresponding 3D available points in depth map represent, when the three-dimensional Euclidean distance of a point to the plane is less than Interior point is apart from tolerance DtolWhen, the interior point that the point is plane is defined, interior point is apart from tolerance DtolCalculate as shown in formula (1);Finally move Except the few tiny plane of interior quantity, and split spatial closeness or close to coplanar plane.
In formula, f is focal length, and b is the baseline length of sensor, and m is linear normalization parameter, Z representative depth values.
Step (1.2) plane is classified:According to obtained dominant plane region, it is assumed that the normal line vector of plane is towards observation Person, calculate in the ratio between point cloud total quantity for putting cloud number and whole scene of plane another side, will be less than the plane of certain threshold value point Class is boundary plane, and the plane more than certain threshold value is categorized as non-boundary plane.Ideally the threshold value is 0, it is contemplated that Influence of noise, it is arranged to 0.01.Plane is classified as shown in Figure 3 with the final effect of segmentation.
Step (2) image over-segmentation:Using the partitioning algorithm based on figure and constrained parameters minimal cut algorithm to pretreatment after Depth map and RGB figures do over-segmentation, obtain different size of regional ensemble R={ r1..., rn}。
With reference to RGB information and depth information, the image of multiple pixel scales from unlike signal passage is carried out first Different degrees of over-segmentation, the other image of region class is obtained, bottom-up group technology is then utilized to area according to provincial characteristics Domain carries out level packet until whole image turns into a region, to obtain the group rank figure for including all size target areas Picture.
The segmentation of step (2.1) based on RGI color spaces:RGB triple channel images are converted into normalization RG passages to add Brightness I passages, i.e. RGI color spaces, it is then different degrees of undue using doing three kinds to RGI images based on the dividing method of figure Cut.
The segmentation of step (2.2) based on the gray-scale map after homomorphic filtering:First RGB image is handled plus homomorphic filtering, place Then reason flow using the dividing method based on figure to the gray-scale map exported after processing as shown in figure 4, carry out three kinds in various degree Over-segmentation.
The segmentation of step (2.3) based on the depth map after hole-filling:Depth map is entered using global optimization color method Row hole-filling, three kinds of different degrees of over-segmentations then are carried out to the depth map after filling up using based on the dividing method of figure.
The segmentation of step (2.4) based on RGB-D hybrid channels:Using the foreground segmentation method of constrained parameters minimal cut, knot RGB image information and depth information are closed, over-segmentation is carried out to the image of RGB-D hybrid channels, this method is based on such as following formula (2) institute The energy theorem shown:
μ in formula, v ∈ N, λ ∈ R, v are the set of all pixels point, edge aggregations of the ε between adjacent pixel.CλFor cost Function, a cost can be produced when assigning label to each pixel.Binary potential function VμvHerein as penalty, when to When similar adjacent pixel assigns different labels, a penalty value will be produced.
Step (2.4.1) calculation cost function Cλ
V in formulafRepresent foreground seeds, vbBackground seed is represented, λ is offset, and formula f is defined as follows formula (4):
f(xμ)=lnpf(xμ)-lnpb(xμ) (4)
P in formulafRepresent that pixel μ belongs to the probability distribution of foreground area, add p after depth informationfIt is defined as follows formula (5) shown in:
D is depth map in formula, and I is RGB image.The representative pixel points of j expression seed regions, these pixels are equal by K- Value-based algorithm (k=5) is selected as regional center, and α is scale factor, and γ is scale factor.
Step (2.4.2) calculates penalty Vμv
The similitude g (μ, υ) of two neighborhood pixels is calculated according to pixel μ and υ gPb values in formula:
σ in formula2For edge sharpening parameter, binary item V is controlledμvSlickness.RGB image and depth map are all carried out GPb is calculated, and its linear combination is risen and is used as the final gPb values of each pixel:
GPb=α gPbr+(1-α)·gPbd (8)
GPb in formularRepresent the gPb values of pixel in RGB figures, gPbdThe gPb values of pixel in depth map are represented, α is set to 0.3。
Step (3) over-segmentation level is grouped:Region merging technique is carried out to complete using four kinds of different measuring similarity modes Regional level is grouped, and obtains the target area of all scale sizes.
Step (3.1) calculates the similarity s (r of all adjacent areas two-by-two firsti, rj) and be added to similarity collection Close in S, find out two region r that similarity is maximum in set SiAnd rj, it is merged into a region rt, and it is added to region In set R;
Step (3.2) removes r from similarity setiAnd rjThe similarity in region adjacent thereto, i.e. S=S s (ri, r*), Calculate new region rtThe similarity in region adjacent thereto, its result is added in similarity set S;
Step (3.3) repeats (3.1) and arrives (3.2) step, until whole image turns into a big region, completes the layer in region Secondary packet, obtain the target area of all scale sizes.
Step (3.1.1) calculates color similarity:The face for the 25bins for obtaining tri- passages of RGB is normalized using L1 norms Color Histogram so that each region obtains the vector of one 75 dimensionThen according to the vector as shown in formula (9) Color similarity between zoning:
Step (3.1.2) calculates texture similarity:The height that variance is 1 is calculated to 8 different directions of each Color Channel This differential, each direction of each Color Channel obtain 10bins histogram so that each region obtains one 240 dimension VectorThen the texture similarity between the zoning as shown in formula (10):
Step (3.1.3) calculates size similarity:Accounting as shown in formula (11) according to each area size in the picture Size similarity between rate zoning:
I refers to whole image in formula.
Step (3.1.4) calculates similarity of coincideing:Calculate the minimum rectangle bounding box B for including two combined regionijWith two The difference of the size in region, then according to identical similar between phase difference accounting rate zoning in the picture as shown in formula (12) Degree:
Step (3.1.5) combines four kinds of similarities:Last similarity s (ri, rj) calculation by more than four kinds of similarities Shown in linear combination such as following formula (13):
s(ri, rj)=a1sc(ri, rj)+a2st(ri, rj)+a3ss(ri, rj)+a4sf(ri, rj) (13)
A in formulai∈ { 0,1 }, represents whether the similarity is used.
Step (4):Object boundary frame matches:To plane domain and the target of non-planar area, it is divided into four kinds of situations and takes Different strategy matchings includes the minimum rectangle bounding box of target, obtains target area boundaries frame.
Step (4.1) is all directly used for plane domain, boundary plane region;To each non-boundary plane, look for The minimum euclidean distance with other non-boundary planes is calculated to its boundary point, the plane split that distance is less than to certain threshold value rises Come, using the non-boundary plane region after split;And positioned at the target of plane domain, only retain the target as caused by RGB image Region.
Step (4.2) is for non-planar area, in addition to the target area too small with non-planar area overlapping area, its He is directly used target area.
Step (4.3) matches object boundary frame:The target area of all uses is converted into shade, to these shades one by one Matching includes the minimum rectangle bounding box including shade, obtains shown in bounding box set B such as formulas (14):
B=BBP+BMPR+BNPR+BPR (14)
BP represents boundary plane region in formula, and MPR represents split plane domain, and NPR represents the mesh positioned at non-planar area Mark, PR represent the target positioned at plane domain.
Then the tiny bounding box in set B is removed, and bounding box is sorted by size, is from top to bottom iterated to calculate Overlapping rate between bounding box, overlap rate Q (bi, bj) computational methods such as following formula (15), then filter off overlapping rate and be more than certain threshold value Bounding box is repeated, obtains optimum target zone boundary frame set to the end, effect is as shown in Figure 5.
B in formulai, bj∈ B, a (bi) represent bounding box biArea.
Technological means disclosed in the present invention program is not limited only to the technological means disclosed in above-mentioned embodiment, in addition to Formed technical scheme is combined by above technical characteristic.

Claims (5)

1. the estimation of indoor scene layout and target area extracting method based on RGB-D images, it is characterised in that:Including following step Suddenly,
Step 1, scene layout's estimation:Depth map is converted into intensive 3D point cloud, by calculating the three-dimensional Euclidean distance between point cloud Carry out plane segmentation division plane domain and non-planar area, and gained plane domain classified, be divided into boundary plane with Non- boundary plane;
Step 2, image over-segmentation:Using the partitioning algorithm based on figure and constrained parameters minimal cut algorithm to pretreated depth Figure does over-segmentation with RGB figures, obtains different size of regional ensemble;
Step 3, the packet of over-segmentation level:Carried out using color, texture, size, four kinds of different measuring similarity modes of coincideing Region merging technique obtains the target area of all scale sizes to complete regional level packet;
Step 4, the matching of object boundary frame:To plane domain and the target of non-planar area, it is divided into boundary plane, non-border is put down In face, plane domain, four kinds of non-planar area situation take different strategy matchings to include the minimum rectangle bounding box of target, obtain To target area boundaries frame.
2. the estimation of indoor scene layout and target area extracting method according to claim 1 based on RGB-D images, its It is characterised by:Step 1 detailed process is,
Step 1.1, plane segmentation:Carry out uniformity on depth map to sample to obtain triangle point set, for each triangle therein Point group uses one candidate plane of RANSAC algorithmic match;Then 3D point cloud space search plane in point, when a point arrive this The three-dimensional Euclidean distance of plane is less than interior point apart from tolerance DtolWhen, the interior point that the point is plane is defined, interior point is apart from tolerance Dtol Calculate such as formula (1);The tiny plane of point in finally removing, and split spatial closeness or close to coplanar plane;
<mrow> <msub> <mi>D</mi> <mrow> <mi>t</mi> <mi>o</mi> <mi>l</mi> </mrow> </msub> <mo>=</mo> <mrow> <mo>(</mo> <mfrac> <mi>m</mi> <mrow> <mi>f</mi> <mi>b</mi> </mrow> </mfrac> <mo>)</mo> </mrow> <mo>*</mo> <msup> <mi>Z</mi> <mn>2</mn> </msup> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>1</mn> <mo>)</mo> </mrow> </mrow>
In formula, f is focal length, and b is the baseline length of sensor, and m is linear normalization parameter, Z representative depth values;
Step 1.2, plane classification:According to obtained dominant plane region, it is assumed that the normal line vector of plane calculates towards observer In the ratio between point cloud total quantity for putting cloud number and whole scene of plane another side, the plane less than threshold value is categorized as border and put down Face, the plane more than threshold value are categorized as non-boundary plane.
3. the estimation of indoor scene layout and target area extracting method according to claim 1 based on RGB-D images, its It is characterised by:Step 2 detailed process is,
Step 2.1, the segmentation based on RGI color spaces:RGB triple channel images are converted into normalization RG passages and add brightness I Passage, i.e. RGI color spaces, over-segmentation then is done to RGI images using based on the dividing method of figure;
Step 2.2, the segmentation based on the gray-scale map after homomorphic filtering:RGB image is handled plus homomorphic filtering, to defeated after processing The gray-scale map gone out carries out over-segmentation using the dividing method based on figure;
Step 2.3, the segmentation based on the depth map after hole-filling:Cavity is carried out to depth map using global optimization color method Fill up, over-segmentation is carried out to the depth map after filling up using based on the dividing method of figure;
Step 2.4, the segmentation based on RGB-D hybrid channels:Using the foreground segmentation method of constrained parameters minimal cut, with reference to RGB Image information and depth information, over-segmentation, the energy theorem based on formula (2) are carried out to the image of RGB-D hybrid channels:
<mrow> <msup> <mi>E</mi> <mi>&amp;lambda;</mi> </msup> <mrow> <mo>(</mo> <mi>X</mi> <mo>)</mo> </mrow> <mo>=</mo> <munder> <mi>&amp;Sigma;</mi> <mrow> <mi>&amp;mu;</mi> <mo>&amp;Element;</mo> <mi>&amp;gamma;</mi> </mrow> </munder> <msub> <mi>C</mi> <mi>&amp;lambda;</mi> </msub> <mrow> <mo>(</mo> <msub> <mi>x</mi> <mi>&amp;mu;</mi> </msub> <mo>)</mo> </mrow> <mo>+</mo> <munder> <mi>&amp;Sigma;</mi> <mrow> <mo>(</mo> <mi>&amp;mu;</mi> <mo>,</mo> <mi>v</mi> <mo>)</mo> <mo>&amp;Element;</mo> <mi>&amp;epsiv;</mi> </mrow> </munder> <msub> <mi>V</mi> <mrow> <mi>&amp;mu;</mi> <mi>v</mi> </mrow> </msub> <mrow> <mo>(</mo> <msub> <mi>x</mi> <mi>&amp;mu;</mi> </msub> <mo>,</mo> <msub> <mi>x</mi> <mi>v</mi> </msub> <mo>)</mo> </mrow> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>2</mn> <mo>)</mo> </mrow> </mrow>
μ in formula, v ∈ N, λ ∈ R, v are the set of all pixels point, edge aggregations of the ε between adjacent pixel, CλFor cost function, A cost, binary potential function V can be produced when assigning label to each pixelμvAs penalty, when to similar adjacent When pixel assigns different labels, a penalty value will be produced.
4. the estimation of indoor scene layout and target area extracting method according to claim 1 based on RGB-D images, its It is characterised by:Step 3 detailed process is,
Step 3.1, the similarity set for calculating all adjacent areas two-by-two, find out two maximum region r of similarityiAnd rj, It is merged into a new region rt, it is added in regional ensemble;
Step 3.2, remove r from similarity setiAnd rjThe similarity in region adjacent thereto, calculate new region rtIt is adjacent thereto The similarity in region, it is added in similarity set;
Step 3.3, repeat step 3.1~3.2, until whole image turns into a big region, the level for completing image is grouped, and is obtained Take the target area of all scale sizes.
5. the estimation of indoor scene layout and target area extracting method according to claim 1 based on RGB-D images, its It is characterised by:Step 4 detailed process is,
Step 4.1, all directly it is used for plane domain, boundary plane region;To each non-boundary plane, its is found Boundary point calculates the minimum euclidean distance with other non-boundary planes, and the plane that distance is less than to threshold value pieces together, using spelling Non- boundary plane region after conjunction;And positioned at the target of plane domain, only retain the target area as caused by RGB image;
Step 4.2, for non-planar area, in addition to the target area too small with non-planar area overlapping area, other mesh Mark region is all directly used;
Step 4.3, matching object boundary frame:The target area of all uses is converted into shade, these shades are matched one by one Comprising the minimum rectangle bounding box including shade, tiny bounding box is then removed, and bounding box is sorted by size, by upper The overlapping rate between bounding box is computed repeatedly under, the bounding box that overlapping rate is more than certain threshold value is filtered off, obtains target area to the end Domain bounding box.
CN201710442910.9A 2017-06-13 2017-06-13 Indoor scene layout estimation and target area extraction method based on RGB-D image Active CN107369158B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201710442910.9A CN107369158B (en) 2017-06-13 2017-06-13 Indoor scene layout estimation and target area extraction method based on RGB-D image

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201710442910.9A CN107369158B (en) 2017-06-13 2017-06-13 Indoor scene layout estimation and target area extraction method based on RGB-D image

Publications (2)

Publication Number Publication Date
CN107369158A true CN107369158A (en) 2017-11-21
CN107369158B CN107369158B (en) 2020-11-13

Family

ID=60306413

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201710442910.9A Active CN107369158B (en) 2017-06-13 2017-06-13 Indoor scene layout estimation and target area extraction method based on RGB-D image

Country Status (1)

Country Link
CN (1) CN107369158B (en)

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108742159A (en) * 2018-04-08 2018-11-06 浙江安精智能科技有限公司 Intelligent control device of water dispenser based on RGB-D cameras and its control method
CN109636814A (en) * 2018-12-18 2019-04-16 联想(北京)有限公司 A kind of image processing method and electronic equipment
CN110263692A (en) * 2019-06-13 2019-09-20 北京数智源科技有限公司 Container switch gate state identification method under large scene
CN111611919A (en) * 2020-05-20 2020-09-01 西安交通大学苏州研究院 Road scene layout analysis method based on structured learning
CN111815696A (en) * 2019-04-11 2020-10-23 曜科智能科技(上海)有限公司 Depth map optimization method, device, equipment and medium based on semantic instance segmentation
WO2020258793A1 (en) * 2019-06-26 2020-12-30 北京市商汤科技开发有限公司 Target detection and training of target detection network
CN113256776A (en) * 2021-06-21 2021-08-13 炫我信息技术(北京)有限公司 Image processing method and device, electronic equipment and computer readable storage medium
CN115272529A (en) * 2022-09-28 2022-11-01 中国海洋大学 Layout-first multi-scale decoupling ocean remote sensing image coloring method and system
CN116740809A (en) * 2023-06-05 2023-09-12 嘉兴米兰映像家具有限公司 Intelligent sofa control method based on user gesture

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20090297049A1 (en) * 2005-07-07 2009-12-03 Rafael Advanced Defense Systems Ltd. Detection of partially occluded targets in ladar images
CN102436654A (en) * 2011-09-02 2012-05-02 清华大学 Adaptive segmentation method of building point cloud
CN104809187A (en) * 2015-04-20 2015-07-29 南京邮电大学 Indoor scene semantic annotation method based on RGB-D data
CN105488809A (en) * 2016-01-14 2016-04-13 电子科技大学 Indoor scene meaning segmentation method based on RGBD descriptor

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20090297049A1 (en) * 2005-07-07 2009-12-03 Rafael Advanced Defense Systems Ltd. Detection of partially occluded targets in ladar images
CN102436654A (en) * 2011-09-02 2012-05-02 清华大学 Adaptive segmentation method of building point cloud
CN104809187A (en) * 2015-04-20 2015-07-29 南京邮电大学 Indoor scene semantic annotation method based on RGB-D data
CN105488809A (en) * 2016-01-14 2016-04-13 电子科技大学 Indoor scene meaning segmentation method based on RGBD descriptor

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
SAURABH GUPTA等: "Perceptual Organization and Recognition of Indoor Scenes from RGB-D Images", 《COMPUTER VISION FOUNDATION》 *

Cited By (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108742159A (en) * 2018-04-08 2018-11-06 浙江安精智能科技有限公司 Intelligent control device of water dispenser based on RGB-D cameras and its control method
CN109636814A (en) * 2018-12-18 2019-04-16 联想(北京)有限公司 A kind of image processing method and electronic equipment
CN111815696B (en) * 2019-04-11 2023-08-22 曜科智能科技(上海)有限公司 Depth map optimization method, device, equipment and medium based on semantic instance segmentation
CN111815696A (en) * 2019-04-11 2020-10-23 曜科智能科技(上海)有限公司 Depth map optimization method, device, equipment and medium based on semantic instance segmentation
CN110263692A (en) * 2019-06-13 2019-09-20 北京数智源科技有限公司 Container switch gate state identification method under large scene
WO2020258793A1 (en) * 2019-06-26 2020-12-30 北京市商汤科技开发有限公司 Target detection and training of target detection network
TWI762860B (en) * 2019-06-26 2022-05-01 大陸商北京市商湯科技開發有限公司 Method, device, and apparatus for target detection and training target detection network, storage medium
CN111611919A (en) * 2020-05-20 2020-09-01 西安交通大学苏州研究院 Road scene layout analysis method based on structured learning
CN113256776A (en) * 2021-06-21 2021-08-13 炫我信息技术(北京)有限公司 Image processing method and device, electronic equipment and computer readable storage medium
CN115272529B (en) * 2022-09-28 2022-12-27 中国海洋大学 Layout-first multi-scale decoupling ocean remote sensing image coloring method and system
CN115272529A (en) * 2022-09-28 2022-11-01 中国海洋大学 Layout-first multi-scale decoupling ocean remote sensing image coloring method and system
CN116740809A (en) * 2023-06-05 2023-09-12 嘉兴米兰映像家具有限公司 Intelligent sofa control method based on user gesture
CN116740809B (en) * 2023-06-05 2024-03-29 嘉兴米兰映像家具有限公司 Intelligent sofa control method based on user gesture

Also Published As

Publication number Publication date
CN107369158B (en) 2020-11-13

Similar Documents

Publication Publication Date Title
CN107369158A (en) The estimation of indoor scene layout and target area extracting method based on RGB D images
CN108665481B (en) Self-adaptive anti-blocking infrared target tracking method based on multi-layer depth feature fusion
CN105389584B (en) Streetscape semanteme marking method based on convolutional neural networks with semantic transfer conjunctive model
CN103810503A (en) Depth study based method for detecting salient regions in natural image
CN106909902B (en) Remote sensing target detection method based on improved hierarchical significant model
Alidoost et al. A CNN-based approach for automatic building detection and recognition of roof types using a single aerial image
CN103337072B (en) A kind of room objects analytic method based on texture and geometric attribute conjunctive model
CN105528575B (en) Sky detection method based on Context Reasoning
CN105261017A (en) Method for extracting regions of interest of pedestrian by using image segmentation method on the basis of road restriction
CN104517095B (en) A kind of number of people dividing method based on depth image
CN107564022A (en) Saliency detection method based on Bayesian Fusion
Asokan et al. Machine learning based image processing techniques for satellite image analysis-a survey
Gupta et al. Automatic trimap generation for image matting
CN113408584B (en) RGB-D multi-modal feature fusion 3D target detection method
Wang et al. An overview of 3d object detection
CN104766065A (en) Robustness prospect detection method based on multi-view learning
CN109086777A (en) A kind of notable figure fining method based on global pixel characteristic
Li et al. Unsupervised road extraction via a Gaussian mixture model with object-based features
CN111401380A (en) RGB-D image semantic segmentation method based on depth feature enhancement and edge optimization
Li et al. Transmission line detection in aerial images: An instance segmentation approach based on multitask neural networks
Bao et al. Unpaved road detection based on spatial fuzzy clustering algorithm
CN113592893B (en) Image foreground segmentation method for determining combination of main body and accurate edge
CN105354547A (en) Pedestrian detection method in combination of texture and color features
Wang et al. A region-line primitive association framework for object-based remote sensing image analysis
Ren et al. Research on infrared small target segmentation algorithm based on improved mask R-CNN

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
CB02 Change of applicant information
CB02 Change of applicant information

Address after: Room 201, building 2, phase II, No.1 Kechuang Road, Yaohua street, Qixia District, Nanjing City, Jiangsu Province, 210003

Applicant after: NANJING University OF POSTS AND TELECOMMUNICATIONS

Address before: 210003 Gulou District, Jiangsu, Nanjing new model road, No. 66

Applicant before: NANJING University OF POSTS AND TELECOMMUNICATIONS

GR01 Patent grant
GR01 Patent grant