CN107832727A - A kind of indoor mall shop feature extracting method - Google Patents
A kind of indoor mall shop feature extracting method Download PDFInfo
- Publication number
- CN107832727A CN107832727A CN201711167137.6A CN201711167137A CN107832727A CN 107832727 A CN107832727 A CN 107832727A CN 201711167137 A CN201711167137 A CN 201711167137A CN 107832727 A CN107832727 A CN 107832727A
- Authority
- CN
- China
- Prior art keywords
- shop
- neutral net
- probability
- decorated style
- picture
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/35—Categorising the entire scene, e.g. birthday party or wedding scene
- G06V20/36—Indoor scenes
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F30/00—Computer-aided design [CAD]
- G06F30/10—Geometric CAD
- G06F30/18—Network design, e.g. design based on topological or interconnect aspects of utility systems, piping, heating ventilation air conditioning [HVAC] or cabling
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/60—Type of objects
- G06V20/62—Text, e.g. of license plates, overlay texts or captions on TV images
- G06V20/63—Scene text, e.g. street names
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Evolutionary Computation (AREA)
- Geometry (AREA)
- Multimedia (AREA)
- Data Mining & Analysis (AREA)
- General Engineering & Computer Science (AREA)
- Computer Hardware Design (AREA)
- Computational Mathematics (AREA)
- Pure & Applied Mathematics (AREA)
- Computer Networks & Wireless Communication (AREA)
- Mathematical Optimization (AREA)
- Life Sciences & Earth Sciences (AREA)
- Artificial Intelligence (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Bioinformatics & Computational Biology (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Evolutionary Biology (AREA)
- Mathematical Analysis (AREA)
- Image Analysis (AREA)
Abstract
The invention discloses a kind of indoor mall shop feature extracting method, include M shop in market;Comprise the following steps:S1, structure shop decorated style identification neutral net prototype;S2, build shop text detection neutral net prototype;S3, the shop image data set established in the market, each image include the text information and decorated style information in shop, image data set are divided into training set and test set;The weights of initialization decorated style identification neutral net and text detection neutral net are distributed by gaussian random, train decorated style to identify neutral net prototype and text detection neutral net prototype to minimize cost function as target, determine the model structure of shop decorated style identification neutral net and text detection neutral net;S4, the identification of shop decorated style, shop text detection;S5, shop Text region;S6, recognition result fusion.The extracting method of the present invention, it is remarkably improved the degree of accuracy of shop identification.
Description
【Technical field】
The present invention relates to computer vision and digital image processing field, and high-level vision feature is based on more particularly to one kind
Indoor mall shop feature extracting method.
【Background technology】
In the indoor environment for lacking portable, inexpensive alignment system (GPS as being used for outdoor positioning), indoor positioning
System (IPS) is always an attractive research topic.Using based on the indoor locating system of infrastructure device including the use of
RFID's, use fluorescent lamp or Wi-Fi access points, the performance impressive of these systems in actual applications.It is another
Aspect, the superiority and challenge that the IPS independent of infrastructure has in itself cause more extensive concern.It is existing
Method proposes the IPS based on computer vision using image retrieval technologies, what these systems can be shot by smart mobile phone
Photo tells the position of user.However, all these methods are required for carrying out offline database structure in advance, this process is suitable
It is time-consuming and expensive.
In recent years, the latest developments in terms of robotics and computer vision have new application again on IPS.It is synchronous fixed
Position and map structuring (SLAM) technology and visual token (VO) are due to can accurately estimate to move and as IPS important technology.
Answered in systems in practice using the monocular SLAM systems of bag of words, monocular VO algorithms and system based on LIDAR
With.However, operation SLAM or VO means that user must be by video camera or laser transceiver recorded video.
In order to solve this problem, there is scholar to propose to carry out shop orientation based on the text recognition method in image.Specifically
For, this system is classified in the shop in image by text identification, and the terrestrial reference as coarse localization (passes through
Shop classification is positioned).This method has flexible scalability, because except the market plan for needing to mark in advance is made
Outside input, it is not necessary to collect the largely data on indoor scene in advance.But only by text identification shop precision
Not high, because in actual environment, many words, which are submerged in noise, to be difficult to detect.
The disclosure of background above technology contents is only used for inventive concept and the technical scheme that auxiliary understands the present invention, and it is not
The prior art of present patent application is necessarily belonged to, shows the applying date of the above in present patent application in no tangible proof
In the case of disclosed, above-mentioned background technology should not be taken to evaluate the novelty and creativeness of the application.
【The content of the invention】
The technical problems to be solved by the invention are:Above-mentioned the deficiencies in the prior art are made up, propose a kind of indoor mall shop
Feature extracting method is spread, is remarkably improved the degree of accuracy of shop identification.
The technical problem of the present invention is solved by following technical scheme:
A kind of indoor mall shop feature extracting method, include M shop in the market;Comprise the following steps:S1, structure
Build shop decorated style identification neutral net prototype, the decorated style identification neutral net prototype include convolutional layer, pond layer,
Excitation layer and full articulamentum;S2, builds shop text detection neutral net prototype, and the text detection neutral net prototype includes
Convolutional layer, pond layer, excitation layer and warp lamination;S3, the shop image data set established in the market, each image include shop
The text information and decorated style information of paving, training set and test set are divided into by image data set;It is distributed by gaussian random
The weights of the decorated style identification neutral net and text detection neutral net are initialized, to minimize cost function as target
The text detection neutral net prototype come in decorated style identification neutral net prototype and step S2 in training step S1, it is determined that
Go out the model structure of shop decorated style identification neutral net and text detection neutral net;S4, shop name is included for one
Claim the picture to be identified with upholstery style, be inputted shop decorated style neutral net and the text detection god respectively
In model structure through network, obtain the picture and belong in the probability and the picture in each shop the area for corresponding to word segment
Domain;S5, by the region input Text region module that word segment is corresponded in the picture obtained in step S4, obtain the picture
NGRAM corresponding to middle word segment is encoded;S6, according to the probability calculation shop decorated style identification probability obtained in step S4,
NGRAM codings according to being obtained in step S5 calculate Text region probability, and the identification in finally corresponding M shop is obtained using weighting
Probability, the shop recognition result of the picture is used as using shop corresponding to maximum probability value in obtained probability.
The beneficial effect that the present invention is compared with the prior art is:
The indoor mall shop feature extracting method of the present invention, StoreFront identification is carried out based on high-level vision feature, passes through structure
Build neutral net and be trained determination with a large amount of shop image data sets and obtain decorated style identification neutral net and word
Neutral net is detected, uses the text in the decorated style information and picture in the shop in above-mentioned neutral net acquisition picture to be measured
Block domain, and the text information (NGRAM codings) in character area identification shop is combined, then by decorated style information and word
Both information of information are merged to identify current shop.The fusion of shop decorated style and text information significantly improves shop
Spread the degree of accuracy of identification so that the indoor mall positioning based on the fusion results also can be more healthy and stronger.
【Brief description of the drawings】
Fig. 1 is the block schematic illustration of the indoor mall shop feature extracting method of the specific embodiment of the invention;
Fig. 2 is the flow chart of the indoor mall shop feature extracting method of the specific embodiment of the invention;
Fig. 3 is the signal of the concrete model structure of the decorated style identification neutral net in the specific embodiment of the invention
Figure;
Fig. 4 is the signal of the concrete model structure of the shop text detection neutral net in the specific embodiment of the invention
Figure;
Fig. 5 is the experiment test that the indoor mall shop feature extracting method in the specific embodiment of the invention is used to position
Result figure.
【Embodiment】
With reference to embodiment and compare accompanying drawing the present invention is described in further details.
As shown in figure 1, the frame diagram of the indoor mall shop feature extracting method of present embodiment, including shop dress
The identification of decorations style, shop text detection merge with identification, recognition result, are carried out positioning four modules according to recognition result.Such as Fig. 2
It is shown, it is the flow chart of the indoor mall shop feature extracting method of present embodiment, comprises the following steps:
S1, structure shop decorated style identification neutral net prototype, the decorated style identification neutral net prototype include
Convolutional layer, pond layer, excitation layer and full articulamentum.
S2, builds shop text detection neutral net prototype, and the text detection neutral net prototype includes convolutional layer, pond
Change layer, excitation layer and warp lamination.
S3, the shop image data set established in the market, text information and affiliated shop of each image including shop
Information, image data set is divided into training set and test set;The decorated style is initialized by gaussian random distribution to identify
The weights of neutral net and text detection neutral net, the decoration wind come using minimizing cost function as target in training step S1
Lattice identify the text detection neutral net prototype in neutral net prototype and step S2, determine shop decorated style identification nerve
The model structure of network and text detection neutral net.
In present embodiment, the shop image data set of foundation includes 2876 RGB shops pictures, covers a market
In all 56 different shops.The mark and affiliated brand, the mark of decorated style of every image character area including shop
Note.In use, the RGB image to be cut into 224x224 image block, and mark each RGB image shop title and
Character area.Shop data set is divided into test set and training set, wherein training set there are 2300, and test set there are 576.
During training, cost function is:L=- ∑sitilog(yi)+λ||W||2。
Wherein, when decorated style identifies neutral net prototype in training step S1, yiRepresent the image in the training set
Shop probability obtained by after decorated style identifies neural metwork training, tiBelong to M for the image in the training set
The actual probabilities in each shop in shop, λ represent the regularization coefficient of cost function, and W represents the decorated style identification nerve
The weights of convolutional layer and full articulamentum in network model.
In training step S2 during shop text detection neutral net prototype, yiRepresent the image in the training set by shop
Each pixel belongs to the probability of character area, t in the image exported after paving text detection neural metwork trainingiFor the training
Each pixel belongs to the actual probabilities of character area in the image of concentration, and λ represents the regularization coefficient of cost function, and W is represented
The weights of convolutional layer and warp lamination in the text detection neural network model.
Usually, λ can take system recognition accuracy soprano on training set by cross-validation experiments, and it is logical
Normal value includes 0.01,0.03,0.1,0.3 etc., regularization coefficient λ=0.01 of cost function in present embodiment.
In the network described in training step S1,8 224x224 fritter is intercepted from each picture of training set at random
And it is sent into network.During interception, the deviation of fritter is generated at random, then the coordinate using this deviation as the fritter upper left corner,
Intercept out a fritter.Network carries out operation, the outputs such as convolution, Chi Hua, activation and full connection to input picture successively to be passed through
The picture of One-hot codings belongs to the probability s of each brands∈R56.This probability is used for calculation cost function simultaneously.
When training the network described in S2, any interception is not carried out to picture, but directly input picture in its entirety, through pulleying
After product, Chi Hua, excitation and deconvolution operation, network exports the two-value picture of a formed objects.Export the picture that picture intermediate value is 1
Vegetarian refreshments represents character area, and 0 represents non-legible region.This two-value picture be used to be compared and calculate with character area mark
Cost function.In this embodiment, the picture number for being input to neutral net described in S1 and S2 every time is 20, and learning rate is
0.001。
The weights initialisation of convolutional neural networks uses one-dimensional gaussian profile, then constantly minimizes cost by algorithm
Function, you can the weights W of the equivalent layer in neutral net specific value is determined, so as to finally give the specific of neutral net
Model structure.Preferably, foregoing cost function is minimized using AdamOptimizer method in present embodiment,
Adam algorithms are more applicable for the network that data volume is big, the number of plies is deep compared to traditional SGD (stochastic gradient descent).Adam algorithms
It is to carry out single order to the gradient of each parameter according to cost function away from estimation and second order away from estimation, so as to which dynamic adjusts each parameter
Learning rate.The advantages of Adam algorithms, is more steady in the parameter in iterative process each time, it is not easy to is trapped in part most
It is excellent.
By training, the concrete model structure for the neutral net for determining to obtain in present embodiment is:
The model structure of decorated style identification neutral net includes the first convolution unit, the individual full articulamentums of N2, wherein the
One convolution unit includes the N1 convolutional layers being sequentially connected in series, connected after each convolutional layer a pond layer and a ReLu excitation
Layer, the output end series connection N2 full articulamentum of first convolution unit, the output dimension of last full articulamentum are
M, for exporting the binary set of M dimensions;Wherein, N1 and N2 value causes the model structure to reach highest standard on training set
True rate.It is trained, it is determined that obtaining N1=5 in present embodiment, N2=3.It is the dress of present embodiment shown in Fig. 3
The concrete model structure of style identification neutral net is adornd, it includes 5 convolutional layers being sequentially connected in series, the size difference of each convolutional layer
For 11x11,5x5,3x3,3x3,3x3, connected behind each convolutional layer a pond layer and a ReLu excitation layer, pond layer
It can be exported with convolutional layer and carry out dimensionality reduction, reduce the complexity of network, improve the generalization ability of network.Pond in present embodiment
Change layer to be operated using L2 Normalization.ReLu excitation layers can solve gradient in deep neural network training process and disappear
The problem of mistake.The output end of the convolutional layer 3 full articulamentums of last series connection.Output of 3 full articulamentums to convolutional layer is carried out
Processing and classification, the output dimension of last full articulamentum is arranged to 56 dimensions, therefore exports the binary set of 56 dimensions.Two-value to
The value of element in amount represents that the picture of input neutral net belongs to the binaryzation result of the probability in each shop in 56 shops.
The binary set of 56 dimension of full articulamentum output is handled in subsequent step S6 by sigmoid functions, obtains picture
Belong to the specific probable value s in each shops, the as recognition result of decorated style.
The model structure of text detection neutral net includes the second convolution unit and N4 warp lamination;The volume Two
Product unit includes the N3 convolutional layers being sequentially connected in series, respectively connected after the convolutional layer of part a pond layer and an excitation layer;Described
The output end series connection N4 warp lamination of two convolution units, last warp lamination are used to export binaryzation picture, schemed
The pixel that piece intermediate value is 1 represents character area, and the pixel being worth for 0 represents non-legible region, wherein, N3 and N4 value make
Obtain the model structure and reach highest accuracy rate on training set.It is trained, it is determined that obtaining above-mentioned specific value.Shown in Fig. 4 i.e.
For the concrete model structure of the shop text detection neutral net of present embodiment, it includes 13 convolutional layers, convolutional layer
The all 3x3 of size, 2x2 pond layer is respectively added after the 2nd, 4,7,10,13 layer, and each connect one after the layer of pond
Individual ReLu excitation layers.Two warp laminations are connected in series after 13 convolutional layers.The output of warp lamination and input neutral net
Picture size it is identical, the value of each of which pixel represents the probability that the pixel belongs to character area.
S4, for a picture to be identified comprising shop title and upholstery style, the shop is inputted respectively
In the model structure for spreading decorated style neutral net and text detection neutral net, the probability that the picture belongs to each shop is obtained
And the region of word segment is corresponded in the picture.
For picture to be identified, it is inputted in decorated style identification neural network model, and neural network model is to input
Picture to be identified carry out convolution, Chi Hua, activation and the processing connected entirely successively, output by One-hot codings the picture
Belong to the probability in each shop.
For picture to be identified, it is inputted in text detection neural network model, and neural network model is treated to input
After identification picture carries out convolution, Chi Hua, excitation and deconvolution processing successively, two with inputting picture formed objects are exported
It is worth picture, the pixel that picture intermediate value is 1 represents character area, and the pixel being worth for 0 represents non-legible region.
S5, by the region input Text region module that word segment is corresponded in the picture obtained in step S4, obtain
NGRAM corresponding to word segment is encoded in the picture.
Specifically, detection obtains corresponding to the area of text information in picture to be identified in text detection network model in step S4
Domain, binaryzation picture is inputted in a Text region module, its two-value picture to the output of text detection neural network model
Text identification processing is carried out, obtains the NGRAM codings of the word in picture to be identified.Text region module can be Max
Jaderberg et al. is in paper Deep Structured Output Learning for Unconstrained Text
The model proposed in Recognition, its input is character area picture, exports and encodes G for the NGRAM of word in pictureN∈
R10000。
S6, according to the probability calculation shop decorated style identification probability obtained in step S4, according to what is obtained in step S5
NGRAM codings calculate Text region probability, the identification probability in finally corresponding M shop are obtained using weighting, with obtained probability
Shop recognition result of the shop as the picture corresponding to middle maximum probability value.
In the step, shop decorated style identification probability is calculated according to such as minor function:ss=fs(a), wherein, a is represented
The output of the decorated style identification neutral net, the binary set of as 56 dimensions, ssBelong to the general of each shop for the picture
Rate, i.e., 56 dimensional vectors being made up of 56 specific probable values, fsFor sigmoid functions.
Text region probability is calculated according to such as minor function:st(j)=∑kGN(k) I (N (k) ∈ S (j)), wherein, S
(j) represent that character string corresponding to the trade name in j-th of shop, k represent the bits of coded in NGRAM coding schemes, be in M shop
Integer in 0 ..., 9999, N (k) represent character string corresponding to k-th of bits of coded in NGRAM codings, and N (k) ∈ S (j) represent word
Symbol string N (k) belongs to S (j), and I is indicative function, GN(k) the NGRAM codings of word recognized in input picture are represented.
After above two probable value is calculated, final identification probability is calculated according to equation below:Y=(1- α)
st+αss, wherein, stRepresent Text region probability, ssRepresent decorated style identification probability;α represents decorated style identification probability
Weights.α value is that the model structure for making the decorated style identify neutral net is taken by cross-validation experiments in training set
Recognition accuracy highest value.Checking obtains α=0.4 in present embodiment.
After final identification probability y is calculated, it is the vector of one group of 56 dimension, by corresponding to maximum probability value in vector
Shop is the shop recognition result as the picture.Further, obtain in picture after the recognition result in shop, market can be combined
Map searches location prediction point corresponding with being identified shop, and using these places as positioning result.Fig. 5 show experiment and surveyed
Test result figure, wherein left side (a) is classified as input picture, image upper right has real shop name label.Right side (b) is classified as output
Positioning result.Five-pointed star pattern represents output, and round dot represents physical location.Experiments verify that the framework of present embodiment
The shop recognition accuracy of lower progress can reach 86.39%.
To sum up, net is built by collecting a large amount of indoor scene images (shop image data set) in present embodiment
Network prototype, minimize cost function and train neural network model, so as to obtain decorated style neutral net and text detection god
Concrete model through network, and obtain the NGRAM codings of word with reference to Text region module.Finally by word and decorated style
Blend, so as to recognize shop is specifically which shop in market.The method of present embodiment can be obviously improved shop
Paving identification and the degree of accuracy of indoor mall positioning.
Above content is to combine specific preferred embodiment further description made for the present invention, it is impossible to is assert
The specific implementation of the present invention is confined to these explanations.For general technical staff of the technical field of the invention,
Some replacements or obvious modification are made on the premise of not departing from present inventive concept, and performance or purposes are identical, should all be considered as
Belong to protection scope of the present invention.
Claims (10)
- A kind of 1. indoor mall shop feature extracting method, it is characterised in that:Include M shop in the market;Including following Step:S1, structure shop decorated style identification neutral net prototype, the decorated style identification neutral net prototype include convolution Layer, pond layer, excitation layer and full articulamentum;S2, build shop text detection neutral net prototype, the text detection nerve net Network prototype includes convolutional layer, pond layer, excitation layer and warp lamination;S3, the shop image data set established in the market, each figure Text information and decorated style information as including shop, training set and test set are divided into by image data set;Pass through height This random distribution initializes the weights of the decorated style identification neutral net and text detection neutral net, to minimize cost Function is the text detection nerve net that target is come in decorated style identification neutral net prototype and step S2 in training step S1 Network prototype, determine the model structure of shop decorated style identification neutral net and text detection neutral net;S4, for one Picture to be identified comprising shop title and upholstery style, be inputted respectively the shop decorated style neutral net and In the model structure of text detection neutral net, obtain the picture and belong in the probability and the picture in each shop to correspond to word Partial region;S5, by the region input Text region module that word segment is corresponded in the picture obtained in step S4, obtain Into the picture, NGRAM corresponding to word segment is encoded;S6, known according to the probability calculation shop decorated style obtained in step S4 Other probability, Text region probability is calculated according to the NGRAM codings obtained in step S5, obtains finally corresponding to M shop using weighting The identification probability of paving, the shop recognition result of the picture is used as using shop corresponding to maximum probability value in obtained probability.
- 2. indoor mall shop according to claim 1 feature extracting method, it is characterised in that:In step S3, cost letter Number is:L=- ∑sitilog(yi)+λ||W||2, wherein, when decorated style identifies neutral net prototype in training step S1, yiTable Shop probability obtained by showing the image in the training set after decorated style identifies neural metwork training, tiFor the instruction Practice the actual probabilities that the image concentrated belongs to each shop in M shop, λ represents the regularization coefficient of cost function, and W represents institute State the weights of convolutional layer and full articulamentum in decorated style identification neural network model;Text detection god in shop in training step S2 During through web original, yiRepresent the image that the image in the training set exports after the text detection neural metwork training of shop In each pixel belong to the probability of character area, tiBelong to character area for each pixel in the image in the training set Actual probabilities, λ represents the regularization coefficient of cost function, W represent in the text detection neural network model convolutional layer and The weights of warp lamination.
- 3. indoor mall shop according to claim 1 feature extracting method, it is characterised in that:In step S3, pass through Adam algorithmic minimizing cost functions determine the model of shop decorated style identification neutral net and text detection neutral net Structure.
- 4. indoor mall shop according to claim 1 feature extracting method, it is characterised in that:In step S3, determine The model structure of decorated style identification neutral net include the first convolution unit, the individual full articulamentums of N2, wherein first Convolution unit includes the N1 convolutional layers being sequentially connected in series, connected after each convolutional layer a pond layer and a ReLu excitation Layer, the output end series connection N2 full articulamentum of first convolution unit, the output dimension of last full articulamentum are M, for exporting the binary set of M dimensions;Wherein, N1 and N2 value causes the model structure to reach highest standard on training set True rate.
- 5. indoor mall shop according to claim 1 feature extracting method, it is characterised in that:In step S3, determine The model structure of the text detection neutral net include the second convolution unit and N4 warp lamination;Second convolution Unit includes the N3 convolutional layers being sequentially connected in series, respectively connected after the convolutional layer of part a pond layer and excitation layer;The volume Two The output end of product unit is connected the N4 warp lamination, and last warp lamination is for exporting binaryzation picture, in picture It is worth and represents character area for 1 pixel, the pixel being worth for 0 represents non-legible region, wherein, N3 and N4 value cause institute State model structure and reach highest accuracy rate on training set.
- 6. indoor mall shop according to claim 1 feature extracting method, it is characterised in that:In step S6, according to such as Shop decorated style identification probability is calculated in minor function:ss=fs(a), wherein, a represents the decorated style identification nerve net The output of network, ssBelong to the probability in each shop, f for the picturesFor sigmoid functions.
- 7. indoor mall shop according to claim 1 feature extracting method, it is characterised in that:In step S6, according to such as Text region probability is calculated in minor function:st(j)=∑kGN(k) I (N (k) ∈ S (j)), wherein, S (j) is represented in M shop Character string corresponding to the trade name in j-th of shop, k represent the bits of coded in NGRAM coding schemes, are 0 ..., 9999, N (k) is represented Character string corresponding to k-th of bits of coded in NGRAM codings, N (k) ∈ S (j) represent that character string N (k) belongs to S (j), and I is the property shown letter Number, GN(k) the NGRAM codings of word recognized in input picture are represented.
- 8. indoor mall shop according to claim 1 feature extracting method, it is characterised in that:In step S6, according to such as Final identification probability is calculated in lower formula:Y=(1- α) st+αss, wherein, stRepresent Text region probability, ssRepresent decoration Style identification probability;α represents the weights of decorated style identification probability.
- 9. indoor mall shop according to claim 8 feature extracting method, it is characterised in that:α value is to pass through intersection Confirmatory experiment takes the model structure for making the decorated style identify neutral net recognition accuracy highest value in training set.
- 10. indoor mall shop according to claim 1 feature extracting method, it is characterised in that:It is further comprising the steps of: S7, the shop recognition result obtained according to step S6, with reference to market map Search and Orientation future position, provide positioning result.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201711167137.6A CN107832727A (en) | 2017-11-21 | 2017-11-21 | A kind of indoor mall shop feature extracting method |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201711167137.6A CN107832727A (en) | 2017-11-21 | 2017-11-21 | A kind of indoor mall shop feature extracting method |
Publications (1)
Publication Number | Publication Date |
---|---|
CN107832727A true CN107832727A (en) | 2018-03-23 |
Family
ID=61652062
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201711167137.6A Pending CN107832727A (en) | 2017-11-21 | 2017-11-21 | A kind of indoor mall shop feature extracting method |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN107832727A (en) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109003133A (en) * | 2018-07-20 | 2018-12-14 | 阿里巴巴集团控股有限公司 | The method and device that shop identifies under a kind of line |
CN109089314B (en) * | 2018-09-30 | 2020-10-02 | 哈尔滨工业大学(深圳) | Indoor positioning method of wifi sequence assisted GPS based on recommendation algorithm |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106557768A (en) * | 2016-11-25 | 2017-04-05 | 北京小米移动软件有限公司 | The method and device is identified by word in picture |
CN106874924A (en) * | 2015-12-14 | 2017-06-20 | 阿里巴巴集团控股有限公司 | A kind of recognition methods of picture style and device |
CN106874296A (en) * | 2015-12-14 | 2017-06-20 | 阿里巴巴集团控股有限公司 | A kind of style recognition methods of commodity and device |
CN107103285A (en) * | 2017-03-24 | 2017-08-29 | 深圳市未来媒体技术研究院 | Face depth prediction approach based on convolutional neural networks |
-
2017
- 2017-11-21 CN CN201711167137.6A patent/CN107832727A/en active Pending
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106874924A (en) * | 2015-12-14 | 2017-06-20 | 阿里巴巴集团控股有限公司 | A kind of recognition methods of picture style and device |
CN106874296A (en) * | 2015-12-14 | 2017-06-20 | 阿里巴巴集团控股有限公司 | A kind of style recognition methods of commodity and device |
CN106557768A (en) * | 2016-11-25 | 2017-04-05 | 北京小米移动软件有限公司 | The method and device is identified by word in picture |
CN107103285A (en) * | 2017-03-24 | 2017-08-29 | 深圳市未来媒体技术研究院 | Face depth prediction approach based on convolutional neural networks |
Non-Patent Citations (6)
Title |
---|
CEES G.M. SNOEK 等: "Early versus Late Fusion in Semantic Video Analysis", 《MM"05》 * |
MAX JADERBERG 等: "Deep Structured Output Learning for Unconstrained Text Recognition", 《ICLR 2015》 * |
MAX JADERBERG 等: "Synthetic Data and Artificial Neural Networks for Natural Scene Text Recognition", 《ARXIV》 * |
SHENLONG WANG 等: "Lost Shopping! Monocular Localization in Large Indoor Spaces", 《2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION》 * |
ZIWEI XU 等: "UTILIZING HIGH-LEVEL VISUAL FEATURE FOR INDOOR SHOPPING MALL NAVIGATION", 《ARXIV》 * |
董海鹰著: "《智能控制理论及应用》", 30 September 2016, 北京:中国铁道出版社 * |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109003133A (en) * | 2018-07-20 | 2018-12-14 | 阿里巴巴集团控股有限公司 | The method and device that shop identifies under a kind of line |
CN109003133B (en) * | 2018-07-20 | 2022-10-14 | 创新先进技术有限公司 | Off-line store identification method and device |
CN109089314B (en) * | 2018-09-30 | 2020-10-02 | 哈尔滨工业大学(深圳) | Indoor positioning method of wifi sequence assisted GPS based on recommendation algorithm |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN110298361B (en) | Semantic segmentation method and system for RGB-D image | |
CN109584248B (en) | Infrared target instance segmentation method based on feature fusion and dense connection network | |
CN109948425B (en) | Pedestrian searching method and device for structure-aware self-attention and online instance aggregation matching | |
Li et al. | Building-a-nets: Robust building extraction from high-resolution remote sensing images with adversarial networks | |
CN105205453B (en) | Human eye detection and localization method based on depth self-encoding encoder | |
CN104462494B (en) | A kind of remote sensing image retrieval method and system based on unsupervised feature learning | |
CN107103277B (en) | Gait recognition method based on depth camera and 3D convolutional neural network | |
CN107871101A (en) | A kind of method for detecting human face and device | |
CN106022363B (en) | A kind of Chinese text recognition methods suitable under natural scene | |
CN107808129A (en) | A kind of facial multi-characteristic points localization method based on single convolutional neural networks | |
CN105956560A (en) | Vehicle model identification method based on pooling multi-scale depth convolution characteristics | |
CN110705566B (en) | Multi-mode fusion significance detection method based on spatial pyramid pool | |
CN111563418A (en) | Asymmetric multi-mode fusion significance detection method based on attention mechanism | |
CN109522883A (en) | A kind of method for detecting human face, system, device and storage medium | |
CN109784288B (en) | Pedestrian re-identification method based on discrimination perception fusion | |
CN105574848A (en) | A method and an apparatus for automatic segmentation of an object | |
CN114155527A (en) | Scene text recognition method and device | |
CN104298974A (en) | Human body behavior recognition method based on depth video sequence | |
CN110163208A (en) | A kind of scene character detecting method and system based on deep learning | |
CN109919992A (en) | Estimate the method for the depth in image | |
CN106407978B (en) | Method for detecting salient object in unconstrained video by combining similarity degree | |
CN110399882A (en) | A kind of character detecting method based on deformable convolutional neural networks | |
CN110223310A (en) | A kind of line-structured light center line and cabinet edge detection method based on deep learning | |
Zhang et al. | Semantic segmentation of very high-resolution remote sensing image based on multiple band combinations and patchwise scene analysis | |
CN108363962A (en) | A kind of method for detecting human face and system based on multi-level features deep learning |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20180323 |