CN105898219B - Object monitor method and device - Google Patents

Object monitor method and device Download PDF

Info

Publication number
CN105898219B
CN105898219B CN201610258363.4A CN201610258363A CN105898219B CN 105898219 B CN105898219 B CN 105898219B CN 201610258363 A CN201610258363 A CN 201610258363A CN 105898219 B CN105898219 B CN 105898219B
Authority
CN
China
Prior art keywords
target sound
sound
sounding
information
mobile terminal
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201610258363.4A
Other languages
Chinese (zh)
Other versions
CN105898219A (en
Inventor
贾伟光
马娜
侯恩星
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Xiaomi Mobile Software Co Ltd
Original Assignee
Beijing Xiaomi Mobile Software Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Xiaomi Mobile Software Co Ltd filed Critical Beijing Xiaomi Mobile Software Co Ltd
Priority to CN201610258363.4A priority Critical patent/CN105898219B/en
Priority to PCT/CN2016/091863 priority patent/WO2017181545A1/en
Publication of CN105898219A publication Critical patent/CN105898219A/en
Priority to EP16197065.2A priority patent/EP3236469B1/en
Priority to US15/378,059 priority patent/US10122916B2/en
Application granted granted Critical
Publication of CN105898219B publication Critical patent/CN105898219B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/64Computer-aided capture of images, e.g. transfer from script file into camera, check of taken image quality, advice or proposal for image composition or decision on when to take image
    • GPHYSICS
    • G08SIGNALLING
    • G08BSIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
    • G08B13/00Burglar, theft or intruder alarms
    • G08B13/18Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength
    • G08B13/189Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems
    • G08B13/194Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems
    • G08B13/196Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems using television cameras
    • G08B13/19663Surveillance related processing done local to the camera
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/18Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
    • GPHYSICS
    • G08SIGNALLING
    • G08BSIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
    • G08B13/00Burglar, theft or intruder alarms
    • G08B13/18Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength
    • G08B13/189Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems
    • G08B13/194Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems
    • G08B13/196Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems using television cameras
    • G08B13/19678User interface
    • G08B13/19682Graphic User Interface [GUI] presenting system data to the user, e.g. information on a screen helping a user interacting with an alarm system
    • GPHYSICS
    • G08SIGNALLING
    • G08BSIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
    • G08B13/00Burglar, theft or intruder alarms
    • G08B13/18Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength
    • G08B13/189Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems
    • G08B13/194Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems
    • G08B13/196Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems using television cameras
    • G08B13/19678User interface
    • G08B13/19684Portable terminal, e.g. mobile phone, used for viewing video remotely
    • GPHYSICS
    • G08SIGNALLING
    • G08BSIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
    • G08B13/00Burglar, theft or intruder alarms
    • G08B13/18Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength
    • G08B13/189Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems
    • G08B13/194Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems
    • G08B13/196Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems using television cameras
    • G08B13/19678User interface
    • G08B13/19689Remote control of cameras, e.g. remote orientation or image zooming control for a PTZ camera
    • GPHYSICS
    • G08SIGNALLING
    • G08BSIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
    • G08B13/00Burglar, theft or intruder alarms
    • G08B13/18Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength
    • G08B13/189Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems
    • G08B13/194Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems
    • G08B13/196Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems using television cameras
    • G08B13/19697Arrangements wherein non-video detectors generate an alarm themselves
    • GPHYSICS
    • G08SIGNALLING
    • G08BSIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
    • G08B21/00Alarms responsive to a single specified undesired or abnormal condition and not otherwise provided for
    • G08B21/02Alarms for ensuring the safety of persons
    • G08B21/0202Child monitoring systems using a transmitter-receiver system carried by the parent and the child
    • G08B21/0205Specific application combined with child monitoring using a transmitter-receiver system
    • G08B21/0208Combination with audio or video communication, e.g. combination with "baby phone" function
    • GPHYSICS
    • G08SIGNALLING
    • G08BSIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
    • G08B25/00Alarm systems in which the location of the alarm condition is signalled to a central station, e.g. fire or police telegraphic systems
    • G08B25/01Alarm systems in which the location of the alarm condition is signalled to a central station, e.g. fire or police telegraphic systems characterised by the transmission medium
    • G08B25/08Alarm systems in which the location of the alarm condition is signalled to a central station, e.g. fire or police telegraphic systems characterised by the transmission medium using communication transmission lines
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/02Feature extraction for speech recognition; Selection of recognition unit
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/06Creation of reference templates; Training of speech recognition systems, e.g. adaptation to the characteristics of the speaker's voice
    • G10L15/063Training
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/08Speech classification or search
    • G10L15/10Speech classification or search using distance or distortion measures between unknown speech and reference templates
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/48Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use
    • G10L25/51Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/63Control of cameras or camera modules by using electronic viewfinders
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/66Remote control of cameras or camera parts, e.g. by remote control devices
    • H04N23/661Transmitting camera control signals through networks, e.g. control via the Internet
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/18Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
    • H04N7/188Capturing isolated or intermittent images triggered by the occurrence of a predetermined event, e.g. an object reaching a predetermined position
    • HELECTRICITY
    • H05ELECTRIC TECHNIQUES NOT OTHERWISE PROVIDED FOR
    • H05KPRINTED CIRCUITS; CASINGS OR CONSTRUCTIONAL DETAILS OF ELECTRIC APPARATUS; MANUFACTURE OF ASSEMBLAGES OF ELECTRICAL COMPONENTS
    • H05K999/00PRINTED CIRCUITS; CASINGS OR CONSTRUCTIONAL DETAILS OF ELECTRIC APPARATUS; MANUFACTURE OF ASSEMBLAGES OF ELECTRICAL COMPONENTS dummy group
    • H05K999/99PRINTED CIRCUITS; CASINGS OR CONSTRUCTIONAL DETAILS OF ELECTRIC APPARATUS; MANUFACTURE OF ASSEMBLAGES OF ELECTRICAL COMPONENTS dummy group dummy group

Abstract

The disclosure is directed to a kind of object monitor method and devices, wherein object monitor method includes: when there are the sounding reasons for when target sound, judging target sound in collected sound;Sounding information is pushed to mobile terminal, which carries sounding reason.The embodiment of the present disclosure by judging the sounding reason of collected target sound, and pushes the sounding information for carrying sounding reason to mobile terminal, user is allowed to obtain the sounding reason of monitored object, reaches monitoring purpose well.

Description

Object monitor method and device
Technical field
This disclosure relates to mobile internet technical field more particularly to a kind of object monitor method and device.
Background technique
With the development of photography technology, there is intelligent camera, intelligent camera be using current newest science and technology at Fruit, most smart point manufacturing process, the video camera that the invaluable experience of comprehensive many years is succeeded in developing.
With networking, digitlization, the extensive use of Gao Qinghua, people are increasing to intelligentized demand, intelligently take the photograph The application of camera is more and more wider, and various characteristics are allowed to the monitoring suitable for all trades and professions moving target.For example, can be to intelligence Mansion, urban road, power department, airport and station etc. are monitored.
It can be seen that intelligent camera can be used for monitoring various objects, therefore, how to be preferably using intelligent camera User provides monitoring service, is one of current technical problem urgently to be solved.
Summary of the invention
To overcome the problems in correlation technique, the disclosure provides a kind of object monitor method and device.
According to the first aspect of the embodiments of the present disclosure, a kind of object monitor method is provided, comprising:
When there are the sounding reasons for when target sound, judging the target sound in collected sound;
Sounding information is pushed to mobile terminal, the sounding information carries the sounding reason.
In one embodiment, after the push sounding information to mobile terminal, the method also includes:
Receive the video acquisition instruction that the mobile terminal is sent;
It is instructed according to the video acquisition to the mobile terminal pushing video information, so that the mobile terminal shows institute State video information.
In one embodiment, it is described according to the video acquisition instruct to the mobile terminal pushing video information it Before, the method also includes:
Obtain the sounding position of the target sound;
It controls camera and turns to the corresponding direction in the sounding position, and acquire the video information.
In one embodiment, the sounding reason of the judgement target sound, comprising:
Feature extraction is carried out to the target sound, obtains target sound characteristic information;
The target sound characteristic information is matched with the multiple sound patterns prestored, and is obtained according to matching result The sounding reason of the target sound, the multiple sound pattern target sound caused by different reasons for identification.
In one embodiment, described that feature extraction is carried out to the target sound, target sound characteristic information is obtained, is wrapped It includes:
Compression processing is carried out to the target sound, obtains voice compressing information;
Feature extraction is carried out to the voice compressing information, obtains the target sound characteristic information.
In one embodiment, the sounding reason of the judgement target sound, comprising:
The target sound is sent to server;
Receive the sounding reason for the target sound that the server returns.
In one embodiment, the method also includes:
If detecting, the current monitor equipment does not access network, broadcasting packet, so that the mobile terminal shows institute State the prompt information that current monitor equipment does not access network;
The title and password of the router of the mobile terminal broadcast are received, and sends connection request to the router, To be connected on the router.
In one embodiment, the method also includes:
For each sound pattern, current sound template is trained.
It is in one embodiment, described that current sound template is trained, comprising:
Acquire the target sound feature samples of preset quantity;
The target sound feature samples of the preset quantity are inputted into current sound template, obtain the mesh of the preset quantity Distortion distance of the sound characteristic sample relative to current sound template is marked, the distortion distance is for indicating the preset quantity Gap between target sound feature samples and the target sound for passing through current sound template output;
If the distortion distance is less than pre-determined distance, stopping is trained the current sound template;
If the distortion distance is more than or equal to pre-determined distance, continue to be trained the current sound template, until The distortion distance is less than pre-determined distance.
In one embodiment, the sounding information also carries the target sound.
According to the second aspect of an embodiment of the present disclosure, a kind of object monitor device is provided, comprising:
Judgment module is configured as when, there are when target sound, judging the hair of the target sound in collected sound Sound reason;
First pushing module is configured as pushing sounding information to mobile terminal, and the sounding information, which carries, described to be sentenced The sounding reason that disconnected module is judged.
In one embodiment, described device further include:
Receiving module is configured as after first pushing module pushes sounding information to mobile terminal, receives institute State the video acquisition instruction of mobile terminal transmission;
Second pushing module is configured as being instructed according to the received video acquisition of the receiving module to the movement Terminal pushing video information, so that the mobile terminal shows the video information.
In one embodiment, described device further include:
Module is obtained, is configured as being instructed according to the video acquisition to the mobile terminal in second pushing module Before pushing video information, the sounding position of the target sound is obtained;
Acquisition module is turned to, being configured as control camera, to turn to the sounding position that the acquisition module obtains corresponding Direction, and acquire the video information.
In one embodiment, the judgment module includes:
Extracting sub-module is configured as carrying out feature extraction to the target sound, obtains target sound characteristic information;
Matched sub-block, is configured as the target sound characteristic information for extracting the extracting sub-module and prestores Multiple sound patterns matched, and the sounding reason of the target sound, the multiple sound are obtained according to matching result Template target sound caused by different reasons for identification.
In one embodiment, the extracting sub-module includes:
Compression unit is configured as carrying out compression processing to the target sound, obtains voice compressing information;
Extraction unit is configured as the voice compressing information obtained to the compression unit and carries out feature extraction, obtains To the target sound characteristic information.
In one embodiment, the judgment module includes:
Sending submodule is configured as sending the target sound to server;
Receiving submodule is configured as receiving the sounding reason for the target sound that the server returns.
In one embodiment, described device further include:
Broadcast module is detected, if being configured as detecting, the current monitor equipment does not access network, broadcasting packet, with The mobile terminal is set to show that the current monitor equipment does not access the prompt information of network;
Link block is received, is configured as receiving the title and password of the router of the mobile terminal broadcast, and to institute It states router and sends connection request, to be connected on the router.
In one embodiment, described device further include:
Training module is configured as being trained current sound template for each sound pattern.
In one embodiment, the training module includes:
Submodule is acquired, the target sound feature samples of acquisition preset quantity are configured as;
Input submodule is configured as the target sound feature of the acquisition collected preset quantity of submodule Sample inputs current sound template, obtains mistake of the target sound feature samples of the preset quantity relative to current sound template True distance, the distortion distance are used to indicate the target sound feature samples of the preset quantity and by the current sound moulds Gap between the target sound of plate output;
First training submodule, if be configured as the distortion distance that the input submodule obtains be less than it is default away from From then stopping is trained the current sound template;
Second training submodule is preset if being configured as the distortion distance that the input submodule obtains and being more than or equal to Distance then continues that the training module is called to be trained the current sound template, until the distortion distance is less than in advance If distance.
In one embodiment, the sounding information also carries the target sound.
According to the third aspect of an embodiment of the present disclosure, a kind of object monitor device is provided, comprising:
Processor;
Memory for storage processor executable instruction;
Wherein, processor is configured as:
When there are the sounding reasons for when target sound, judging the target sound in collected sound;
Sounding information is pushed to mobile terminal, the sounding information carries the sounding reason.
The technical scheme provided by this disclosed embodiment can include the following benefits: by judging collected target The sounding reason of sound, and the sounding information for carrying sounding reason is pushed to mobile terminal, user is obtained and is supervised The sounding reason for controlling object reaches monitoring purpose well.
By allowing mobile terminal to show the video information to mobile terminal pushing video information, to facilitate use The video information is remotely checked at family.
Video information is acquired by obtaining the sounding position of target sound, implementation is simple.
Pass through the target sound characteristic information that will be extracted and the trained multiple sound patterns progress prestored Match, obtain the sounding reason of target sound, implementation is simple, and obtained sounding reason accuracy rate is high.
By to target sound carry out compression processing, reach reduction system operation and storage burden and removal target sound The purpose of sound correlation.
It also may determine that the sounding reason of target sound by interacting with server, it can be seen that, it can be by more Kind mode judges the sounding reason of target sound.
By interacting with mobile terminal with couple in router, so that establishing binding relationship for the two provides condition, into It and is subsequent to be monitored the condition of offer.
By being trained to sound pattern, to judge that the sounding reason of target sound provides condition.
By the way that the target sound feature samples of preset quantity are inputted current sound template, distortion distance is obtained, and will lose True distance is compared with pre-determined distance, and to complete the training process of sound pattern, implementation is simple, the sound mould trained Plate can accurately identify the sounding reason of target sound.
Target sound is also carried by describing sounding information, illustrates the multiplicity of the sounding information pushed to mobile terminal Property.
It should be understood that above general description and following detailed description be only it is exemplary and explanatory, not The disclosure can be limited.
Detailed description of the invention
The drawings herein are incorporated into the specification and forms part of this specification, and shows and meets implementation of the invention Example, and be used to explain the principle of the present invention together with specification.
Fig. 1 is a kind of flow chart of object monitor method shown according to an exemplary embodiment.
Fig. 2 is the flow chart of another object monitor method shown according to an exemplary embodiment.
Fig. 3 is a kind of scene figure of object monitor method shown according to an exemplary embodiment.
Fig. 4 is a kind of flow chart of sounding reason for judging target sound shown according to an exemplary embodiment.
Fig. 5 is the flow chart that a kind of pair of current sound template shown according to an exemplary embodiment is trained.
Fig. 6 is a kind of block diagram of object monitor device shown according to an exemplary embodiment.
Fig. 7 is the block diagram of another object monitor device shown according to an exemplary embodiment.
Fig. 8 is the block diagram of another object monitor device shown according to an exemplary embodiment.
Fig. 9 is the block diagram of another object monitor device shown according to an exemplary embodiment.
Figure 10 is the block diagram of another object monitor device shown according to an exemplary embodiment.
Figure 11 is the block diagram of another object monitor device shown according to an exemplary embodiment.
Figure 12 is the block diagram of another object monitor device shown according to an exemplary embodiment.
Figure 13 is the block diagram of another object monitor device shown according to an exemplary embodiment.
Figure 14 is the block diagram of another object monitor device shown according to an exemplary embodiment.
Figure 15 is a kind of block diagram suitable for object monitor device shown according to an exemplary embodiment.
Specific embodiment
Example embodiments are described in detail here, and the example is illustrated in the accompanying drawings.Following description is related to When attached drawing, unless otherwise indicated, the same numbers in different drawings indicate the same or similar elements.Following exemplary embodiment Described in embodiment do not represent all embodiments consistented with the present invention.On the contrary, they be only with it is such as appended The example of device and method being described in detail in claims, some aspects of the invention are consistent.
Fig. 1 is a kind of flow chart of object monitor method shown according to an exemplary embodiment, as shown in Figure 1, this is right As monitoring method can be applied in monitoring device, which may include but is not limited to intelligent camera, this method packet Include following steps S101-S102:
In step s101, when there are the sounding reasons for when target sound, judging the target sound in collected sound.
Wherein it is possible to acquire the sound of monitored object by the microphone being built in video camera, monitored object can To include but is not limited to children, old man etc., target sound can refer to crying etc..
In this embodiment, intelligent camera, can be special by extracting crying after collecting target sound such as crying Reference breath, and the crying characteristic information of extraction is matched with trained multiple sound patterns to obtain the hair of crying Sound reason.
In addition, intelligent camera can also by with server interact to obtain crying sounding reason.For example, intelligence Energy video camera can send target sound to server, and server can also pass through after receiving target sound extracts crying spy Reference breath, and the crying characteristic information of extraction is matched with trained multiple sound patterns to obtain the hair of crying Then sound reason returns to the sounding reason of crying to intelligent camera.
In step s 102, sounding information is pushed to mobile terminal, sounding information carries sounding reason.
In this embodiment, it after judging the sounding reason of target sound, can be pushed to mobile terminal such as mobile phone Carry the sounding information of sounding reason, with prompt user children cry the reason of.
In addition, target sound can also be carried in the sounding information, for example, can be pushed to mobile phone children crying and The reason of sobbing.
Above-mentioned object monitor embodiment of the method, by judging the sounding reason of collected target sound, and eventually to movement End push carries the sounding information of sounding reason, allows user to obtain the sounding reason of monitored object, reaches fine Monitoring purpose.
Fig. 2 is the flow chart of another object monitor method shown according to an exemplary embodiment, as shown in Fig. 2, After above-mentioned steps S102, this method can also include the following steps:
In step s 103, the video acquisition instruction that mobile terminal is sent is received.
In this embodiment, current monitor equipment be for example built-in with the intelligent camera of Wireless Fidelity (WIFI) chip to After mobile terminal pushes sounding information, if the user desired that checking the video information of monitored object, such as the video of children Information then can send video acquisition instruction to intelligent camera by the application program (APP) on mobile terminal.
In this embodiment, intelligent camera and above-mentioned APP can establish binding relationship by same user account.And The two is established before binding relationship, and intelligent camera can be linked on same router with the mobile terminal, intelligent camera The process of couple in router can be with are as follows:
Intelligent camera, which detects, oneself does not access network, then broadcasting packet, so that mobile terminal shows intelligent camera The prompt information of network is not accessed, and user can broadcast by mobile terminal to intelligent camera after seeing the prompt information The title and password of the router oneself connected, intelligent camera is after the title and password for receiving router, Ke Yixiang The router sends connection request, to be connected on the router.
It is located in same local area network since intelligent camera and mobile terminal are connected to same router, APP can To obtain the binding state of intelligent camera by scanning, and show the binding state of intelligent camera.Wherein, binding state packet It includes and has bound and unbound two states.
When the binding state of intelligent camera is unbound, if user wants to control the intelligent camera, can touch Bindings are sent out, then the APP sends the mark and user account such as millet account for carrying intelligent camera to server Request, in this way, server can bind intelligent camera and user account according to the request.
After the intelligent camera and mobile terminal APP are bound by same user account, user can pass through the movement Terminal APP sends control instruction such as video acquisition to intelligent camera and instructs.
It in step S104, is instructed according to the video acquisition to mobile terminal pushing video information, so that mobile terminal is aobvious Show video information.
In this embodiment, the sounding position of the available target sound of intelligent camera, and control camera and turn to hair The corresponding direction in sound position, then acquires video information.The video information acquisition mode is simple, it is easy to accomplish.
After intelligent camera is collected video information, if receiving the video acquisition instruction from mobile terminal, It can be to the mobile terminal pushing video information, to facilitate user remotely to check the video information of monitored object such as children.
The disclosure is illustrated below with reference to Fig. 3, as shown in figure 3, intelligent camera 31 passes through built-in wheat Gram wind collects in sound there are after crying, by extracting crying characteristic information, and by the crying characteristic information of extraction with Trained multiple sound patterns are matched to obtain the sounding reason of crying.Since intelligent camera 31 and mobile phone 32 connect It is located in same local area network to same router 33, therefore, intelligent camera 31 can pass through with the APP being located in mobile phone 32 Active user's account establishes binding relationship, therefore intelligent camera 31 can push sounding information to mobile phone 32 by server 34, The sounding information carries the sounding reason of crying, if user wishes to check the video information that intelligent camera 31 is shot at this time, Then video acquisition instruction can be sent to intelligent camera 31 by mobile phone 32, intelligent camera 31 is receiving the video acquisition After instruction, the video information obtained can be pushed to mobile phone 32, user is allowed remotely to check video information.
Above-described embodiment, by allowing mobile terminal to show the video information to mobile terminal pushing video information, To facilitate user remotely to check the video information.
Fig. 4 is a kind of flow chart of sounding reason for judging target sound shown according to an exemplary embodiment, the reality Apply example to be described from monitoring device such as intelligent camera pusher side, can also be described from server side, i.e., monitoring device and Server can be judged by the following manner the sounding reason of target sound, as shown in figure 4, this method comprises the following steps:
In step S401, for each sound pattern, current sound template is trained.
In this embodiment, it is necessary first to each sound pattern in multiple sound patterns is trained, with reach by After target sound inputs sound pattern, sounding reason can be identified.Wherein, multiple sound patterns for identification draw by different reasons The target sound risen.
Wherein, as shown in figure 5, the process being trained to current sound template may include steps of:
In step S4011, the target sound feature samples of preset quantity are acquired.
In step S4012, the target sound feature samples of preset quantity are inputted into current sound template, obtain present count Distortion distance of the target sound feature samples of amount relative to current sound template.
Distortion distance is used to indicate the target sound feature samples of preset quantity and the mesh exported by current sound template Mark the gap between sound.
In step S4013, judge whether distortion distance is less than pre-determined distance, if distortion distance is less than pre-determined distance, Stopping is trained current sound template, if distortion distance is more than or equal to pre-determined distance, turns to and executes step S4011.
Wherein, distortion distance is smaller, then shows the target sound feature samples of preset quantity and by current sound template Gap between the target sound of output is smaller.
When distortion distance is less than pre-determined distance, then otherwise training of the completion to current sound template continues to train, until Distortion distance is less than pre-determined distance.
In step S402, feature extraction is carried out to target sound, obtains target sound characteristic information.
In this embodiment, if target sound such as crying data volume is bigger, first target sound can be pressed Contracting processing, to reduce the operation and storage burden of system, in addition, compressed target sound can remove the correlation of target sound Property, such as the correlation of removal crying, to achieve the purpose that express crying signal with less characteristic information.Certainly, if mesh Mark voice data amount is little and has removed correlation, then does not need to carry out compression processing.Wherein it is possible to using existing compression skill Art such as lossless compressiong carries out compression processing to target sound.
In this embodiment, to target sound carry out compression processing can be to compression after obtaining voice compressing information Acoustic information carries out feature extraction, obtains target sound characteristic information.
In step S403, target sound characteristic information is matched with the multiple sound patterns prestored, and according to The sounding reason of target sound is obtained with result.
In this embodiment it is possible to by the target sound characteristic information extracted and the trained multiple sound prestored Sound template is matched, to obtain the sounding reason of target sound.
Above-described embodiment passes through the target sound characteristic information that will be extracted and the trained multiple sound prestored Template is matched, and the sounding reason of target sound is obtained, and implementation is simple, and obtained sounding reason accuracy rate is high.
Corresponding with aforementioned object monitoring method embodiment, the disclosure additionally provides object monitor Installation practice.
Fig. 6 is a kind of block diagram of object monitor device shown according to an exemplary embodiment, as shown in fig. 6, the object Monitoring device includes: judgment module 61 and the first pushing module 62.
Judgment module 61 is configured as when there are the sounding originals for when target sound, judging target sound in collected sound Cause.
Wherein, judgment module 61 can acquire the sound of monitored object, quilt by the microphone being built in video camera Monitored object may include but be not limited to children, old man etc., and target sound can refer to crying etc..
In this embodiment, intelligent camera, can be special by extracting crying after collecting target sound such as crying Reference breath, and the crying characteristic information of extraction is matched with trained multiple sound patterns to obtain the hair of crying Sound reason.
In addition, intelligent camera can also by with server interact to obtain crying sounding reason.For example, intelligence Energy video camera can send target sound to server, and server can also pass through after receiving target sound extracts crying spy Reference breath, and the crying characteristic information of extraction is matched with trained multiple sound patterns to obtain the hair of crying Then sound reason returns to the sounding reason of crying to intelligent camera.
First pushing module 62 is configured as pushing sounding information to mobile terminal, and sounding information carries judgment module 61 The sounding reason judged.
In this embodiment, it after judging the sounding reason of target sound, can be pushed to mobile terminal such as mobile phone Carry the sounding information of sounding reason, with prompt user children cry the reason of.
In addition, target sound can also be carried in the sounding information, for example, can be pushed to mobile phone children crying and The reason of sobbing.
For device as shown in FIG. 6 for realizing above-mentioned method flow as shown in Figure 1, the related content being related to describes phase Together, it does not repeat herein.
Above-described embodiment is carried by judging the sounding reason of collected target sound, and to mobile terminal push The sounding information of sounding reason allows user to obtain the sounding reason of monitored object, reaches monitoring purpose well.
Fig. 7 is the block diagram of another object monitor device shown according to an exemplary embodiment, as shown in fig. 7, upper On the basis of stating embodiment illustrated in fig. 6, which may also include that receiving module 63 and the second pushing module 64.
Receiving module 63 is configured as after the first pushing module 62 pushes sounding information to mobile terminal, is received and is moved The video acquisition instruction that dynamic terminal is sent.
In this embodiment, current monitor equipment be for example built-in with the intelligent camera of Wireless Fidelity (WIFI) chip to After mobile terminal pushes sounding information, if the user desired that checking the video information of monitored object, such as the video of children Information then can send video acquisition instruction to intelligent camera by the application program (APP) on mobile terminal.
After the intelligent camera and mobile terminal APP are bound by same user account, user can pass through the movement Terminal APP sends control instruction such as video acquisition to intelligent camera and instructs.
Second pushing module 64 is configured as being instructed according to the received video acquisition of receiving module 63 to mobile terminal push Video information, so that mobile terminal shows video information.
For device as shown in Figure 7 for realizing above-mentioned method flow as shown in Figure 2, the related content being related to describes phase Together, it does not repeat herein.
Above-described embodiment, by allowing mobile terminal to show the video information to mobile terminal pushing video information, To facilitate user remotely to check the video information.
Fig. 8 is the block diagram of another object monitor device shown according to an exemplary embodiment, as shown in figure 8, upper On the basis of stating embodiment illustrated in fig. 7, which can also include: to obtain module 65 and steering acquisition module 66.
Module 65 is obtained to be configured as being instructed according to video acquisition to mobile terminal pushing video in the second pushing module 64 Before information, the sounding position of target sound is obtained.
It turns to acquisition module 66 and is configured as the corresponding side in sounding position that control camera turns to the acquisition acquisition of module 65 To, and acquire video information.
In this embodiment, the sounding position of the available target sound of intelligent camera, and control camera and turn to hair The corresponding direction in sound position, then acquires video information.The video information acquisition mode is simple, it is easy to accomplish.
For device as shown in Figure 8 for realizing above-mentioned method flow as shown in Figure 2, the related content being related to describes phase Together, it does not repeat herein.
Above-described embodiment acquires video information by obtaining the sounding position of target sound, and implementation is simple.
Fig. 9 is the block diagram of another object monitor device shown according to an exemplary embodiment, as shown in figure 9, upper On the basis of stating embodiment illustrated in fig. 6, judgment module 61 can include: extracting sub-module 611 and matched sub-block 612.
Extracting sub-module 611 is configured as carrying out feature extraction to target sound, obtains target sound characteristic information.
Matched sub-block 612 be configured as by target sound characteristic information that extracting sub-module 611 is extracted with prestore Multiple sound patterns are matched, and obtain the sounding reason of target sound according to matching result, and multiple sound patterns are for knowing Target sound caused by not different reasons.
In this embodiment it is possible to by the target sound characteristic information extracted and the trained multiple sound prestored Sound template is matched, to obtain the sounding reason of target sound.
For device as shown in Figure 9 for realizing above-mentioned method flow as shown in Figure 4, the related content being related to describes phase Together, it does not repeat herein.
Above-described embodiment passes through the target sound characteristic information that will be extracted and the trained multiple sound prestored Template is matched, and the sounding reason of target sound is obtained, and implementation is simple, and obtained sounding reason accuracy rate is high.
Figure 10 is the block diagram of another object monitor device shown according to an exemplary embodiment, as shown in Figure 10, On the basis of above-mentioned embodiment illustrated in fig. 9, extracting sub-module 611 can include: compression unit 6111 and extraction unit 6112.
Compression unit 6111 is configured as carrying out compression processing to target sound, obtains voice compressing information.
Extraction unit 6112 is configured as the voice compressing information obtained to compression unit 6111 and carries out feature extraction, obtains Target sound characteristic information.
For realizing above-mentioned method flow as shown in Figure 4, the related content being related to is described device as shown in Figure 10 It is identical, it does not repeat herein.
In this embodiment, if target sound such as crying data volume is bigger, first target sound can be pressed Contracting processing, to reduce the operation and storage burden of system, in addition, compressed target sound can remove the correlation of target sound Property, such as the correlation of removal crying, to achieve the purpose that express crying signal with less characteristic information.Certainly, if mesh Mark voice data amount is little and has removed correlation, then does not need to carry out compression processing.Wherein it is possible to using existing compression skill Art such as lossless compressiong carries out compression processing to target sound.
Above-described embodiment, by target sound carry out compression processing, reach operation and the storage burden of reduction system with And the purpose of removal target sound correlation.
Figure 11 is the block diagram of another object monitor device shown according to an exemplary embodiment, as shown in figure 11, On the basis of above-mentioned embodiment illustrated in fig. 6, judgment module 61 can include: sending submodule 613 and receiving submodule 614.
Sending submodule 613 is configured as sending target sound to server.
Receiving submodule 614 is configured as receiving the sounding reason for the target sound that server returns.
Object monitor device such as intelligent camera can be former by the sounding for interacting to obtain crying with server Cause.For example, intelligent camera can send target sound to server, server can also pass through after receiving target sound Crying characteristic information is extracted, and the crying characteristic information of extraction is matched to come with trained multiple sound patterns To the sounding reason of crying, the sounding reason of crying is then returned to intelligent camera.
For realizing above-mentioned method flow as shown in Figure 1, the related content being related to is described device as shown in figure 11 It is identical, it does not repeat herein.
Above-described embodiment also may determine that the sounding reason of target sound by interacting with server, it can be seen that, The sounding reason of target sound can be judged in several ways.
Figure 12 is the block diagram of another object monitor device shown according to an exemplary embodiment, as shown in figure 12, On the basis of above-mentioned embodiment illustrated in fig. 7, which may also include that detection broadcast module 67 and receives link block 68.
If detection broadcast module 67 is configured as detecting current monitor equipment and does not access network, broadcasting packet, so that Mobile terminal shows that current monitor equipment does not access the prompt information of network.
It receives link block 68 to be configured as receiving the title and password of the router of mobile terminal broadcast, and to router Connection request is sent, to be connected on router.
In this embodiment, intelligent camera and above-mentioned APP can establish binding relationship by same user account.And The two is established before binding relationship, and intelligent camera can be linked on same router with the mobile terminal, intelligent camera The process of couple in router can be with are as follows:
Intelligent camera, which detects, oneself does not access network, then broadcasting packet, so that mobile terminal shows intelligent camera The prompt information of network is not accessed, and user can broadcast by mobile terminal to intelligent camera after seeing the prompt information The title and password of the router oneself connected, intelligent camera is after the title and password for receiving router, Ke Yixiang The router sends connection request, to be connected on the router.
It is located in same local area network since intelligent camera and mobile terminal are connected to same router, APP can To obtain the binding state of intelligent camera by scanning, and show the binding state of intelligent camera.Wherein, binding state packet It includes and has bound and unbound two states.
When the binding state of intelligent camera is unbound, if user wants to control the intelligent camera, can touch Bindings are sent out, then the APP sends the mark and user account such as millet account for carrying intelligent camera to server Request, in this way, server can bind intelligent camera and user account according to the request.
For realizing above-mentioned method flow as shown in Figure 2, the related content being related to is described device as shown in figure 12 It is identical, it does not repeat herein.
Above-described embodiment, by interacting with mobile terminal with couple in router, to establish binding relationship for the two Offer condition, and then the condition of offer is monitored to be subsequent.
Figure 13 is the block diagram of another object monitor device shown according to an exemplary embodiment, as shown in figure 13, On the basis of above-mentioned embodiment illustrated in fig. 9, which may also include that training module 69.
Training module 69 is configured as being trained current sound template for each sound pattern.
In this embodiment, it is necessary first to each sound pattern in multiple sound patterns is trained, with reach by After target sound inputs sound pattern, sounding reason can be identified.Wherein, multiple sound patterns for identification draw by different reasons The target sound risen.
For realizing above-mentioned method flow as shown in Figure 4, the related content being related to describes device as shown in fig. 13 that It is identical, it does not repeat herein.
Above-described embodiment, by being trained to sound pattern, to judge that the sounding reason of target sound provides condition.
Figure 14 is the block diagram of another object monitor device shown according to an exemplary embodiment, as shown in figure 14, On the basis of above-mentioned embodiment illustrated in fig. 13, training module 69 can include: acquisition submodule 691, input submodule 692, first Training submodule 693 and the second training submodule 694.
Acquisition submodule 691 is configured as the target sound feature samples of acquisition preset quantity.
Input submodule 692 is configured as that the target sound feature sample of the collected preset quantity of submodule 691 will be acquired This input current sound template, obtain the target sound feature samples of preset quantity relative to current sound template distortion away from From distortion distance is used to indicate the target sound feature samples of preset quantity and the target sound exported by current sound template Between gap.
If the first training submodule 693, which is configured as the distortion distance that input submodule 692 obtains, is less than pre-determined distance, Stopping is trained current sound template.
If the second training submodule 694 be configured as distortion distance that input submodule 692 obtains be more than or equal to it is default away from From, then continue call training module current sound template is trained, until distortion distance be less than pre-determined distance.
Wherein, distortion distance is smaller, then shows the target sound feature samples of preset quantity and by current sound template Gap between the target sound of output is smaller.
When distortion distance is less than pre-determined distance, then otherwise training of the completion to current sound template continues to train, until Distortion distance is less than pre-determined distance.
For realizing above-mentioned method flow as shown in Figure 5, the related content being related to is described device as shown in figure 14 It is identical, it does not repeat herein.
Above-described embodiment is distorted by the way that the target sound feature samples of preset quantity are inputted current sound template Distance, and distortion distance is compared with pre-determined distance, to complete the training process of sound pattern, implementation is simple, instruction The sound pattern practised can accurately identify the sounding reason of target sound.
About the device in above-described embodiment, the concrete mode that wherein modules, submodule execute operation is having It closes and is described in detail in the embodiment of this method, no detailed explanation will be given here.
Figure 15 is a kind of block diagram suitable for object monitor device shown according to an exemplary embodiment.For example, device 1500 can be mobile phone, computer, digital broadcasting terminal, messaging device, game console, tablet device, medical treatment Equipment, body-building equipment, personal digital assistant, aircraft etc..
Referring to Fig.1 5, device 1500 may include following one or more components: processing component 1502, memory 1504, Power supply module 1506, multimedia component 1508, audio component 1510, the interface 1512 of input/output (I/O), sensor module 1514 and communication component 1516.
The integrated operation of the usual control device 1500 of processing component 1502, such as with display, telephone call, data communication, Camera operation and record operate associated operation.Processing element 1502 may include one or more processors 1520 to execute Instruction, to perform all or part of the steps of the methods described above.In addition, processing component 1502 may include one or more moulds Block, convenient for the interaction between processing component 1502 and other assemblies.For example, processing component 1502 may include multi-media module, To facilitate the interaction between multimedia component 1508 and processing component 1502.
Memory 1504 is configured as storing various types of data to support the operation in equipment 1500.These data Example includes the instruction of any application or method for operating on device 1500, contact data, telephone book data, Message, picture, video etc..Memory 1504 can by any kind of volatibility or non-volatile memory device or they Combination is realized, such as static random access memory (SRAM), electrically erasable programmable read-only memory (EEPROM), it is erasable can Program read-only memory (EPROM), programmable read only memory (PROM), read-only memory (ROM), magnetic memory, flash memory Reservoir, disk or CD.
Power supply module 1506 provides electric power for the various assemblies of device 1500.Power supply module 1506 may include power management System, one or more power supplys and other with for device 1500 generate, manage, and distribute the associated component of electric power.
Multimedia component 1508 includes the screen of one output interface of offer between described device 1500 and user.? In some embodiments, screen may include liquid crystal display (LCD) and touch panel (TP).If screen includes touch panel, Screen may be implemented as touch screen, to receive input signal from the user.Touch panel includes that one or more touch passes Sensor is to sense the gesture on touch, slide, and touch panel.The touch sensor can not only sense touch or sliding is dynamic The boundary of work, but also detect duration and pressure associated with the touch or slide operation.In some embodiments, more Media component 1508 includes a front camera and/or rear camera.When equipment 1500 is in operation mode, as shot mould When formula or video mode, front camera and/or rear camera can receive external multi-medium data.Each preposition camera shooting Head and rear camera can be a fixed optical lens system or have focusing and optical zoom capabilities.
Audio component 1510 is configured as output and/or input audio signal.For example, audio component 1510 includes a wheat Gram wind (MIC), when device 1500 is in operation mode, when such as call mode, recording mode, and voice recognition mode, microphone quilt It is configured to receive external audio signal.The received audio signal can be further stored in memory 1504 or via communication Component 1516 is sent.In some embodiments, audio component 1510 further includes a loudspeaker, is used for output audio signal.
I/O interface 1512 provides interface, above-mentioned peripheral interface module between processing component 1502 and peripheral interface module It can be keyboard, click wheel, button etc..These buttons may include, but are not limited to: home button, volume button, start button and Locking press button.
Sensor module 1514 includes one or more sensors, and the state for providing various aspects for device 1500 is commented Estimate.For example, sensor module 1514 can detecte the state that opens/closes of equipment 1500, the relative positioning of component, such as institute The display and keypad that component is device 1500 are stated, sensor module 1514 can be with detection device 1500 or device 1,500 1 The position change of a component, the existence or non-existence that user contacts with device 1500,1500 orientation of device or acceleration/deceleration and dress Set 1500 temperature change.Sensor module 1514 may include proximity sensor, be configured in not any physics It is detected the presence of nearby objects when contact.Sensor module 1514 can also include optical sensor, as CMOS or ccd image are sensed Device, for being used in imaging applications.In some embodiments, which can also include acceleration sensing Device, gyro sensor, Magnetic Sensor, pressure sensor or temperature sensor.
Communication component 1516 is configured to facilitate the communication of wired or wireless way between device 1500 and other equipment.Dress The wireless network based on communication standard, such as WiFi can be accessed by setting 1500,2G or 3G or their combination.It is exemplary at one In embodiment, communication component 1516 receives broadcast singal or broadcast correlation from external broadcasting management system via broadcast channel Information.In one exemplary embodiment, the communication component 1516 further includes near-field communication (NFC) module, to promote short distance Communication.For example, radio frequency identification (RFID) technology, Infrared Data Association (IrDA) technology, ultra wide band can be based in NFC module (UWB) technology, bluetooth (BT) technology and other technologies are realized.
In the exemplary embodiment, device 1500 can be by one or more application specific integrated circuit (ASIC), number Signal processor (DSP), digital signal processing appts (DSPD), programmable logic device (PLD), field programmable gate array (FPGA), controller, microcontroller, microprocessor or other electronic components are realized, for executing the above method.
In the exemplary embodiment, a kind of non-transitorycomputer readable storage medium including instruction, example are additionally provided It such as include the memory 1504 of instruction, above-metioned instruction can be executed by the processor 1520 of device 1500 to complete the above method.Example Such as, the non-transitorycomputer readable storage medium can be ROM, random access memory (RAM), CD-ROM, tape, soft Disk and optical data storage devices etc..
Those skilled in the art will readily occur to its of the disclosure after considering specification and practicing disclosure disclosed herein Its embodiment.This application is intended to cover any variations, uses, or adaptations of the disclosure, these modifications, purposes or Person's adaptive change follows the general principles of this disclosure and including the undocumented common knowledge in the art of the disclosure Or conventional techniques.The description and examples are only to be considered as illustrative, and the true scope and spirit of the disclosure are by following Claim is pointed out.
It should be understood that the present disclosure is not limited to the precise structures that have been described above and shown in the drawings, and And various modifications and changes may be made without departing from the scope thereof.The scope of the present disclosure is only limited by the accompanying claims.

Claims (19)

1. a kind of object monitor method, which is characterized in that the method is applied in monitoring device, which comprises
When there are the sounding reasons for when target sound, judging the target sound in collected sound;
Sounding information is pushed to mobile terminal, the sounding information carries the sounding reason;
The sounding reason of the judgement target sound, comprising:
Feature extraction is carried out to the target sound, obtains target sound characteristic information;
The target sound characteristic information is matched with the trained obtained multiple sound patterns prestored, and according to matching As a result the sounding reason of the target sound is obtained, the multiple sound pattern target sound caused by different reasons for identification Sound.
2. object monitor method according to claim 1, which is characterized in that push sounding information to mobile terminal described Later, the method also includes:
Receive the video acquisition instruction that the mobile terminal is sent;
It is instructed according to the video acquisition to the mobile terminal pushing video information, so that the mobile terminal shows the view Frequency information.
3. object monitor method according to claim 2, which is characterized in that it is described according to the video acquisition instruct to Before the mobile terminal pushing video information, the method also includes:
Obtain the sounding position of the target sound;
It controls camera and turns to the corresponding direction in the sounding position, and acquire the video information.
4. object monitor method according to claim 1, which is characterized in that described to be mentioned to target sound progress feature It takes, obtains target sound characteristic information, comprising:
Compression processing is carried out to the target sound, obtains voice compressing information;
Feature extraction is carried out to the voice compressing information, obtains the target sound characteristic information.
5. object monitor method according to claim 1, which is characterized in that the sounding of the judgement target sound is former Cause, comprising:
The target sound is sent to server;
Receive the sounding reason for the target sound that the server returns.
6. object monitor method according to claim 2, which is characterized in that the method also includes:
If detecting, the monitoring device does not access network, broadcasting packet, so that the mobile terminal shows that the monitoring is set The standby prompt information for not accessing network;
The title and password of the router of the mobile terminal broadcast are received, and sends connection request to the router, with even It is connected on the router.
7. object monitor method according to claim 1, which is characterized in that the method also includes:
For each sound pattern, current sound template is trained.
8. object monitor method according to claim 7, which is characterized in that it is described that current sound template is trained, Include:
Acquire the target sound feature samples of preset quantity;
The target sound feature samples of the preset quantity are inputted into current sound template, obtain the target sound of the preset quantity Distortion distance of the sound feature samples relative to current sound template, the distortion distance are used to indicate the target of the preset quantity Gap between sound characteristic sample and the target sound for passing through current sound template output;
If the distortion distance is less than pre-determined distance, stopping is trained the current sound template;
If the distortion distance is more than or equal to pre-determined distance, continue to be trained the current sound template, until described Distortion distance is less than pre-determined distance.
9. object monitor method according to any one of claims 1 to 8, which is characterized in that the sounding information also carries There is the target sound.
10. a kind of object monitor device, which is characterized in that described device is applied in monitoring device, and described device includes:
Judgment module is configured as when there are the sounding originals for when target sound, judging the target sound in collected sound Cause;
First pushing module is configured as pushing sounding information to mobile terminal, and the sounding information carries the judgement mould The sounding reason that block is judged;
The judgment module includes:
Extracting sub-module is configured as carrying out feature extraction to the target sound, obtains target sound characteristic information;
Matched sub-block, the warp for being configured as the target sound characteristic information for extracting the extracting sub-module and prestoring Multiple sound patterns that training obtains are matched, and obtain the sounding reason of the target sound according to matching result, described Multiple sound patterns target sound caused by different reasons for identification.
11. object monitor device according to claim 10, which is characterized in that described device further include:
Receiving module is configured as after first pushing module pushes sounding information to mobile terminal, receives the shifting The video acquisition instruction that dynamic terminal is sent;
Second pushing module is configured as being instructed according to the received video acquisition of the receiving module to the mobile terminal Pushing video information, so that the mobile terminal shows the video information.
12. object monitor device according to claim 11, which is characterized in that described device further include:
Module is obtained, is configured as being instructed according to the video acquisition to mobile terminal push in second pushing module Before video information, the sounding position of the target sound is obtained;
Acquisition module is turned to, control camera is configured as and turns to the corresponding side in the sounding position that the acquisition module obtains To, and acquire the video information.
13. object monitor device according to claim 10, which is characterized in that the extracting sub-module includes:
Compression unit is configured as carrying out compression processing to the target sound, obtains voice compressing information;
Extraction unit is configured as the voice compressing information obtained to the compression unit and carries out feature extraction, obtains institute State target sound characteristic information.
14. object monitor device according to claim 10, which is characterized in that the judgment module includes:
Sending submodule is configured as sending the target sound to server;
Receiving submodule is configured as receiving the sounding reason for the target sound that the server returns.
15. object monitor device according to claim 11, which is characterized in that described device further include:
Broadcast module is detected, if being configured as detecting that the monitoring device does not access network, broadcasting packet, so that the shifting Dynamic terminal shows that the monitoring device does not access the prompt information of network;
Link block is received, is configured as receiving the title and password of the router of the mobile terminal broadcast, and to the road Connection request is sent by device, to be connected on the router.
16. object monitor device according to claim 10, which is characterized in that described device further include:
Training module is configured as being trained current sound template for each sound pattern.
17. object monitor device according to claim 16, which is characterized in that the training module includes:
Submodule is acquired, the target sound feature samples of acquisition preset quantity are configured as;
Input submodule is configured as the target sound feature samples of the acquisition collected preset quantity of submodule Input current sound template, obtain the target sound feature samples of the preset quantity relative to current sound template distortion away from From the target sound feature samples that, the distortion distance is used to indicate the preset quantity with it is defeated by the current sound template The gap between target sound out;
First training submodule, if being configured as the distortion distance that the input submodule obtains is less than pre-determined distance, Stopping is trained the current sound template;
Second training submodule, if be configured as the distortion distance that the input submodule obtains be more than or equal to it is default away from From, then continue that the training module is called to be trained the current sound template, until the distortion distance be less than it is default Distance.
18. object monitor device according to any one of claims 10 to 17, which is characterized in that the sounding information is also taken With the target sound.
19. a kind of object monitor device characterized by comprising
Processor;
Memory for storage processor executable instruction;
Wherein, the processor is configured to:
When there are the sounding reasons for when target sound, judging the target sound in collected sound;
Sounding information is pushed to mobile terminal, the sounding information carries the sounding reason;
The sounding reason of the judgement target sound, comprising:
Feature extraction is carried out to the target sound, obtains target sound characteristic information;
The target sound characteristic information is matched with the trained obtained multiple sound patterns prestored, and according to matching As a result the sounding reason of the target sound is obtained, the multiple sound pattern target sound caused by different reasons for identification Sound.
CN201610258363.4A 2016-04-22 2016-04-22 Object monitor method and device Active CN105898219B (en)

Priority Applications (4)

Application Number Priority Date Filing Date Title
CN201610258363.4A CN105898219B (en) 2016-04-22 2016-04-22 Object monitor method and device
PCT/CN2016/091863 WO2017181545A1 (en) 2016-04-22 2016-07-27 Object monitoring method and device
EP16197065.2A EP3236469B1 (en) 2016-04-22 2016-11-03 Object monitoring method and device
US15/378,059 US10122916B2 (en) 2016-04-22 2016-12-14 Object monitoring method and device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201610258363.4A CN105898219B (en) 2016-04-22 2016-04-22 Object monitor method and device

Publications (2)

Publication Number Publication Date
CN105898219A CN105898219A (en) 2016-08-24
CN105898219B true CN105898219B (en) 2019-05-21

Family

ID=56704587

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201610258363.4A Active CN105898219B (en) 2016-04-22 2016-04-22 Object monitor method and device

Country Status (4)

Country Link
US (1) US10122916B2 (en)
EP (1) EP3236469B1 (en)
CN (1) CN105898219B (en)
WO (1) WO2017181545A1 (en)

Families Citing this family (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107122807B (en) * 2017-05-24 2021-05-21 努比亚技术有限公司 Home monitoring method, server and computer readable storage medium
CN107705786A (en) * 2017-09-27 2018-02-16 努比亚技术有限公司 A kind of method of speech processing, device and computer-readable recording medium
CN108234944A (en) * 2017-12-29 2018-06-29 佛山市幻云科技有限公司 Children's monitoring method, device, server and system based on crying identification
CN111128227B (en) * 2019-12-30 2022-06-17 云知声智能科技股份有限公司 Sound detection method and device
CN113488078B (en) * 2020-04-24 2024-03-29 海信集团有限公司 Health state monitoring method and equipment
CN113467258A (en) * 2020-04-24 2021-10-01 海信集团有限公司 Intelligent monitoring method and equipment thereof
CN113689663B (en) * 2021-08-05 2022-10-04 浙江晨逸新声物联科技有限公司 System and method for monitoring infant emergency

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105407316A (en) * 2014-08-19 2016-03-16 北京奇虎科技有限公司 Implementation method for intelligent camera system, intelligent camera system, and network camera
CN105489217A (en) * 2015-12-30 2016-04-13 武传胜 Method and system of monitoring the aged through sound by means of internet of things

Family Cites Families (23)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP3564501B2 (en) * 2001-03-22 2004-09-15 学校法人明治大学 Infant voice analysis system
US20030018975A1 (en) * 2001-07-18 2003-01-23 Stone Christopher J. Method and system for wireless audio and video monitoring
JP2004153732A (en) * 2002-11-01 2004-05-27 Toshiba Eng Co Ltd Nursing home monitoring system
JP4491462B2 (en) 2003-05-21 2010-06-30 コーニンクレッカ フィリップス エレクトロニクス エヌ ヴィ Monitor system capable of generating audible messages
US7266347B2 (en) * 2003-06-10 2007-09-04 Gross John N Remote monitoring device and process
JP4861723B2 (en) * 2006-02-27 2012-01-25 池上通信機株式会社 Monitoring system
US8041540B2 (en) * 2009-12-09 2011-10-18 General Electric Company System, device, and method for acoustic and visual monitoring of a wind turbine
US9020622B2 (en) * 2010-06-17 2015-04-28 Evo Inc. Audio monitoring system and method of use
US20120330499A1 (en) * 2011-06-23 2012-12-27 United Technologies Corporation Acoustic diagnostic of fielded turbine engines
TWI474315B (en) * 2012-05-25 2015-02-21 Univ Nat Taiwan Normal Infant cries analysis method and system
US20140086419A1 (en) * 2012-09-27 2014-03-27 Manjit Rana Method for capturing and using audio or sound signatures to analyse vehicle accidents and driver behaviours
KR101974482B1 (en) * 2013-04-05 2019-05-02 삼성전자주식회사 Electronic stethoscopy apparatus, automatic diagnostic apparatus and method for diagnosing automatically
CN103280220B (en) * 2013-04-25 2015-11-18 北京大学深圳研究生院 A kind of real-time recognition method for baby cry
KR102195897B1 (en) 2013-06-05 2020-12-28 삼성전자주식회사 Apparatus for dectecting aucoustic event, operating method thereof, and computer-readable recording medium having embodied thereon a program which when executed by a computer perorms the method
CN103489282B (en) * 2013-09-24 2016-01-20 华南理工大学 A kind of baby monitor and vagitus recognition methods that can identify vagitus
WO2015157440A1 (en) * 2014-04-08 2015-10-15 Assaf Glazer Systems and methods for configuring baby monitor cameras to provide uniform data sets for analysis
US9159217B1 (en) * 2014-04-10 2015-10-13 Twin Harbor Labs, LLC Methods and apparatus notifying a remotely located user of the operating condition of a household appliance
US9538959B2 (en) * 2014-08-03 2017-01-10 Morpheus, Llc System and method for human monitoring
CN104732689B (en) * 2015-03-18 2017-10-20 苏州科技大学 Household burglary-resisting system and control method based on Internet of Things and mobile phone terminal
CN105208110B (en) * 2015-08-31 2019-11-01 小米科技有限责任公司 Resource control method and device
US10134422B2 (en) * 2015-12-01 2018-11-20 Qualcomm Incorporated Determining audio event based on location information
US9843877B2 (en) * 2015-12-31 2017-12-12 Ebay Inc. Sound recognition
US9858789B2 (en) * 2016-01-19 2018-01-02 Vivint, Inc. Occupancy-targeted baby monitoring

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105407316A (en) * 2014-08-19 2016-03-16 北京奇虎科技有限公司 Implementation method for intelligent camera system, intelligent camera system, and network camera
CN105489217A (en) * 2015-12-30 2016-04-13 武传胜 Method and system of monitoring the aged through sound by means of internet of things

Also Published As

Publication number Publication date
US10122916B2 (en) 2018-11-06
EP3236469B1 (en) 2021-05-19
CN105898219A (en) 2016-08-24
EP3236469A1 (en) 2017-10-25
WO2017181545A1 (en) 2017-10-26
US20170310885A1 (en) 2017-10-26

Similar Documents

Publication Publication Date Title
CN105898219B (en) Object monitor method and device
KR101945682B1 (en) Method, apparatus, program and recording medium for implementing touch key and fingerprint identification and terminal device
CN105828201B (en) Method for processing video frequency and device
CN105204742B (en) Control method, device and the terminal of electronic equipment
CN104537365B (en) Touch key-press and fingerprint recognition implementation method, device and terminal device
CN104159218B (en) Internetwork connection establishing method and device
CN105468767B (en) Card information acquisition methods and device
CN105069426B (en) Similar pictures judgment method and device
CN105160320B (en) Fingerprint identification method, device and mobile terminal
CN106162211B (en) The method and device being broadcast live based on live streaming application
CN103955275A (en) Application control method and device
CN107766820A (en) Image classification method and device
CN104933419A (en) Method and device for obtaining iris images and iris identification equipment
CN106303260A (en) Photographic head changing method, device and terminal unit
CN106303198A (en) Photographing information acquisition methods and device
CN103955274A (en) Application control method and device
CN107948660A (en) The method and device of Video coding adaptation
CN104010231A (en) Data synchronizing method and device
CN107872620A (en) video recording method and device
CN106572306A (en) Image shooting method and electronic equipment
CN106453981B (en) Electronic equipment method for processing voice messages and device
CN109740490A (en) Auth method and device
CN106133736B (en) Fingerprint authentication method and device
CN106126060A (en) Intelligent home furnishing control method and device
CN104486390B (en) The method and device of data sharing

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant