CN111526292A - Audio and video mixing monitoring method, equipment and storage device - Google Patents

Audio and video mixing monitoring method, equipment and storage device Download PDF

Info

Publication number
CN111526292A
CN111526292A CN202010363581.0A CN202010363581A CN111526292A CN 111526292 A CN111526292 A CN 111526292A CN 202010363581 A CN202010363581 A CN 202010363581A CN 111526292 A CN111526292 A CN 111526292A
Authority
CN
China
Prior art keywords
monitoring
audio
video
target
acquiring
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202010363581.0A
Other languages
Chinese (zh)
Other versions
CN111526292B (en
Inventor
陈明珠
杨增启
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zhejiang Dahua Technology Co Ltd
Original Assignee
Zhejiang Dahua Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zhejiang Dahua Technology Co Ltd filed Critical Zhejiang Dahua Technology Co Ltd
Priority to CN202010363581.0A priority Critical patent/CN111526292B/en
Publication of CN111526292A publication Critical patent/CN111526292A/en
Application granted granted Critical
Publication of CN111526292B publication Critical patent/CN111526292B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/18Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/10Terrestrial scenes
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L17/00Speaker identification or verification
    • G10L17/22Interactive procedures; Man-machine interfaces
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/48Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use
    • G10L25/51Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination
    • G10L25/60Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination for measuring the quality of voice signals
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/695Control of camera direction for changing a field of view, e.g. pan, tilt or based on tracking of objects

Abstract

The invention discloses an audio and video mixing monitoring method, equipment and a storage device, wherein the audio and video mixing monitoring method comprises the following steps: acquiring a monitoring target in a monitoring image shot by video monitoring equipment, and carrying out video monitoring on the monitoring target; acquiring peripheral objects within a preset range from the monitored target, and identifying the target; obtaining an evaluation function of the peripheral object according to the identification result of the target identification, wherein the evaluation function is used for evaluating the energy of the object reflected audio signal; and acquiring a first intermediate object according to the evaluation function, and controlling audio monitoring equipment to perform audio monitoring on the first intermediate object. By the mode, the method can quickly and accurately select the monitoring intermedium to carry out audio monitoring and output high-quality audio and video monitoring.

Description

Audio and video mixing monitoring method, equipment and storage device
Technical Field
The application relates to the field of security monitoring, in particular to an audio and video mixing monitoring method, equipment and a storage device.
Background
In the field of security monitoring, video monitoring technology has been widely applied, however, in some crime evidence obtaining scenes, it is often not enough to rely on video monitoring information, and it is desirable to acquire accurate audio information to further acquire detailed information of a criminal offender. Therefore, besides installing the video monitoring device, an audio monitoring device is also needed to be installed for monitoring.
The following two technical schemes are generally adopted in the existing audio monitoring technology:
in the first scheme, laser monitoring is adopted. The voice signal can cause tiny vibration of surrounding objects, the vibration characteristics of the objects are related to the vibration characteristics of the sound source, and therefore, when the laser is irradiated on the target objects, the voice signal is loaded on the laser beam; the photoelectric receiving system collects the light wave signals reflected from the target object, and the voice signals can be recovered through photoelectric conversion, signal amplification and filtering and voice demodulation.
And in the second scheme, equipment such as a microphone is adopted for monitoring. The camera is added with a microphone and the like for sound collection, namely, a microphone sound head is added with an amplifying circuit to output sound, and the collected sound in the environment is coded and decoded and then is output together with a video signal.
In the first scheme, laser monitoring is adopted, so that the quality of audio and video is generally difficult to be considered simultaneously. In addition, in the laser monitoring process, single-point or multi-point monitoring is generally completed by methods such as trial and error or a method of arranging a monitoring intermediary in advance, so that the laser monitoring method is difficult to adapt to complex and variable scenes. In addition, the process of finding the intermediaries in some scenes by finding the best intermediaries through a trial and error method is too long, and particularly when the moving target is monitored, the target is already moved to a new scene after all the intermediaries around the target are traversed, so that the monitoring effect is poor.
In the second scheme, a microphone is adopted for monitoring, so that the directivity is poor, and accurate monitoring on any interested monitored object is difficult to complete; and the method is easily interfered by environmental noise, so that the problems of short monitoring distance and useless monitoring data redundancy are caused, and the audio and video monitoring quality cannot be considered at the same time.
Therefore, it is necessary to provide an audio/video mixing monitoring method, device and storage apparatus to solve the above technical problems.
Disclosure of Invention
The application provides an audio and video mixing monitoring method, equipment and a storage device, which can achieve the purposes of quickly and accurately selecting a monitoring intermediary object to carry out audio monitoring and outputting high-quality audio monitoring.
In order to solve the technical problem, the application adopts a technical scheme that: the audio and video mixing monitoring method comprises the following steps:
acquiring a monitoring target in a monitoring image shot by video monitoring equipment, and carrying out video monitoring on the monitoring target;
acquiring peripheral objects within a preset range from the monitored target, and identifying the target;
obtaining an evaluation function of the peripheral object according to the identification result of the target identification, wherein the evaluation function is used for evaluating the energy of the object reflected audio signal;
and acquiring a first intermediate object according to the evaluation function, and controlling audio monitoring equipment to perform audio monitoring on the first intermediate object.
In order to solve the above technical problem, another technical solution adopted by the present application is: providing an audio and video mixing monitoring device, which comprises a processor and a memory coupled with the processor, wherein the memory stores program instructions for realizing the audio and video mixing monitoring method; the processor is used for executing the program instructions stored in the memory so as to perform audio and video mixing monitoring on the monitoring.
In order to solve the above technical problem, the present application adopts another technical solution that: a storage device is provided, which stores a program file capable of realizing the audio and video mixing monitoring method.
The beneficial effect of this application is:
according to the audio and video mixing monitoring method, the device and the storage device, the evaluation function of the peripheral object in the preset range of the monitoring target is obtained, the first intermediate object is obtained according to the evaluation function, the audio monitoring device is controlled to perform audio monitoring on the first intermediate object, the selection accuracy of the first intermediate object can be improved, and the audio monitoring quality is improved;
furthermore, target recognition is carried out on the peripheral object, the material attribute of the peripheral object is obtained, the evaluation function is obtained through the material attribute, calculation is simple, operability is high, accuracy of the evaluation function is improved, and selection accuracy of the first intermediate is improved.
Furthermore, a distance weight factor is introduced into the calculation of the evaluation function, so that the accuracy of the evaluation function is improved, and the selection accuracy of the first intermediate is further improved;
further, an evaluation function is obtained by dividing the pickup area, and the speed of selecting the first medium object is increased;
further, an intermediary database is established, the evaluation function is stored in the database, and when a first intermediary is selected, the first intermediary is directly selected from the intermediary database, so that the speed of selecting the first intermediary is further improved;
furthermore, an evaluation function is calculated theoretically, so that the accuracy of the evaluation function is improved, and the selection accuracy of the first intermediate is further improved;
further, the audio monitoring is carried out by selecting the monitoring area of the first intermediate object, so that the audio monitoring quality is improved;
furthermore, the monitoring quality is improved by performing video monitoring through linkage of the panoramic camera and the spherical camera, when the first distance of the monitoring target is greater than the pickup distance, only video monitoring is adopted, audio monitoring is not adopted, and system resources are saved.
Drawings
Fig. 1 is a schematic structural diagram of an audio/video mixing monitoring system to which an embodiment of the present invention is applied;
fig. 2 is a schematic flow chart of an audio/video mixing monitoring method according to a first embodiment of the present invention;
FIG. 3 is a schematic flow chart of acquiring a monitored area according to a first embodiment of the present invention;
FIG. 4 is a flowchart illustrating a process of obtaining a voiceprint database according to a first embodiment of the invention;
FIG. 5 is a schematic flow chart of obtaining an evaluation function according to a second embodiment of the present invention;
FIG. 6 is a schematic flow chart of obtaining an evaluation function according to a third embodiment of the present invention;
fig. 7 is a schematic diagram of dividing a sound pickup area according to a third embodiment of the present invention;
FIG. 8 is a schematic flow chart of obtaining an evaluation function according to a fourth embodiment of the present invention;
FIG. 9 is a schematic diagram of a model of an audio monitoring device emitting laser light to surrounding objects according to a fourth embodiment of the present invention;
FIG. 10 is a schematic diagram of a model of an audio monitoring device receiving laser reflections according to a fourth embodiment of the present invention;
fig. 11 is a schematic structural diagram of an audio/video mixing monitoring apparatus according to an embodiment of the present invention;
fig. 12 is a schematic structural diagram of an audio/video mixing monitoring device according to an embodiment of the present invention;
fig. 13 is a schematic structural diagram of a memory device according to an embodiment of the invention.
Detailed Description
The technical solutions in the embodiments of the present application will be clearly and completely described below with reference to the drawings in the embodiments of the present application, and it is obvious that the described embodiments are only a part of the embodiments of the present application, and not all of the embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present application.
The terms "first", "second" and "third" in this application are used for descriptive purposes only and are not to be construed as indicating or implying relative importance or implying any indication of the number of technical features indicated. Thus, a feature defined as "first," "second," or "third" may explicitly or implicitly include at least one of the feature. In the description of the present application, "plurality" means at least two, e.g., two, three, etc., unless explicitly specifically limited otherwise. All directional indications (such as up, down, left, right, front, and back) in the embodiments of the present application are only used to explain the relative positional relationship between the components, the movement, and the like in a specific posture (as shown in the drawings), and if the specific posture is changed, the directional indication is changed accordingly. Furthermore, the terms "include" and "have," as well as any variations thereof, are intended to cover non-exclusive inclusions. For example, a process, method, system, article, or apparatus that comprises a list of steps or elements is not limited to only those steps or elements listed, but may alternatively include other steps or elements not listed, or inherent to such process, method, article, or apparatus.
Reference herein to "an embodiment" means that a particular feature, structure, or characteristic described in connection with the embodiment can be included in at least one embodiment of the application. The appearances of the phrase in various places in the specification are not necessarily all referring to the same embodiment, nor are separate or alternative embodiments mutually exclusive of other embodiments. It is explicitly and implicitly understood by one skilled in the art that the embodiments described herein can be combined with other embodiments.
Referring to fig. 1, fig. 1 is a schematic structural diagram of an audio/video hybrid monitoring system to which an embodiment of the present invention is applied. The audio and video mixed monitoring system comprises video monitoring equipment, audio monitoring equipment and a control module, wherein the video monitoring equipment is used for carrying out video monitoring on a monitored target, the audio monitoring equipment is used for carrying out audio monitoring on the monitored target, the control module executes the audio and video mixed monitoring method to the video monitoring equipment and the audio monitoring equipment, wherein the control module can be arranged in the video monitoring equipment or the audio monitoring equipment and can also be respectively arranged in the video monitoring equipment and the audio monitoring equipment and can also be arranged in the video monitoring equipment and a server outside the audio monitoring equipment, and the video monitoring equipment, the audio monitoring equipment and the control module can be communicated with each other. In this embodiment, video monitoring equipment includes the panoramic camera, the panoramic camera is used for shooing the surveillance image, audio monitoring equipment includes the laser sound pickup, thereby the laser sound pickup is used for transmitting laser and receives reflection laser and draws audio information, just the panoramic camera with the laser sound pickup all is located cloud platform 1, just the exit angle that the laser beam was launched to the laser sound pickup is coaxial with on the cloud platform 1 the optical axis of panoramic camera.
In order to make the quality of the video monitoring picture higher and the details of the picture clearer, in another embodiment, the video monitoring device may further include a dome camera, the dome camera is located on the pan-tilt 2, the panoramic camera may take a panoramic image in the monitoring environment, and the dome camera may zoom the detailed image of the object in the monitoring scene.
Referring to fig. 2, fig. 2 is a schematic flow chart of an audio/video mixing monitoring method according to an embodiment of the present invention. It should be noted that the method of the present invention is not limited to the flow sequence shown in fig. 2 if the results are substantially the same. As shown in fig. 2, the method comprises the steps of:
step S101: acquiring a monitoring target in a monitoring image shot by video monitoring equipment, and carrying out video monitoring on the monitoring target;
in this embodiment, the video monitoring device includes a panoramic camera, the panoramic camera performs panoramic video monitoring shooting on a monitored environment, and identifies a monitored target in a shot image through a target identification technology, where the monitored target may be an object such as a person or a vehicle, and in step S101, the monitored target identified through the target identification technology further includes a first coordinate of the monitored target.
In another embodiment, in order to improve the quality of video surveillance, the video surveillance device may further include a dome camera, where the dome camera controls the pan-tilt 2 to rotate to the position aligned with the surveillance target according to the first coordinate of the surveillance target, so as to perform video surveillance with higher picture details on the surveillance target.
Step S102: and acquiring peripheral objects within a preset range from the monitored target, and identifying the target.
Specifically, the video monitoring device monitors and shoots a distance within a preset monitoring target range to generate a monitoring graph, all objects in the monitoring graph can be identified through a target identification technology, the target identification technology can identify the types of the peripheral objects within the preset range, such as vehicles, people, trees, buildings and the like, and the second coordinates of the peripheral objects, the preset range can be set manually, and the preset range can also be set synchronously to be the monitoring range of the audio monitoring device.
Step S103: and acquiring an evaluation function of the peripheral object according to the recognition result of the target recognition.
It should be noted that the evaluation function is used to evaluate the energy of the object reflecting the audio signal, in this embodiment, the audio monitoring device may include a laser pickup, and the laser pickup may obtain the evaluation function of the peripheral object by emitting laser to the peripheral objects of different types and receiving the echo energy of the reflected laser after the peripheral objects of different types are pressed by the sound source to vibrate, so that the audio monitoring device may determine the echo energy of the objects of different types according to the second coordinate control rotation of the pan/tilt head 1 before the audio/video hybrid monitoring system is installed, so as to obtain the evaluation function of the peripheral objects of different types.
Step S104: and acquiring a first intermediate object according to the evaluation function, and controlling audio monitoring equipment to perform audio monitoring on the first intermediate object.
It should be noted that the first intermediate object is an object with the largest energy of the reflected audio signals in the peripheral objects, that is, the peripheral object with the largest evaluation function value is selected as the first intermediate object for audio monitoring.
Further, referring to fig. 3 together for better audio monitoring effect, fig. 3 is a schematic flowchart of a process of acquiring a monitoring area according to a first embodiment of the present invention, and in step S104, the method for controlling an audio monitoring device to perform audio monitoring on the first intermediary object may further include the following steps:
step S104 a: uniformly dividing the first interposer into M vibration areas;
specifically, the first interposer obtained in step S104 is uniformly divided into M vibration regions.
Step S104 b: obtaining vibration voltage generated after each vibration area is pressed by a sound source;
further, the cradle head 1 is controlled to align the audio monitoring device to the M vibration areas on the first intermediary, and vibration voltages generated after the vibration areas are compressed by a sound source are respectively acquired.
Step S104 c: selecting the vibration region with the largest voltage value in the vibration voltages of the M vibration regions as a monitoring region;
in this embodiment, the vibration region with the largest vibration voltage is selected from the vibration regions to serve as the monitoring region for audio monitoring, and the monitoring coordinates of the monitoring region are obtained.
Step S104 d: and controlling the audio monitoring equipment to perform audio monitoring on the monitoring area.
Finally, the audio monitoring equipment aims at the monitoring area to carry out audio monitoring through the rotation of the holder 1 according to the monitoring coordinate, and specifically, when the audio monitoring is carried out, the vibration voltage needs to be subjected to signal processing and amplification to be restored into sound information; extracting matched target sound in a voiceprint database according to the sound information; and finally, carrying out audio monitoring on the target sound.
Referring to fig. 4, fig. 4 is a schematic flow chart of acquiring a voiceprint database according to a first embodiment of the invention.
Step S104 a': acquiring first sound information of the monitoring target in different scenes;
before establishing the voiceprint database, first sound information of the monitoring target under different scenes can be obtained, wherein the different scenes can comprise an indoor scene, an outdoor scene, a scene with large volume and a scene with small volume;
step S104 b': filtering the noise and silence of the first sound information to obtain second sound information;
and filtering the part of invalid sound information such as noise, silence and the like of the first sound information, inputting the filtered first sound information into a neural network, and extracting to obtain second sound information, wherein the second sound information is sound with extracted sound characteristic information.
Step S104 c': and acquiring sound characteristics according to the second sound information, performing learning training, and storing the sound characteristics to the voiceprint database.
Through the learning training of the neural network, more accurate sound characteristics can be extracted, and finally, the sound characteristics are stored in the voiceprint database.
In order to improve the quality of audio monitoring and save system resources, in another embodiment, before the audio monitoring device performs audio monitoring, the method may further include:
acquiring a pickup distance of the audio monitoring equipment and a first distance from the audio monitoring equipment to the monitoring target;
it should be noted that, because the pickup distance of the audio monitoring device is limited, when the first distance of the monitoring target is greater than the pickup distance, the audio monitoring effect of the audio monitoring device may not be ideal, and in order to save system resources and reduce system performance loss, it may be determined first whether the monitoring target is within the pickup distance.
Specifically, before the audio/video hybrid monitoring system is installed and monitored, the system may be initialized, such as initializing a storage space, a video monitoring device, an audio monitoring device, and the like, and in the system initialization process, the device parameters of the video monitoring device and the audio monitoring device, including the pickup distance S1 of the audio monitoring device, may be obtained.
When the audio monitoring equipment receives the first coordinate of the monitoring target monitored by the video monitoring equipment, the audio monitoring equipment can align the laser pickup of the audio monitoring equipment to the monitoring target through the rotation of the holder 1, open laser and emit the laser to the monitoring target through recordingCalculating a first distance S2 from the audio monitoring device to the monitoring target by the time t of the monitoring target reflecting and receiving the reflected laser, that is: s2=VLaserT/2, wherein VLaserThe velocity in air is approximately 3 x 108m/s。
When the first distance is smaller than the pickup distance, performing video monitoring by using the video monitoring equipment, and performing audio monitoring by using the audio monitoring equipment;
and when the first distance is greater than the pickup distance, only using the video monitoring equipment to carry out video monitoring.
According to the audio and video hybrid monitoring method, the evaluation function of the peripheral object in the preset monitoring target range is obtained, the first intermediate object is obtained according to the evaluation function, and the audio monitoring equipment is controlled to perform audio monitoring on the first intermediate object, so that the selection accuracy of the first intermediate object can be improved, the audio monitoring quality can be improved, and the user experience can be improved;
furthermore, the evaluation function is obtained by carrying out target recognition on the peripheral object, the calculation is simple, the operability is strong, the first intermediate object can be intelligently selected for video monitoring, and the user experience is improved.
Further, the audio monitoring is carried out by selecting the monitoring area of the first intermediate object, so that the audio monitoring quality is improved;
furthermore, the monitoring quality is improved by performing video monitoring through linkage of the panoramic camera and the spherical camera, and only video monitoring is adopted and audio monitoring is not adopted when the first distance of the monitored target is greater than the pickup distance, so that system resources are saved.
Referring to fig. 5, fig. 5 is a schematic flow chart illustrating obtaining an evaluation function according to a second embodiment of the invention. It should be noted that the method of the present invention is not limited to the flow sequence shown in fig. 5 if the results are substantially the same. As shown in fig. 5, the obtaining of the evaluation function of the peripheral object according to the recognition result of the target recognition may further include:
step S201: acquiring output voltage generated after the peripheral object in the preset range is compressed by a sound source and distance weight factors of the peripheral object;
it should be noted that, because the objects made of different materials are subjected to different vibrations generated after being compressed by the sound source, and therefore the reflected laser energy is different, in this embodiment, the target identification further needs to compare the surrounding objects in the monitoring graph with sample data in a texture library by a texture identification technology, so as to identify the material attribute of each of the surrounding objects.
In step S201, firstly, a sound source needs to be arranged within the preset range to simulate the monitored target to sound, the audio monitoring device obtains an output voltage generated after the peripheral object within the preset range is compressed by the sound source, and obtains an evaluation function of the peripheral object according to the material property according to the output voltage.
In order to avoid the situation that the peripheral object with a large acoustic reactance may appear only once in a range close to the audio monitoring equipment, so that the average voltage of the peripheral object with the large acoustic reactance calculated in the above step is larger than the average voltage of other peripheral objects with a small acoustic reactance far from the audio monitoring equipment, and the wrong selection is caused; a distance weight factor is further introduced into the calculation of the evaluation function, wherein the distance weight factor represents the weight of the energy of the audio signal reflected by the objects at different distances from the audio monitoring equipment, and the method comprises the following steps:
step S201 a: acquiring echo energy of the peripheral object in the preset range;
in this embodiment, the rotation of the pan/tilt head 1 may be controlled according to the second coordinate of the peripheral object, so that the audio monitoring device respectively emits laser to the peripheral object, and obtains the echo energy reflected by the peripheral object.
Step S201 b: acquiring the monitoring distance between the peripheral object and the audio monitoring equipment in the preset range;
the method for obtaining the monitoring distance is similar to the method for calculating the first distance S2, and therefore, the description thereof is omitted.
Step S201 c: and acquiring the distance weight factor of the peripheral object according to the monitoring distance and the echo energy.
In this embodiment, the distance weighting factor may be obtained by the listening distance of the peripheral object and a ratio of the listening distance to the echo energy.
Step S202: and calculating the evaluation function of the peripheral objects with the same material attribute according to the output voltage and the distance weight factor.
According to the law of constant laser power and conservation of energy, the evaluation function of the peripheral object with the same material property can be obtained by calculating the average value again after the average voltage of the peripheral object with the same material property is multiplied by the corresponding distance weight factor.
According to the audio and video hybrid monitoring method, the target recognition is carried out on the peripheral object, the material attribute of the peripheral object is obtained, the evaluation function is obtained through the material attribute, the calculation is simple, the operability is high, the accuracy of the evaluation function is improved, and the selection accuracy of the first intermediate object is further improved.
Furthermore, a distance weight factor is introduced into the calculation of the evaluation function, so that the accuracy of the evaluation function is improved, and the selection accuracy of the first intermediate is further improved.
To make the calculation of the evaluation function more accurate and faster, please refer to fig. 6, fig. 6 is a schematic flow chart of obtaining the evaluation function according to the third embodiment of the present invention, it should be noted that, if there is substantially the same result, the method of the present invention is not limited to the flow sequence shown in fig. 6. As shown in fig. 6, the obtaining of the evaluation function of the peripheral object according to the recognition result of the target recognition may further include:
step S301: dividing the preset range into N pickup areas, wherein N is more than or equal to 1;
referring to fig. 7, fig. 7 is a schematic diagram illustrating dividing a sound pickup area according to a third embodiment of the present invention. In this embodiment, the preset range is set as the range of the sound pickup distance of the audio monitoring device, the sound pickup distance of the audio monitoring device is divided into N segments, the preset range is divided into N sound pickup regions, the schematic diagram of the divided sound pickup regions shown in fig. 7 may be established according to the types of the peripheral objects and the second coordinates corresponding to the peripheral objects, and the preset range is divided into the sound pickup regions [0, S1/N ], [ S1/N, 2S 1/N ] }. -. [ (N-1) × S1/N, S1], respectively.
Step S302: averaging the output voltages of the peripheral objects with the same material attribute in the ith pick-up area to obtain a first voltage of the peripheral objects with the same material attribute in the ith pick-up area, wherein i is more than or equal to 1 and less than or equal to N;
specifically, the output voltage generated after the peripheral object in the preset range is pressed by the sound source is recorded as
Figure BDA0002475816720000111
And the converted output voltage corresponding to the reflection of the laser energy back from the material property y in the ith pickup area is represented. If a plurality of peripheral objects with the same material attribute exist in the same sound-collecting area, averaging is carried out
Figure BDA0002475816720000112
Thereby acquiring the first voltage in the ith pickup region of the peripheral object having the same material property:
Figure BDA0002475816720000113
for example, when 5 cardboards with different sizes are identified in the monitoring scene within the distance range of [0, S1/N ], the first voltages corresponding to the cardboards are:
Figure BDA0002475816720000114
therefore, the average voltage of the peripheral objects having the same material property in the N sound pickup areas within the preset range may be obtained according to the first voltage, for example:
Figure BDA0002475816720000115
Figure BDA0002475816720000121
step S303: acquiring a region distance weight factor of the peripheral object with the same material property in the ith sound-picking region according to the distance weight factor;
similarly, to avoid that the peripheral object with a certain material property y having a large acoustic reactance may appear only once in a range close to the audio monitoring device, and the average voltage calculated in the above step is larger than other peripheral objects with a small acoustic reactance far from the audio monitoring device, which may cause a wrong selection, it is further necessary to obtain a region distance weight factor f (d) of the peripheral object with the same material property in the same sound pickup region. The calculation method of the region distance weighting factor f (d) is similar to that of step S201c in the second embodiment, and is not repeated here.
Step S304: and calculating evaluation functions of the peripheral objects with the same material attribute in the N sound-picking areas according to the first voltage and the area distance weighting factor.
Further, according to the law of constant laser power and energy conservation, the following can be obtained:
f(D1)*U1=f(D2)*U2=............=f(Dn)*Un
wherein U1 to Un are the average voltages of the peripheral objects with the same material property y in the distance segments of N1 to Nn, respectively, and f (D1) to f (dn) are the region distance weighting factors of the pickup region [0, S1/N ], [ S1/N,2 × S1/N ], [ (N-1)' S1/N, S1 ]. Since U1> U2.. ·> Un, the solved weight factor relationship is f (D1) < f (D2.......... · < f (dn)).
In summary, the evaluation function of the peripheral objects having the same material property y is obtained as follows:
Figure BDA0002475816720000122
wherein y represents the material property, such as cardboard, glass door/window, plastic bottle, tree, etc., as mentioned above.
In this embodiment, when the monitoring target moves, for example, from the N1-N2 sound pickup area to the N2-N3 sound pickup area, it is only necessary to switch to the N2-N3 sound pickup area to calculate the evaluation functions of the peripheral objects in the sound pickup area, and it is not necessary to traverse all the peripheral objects again.
According to the audio and video mixing monitoring method, the evaluation function is obtained by dividing the pickup area, and the speed of selecting the first intermediate object is increased.
In the above embodiment, an evaluation function of the peripheral object having the same material property is obtained by performing actual measurement through arranging a sound source, performing laser measurement, and the like in advance, please refer to fig. 8, 9, and 10, fig. 8 is a schematic flow chart of obtaining the evaluation function according to a fourth embodiment of the present invention, fig. 9 is a schematic model diagram of transmitting laser to the peripheral object by an audio monitoring device according to the fourth embodiment of the present invention, and fig. 10 is a schematic model diagram of receiving laser reflection by the audio monitoring device according to the fourth embodiment of the present invention, in this embodiment, evaluation functions of peripheral objects having different material properties may be obtained through theoretical calculation, which includes:
step S401: acquiring incident energy of laser emitted by the laser pickup to the peripheral object according to physical parameters of the laser pickup, wherein the physical parameters comprise a laser radiation angle, laser emission energy, laser wavelength and a laser transmission attenuation coefficient;
specifically, the laser is generated under the influence of particles such as dust in the air in the process of transmitting in the airDiffuse reflection and the like cause attenuation of laser energy in situ, and the attenuation of energy becomes more severe with increasing propagation distance. The audio monitoring equipment aims at the monitored peripheral object, laser radiation angle is theta, laser is emitted to the peripheral object with distance L from the audio monitoring equipment, and the size of laser spot generated when the laser strikes the peripheral object is SLight spotLaser emission energy of PLaserThe wavelength is λ, and in this embodiment, the commonly used laser monitoring wavelengths are 905nm and 1540 nm; the laser transmission attenuation coefficient phi (lambda) of the laser in the atmosphere is related to the wavelength, so that the transmittance Q (lambda) of the laser and the spot area S of the peripheral object are obtainedLight spotRespectively as follows:
q (λ) ═ exp (-L × Φ (λ)) formula 1
SLight spot=πL2tg2(theta/2) formula 2
Wherein, the distance L from the peripheral object to the audio monitoring device can be obtained by laser ranging:
L=Vlaser*tFlying[ 2 ] formula 3
In addition, the energy attenuation in the laser propagation process is related to the area of the light spots on the peripheral object and is inversely proportional to the area of the light spots on the peripheral object, so that the incident energy P of the peripheral object at the distance L from the audio monitoring equipment can be obtainedIncident lightComprises the following steps:
Pincident light=PLaser*Q(λ)/SLight spotFormula 4
The following formulas 1 to 4 can be obtained:
Figure BDA0002475816720000141
wherein, PLaser、VLaserThe variables are known, the time of flight t of the laser to said peripheral objectFlyingThe distance L between the peripheral object and the laser source is positively correlated, and the distance can be obtained through laser ranging.
Step S402: and acquiring the reflection energy of the peripheral object as the evaluation function according to the incident energy and the material property of the peripheral object.
Further, the incidence angle α of the laser hitting the peripheral object, the refraction angle β of the laser reflected by the peripheral object, the energy density ρ of the laser beam, and the spot area S obtained by projecting the laser beam after hitting the surface of the peripheral object are1I.e. incident beam area is S1Then the reflection area S2And S1Equal, refracted beam area S3(ii) a Incident energy of PIncident light. Then there are:
Figure BDA0002475816720000142
Pincident light=ρ*S1Formula 7
PReflection=PIncident light-PRefractionFormula 8
The reflected energy of the peripheral object can be obtained by combining the formulas 6-8:
Figure BDA0002475816720000143
further, according to the law of refraction:
Figure BDA0002475816720000144
nintermediary articleThe refractive index of the material property of the peripheral object, since the refractive index of light in air is approximately 1, nAir (a)1, the reflection energy of the peripheral object can be obtained as follows:
Figure BDA0002475816720000151
in the above formula 11, PIncident lightThe result is shown in formula 5 in step S401, and for all the peripheral objects in the preset range, the pan/tilt head 1 drives the audio monitoring device to rotate and emit laser beams aiming at the peripheral objects, so that the incident angle α of the laser beams irradiating the peripheral objects is substantially fixedThe refractive index of nature is also empirical.
Therefore, the evaluation function formula of the laser reflection energy can be obtained as follows:
Figure BDA0002475816720000152
for a given audio monitoring system, the main variable in the above equation is tFlyingAnd nMedium. Wherein t isFlyingIs related to the distance from the surrounding object to the audio monitoring device and has obtained t when ranging the surrounding objectFlyingValue, and nMediumThe evaluation function may be obtained based on the material properties of the peripheral object, which are related to refractive indices of different material properties, for example, a refractive index of glass is 1.5.
It should be noted that the above-mentioned methods of obtaining the evaluation function through actual measurement and obtaining the evaluation function through theoretical calculation can also mutually check the accuracy of the calculation of the evaluation function.
According to the audio and video mixing monitoring method, the evaluation function is calculated theoretically, the accuracy of the evaluation function is improved, and the selection accuracy of the first intermediate is further improved.
After the evaluation function is obtained in the foregoing embodiments, to quickly and accurately select the first intermediary object, in another embodiment, the obtaining the first intermediary object according to the evaluation function may include:
establishing an intermediary database, storing the evaluation functions of the peripheral objects according to the material attributes, and arranging and storing the evaluation functions according to the sequence of the function values of the evaluation functions from large to small;
obtaining a medium object set of the peripheral objects within a preset range from the monitoring target; obtaining the first interposer arranged in a first bit according to the material property of the set of interposers. When a plurality of the peripheral objects with the same material property exist in the preset range, the peripheral object with the minimum distance from the monitored target can be selected as a first intermediate object.
By establishing the intermediary database, storing the evaluation function into the database and directly selecting the first intermediary from the intermediary database when selecting the first intermediary, the speed of selecting the first intermediary is further improved.
Fig. 11 is a schematic structural diagram of an audio/video mixing monitoring apparatus according to an embodiment of the present invention. As shown in fig. 11, the apparatus includes a video monitoring module 11, an audio monitoring module 12, and a control module 13, where the video monitoring module 11, the audio monitoring module 12, and the control module 13 may communicate with each other, and the control module 13 may be disposed on the video monitoring module 11 or the audio monitoring module 12, or may be disposed on a server outside the video monitoring module 11 and the audio monitoring module 12.
The video monitoring module 11 is used for performing video monitoring.
The audio monitoring module 12 is used for audio monitoring.
The control module 13 is configured to acquire a monitoring target in a monitoring image captured by a video monitoring device, and control the video monitoring device to perform video monitoring on the monitoring target;
optionally, the control module 13 may also be configured to acquire a peripheral object within a monitoring range and perform target identification on the peripheral object; and obtaining an evaluation function of the peripheral object according to the identification result, obtaining a first intermediate object according to the evaluation function, and controlling the audio monitoring equipment to perform audio monitoring on the first intermediate object.
It can be understood that specific ways for implementing the functions of the modules of the audio/video hybrid monitoring device may refer to specific steps corresponding to the above embodiments, and therefore, the detailed description thereof is omitted here.
Referring to fig. 12, fig. 12 is a schematic structural diagram of an audio/video mixing monitoring device according to an embodiment of the present invention. As shown in fig. 12, the av mixing monitor 60 includes a processor 61 and a memory 62 coupled to the processor 61.
The memory 62 stores program instructions for implementing the audio/video mixing monitoring method according to any of the above embodiments.
The processor 61 is used for executing the program instructions stored in the memory 62 to perform audio-video hybrid monitoring on the monitoring target.
The processor 61 may also be referred to as a CPU (Central Processing Unit). The processor 61 may be an integrated circuit chip having signal processing capabilities. The processor 61 may also be a general purpose processor, a Digital Signal Processor (DSP), an Application Specific Integrated Circuit (ASIC), a Field Programmable Gate Array (FPGA) or other programmable logic device, discrete gate or transistor logic, discrete hardware components. A general purpose processor may be a microprocessor or the processor may be any conventional processor or the like.
Referring to fig. 13, fig. 13 is a schematic structural diagram of a memory device according to an embodiment of the invention. The storage device of the embodiment of the present invention stores a program file 71 capable of implementing all the above-mentioned audio/video hybrid monitoring methods, where the program file 71 may be stored in the storage device in the form of a software product, and includes several instructions to enable a computer device (which may be a personal computer, a server, or a network device, etc.) or a processor (processor) to execute all or part of the steps of the methods according to the embodiments of the present application. The aforementioned storage device includes: various media capable of storing program codes, such as a usb disk, a mobile hard disk, a Read-Only Memory (ROM), a Random Access Memory (RAM), a magnetic disk or an optical disk, or terminal devices, such as a computer, a server, a mobile phone, and a tablet.
In the several embodiments provided in the present application, it should be understood that the disclosed system, apparatus and method may be implemented in other manners. For example, the above-described apparatus embodiments are merely illustrative, and for example, a division of a unit is merely a logical division, and an actual implementation may have another division, for example, a plurality of units or components may be combined or integrated into another system, or some features may be omitted, or not executed. In addition, the shown or discussed mutual coupling or direct coupling or communication connection may be an indirect coupling or communication connection through some interfaces, devices or units, and may be in an electrical, mechanical or other form.
In addition, functional units in the embodiments of the present invention may be integrated into one processing unit, or each unit may exist alone physically, or two or more units are integrated into one unit. The integrated unit can be realized in a form of hardware, and can also be realized in a form of a software functional unit. The above embodiments are merely examples and are not intended to limit the scope of the present disclosure, and all modifications, equivalents, and flow charts using the contents of the specification and drawings of the present disclosure or those directly or indirectly applied to other related technical fields are intended to be included in the scope of the present disclosure.

Claims (14)

1. An audio and video mixing monitoring method is characterized by comprising the following steps:
acquiring a monitoring target in a monitoring image shot by video monitoring equipment, and carrying out video monitoring on the monitoring target;
acquiring peripheral objects within a preset range from the monitored target, and identifying the target;
obtaining an evaluation function of the peripheral object according to the identification result of the target identification, wherein the evaluation function is used for evaluating the energy of the object reflected audio signal;
and acquiring a first intermediate object according to the evaluation function, and controlling audio monitoring equipment to perform audio monitoring on the first intermediate object.
2. The audio-video mixing monitoring method according to claim 1, wherein the recognition result includes material properties of the peripheral object.
3. The audio-video mixing monitoring method according to claim 2, wherein the obtaining of the evaluation function of the peripheral object according to the recognition result of the target recognition comprises:
acquiring output voltage generated after the peripheral object in the preset range is pressed by a sound source and distance weight factors of the peripheral object, wherein the distance weight factors represent weights of audio signal energy reflected by objects at different distances from the audio monitoring equipment;
and calculating the evaluation function of the peripheral objects with the same material attribute according to the output voltage and the distance weight factor.
4. The audio-video mixing monitoring method according to claim 3, wherein before calculating the evaluation function of the peripheral objects having the same material property according to the output voltage and the distance weighting factor, the method further comprises:
dividing the preset range into N pickup areas, wherein N is more than or equal to 1;
averaging the output voltages of the peripheral objects with the same material attribute in the ith pick-up area to obtain a first voltage of the peripheral objects with the same material attribute in the ith pick-up area, wherein i is more than or equal to 1 and less than or equal to N;
acquiring a region distance weight factor of the peripheral object with the same material property in the ith sound-picking region according to the distance weight factor;
and calculating evaluation functions of the peripheral objects with the same material attribute in the N sound-picking areas according to the first voltage and the area distance weighting factor.
5. The audio-video mixing monitoring method according to claim 3, wherein the obtaining of the distance weighting factor of the peripheral object comprises:
acquiring echo energy of the peripheral object in the preset range;
acquiring the monitoring distance between the peripheral object and the audio monitoring equipment in the preset range;
and acquiring the distance weight factor of the peripheral object according to the monitoring distance and the echo energy.
6. The audio-video mixing monitoring method according to claim 2, wherein the obtaining of the evaluation function of the peripheral object according to the recognition result of the target recognition comprises:
acquiring incident energy of laser emitted by the laser pickup to the peripheral object according to physical parameters of the audio monitoring equipment, wherein the physical parameters comprise a laser radiation angle, laser emission energy, laser wavelength and a laser transmission attenuation coefficient;
and acquiring the reflection energy of the peripheral object as the evaluation function according to the incident energy and the material property of the peripheral object.
7. The audio-video mixing monitoring method according to claim 2, wherein the obtaining a first mediator according to the evaluation function includes:
establishing an intermediary database, storing the evaluation functions of the peripheral objects according to the material attributes, and arranging and storing the evaluation functions according to the sequence of the function values of the evaluation functions from large to small;
obtaining a medium object set of the peripheral objects within the preset range;
obtaining the first interposer arranged in a first bit according to the material property of the set of interposers.
8. The audio-video mixing monitoring method according to claim 1, wherein controlling the audio monitoring device to perform audio monitoring on the first interposer includes:
uniformly dividing the first interposer into M vibration areas;
obtaining vibration voltage generated after each vibration area is pressed by a sound source;
selecting the vibration region with the largest voltage value in the vibration voltages of the M vibration regions as a monitoring region;
and controlling the audio monitoring equipment to perform audio monitoring on the monitoring area.
9. The audio-video mixing monitoring method according to claim 8, wherein after controlling the audio monitoring device to perform audio monitoring on the first intermediary, the method further comprises:
restoring the vibration voltage of the monitoring area into sound information;
acquiring a voiceprint database, and extracting target sound from the voiceprint database according to the sound information;
and carrying out audio monitoring on the target sound.
10. The audio-video mixing monitoring method according to claim 9, wherein the obtaining a voiceprint database comprises:
acquiring first sound information of the monitoring target in different scenes;
filtering the noise and silence of the first sound information to obtain second sound information;
and acquiring sound characteristics according to the second sound information, performing learning training, and storing the sound characteristics to the voiceprint database.
11. The audio-video hybrid monitoring method according to claim 1, wherein the video monitoring device comprises a panoramic camera and a dome camera, and the video monitoring of the monitored object further comprises:
acquiring the monitoring target from a monitoring image shot by the panoramic camera;
and controlling the dome camera to carry out video monitoring on the monitored target according to the monitored target.
12. The audio-video mixing monitoring method according to claim 1, further comprising:
acquiring a pickup distance of the audio monitoring equipment and a first distance from the audio monitoring equipment to the monitoring target;
when the first distance is smaller than the pickup distance, performing video monitoring by using the video monitoring equipment, and performing audio monitoring by using the audio monitoring equipment;
and when the first distance is greater than the pickup distance, using the video monitoring equipment to perform video monitoring.
13. An audio-video mixing monitoring device, the audio-video mixing monitoring device comprising a processor, a memory coupled to the processor, wherein,
the memory stores program instructions for implementing an audio-video mixing monitoring method as claimed in any one of claims 1-12;
the processor is configured to execute the program instructions stored in the memory to perform audio and video surveillance of a surveillance target.
14. A storage device, characterized in that a program file capable of implementing the audio-video mixing monitoring method according to any one of claims 1 to 12 is stored.
CN202010363581.0A 2020-04-30 2020-04-30 Audio and video mixing monitoring method, equipment and storage device Active CN111526292B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010363581.0A CN111526292B (en) 2020-04-30 2020-04-30 Audio and video mixing monitoring method, equipment and storage device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010363581.0A CN111526292B (en) 2020-04-30 2020-04-30 Audio and video mixing monitoring method, equipment and storage device

Publications (2)

Publication Number Publication Date
CN111526292A true CN111526292A (en) 2020-08-11
CN111526292B CN111526292B (en) 2021-08-03

Family

ID=71912083

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010363581.0A Active CN111526292B (en) 2020-04-30 2020-04-30 Audio and video mixing monitoring method, equipment and storage device

Country Status (1)

Country Link
CN (1) CN111526292B (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112383720A (en) * 2020-11-11 2021-02-19 杭州海康威视数字技术股份有限公司 Image data processing equipment and image data processing method

Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6317237B1 (en) * 1997-07-31 2001-11-13 Kyoyu Corporation Voice monitoring system using laser beam
CN103163527A (en) * 2013-04-02 2013-06-19 哈尔滨工业大学 Laser monitoring device based on charge-coupled device (CCD) auxiliary positioning and double-optical wedge light beam regulation and monitoring method by using same
US20150326968A1 (en) * 2014-05-08 2015-11-12 Panasonic Intellectual Property Management Co., Ltd. Directivity control apparatus, directivity control method, storage medium and directivity control system
CN107274910A (en) * 2017-05-17 2017-10-20 宁波桑德纳电子科技有限公司 The supervising device and audio/video linkage method of a kind of audio/video linkage
CN109429031A (en) * 2017-08-21 2019-03-05 中兴通讯股份有限公司 Infrared monitoring device, corresponding infrared light method of adjustment and storage medium
CN110176117A (en) * 2019-06-17 2019-08-27 广东翔翼科技信息有限公司 A kind of monitoring device and monitoring method of Behavior-based control identification technology
CN110347187A (en) * 2019-08-09 2019-10-18 北京机械设备研究所 A kind of target detection tracing System and method for based on sound and image information
CN110636262A (en) * 2019-09-17 2019-12-31 北京声智科技有限公司 Monitoring method and monitoring device
CN111006754A (en) * 2019-12-31 2020-04-14 杭州爱莱达科技有限公司 Scanning type multi-channel coherent laser pickup method and device
CN111050130A (en) * 2019-12-12 2020-04-21 深圳市大拿科技有限公司 Camera control method and device and storage medium

Patent Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6317237B1 (en) * 1997-07-31 2001-11-13 Kyoyu Corporation Voice monitoring system using laser beam
CN103163527A (en) * 2013-04-02 2013-06-19 哈尔滨工业大学 Laser monitoring device based on charge-coupled device (CCD) auxiliary positioning and double-optical wedge light beam regulation and monitoring method by using same
US20150326968A1 (en) * 2014-05-08 2015-11-12 Panasonic Intellectual Property Management Co., Ltd. Directivity control apparatus, directivity control method, storage medium and directivity control system
CN107274910A (en) * 2017-05-17 2017-10-20 宁波桑德纳电子科技有限公司 The supervising device and audio/video linkage method of a kind of audio/video linkage
CN109429031A (en) * 2017-08-21 2019-03-05 中兴通讯股份有限公司 Infrared monitoring device, corresponding infrared light method of adjustment and storage medium
CN110176117A (en) * 2019-06-17 2019-08-27 广东翔翼科技信息有限公司 A kind of monitoring device and monitoring method of Behavior-based control identification technology
CN110347187A (en) * 2019-08-09 2019-10-18 北京机械设备研究所 A kind of target detection tracing System and method for based on sound and image information
CN110636262A (en) * 2019-09-17 2019-12-31 北京声智科技有限公司 Monitoring method and monitoring device
CN111050130A (en) * 2019-12-12 2020-04-21 深圳市大拿科技有限公司 Camera control method and device and storage medium
CN111006754A (en) * 2019-12-31 2020-04-14 杭州爱莱达科技有限公司 Scanning type multi-channel coherent laser pickup method and device

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112383720A (en) * 2020-11-11 2021-02-19 杭州海康威视数字技术股份有限公司 Image data processing equipment and image data processing method
CN112383720B (en) * 2020-11-11 2023-03-07 杭州海康威视数字技术股份有限公司 Image data processing device and image data processing method

Also Published As

Publication number Publication date
CN111526292B (en) 2021-08-03

Similar Documents

Publication Publication Date Title
US11412108B1 (en) Object recognition techniques
Sami et al. Spying with your robot vacuum cleaner: eavesdropping via lidar sensors
JP7062878B2 (en) Information processing method and information processing equipment
US10339913B2 (en) Context-based cancellation and amplification of acoustical signals in acoustical environments
WO2022036980A1 (en) Pose determination method and apparatus, electronic device, storage medium, and program
CN108668077A (en) Camera control method, device, mobile terminal and computer-readable medium
CN113281706B (en) Target positioning method, device and computer readable storage medium
CN110278383A (en) Focus method, device and electronic equipment, storage medium
KR20140144410A (en) Beamforming method and apparatus for sound signal
CN110688957A (en) Living body detection method and device applied to face recognition and storage medium
CN111368811B (en) Living body detection method, living body detection device, living body detection equipment and storage medium
CN108737934B (en) Intelligent sound box and control method thereof
CN110876036B (en) Video generation method and related device
CN105208287A (en) Photographing method and device
CN109145150A (en) target matching method and device, electronic equipment and storage medium
US9558563B1 (en) Determining time-of-fight measurement parameters
CN111526292B (en) Audio and video mixing monitoring method, equipment and storage device
Zhang et al. Ambiear: mmwave based voice recognition in nlos scenarios
CN112423191A (en) Video call device and audio gain method
US10921446B2 (en) Collaborative mapping of a space using ultrasonic sonar
CN111208970B (en) Audio playing method and audio playing device
CN115713616B (en) House source space model generation method and device, terminal equipment and storage medium
CN112202655A (en) Intelligent electric appliance, image recognition method, electronic device and storage medium
CN108600623B (en) Refocusing display method and terminal device
WO2023164814A1 (en) Media apparatus and control method and device therefor, and target tracking method and device

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant