WO2019184275A1 - Procédé, dispositif et système de traitement d'image - Google Patents

Procédé, dispositif et système de traitement d'image Download PDF

Info

Publication number
WO2019184275A1
WO2019184275A1 PCT/CN2018/106752 CN2018106752W WO2019184275A1 WO 2019184275 A1 WO2019184275 A1 WO 2019184275A1 CN 2018106752 W CN2018106752 W CN 2018106752W WO 2019184275 A1 WO2019184275 A1 WO 2019184275A1
Authority
WO
WIPO (PCT)
Prior art keywords
image
video frame
label
collection device
tag
Prior art date
Application number
PCT/CN2018/106752
Other languages
English (en)
Chinese (zh)
Inventor
金海善
林圣拿
何溯
杨俊�
Original Assignee
杭州海康威视系统技术有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from CN201810272370.9A external-priority patent/CN109274926B/zh
Application filed by 杭州海康威视系统技术有限公司 filed Critical 杭州海康威视系统技术有限公司
Publication of WO2019184275A1 publication Critical patent/WO2019184275A1/fr

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T19/00Manipulating 3D models or images for computer graphics
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/18Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast

Definitions

  • the present application relates to the field of video surveillance technologies, and in particular, to an image processing method, device, and system.
  • image acquisition devices are provided in many scenes, and related personnel can monitor the scene through video frame images collected by the device.
  • the display content only includes the image itself and the acquisition time of the image.
  • the user who views the video frame image he can only familiarize himself with the real environment corresponding to the video frame image, and understand the specific content contained in the image based on the real environment. It can be seen that this image display method is not intuitive and the display effect is poor.
  • An object of the embodiments of the present application is to provide an image processing method, device, and system, which improve the display effect of a video frame image.
  • an image processing method including:
  • the video frame image after the tag is displayed according to the preset display rule.
  • the video frame image is a panoramic image
  • the first collecting device is configured to correspond to at least one second collecting device
  • the second collecting device performs image capturing on the sub-scene corresponding to the panoramic image
  • the method further includes:
  • the step of determining at least one target location in the video frame image includes:
  • the first collection device is an augmented reality AR panoramic camera.
  • the step of generating a label according to the image of the sub-scene includes:
  • the step of adding the target information in the sub-scene image to the content of the label includes:
  • Identifying the sub-scene image determining target information in the sub-scene image according to the recognition result; adding the target information to the content of the label;
  • the step of displaying the tagged video frame image according to the preset display rule includes:
  • the content of the added tag is displayed.
  • the step of displaying the tagged video frame image according to the preset display rule includes:
  • the video frame image after the tag is added, and the content of the added tag.
  • display the contents of the added tags including:
  • the method further includes:
  • the clicked label is determined as the target label
  • the content of the target tag is displayed in the video frame image.
  • the method before the step of determining the at least one target location in the video frame image, the method further includes:
  • the step of determining at least one target location in the video frame image includes:
  • a target location of the added tag is determined according to the tag addition instruction.
  • the step of displaying the tagged video frame image according to the preset display rule includes:
  • Determining a layer display strategy and determining, according to the layer display strategy, a current display layer and a display manner of the current display layer;
  • the label corresponding to the current display layer is displayed.
  • the step of acquiring the sub-scene image collected by the second collection device includes:
  • the step of generating a label includes:
  • the step of detecting whether an abnormal event occurs in the panoramic image comprises:
  • the step of determining the target second collection device corresponding to the abnormal event includes:
  • the method further includes:
  • the step of displaying the tagged video frame image according to the preset display rule includes:
  • the label is displayed in the video frame image in a preset alarm mode.
  • an embodiment of the present application further discloses an image processing apparatus, including: a processor and a memory;
  • a memory for storing a computer program
  • the processor when used to execute the program stored on the memory, implements the following steps:
  • the video frame image after the tag is displayed according to the preset display rule.
  • the video frame image is a panoramic image
  • the first collecting device is configured to correspond to at least one second collecting device
  • the second collecting device performs image capturing on the sub-scene corresponding to the panoramic image
  • the processor is further configured to implement the following steps:
  • processor is further configured to implement the following steps:
  • processor is further configured to implement the following steps:
  • Identifying the sub-scene image determining target information in the sub-scene image according to the recognition result; adding the target information to the content of the label;
  • processor is further configured to implement the following steps:
  • the content of the added tag is displayed.
  • processor is further configured to implement the following steps:
  • the video frame image after the tag is added, and the content of the added tag.
  • processor is further configured to implement the following steps:
  • processor is further configured to implement the following steps:
  • the clicked label is determined as the target label
  • the content of the target tag is displayed in the video frame image.
  • processor is further configured to implement the following steps:
  • a target location of the added tag is determined according to the tag addition instruction.
  • processor is further configured to implement the following steps:
  • Determining a layer display strategy and determining, according to the layer display strategy, a current display layer and a display manner of the current display layer;
  • the label corresponding to the current display layer is displayed.
  • processor is further configured to implement the following steps:
  • processor is further configured to implement the following steps:
  • processor is further configured to implement the following steps:
  • processor is further configured to implement the following steps:
  • the label is displayed in the video frame image in a preset alarm mode.
  • an embodiment of the present application further discloses an image processing system, including: a first collection device and an image processing device, where
  • the first collecting device is configured to collect a video frame image, and send the collected video frame image to the image processing device;
  • the image processing device is configured to determine, according to a video frame image acquired by the first collection device, at least one target location in the video frame image; add a label at each determined target location, the label is based on user input
  • the content or the image acquired by the second collection device is generated; and the video frame image after the tag is added is displayed according to the preset display rule.
  • the system further includes: at least one second collection device,
  • the second collection device is configured to perform image collection on a sub-scene corresponding to the panoramic image, where the panoramic image is a video frame image collected by the first collection device;
  • the image processing device is further configured to acquire a sub-scene image acquired by the second collection device; generate a label according to the sub-scene image; and determine, according to the calibration information of the first collection device and the second collection device acquired in advance The label corresponding to the second collection device is at a target position in the panoramic image.
  • the first collection device is an augmented reality AR panoramic camera.
  • an embodiment of the present application further discloses a computer readable storage medium, where the computer readable storage medium stores a computer program, and when the computer program is executed by a processor, implement any one of the foregoing image processing methods. .
  • an embodiment of the present application further discloses an executable program code for being executed to execute any of the image processing methods described above.
  • the label can help the user understand the specific content included in the video frame image, and therefore, adding the tagged video
  • the frame image can display the image content more intuitively, and the display effect is better.
  • FIG. 1 is a schematic diagram of a first process of an image processing method according to an embodiment of the present disclosure
  • FIG. 1 is a schematic diagram of a display interface according to an embodiment of the present application.
  • FIG. 1b is a schematic diagram of another display interface provided by an embodiment of the present application.
  • FIG. 2 is a second schematic flowchart of an image processing method according to an embodiment of the present application.
  • FIG. 2a is a schematic diagram of an application scenario provided by an embodiment of the present application.
  • FIG. 3 is a schematic diagram of a third process of an image processing method according to an embodiment of the present disclosure.
  • FIG. 4 is a schematic structural diagram of an image processing apparatus according to an embodiment of the present application.
  • FIG. 4b is a schematic structural diagram of another image processing device according to an embodiment of the present disclosure.
  • FIG. 5 is a schematic structural diagram of an image processing system according to an embodiment of the present application.
  • an embodiment of the present application provides an image processing method, device, and system.
  • the method can be applied to various image processing devices, and is not specifically limited.
  • FIG. 1 is a schematic flowchart of an image processing method according to an embodiment of the present disclosure, including:
  • S101 Determine, for the video frame image acquired by the first collection device, at least one target location in the video frame image.
  • the processing object of the embodiment of the present application is a video frame image, and the image provided by the embodiment of the present application may be processed for each frame of the video.
  • the target location there are several ways to determine the target location. For example, one or more fixed positions may be set in advance in the video frame image as the target position; for example, the intermediate position of the video frame image may be set as the target position.
  • the user-specified location may be determined as the target location according to the user instruction. It should be noted that, for the same video, or multiple videos of the same scene, the user may only send an instruction once, according to the instruction, may be in this paragraph or The target position is determined in each of the plurality of pieces of video.
  • the installation location of the first collection device is generally fixed, and the scene corresponding to the captured video frame image is also substantially unchanged. Therefore, in each video frame image, the difference in the screen content corresponding to the preset position is usually The difference in the screen content corresponding to the position specified in the above user command is usually not large, and the target position can be determined in the plurality of video frame images according to an instruction sent by the user.
  • the way to determine the target location can also be other, and is not limited.
  • S102 Add a label at each determined target position, and the label is generated according to an input instruction or an image acquired by the second collection device.
  • the input command can be a tagged instruction entered by the user.
  • the second collection device may be an acquisition device that is configured in the same scenario as the first collection device. For example, the first collection device performs image collection for the scene A, and the second collection device performs image collection for the sub-scenario A1 in the scene A.
  • the label may include a "tag symbol” and a "tag content”.
  • the "tag symbol” may be an arrow, a triangle, etc.
  • the "tag symbol” is for marking a position in the video frame image.
  • the specific format of the label is not limited; the content of the label may be an image collected by other collection devices, or may be some image analysis data, or may be associated data of the scene at the label, and the like, and is not limited.
  • the image analysis data may be a face recognition result, a vehicle recognition result, or the like
  • the associated data of the scene may be an introduction content of the scene, or if the scene is a traffic bayonet, the associated data may be traffic flow data or the like.
  • the tag may also include a "tag name", for example, may be some simple text information, such as "some building", “some park” and the like.
  • the input instruction is the text information “some building” input by the user and the specific introduction of the building
  • a label may be generated, and the label symbol may be an arrow, and the label name may be the text information “some building”, the label
  • the content can be a specific introduction to the building.
  • the target location is a traffic bayonet
  • the label content added at the traffic bayonet may be video data collected at the bayonet, a captured image at the bayonet, traffic flow data at the bayonet, and the like.
  • the user can design his own label according to his own needs. Specifically, the user may click on a certain position in the video frame image and input some text or image content; the device performing the scheme may generate a corresponding label according to the content input by the user, and after the video frame image and the subsequent In the video frame image, the location clicked by the user is determined as the target location, and the generated tag is added at the target location.
  • the label may be generated according to the image collected by the other collection device.
  • the first collection device performs image collection for the scene A
  • the second collection device performs image collection for the sub-scenario A1 in the scene A.
  • the tag may be generated according to the image collected by the second collecting device, and the position corresponding to the sub-scene A1 is determined as the target position in S101, and the tag corresponding to the sub-scene A1 is added at the target position.
  • the label added in the video frame image includes both the label generated according to the user's needs and the label generated according to the image collected by other collection devices, so that the label type is more abundant.
  • the tag may include a "tag symbol” and a "tag content”.
  • the "tag symbol” and the “tag content” may be separately displayed.
  • the "tag symbol” may be added to the video frame.
  • the "tag content” is displayed in an area other than the video frame image, so that the content of the tag does not cover the video frame image, and the display effect is better.
  • the tag further includes a "tag name”
  • the "tag name” may be displayed in the video frame image, or may be displayed in an area outside the video frame image, which is not limited.
  • the video frame image after the tag is added may be displayed, and in the second area, the content of the added tag may be displayed.
  • the first area and the second area may be different areas of the same display device, or may be a display area in an adjacent display device, which is not limited.
  • the video frame image after adding the label and the content of the added label are displayed.
  • the video frame image after the tag is added may be displayed in the main screen area, and the content of the added tag is displayed in the small screen area.
  • the small screen area may be located at any position on the right side, the left side, the upper side, and the lower side of the main screen area, and is not limited.
  • the "content of the tag” can be of various types, such as video data, captured images, image analysis data, and the like, and different types of data can be displayed in different areas.
  • the video data and the captured image may be displayed in the small screen area or the second area in the above picture
  • the image analysis data may be displayed in the video frame image, etc., and the specific display manner is not limited.
  • the specific shape, color, transparency, and specific type of "tag content” of the "tag symbol” may be set in advance or may be changed according to user selection.
  • the current display label may be determined; and the content of the current display label is displayed.
  • the display order can be set, and the current display label is determined according to the order.
  • the display order can be determined randomly, or can be set according to the importance degree of each label.
  • the label corresponding to the display instruction is determined as the current display label, and the like, and is not limited.
  • the clicked tag may be determined as the target tag; the content of the target tag is displayed in the video frame image.
  • the content of the label can be directly displayed in the video frame image.
  • a layer classification policy may be preset, and according to the policy, a layer category corresponding to each label is determined. In other words, it is to divide each label into different layer categories. For example, you can divide labels into intersection label layers, bayonet label layers, area label layers, building label layers, and more.
  • the layer display strategy can be determined based on user instructions.
  • the layer display strategy can include the current display layer and how the current display layer is displayed.
  • the user instruction only includes the current display layer information, and the device determines the current display layer according to the user instruction.
  • the device stores the display manner corresponding to each layer, so that the device can further determine the current display image.
  • the display mode of the layer in the second case, the user instruction includes the current display layer information and the display mode information, and the device can determine the current display layer and the display mode of the current display layer according to the user instruction, which are all reasonable.
  • the display mode may include: flashing display, jitter display, static display, etc., and is not limited.
  • the label is displayed separately from the content of the label, and the display manner may include the manner in which the label is displayed, or the manner in which the label content is displayed, for example, a building label.
  • the corresponding display manner of the layer may be: the label is displayed in the video frame image, and the corresponding label content is flashed in other areas (the second area or the picture-in-picture area).
  • the detailed image corresponding to the video frame image collected by the first collection device may be acquired, and after S101, according to the pixel point correspondence between the detail image and the video frame image acquired in advance, Determining that the target location corresponds to the location in the detail image as the to-be-processed location; adding the label added at the target location to the to-be-processed location corresponding to the target location; in this embodiment, S103 may include: The video frame image after the tag is added and the detailed image after the tag is displayed according to the preset display rule.
  • the video frame image acquired in S101 may be a panoramic image, and in addition, a detailed image corresponding to the panoramic image may be acquired, and the panoramic view is obtained according to a pixel point correspondence relationship between the panoramic image and the detailed image.
  • the label added to the image corresponds to the detail image, and the label is also added in the detail image.
  • the third collection device may be disposed outside the first collection device, where the first collection device and the third collection device perform image collection for the same scene, the first collection device collects the panoramic image, and the third collection device collects the detailed image.
  • the third collecting device can be a ball machine, the ball machine can be rotated, and detailed images of different viewing angles can be collected.
  • the pixel point correspondence between the panoramic image and the detail image may be obtained according to calibration information between the first collection device and the third collection device.
  • the dome camera can collect detailed images corresponding to the four regions, which are the detail image B1 and the detail image B2, respectively.
  • the four detail images can be displayed in turn in a preset order.
  • the currently displayed detail image is B1
  • 10 target positions are determined in area 1
  • labels are added for the 10 target positions, correspondingly, there are also 10 pending positions in the detail image B1
  • Add the same 10 labels to these 10 pending locations since the number of tags is large, only a part of the tags may be displayed in the area 1 of the panoramic image A, and the 10 tags are displayed in the detail image B1.
  • the video frame image after the label is added may be displayed in the first area, and the detailed image after the label is added may be displayed in the third area; or, the label may be displayed in the form of picture-in-picture The video frame image and the detailed image after the tag is added.
  • the display label described here only displays the "tag symbol” and displays the "tag content” in another area.
  • the image of the added video frame may be displayed, in the second area, the content of the added label is displayed, and in the third area, the detailed image after the label is added.
  • the first area, the second area, and the third area mentioned herein may be different areas of the same display device, or may be display areas in different display devices.
  • the video frame image after the tag is added, the detailed image after the tag is added, and the content of the added tag can be displayed in the form of picture-in-picture.
  • the video frame image after adding the label is displayed in the main screen area
  • the detailed image after adding the label is displayed in the small screen area in the lower left corner
  • the content of the added label is displayed in the small screen area on the right side.
  • the "tag name” may be displayed in the video frame image, or may be displayed in an area outside the video frame image, which is not limited.
  • FIG. 2 is a second schematic flowchart of an image processing method according to an embodiment of the present disclosure. The embodiment shown in FIG. 2 is based on the embodiment shown in FIG.
  • S201 Acquire a sub-scene image collected by the second collection device.
  • the video frame image collected by the first collection device is a panoramic image
  • the first collection device corresponds to at least one second collection device
  • the second collection device performs image collection on the sub-scene corresponding to the panoramic image.
  • the image collected by the second collection device is a sub-scene image.
  • the first collection device may be an augmented reality AR panoramic camera, so that the collected panoramic image is better.
  • the first collecting device may also be a plurality of guns, and the images collected by the plurality of guns are spliced to obtain a panoramic image.
  • the second collection device can be an ordinary camera, such as a ball machine, a capture machine, and the like. If the second collection device is a dome camera, the sub-scene image may be a surveillance video image. If the second collection device is a capture camera, the sub-scene image may be a snapshot image, and the like, which is not limited.
  • a large scene A includes four sub-scenes: A1, A2, A3, and A4.
  • the first collection device performs image collection on scene A
  • the second collection device 1 performs A1 on A1.
  • Image acquisition the second collection device 2 performs image acquisition on A2
  • the second collection device 3 performs image acquisition on A3, and the second collection device 4 performs image acquisition on A4.
  • the first collection device and the second collection device may be the same device, such as an AR eagle eye device, and the AR eagle eye device has an augmented reality function.
  • the AR eagle eye device may be integrated with a plurality of camera lenses and one In the ball machine lens, the image obtained by splicing the plurality of camera lenses can be used as a panoramic image, and the image captured by the camera lens is used as a sub-scene image.
  • the AR Hawkeye device can also be provided with a platform for scheduling and managing the plurality of camera lenses and a dome camera lens.
  • the second collection device sends the collected sub-scene image to the device that executes the solution in real time.
  • the device that executes the solution acquires the sub-scene image from the second collection device after receiving the user instruction.
  • the device that executes the solution acquires the sub-scene image from the second collection device corresponding to the abnormal event after detecting an abnormal event in the video frame image (panoramic image) of the S101.
  • the abnormal event may be a traffic accident, a robbery event, etc., and is not limited.
  • the embodiment of the present application does not limit the timing of acquiring a sub-scene image.
  • the label may include a "tag symbol” and a "tag content”.
  • the "tag symbol” may be an arrow, a triangle, etc., and the "tag symbol” is for marking a position in the video frame image.
  • the label the specific form is not limited; the "content of the label” may include the sub-scene image.
  • the tag may also include a "tag name", for example, may be some simple text information, such as "some building", “some park” and the like.
  • the sub-scene image and/or the target information in the sub-scene image may be added to the content of the tag.
  • the tag contains the target information in the sub-scene image.
  • the target information may include vehicle information in the image, such as a license plate number, a vehicle body color, etc., and may also include road information, such as traffic flow in the road; or
  • vehicle information in the image such as a license plate number, a vehicle body color, etc.
  • road information such as traffic flow in the road
  • the target information may be abnormal event information, such as a traffic accident.
  • the target information may be character information in the image, such as height, gender, etc.; or, in the third scheme, the target information may be abnormal event information, such as Robbery, fire, etc.
  • the method for obtaining the target information is different.
  • the device that executes the solution may identify the sub-scene image acquired by the S 201, and determine the target information in the sub-scene image according to the recognition result;
  • the second collection device may have an image recognition function, and the second collection device sends the identified target information to the device;
  • the server connected to the second collection device identifies the sub-scene image, and The identified target information is sent to the device; these methods are reasonable.
  • the tag contains both the sub-scene image and the target information in the sub-scene image.
  • the target information can be understood as an introduction or description of the sub-scene image, and the target information can be set around the sub-scene image so that the user can better understand what is happening in the sub-scene image.
  • S101 may be S101A: determining, according to the calibration information of the first collection device and the second collection device that are acquired in advance, a target position of the label corresponding to the second collection device in the panoramic image.
  • the calibration relationship can be understood as a relationship between the panoramic image coordinate system and the sub-scene image coordinate system. Conversion relationship. For example, there is a position X in the sub-scene A1, the pixel coordinate point of the position X in the panoramic image is (x1, y1), and the pixel coordinate point in the sub-scene image acquired by the second acquisition device 1 is (x2, Y2), the calibration relationship is the conversion relationship between (x1, y1) and (x2, y2).
  • related information (calibration information) of the calibration relationship may be acquired in advance, and the calibration information may be used to determine a position of the label of the second collection device in the panoramic image.
  • a third collection device is further disposed in addition to the first collection device and the second collection device.
  • the first collection device is a plurality of guns
  • the panoramic image is acquired
  • the second collection device is a capture camera.
  • the captured image is captured as a sub-scene image
  • the third acquisition device acquires a detailed image.
  • Determining at least one target position in the panoramic image and determining, according to the calibration information between the first collection device and the third collection device, the target position corresponding to the position in the detail image as the to-be-processed position;
  • the panoramic image after the label is added, the detailed image after the label is added, and the content of the added label are displayed.
  • the images collected by different devices can only be displayed separately (there is no relationship between the images). If the user needs to pay attention to the images collected by multiple devices, you need to switch back and forth between the images collected by the multiple devices. complex.
  • the first collecting device collects the panoramic image
  • the second collecting device collects the image of the sub-scene in the panoramic image to generate a sub-scene image; generates a label according to the sub-scene image, and adds the label to the label.
  • the panoramic image the panoramic image after the label is displayed is displayed; thus, the solution displays the image (the panoramic image) collected by the first collection device and the image (label) collected by the second collection device, and the user does not display If you need to switch, you can pay attention to the images collected by multiple devices, and the operation is simple.
  • an abnormal event may be detected in the panoramic image collected by the first collection device; if yes, the target second collection device corresponding to the abnormal event is determined; and the sub-scene image collected by the target second collection device is acquired.
  • the abnormality model may be preset: according to the above description, the abnormal events may include traffic accidents, robberies, fires, etc., and these abnormal events may be simulated in advance to generate corresponding abnormal models.
  • the panoramic image is then matched with the preset anomaly model. If the matching is successful, an abnormal event occurs in the panoramic image.
  • the position where the match is successful is the position of the abnormal event in the panoramic image.
  • the abnormal event alarm information sent by the other device or the user for the panoramic image may be received, and the alarm information is received, and an abnormal event occurs in the panoramic image.
  • the device that implements the solution can communicate with other devices, and other devices can send abnormal event alarm information to the device after determining that an abnormal event occurs in the panoramic image.
  • the user can also send an abnormal event alarm message to the device, which is also reasonable.
  • the abnormal event alarm information may carry the position of the abnormal event in the panoramic image.
  • the calibration relationship there is a calibration relationship between the first collection device and the four second collection devices.
  • related information calibration information
  • the calibration information can determine the target second collection device corresponding to the above-mentioned "position of the abnormal event in the panoramic image", that is, the second collection device that performs image acquisition for the sub-scene where the abnormal event is located.
  • S202 is: generating a label corresponding to the abnormal event according to the sub-scene image.
  • the focus area may be divided in the panoramic image in advance, and when an abnormal event occurs in the panoramic image is detected, it may be determined whether the position of the abnormal event in the panoramic image is at a preset. The focus area; if so, the label is displayed in the video frame image in a preset alarm mode.
  • intersection A in the panoramic image is an area that needs to be focused
  • the intersection A is set as the focus area in the panoramic image in advance; if an abnormal event occurs in the panoramic image, and the abnormal event occurs at the intersection In A, the label is displayed in the video frame image in a preset alarm mode.
  • the content of the label and the label are separately displayed may be displayed in the second area or the picture-in-picture area by an alarm method, for example, the color change of the pop-up window, the pop-up window shake, etc.
  • an alarm method for example, the color change of the pop-up window, the pop-up window shake, etc. The specific is not limited.
  • FIG. 3 is a third schematic flowchart of an image processing method according to an embodiment of the present disclosure. The embodiment shown in FIG. 3 is based on the embodiment shown in FIG.
  • S301 Receive a label adding instruction sent by a user.
  • the user can click on a target such as a building or an intersection in a video frame image, and then input the content related to the target (target content), and the target content may include text information (such as a building name). , or other relevant instructions), or can also contain images.
  • a target such as a building or an intersection in a video frame image
  • the target content may include text information (such as a building name). , or other relevant instructions), or can also contain images.
  • the tag addition instruction can carry the target location (the location clicked by the user) and the target content (the content, text or image input by the user).
  • the user may also obtain the sub-scene image collected by the second collection device, and use the acquired sub-scene image as the target content, or the user may select the sub-scene image and the target information in the sub-scene image (as shown in FIG. 2
  • the target information in the embodiment has the same meaning and will not be described again as the target content.
  • S302 Generate a label according to the label adding instruction.
  • the label may include a "tag symbol” and a "tag content”.
  • the "tag symbol” may be an arrow, a triangle, etc.
  • the "tag symbol” is for marking a position in the video frame image.
  • the specific form of the label is not limited; in this embodiment, the target content input by the user may be used as the content of the label.
  • the tag may also include a "tag name", for example, may be some simple text information, such as “some building”, “some park” and the like. It is also possible to use part of the content input by the above user as the name of the tag.
  • S101 is S101B: determining the target position of the added tag according to the tag adding instruction.
  • the target location is the location that the above user clicks.
  • the location and content of the label are determined by the user, that is, the user can design his own label according to his own needs, and the user experience is better.
  • the embodiment of the present application further provides an image processing device.
  • the embodiment of the present application further provides an image processing device, as shown in FIG. 4a, comprising: a processor 401 and a memory 402;
  • the processor 401 is configured to implement any of the above image processing methods when executing a program stored on the memory 402.
  • FIG. 4b is a schematic structural diagram of another image processing apparatus according to an embodiment of the present disclosure, including: a housing 501, a processor 502, a memory 503, a circuit board 504, and a power supply circuit 505, wherein the circuit board 504 is disposed in the housing 501.
  • the processor 502 and the memory 503 are disposed on the circuit board 504;
  • the power supply circuit 505 is configured to supply power to the respective circuits or devices of the image processing apparatus;
  • the memory 503 is configured to store executable program code;
  • the processor 502 passes The executable program code stored in the memory 503 is read to execute a program corresponding to the executable program code for performing the following steps:
  • the video frame image after the tag is displayed according to the preset display rule.
  • the video frame image is a panoramic image
  • the first collection device corresponds to at least one second collection device
  • the second collection device performs image collection on the sub-scene corresponding to the panoramic image
  • the processor is further configured to implement the following steps:
  • the processor is further configured to implement the following steps:
  • the processor is further configured to implement the following steps:
  • Identifying the sub-scene image determining target information in the sub-scene image according to the recognition result; adding the target information to the content of the label;
  • the processor is further configured to implement the following steps:
  • the content of the added tag is displayed.
  • the processor is further configured to implement the following steps:
  • the video frame image after the tag is added, and the content of the added tag.
  • the processor is further configured to implement the following steps:
  • the processor is further configured to implement the following steps:
  • the clicked label is determined as the target label
  • the content of the target tag is displayed in the video frame image.
  • the processor is further configured to implement the following steps:
  • a target location of the added tag is determined according to the tag addition instruction.
  • the processor is further configured to implement the following steps:
  • Determining a layer display strategy and determining, according to the layer display strategy, a current display layer and a display manner of the current display layer;
  • the label corresponding to the current display layer is displayed.
  • the processor is further configured to implement the following steps:
  • the processor is further configured to implement the following steps:
  • the processor is further configured to implement the following steps:
  • the processor is further configured to implement the following steps:
  • the label is displayed in the video frame image in a preset alarm mode.
  • the processor is further configured to implement the following steps:
  • the video frame image after the tag is added and the detailed image after the tag is displayed according to the preset display rule.
  • the processor is further configured to implement the following steps:
  • the first area displaying the video frame image after the label is added; in the third area, displaying the detailed image after adding the label;
  • the video frame image after the tag is added, and the detail image after the tag is added.
  • the label can help the user understand the specific content included in the video frame image, and therefore, after adding the label
  • the video frame image can display the image content more intuitively, and the display effect is better.
  • the embodiment of the present application further provides an image processing system, where the system may include: a first collection device and an image processing device, where
  • the first collecting device is configured to collect a video frame image, and send the collected video frame image to the image processing device;
  • the image processing device is configured to determine, according to a video frame image acquired by the first collection device, at least one target location in the video frame image; adding a label at each determined target location, the label according to the input instruction Or the image collected by the second collection device is generated; and the video frame image after the label is added is displayed according to the preset display rule.
  • the system further includes: at least one second collection device (second acquisition device 1, second collection device 2, second collection device 3, and second collection device 4) ,
  • the second collection device is configured to perform image collection on a sub-scene corresponding to the panoramic image, where the panoramic image is a video frame image collected by the first collection device;
  • the image processing device is further configured to acquire a sub-scene image acquired by the second collection device; generate a label according to the sub-scene image; and determine, according to the calibration information of the first collection device and the second collection device acquired in advance The label corresponding to the second collection device is at a target position in the panoramic image.
  • the image processing device in this embodiment may be a platform device, which may acquire resources from multiple collection devices, display images, and interact with users.
  • the first collection device is an augmented reality AR panoramic camera.
  • the image processing device can also be used to:
  • the image processing device can also be used to:
  • Identifying the sub-scene image determining target information in the sub-scene image according to the recognition result; adding the target information to the content of the label;
  • the image processing device can also be used to:
  • the content of the added tag is displayed.
  • the image processing device can also be used to:
  • the video frame image after the tag is added, and the content of the added tag.
  • the image processing device can also be used to:
  • the image processing device can also be used to:
  • the clicked label is determined as the target label
  • the content of the target tag is displayed in the video frame image.
  • the image processing device can also be used to:
  • a target location of the added tag is determined according to the tag addition instruction.
  • the image processing device can also be used to:
  • Determining a layer display strategy and determining, according to the layer display strategy, a current display layer and a display manner of the current display layer;
  • the label corresponding to the current display layer is displayed.
  • the image processing device can also be used to:
  • the image processing device can also be used to:
  • the image processing device can also be used to:
  • the image processing device can also be used to:
  • the label is displayed in the video frame image in a preset alarm mode.
  • the system may further include: a third collection device;
  • the third collection device is configured to collect a detailed image corresponding to the panoramic image, where the panoramic image is a video frame image collected by the first collection device;
  • the image processing device is further configured to acquire a detail image collected by the third collection device; and determine, according to a pixel point correspondence relationship between the detail image and the video frame image, the target location is corresponding to a position in the detail image as a to-be-processed location; adding a tag added at the target location to a to-be-processed location corresponding to the target location; according to a preset display rule, the tagged video frame image, and The detailed image after the label is added for display.
  • the image processing device acquires a video frame image collected by the first collection device, adds a label to a target position in the video frame image, and then displays the video frame image after the label is added; Help users understand the specific content contained in the video frame image. Therefore, the video frame image after the label is added can display the image content more intuitively, and the display effect is better.
  • the embodiment of the present application further provides a computer readable storage medium, where the computer readable storage medium stores a computer program, and when the computer program is executed by the processor, implements any of the above image processing methods.
  • the embodiment of the present application also provides an executable program code for being executed to execute any of the image processing methods described above.

Landscapes

  • Engineering & Computer Science (AREA)
  • Computer Graphics (AREA)
  • Computer Hardware Design (AREA)
  • General Engineering & Computer Science (AREA)
  • Software Systems (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Television Signal Processing For Recording (AREA)

Abstract

La présente invention concerne, selon un mode de réalisation, un procédé, un dispositif et un système de traitement d'image. Le procédé consiste à : ajouter une balise au niveau d'une position cible dans une image de trame vidéo et afficher l'image de trame vidéo additionnée de la balise. La balise peut aider un utilisateur à comprendre un contenu spécifique présent dans l'image de trame vidéo. L'image de trame vidéo additionnée de la balise peut présenter de manière plus intuitive son contenu, ce qui permet d'obtenir un meilleur effet de présentation.
PCT/CN2018/106752 2018-03-29 2018-09-20 Procédé, dispositif et système de traitement d'image WO2019184275A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201810272370.9 2018-03-29
CN201810272370.9A CN109274926B (zh) 2017-07-18 2018-03-29 一种图像处理方法、设备及系统

Publications (1)

Publication Number Publication Date
WO2019184275A1 true WO2019184275A1 (fr) 2019-10-03

Family

ID=68062694

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2018/106752 WO2019184275A1 (fr) 2018-03-29 2018-09-20 Procédé, dispositif et système de traitement d'image

Country Status (1)

Country Link
WO (1) WO2019184275A1 (fr)

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102457718A (zh) * 2010-10-14 2012-05-16 霍尼韦尔国际公司 在视频监控中利用用户输入对视频数据加图形书签
CN103929618A (zh) * 2014-04-18 2014-07-16 卢旭东 一种户外广告牌状态标识系统的运行控制方法
CN104285244A (zh) * 2012-05-23 2015-01-14 高通股份有限公司 用于标注的图像驱动的视图管理
US20170364747A1 (en) * 2016-06-15 2017-12-21 International Business Machines Corporation AUGEMENTED VIDEO ANALYTICS FOR TESTING INTERNET OF THINGS (IoT) DEVICES

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102457718A (zh) * 2010-10-14 2012-05-16 霍尼韦尔国际公司 在视频监控中利用用户输入对视频数据加图形书签
CN104285244A (zh) * 2012-05-23 2015-01-14 高通股份有限公司 用于标注的图像驱动的视图管理
CN103929618A (zh) * 2014-04-18 2014-07-16 卢旭东 一种户外广告牌状态标识系统的运行控制方法
US20170364747A1 (en) * 2016-06-15 2017-12-21 International Business Machines Corporation AUGEMENTED VIDEO ANALYTICS FOR TESTING INTERNET OF THINGS (IoT) DEVICES

Similar Documents

Publication Publication Date Title
CN109274926B (zh) 一种图像处理方法、设备及系统
US10043079B2 (en) Method and apparatus for providing multi-video summary
CN104137154B (zh) 用于管理视频数据的系统和方法
US20110109747A1 (en) System and method for annotating video with geospatially referenced data
CN110536074B (zh) 一种智能巡检系统、巡检方法
US8929596B2 (en) Surveillance including a modified video data stream
CN110557603B (zh) 移动目标的监控方法、装置和可读存储介质
CN110136091B (zh) 图像处理方法及相关产品
KR101652856B1 (ko) Cctv에서 관제 이벤트에 기초한 사용자 인터페이스 화면 제공 장치
CN101272483B (zh) 用于管理运动监视相机的系统和方法
EP3062506B1 (fr) Procédé et appareil de commutation d'image
CN112162683A (zh) 一种图像放大方法、装置及存储介质
JP2019125053A (ja) 情報端末装置、情報処理システムおよび表示制御プログラム
KR100653825B1 (ko) 변화검출방법 및 장치
JP4632362B2 (ja) 情報出力システム、情報出力方法およびプログラム
WO2019184275A1 (fr) Procédé, dispositif et système de traitement d'image
KR101842564B1 (ko) 다중 영상에 대한 집중 영상 감시 방법 및 이를 위한 집중 영상 관리 서버, 집중 영상 감시 시스템, 컴퓨터 프로그램, 그 기록매체
CN110737385A (zh) 一种视频鼠标交互方法、智能终端及存储介质
CN113905211B (zh) 一种视频巡逻方法、装置、电子设备及存储介质
US20210375109A1 (en) Team monitoring
US20230162591A1 (en) Interactive kiosk with emergency call module
US20030112415A1 (en) Apparatus for projection and capture of a display interface
KR20200073669A (ko) 영상 정보 관리 방법, 영상 정보 관리 장치 및 컴퓨터 프로그램
KR102398280B1 (ko) 관심 영역에 대한 영상을 제공하기 위한 장치 및 방법
CN114677163A (zh) 一种广告交互方法、装置、介质及设备

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 18911456

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 18911456

Country of ref document: EP

Kind code of ref document: A1

122 Ep: pct application non-entry in european phase

Ref document number: 18911456

Country of ref document: EP

Kind code of ref document: A1

122 Ep: pct application non-entry in european phase

Ref document number: 18911456

Country of ref document: EP

Kind code of ref document: A1

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 19.05.2021)