CN112055179A - Video playing method and device - Google Patents

Video playing method and device Download PDF

Info

Publication number
CN112055179A
CN112055179A CN202010953413.7A CN202010953413A CN112055179A CN 112055179 A CN112055179 A CN 112055179A CN 202010953413 A CN202010953413 A CN 202010953413A CN 112055179 A CN112055179 A CN 112055179A
Authority
CN
China
Prior art keywords
target
video
target image
video playing
attribute information
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202010953413.7A
Other languages
Chinese (zh)
Inventor
苏琳
晋兆龙
索凯
简宏伟
朱聪聪
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Suzhou Keda Technology Co Ltd
Original Assignee
Suzhou Keda Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Suzhou Keda Technology Co Ltd filed Critical Suzhou Keda Technology Co Ltd
Priority to CN202010953413.7A priority Critical patent/CN112055179A/en
Publication of CN112055179A publication Critical patent/CN112055179A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/18Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/44Arrangements for executing specific programs
    • G06F9/451Execution arrangements for user interfaces
    • G06F9/454Multi-language systems; Localisation; Internationalisation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/431Generation of visual interfaces for content selection or interaction; Content or additional data rendering
    • H04N21/4312Generation of visual interfaces for content selection or interaction; Content or additional data rendering involving specific graphical features, e.g. screen layout, special fonts or colors, blinking icons, highlights or animations
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/45Management operations performed by the client for facilitating the reception of or the interaction with the content or administrating data related to the end-user or to the client device itself, e.g. learning user preferences for recommending movies, resolving scheduling conflicts
    • H04N21/458Scheduling content for creating a personalised stream, e.g. by combining a locally stored advertisement with an incoming stream; Updating operations, e.g. for OS modules ; time-related management operations
    • H04N21/4586Content update operation triggered locally, e.g. by comparing the version of software modules in a DVB carousel to the version stored locally
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/472End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content

Abstract

The invention provides a video playing method and a video playing device, wherein the method comprises the following steps: providing a video playing interface, wherein the video playing interface comprises: the video playing area is used for playing videos; and a target image analysis display area for displaying a plurality of target images acquired from the video played in the video playing area and target attribute information of at least one target image in the plurality of target images. The video playing method and the video playing device provided by the invention realize video playing, so that the auxiliary staff can track the target and the target attribute and carry out related operation in the video playing interface.

Description

Video playing method and device
Technical Field
The present invention relates to the field of video processing, and in particular, to a video playing method and apparatus.
Background
Since the eighties of the last century, the development of video surveillance has undergone changes in analog technology, digital technology, and network technology. The video monitoring system is used as a core part of the development of the security field, is relied on by various fields such as judicial sciences, traffic, education, medical treatment, finance, real estate, agriculture and the like, effectively prevents illegal events, improves the evidence obtaining efficiency of the illegal events, and provides a good development environment for the construction of various industries. With the development and progress of the human civilization society, a high-performance video monitoring system is an indispensable important technology in the modern world and the future world.
In a video monitoring system, a video monitoring interface is always used as the most intuitive part of visual management. However, a conventional video monitoring interface generally only supports playing of a video, but actually, a large amount of information exists in the video, and if an operator wants to obtain a certain target in the video, the operator needs to look at the interface all the time to perform a visual search.
Because human energy is limited and visual fatigue is easy to generate, and under the condition of special light and the like, visual errors are inevitable, a video playing interface capable of simultaneously supporting the display of video analysis targets and attributes thereof is a trend pursued by the development of video monitoring systems. The information displayed in the video monitoring system is of various kinds, but in real-world applications, only one or two kinds (such as vehicle and personnel) of analysis targets are concerned by operators, so that the video played by the traditional monitoring system interface contains a large amount of invalid information, and a great burden is added to the work of locating the targets in the video through human eyes.
Therefore, the technical staff in the field needs to solve the problem of how to realize video playing so as to assist the staff in tracking the target and the target attribute and performing related operations in the video playing interface.
Disclosure of Invention
In order to overcome the defects in the prior art, the invention provides a video playing method and a video playing device, which assist workers in tracking targets and target attributes and performing related operations in a video playing interface.
According to an aspect of the present invention, there is provided a video playing method, including:
providing a video playing interface, wherein the video playing interface comprises:
the video playing area is used for playing videos; and
and the target image analysis display area is used for displaying a plurality of target images acquired from the video played in the video playing area and target attribute information of at least one target image in the plurality of target images.
In some embodiments of the present invention, a target frame moving along with a target is displayed in a video played in the video playing area, and the target image is captured according to a preset capturing rule according to the target frame.
In some embodiments of the present invention, the target image is updated as the target within the target frame changes, and the target attribute information of the target image is updated as the target image is updated.
In some embodiments of the present invention, when the video playing area plays the offline video, the target image shown in the target image analysis showing area and the target attribute information of the target image are stored in a configuration file, wherein the target image associated with the video frame number currently played by the offline video and the target attribute information of the target image are read from the configuration file to be shown in the target image analysis showing area.
In some embodiments of the present invention, the target attribute information of the target image is displayed in order of attribute confidence from high to low.
In some embodiments of the present invention, the video playing interface further comprises:
the target image display area is used for displaying a plurality of target images in different areas according to target types, and the number of the target images displayed in the target image display area is greater than or equal to the number of the target images displayed in the target image analysis display area.
In some embodiments of the present invention, the video playing interface further comprises:
and the target detail display window is triggered and displayed through the operation on the target image, the target detail display window is used for displaying the target attribute information of the operated target image, and the number of the target attribute information of the target detail display window is more than or equal to the number of the target attribute information displayed in the target image analysis display area.
In some embodiments of the present invention, after the target attribute information of the target detail display window is edited, the target attribute information is updated to the corresponding target attribute information displayed in the target image analysis display area.
In some embodiments of the invention, the target detail display window displays the operated target image and target images of other target types associated with the operated target image.
In some embodiments of the present invention, after providing the video playing interface, the method further includes: receiving a language category selected by a user; searching a translation text corresponding to the selected language type from an array; and switching the characters of the video playing interface into the searched translation text.
According to another aspect of the present invention, there is also provided a video playback apparatus, including:
an interface providing module configured to provide a video playing interface, the video playing interface comprising:
the video playing area is used for playing videos; and
and the target image analysis display area is used for displaying a plurality of target images acquired from the video played in the video playing area and target attribute information of at least one target image in the plurality of target images.
Therefore, compared with the prior art, the scheme provided by the invention has the following advantages:
according to the invention, through the video playing interface comprising the video playing area and the target image analysis display area, a plurality of target images acquired from the video played in the video playing area and the target attribute information of at least one target image in the plurality of target images are displayed while the video is played, so that the video playing interface assists workers to track the target and the target attribute and carry out related operations.
Drawings
The above and other features and advantages of the present invention will become more apparent by describing in detail exemplary embodiments thereof with reference to the attached drawings.
Fig. 1 is a schematic diagram illustrating a video playing interface provided by a video playing method according to an embodiment of the present invention.
FIG. 2 shows a schematic diagram of a target detail display window according to an embodiment of the invention.
Fig. 3 is a schematic diagram illustrating a target detail display window displayed in a video playing interface according to an embodiment of the present invention.
Fig. 4 is a block diagram of a video playback apparatus according to an embodiment of the present invention.
Fig. 5 schematically illustrates a computer-readable storage medium in an exemplary embodiment of the disclosure.
Fig. 6 schematically illustrates an electronic device in an exemplary embodiment of the disclosure.
Detailed Description
Example embodiments will now be described more fully with reference to the accompanying drawings. Example embodiments may, however, be embodied in many different forms and should not be construed as limited to the examples set forth herein; rather, these embodiments are provided so that this disclosure will be thorough and complete, and will fully convey the concept of example embodiments to those skilled in the art. The described features, structures, or characteristics may be combined in any suitable manner in one or more embodiments.
Furthermore, the drawings are merely schematic illustrations of the invention and are not necessarily drawn to scale. The same reference numerals in the drawings denote the same or similar parts, and thus their repetitive description will be omitted. Some of the block diagrams shown in the figures are functional entities and do not necessarily correspond to physically or logically separate entities. These functional entities may be implemented in the form of software, or in one or more hardware modules or integrated circuits, or in different networks and/or processor devices and/or microcontroller devices.
The flow charts shown in the drawings are merely illustrative and do not necessarily include all of the steps. For example, some steps may be decomposed, and some steps may be combined or partially combined, so that the actual execution sequence may be changed according to the actual situation.
In order to overcome the defects in the prior art, the invention provides a video playing method and device, which assist a worker in tracking a target and a target attribute and performing related operations in a video playing interface.
The video playing method provided by the invention comprises the step of providing a video playing interface. The following describes the video playing interface provided in the video playing method provided by the present invention with reference to fig. 1 to fig. 3, respectively. Fig. 1 is a schematic diagram illustrating a video playing interface provided by a video playing method according to an embodiment of the present invention. FIG. 2 shows a schematic diagram of a target detail display window according to an embodiment of the invention. Fig. 3 is a schematic diagram illustrating a target detail display window displayed in a video playing interface according to an embodiment of the present invention.
The video playing interface 1 comprises a video playing area 11 and a target image analysis and display area 12. The video playing area 11 is used for playing video. The target image analysis display area 12 is configured to display a plurality of target images 121 acquired from the video played by the video playing area 11 and target attribute information 122 of at least one of the plurality of target images. Specifically, the target image analysis display area 12 is used for displaying the target attribute information 122 of the selected at least one target image 121.
Therefore, in the video playing method provided by the invention, through the video playing interface comprising the video playing area and the target image analysis display area, a plurality of target images acquired from the video played in the video playing area and the target attribute information of at least one target image in the plurality of target images are displayed while the video is played, so that the video playing interface assists workers to track the target and the target attribute and perform related operations.
In some embodiments of the present invention, the video playing area 11 plays a video in which target boxes 111, 112, and 113 moving with a target are displayed. Furthermore, the target frame of the selected target type can be displayed in the video playing process, so that the problems that in the traditional video playing process, observers need to position target analysis results by naked eyes to absorb a large amount of invalid information, the visual fatigue of human eyes is increased, and the target is positioned and analyzed inaccurately are solved. For example, in some videos, object boxes of object types such as pedestrians, faces, vehicles, etc. may be selected so that these selected object types may be highlighted while objects in the environment (such as trees, trash cans, etc.) are faded.
Specifically, in the above embodiment, when the video playing area plays a real-time video, the target recognition algorithm is run in real time, so that the recognized target box is displayed in the video in real time. Further, the display parameters of the target frame for each frame may be saved by video frame. The display parameters may include, for example, the display form (e.g., different object categories display the object frame in different colors, line shapes; when there are fewer objects, the object frame may also be displayed in different colors, line shapes per object) and the display position (coordinates in the video image). For example, video frame numbers (used to identify different video frames) and display parameter associations may be stored in a configuration file. The video frame number and the display parameter can be stored in a key value mapping mode, the video frame number is stored as a key, and the display parameter is stored as a value, so that the association relationship between the video frame number and the display parameter is represented. Therefore, the configuration file can be downloaded and read when the offline video is played, so that the target frame can be displayed in the offline video. Specifically, in the video playing process, the client reads the configuration file information, analyzes the SEI information (supplemental enhancement information) of each frame, belongs to the code stream category, and provides a method for adding additional information into the video code stream), can scale the position coordinates of the target appearing in the video according to the size of the current playing window according to a certain proportion to form a target frame, so that the target frame is drawn into the current frame code stream, and a user can obviously observe the position of the target in the video viewing process, and can see the displacement condition of the captured target in the video along with the video playing.
In some embodiments of the present invention, the target image of the target image analysis display area 12 may be captured according to a preset capturing rule according to the target frame. The preset snapshot rules can be set as required. For example, the snapshot rule may be set to recognize a target, i.e., a snapshot; the target executes the set action, namely snapshot; capturing every other preset period after the target is identified; the invention can set various different snapshot rules through manual operation for snapshot and the like, which is not described herein.
In the above embodiment, when the video playing area plays a real-time video, the target image is updated along with the change of the target in the target frame, and the target attribute information of the target image is updated along with the update of the target image. Specifically, the attribute recognition is updated due to the movement of the object within the object frame, the view angle conversion, and the like. For example, when the subject is a pedestrian, whose angle of view is photographed changes from the back side to the front side, more accurate subject attribute information will be obtained, and therefore the update of the subject attribute information can be performed accordingly. Further, when the identified attribute information is not identified during the update of the target attribute information, the target attribute information is not updated and the previously identified attribute information is retained. Further, in the present embodiment, when a target image of the same target is captured multiple times, a target image may be determined according to the size, definition, angle of view, and the like of the target in the target image to be displayed in the target image analysis display area 12. Therefore, the staff is further assisted to accurately know the image and the attribute of the target. Further, the target image and the corresponding target attribute may be saved according to the video frame, so that the display of the target image analysis display area 12 may be realized in an offline video playing mode in a configuration file manner.
Specifically, when the video playing area plays an offline video, the target image displayed in the target image analysis display area and the target attribute information of the target image are stored in a configuration file in association with the video frame number, wherein when the offline video is played to the video frame number, the target image associated with the video frame number and the target attribute information of the target image are read from the configuration file to be displayed in the target image analysis display area. Further, when an operator performs a video playing operation, a video playing interface reads a video information configuration file from the server, that is, the client sends a video task id (task id) to be played to the server, the server receives the task id and then judges whether the configuration file corresponding to the video is downloaded (the first played video configuration file is not downloaded, so that the judgment can be made according to the playTimes of the video), and if the configuration file is downloaded, the stored information can be directly sent; if not, the configuration file is downloaded to the local and then the corresponding configuration file information is sent. The configuration file stores information of each frame of video, which includes frame data played by the video, analyzed target snap picture paths in the video and different types of target attribute information (for example, pedestrian and vehicle, pedestrian attribute information includes gender, height, age, hair style, hair color, coat texture, backpack, hat color, mask, riding, pedestrian direction, personnel speed, ethnic attributes, etc., and vehicle information includes license plate number, vehicle type, vehicle size, whether to make a call, vehicle direction, vehicle color, manufacturer, unwashed seat belt, vehicle brand, color shade, style, sun visor, vehicle brand, license plate color, vehicle speed, etc., but the invention is not limited thereto). By using the mapping technology, the video frame number is set as a key, the target image and the target attribute information of each frame are used as values, and the video frame number is associated and matched with the corresponding target attribute information, namely, the code stream information corresponding to each frame can be output in the video playing process. Specifically, in some embodiments, the target image and target attribute information in the configuration file are the target image and target attribute information obtained corresponding to the frame. In some variations, the target image and target attribute information may be updated as previously described, whereby the target image and target attribute information associated with a video frame number may be a higher quality target image and more complete target attribute information for the same target after updating. In still other variations, the target frame in the played video and the target image in the target image analysis display area 12 may be associated by connecting lines, adding the same mark, and the like.
In some embodiments of the present invention, the target attribute information of the target image may be displayed in order of attribute confidence from high to low. The confidence level indicates the degree of credibility of the target attribute information. In some embodiments, the target image analysis presentation area 12 displays only a portion of the target attribute information, and therefore, N target attribute information with higher confidence may be displayed in the target image analysis presentation area 12, where N is an integer greater than 1. The confidence level may be obtained through various algorithm calculations, for example, the confidence level of the attribute may be determined through a classification probability value in a machine learning model of attribute identification, and the invention is not limited thereto. Further, the confidence level may also be an attribute priority set by a human.
Specifically, the above embodiment supports dynamic display of the target image, and supports dynamic display of partial target attribute information corresponding to the current target image slice, and the target attribute information is sequentially output from high to low according to the priority level of the confidence level. The target image and target attribute binding output mode improves the readability of target attribute information, solves the problem that human eyes are difficult to distinguish certain target attributes, and improves the efficiency and accuracy of obtaining the target attribute information.
Specifically, in the above embodiments, a list view (list display) method in Qt (cross-platform C + + graphical user interface application development framework) may be used to define a target attribute detail display list, a delete (attribute list template) in ListView is defined as two columns, where the left side is an attribute title and the right side is specific attribute information corresponding to the title, an objnfomodel is defined for storing each piece of target picture information and attribute detail, and the objnfomodel reads the target information from the server configuration file by using an append () function, thereby completing the information storage work. Defining a showIndex (display index) variable and initializing to 0, starting a Timer and performing showIndex accumulation operation, typesetting attribute titles in a target attribute detail list delegat in sequence from big to small according to attribute references, sequentially reading target attribute details in the current objInfoModel according to the numerical value of the showIndex and displaying the target attribute details in an attribute information display table (when the showIndex is n, correspondingly reading current objInfoModel attribute information corresponding to the nth row attribute title of the target attribute detail list), so that a dynamic display effect of the target details can be realized, the dynamic display mode can attract human eye attention more easily, and the information obtaining efficiency and accuracy of operators can be improved.
In some embodiments, the confidence level may be calculated in real-time, and as the target attribute information is updated with the target image, the confidence level of the updated target attribute may be recalculated, updating the rank order of the target attribute information as well. In other embodiments, whether to update the target image may be determined according to the confidence level of each target attribute information of the target image (for example, when the number of target attribute information with increased confidence levels is greater than a predetermined threshold, the target image is updated, when the confidence level of one or more set target attribute information is increased, the target image is updated, when the average confidence level is increased, the target image is updated, and the present invention is not limited thereto), and after the target image is updated, whether to update the sequence of the target attributes is determined.
In some embodiments of the present invention, the video playing interface 1 further includes a target image display area 13. The target image display area 13 is used for displaying a plurality of target images 133 and 134 in regions according to target types. For example, in fig. 1, the area 131 and the area 132 are divided by the object type, and the object image 133 belongs to the object type of the area 131 and is thus displayed in the area 131; the target image 134 belongs to the target type of the area 132 and is therefore displayed in the area 132.
Specifically, the target image display area 13 may display a target image related to a previous frame in addition to the target image related to the current frame played by the video playing area 11. In other words, the target image display area 13 is used to record more complete captured target images, and thus the number of target images displayed in the target image display area 13 is equal to or greater than the number of target images displayed in the target image analysis display area 12.
In the embodiment shown in fig. 1, the target image display area 13 displays target images of different target types (for example, people, human faces, automobiles, and non-automobiles) in rows, and the user can select different types of snapshot results to be displayed according to the desired types, so that the target search efficiency can be improved. Specifically, in some specific embodiments, the played video includes different types of objects such as people, human faces, motor vehicles, non-motor vehicles, and the like, the object image display area 13 of this embodiment may adopt a GridView (a control that displays content in a two-dimensional scrollable grid) display manner, define an object display style template snapPicDelegate, set styles such as width, height, and object frame of a snapshot object image, respectively define people (perSnapPicModel), faces (faceSnapPicModel), and motor vehicles (vehSnapPicModel), in the non-motor vehicle (novehSnapPicModel) model, the object types ObjType of personnel, human faces, motor vehicles and non-motor vehicles can be respectively set to be 1, 16, 2 and 32, different object type models can read corresponding object information in a server configuration file by utilizing an apend () function according to the ObjType and display the corresponding object information in different types of object image lists, and the effect of displaying different types of object images in rows can be realized.
In some embodiments of the present invention, the video playing interface 1 further comprises a target detail display window 14. The target detail display window 14 triggers display via operation on the target image (as shown in fig. 3). The target detail display window 14 is used for displaying the target attribute information 143 of the operated target image 141, and the number of the target attribute information of the target detail display window 14 is greater than or equal to the number of the target attribute information displayed in the target image analysis display area 12.
Specifically, the target detail display window 14 may trigger display based on an operation (such as double-clicking or setting a touch manner) on the target image in the target image presentation area 13. The target detail display window 14 may also trigger display based on an operation (such as double-clicking or setting a touch manner) on the target image of the target image analysis presentation area 12. The invention is not so limited.
In some embodiments of the present invention, the target attribute information of the target detail display window 14 is edited and then updated to the corresponding target attribute information displayed in the target image analysis display area 12.
Specifically, the target detail display window 14 may display all attribute information of the target, and support interactive modification of target attribute information in the target detail, thereby improving accuracy of data in the database. In the process of dynamically displaying the target attribute, because the efficiency of acquiring information by the interface space and the human eyes is limited, all information of the analyzed target attribute cannot be displayed, and only part of the information can be displayed, the display of the target detail display window 14 can be realized by a createObject method, so that all target attribute information corresponding to the target image is displayed. In the target detail display window 14, in order to clearly and intuitively display all the target attribute information of the target image, a column definition attribute information table may be used, and attribute titles and corresponding attribute contents are defined in the table (if the target attribute is too much, a ScrollViewStyle (scroll display form) may be used to establish a scroll-type drop-down box in the target detail display window 14). In order to compensate for the algorithm error and improve the accuracy of the configuration file, the embodiment supports manual editing and storing of the target attribute information, that is, the client transmits the modified information to the server through rpc (Remote procedure call protocol) asynchronous interface, and stores the modified information in the configuration file, so as to improve the accuracy of the target analysis attribute in the video monitoring query system. Further, in the scene of playing the offline video, when the target detail display window 14 is closed, an operation of re-reading the target image associated with the currently played video frame number and the target attribute information of the target image from the configuration file is triggered, so that the corresponding target attribute information displayed in the target image analysis display area 12 can be updated together. In the scene of playing the real-time video, since the configuration file is not completely generated, the configuration file may be modified based on the editing of the target attribute information of the target detail display window 14, and the corresponding target attribute information displayed in the target image analysis display area 12 may be updated directly by, for example, parameter transmission.
Further, in some embodiments of the present invention, the target detail display window displays the operated target image and target images of other target types associated with the operated target image. For example, the target image 141 of the pedestrian type and the target image 142 of the face of the pedestrian can be displayed together, so that the operator can more conveniently acquire various kinds of information of the pedestrian. Specifically, after the operated target image is determined, the target frame corresponding to the operated target image may be determined from the video, and the target frames of other target types may be searched in the video frame capturing the target image, and the target images of other target types associated with the target image may be determined according to the geometric relationship (e.g., the degree of overlap, whether the target image is included, etc.) of the target frames. The invention is not limited in this respect and may also be determined by the geometric relationship of the object frame to the object frames of the object images of other object types in a plurality of video frames.
In the foregoing embodiment, when it is necessary to associate the target image 141 of a pedestrian type with the target image 142 of the face of the pedestrian, when there is a containment relationship between the target frame corresponding to the operated target image and the target frame of the target image of another other target image type, the two target images are associated. In other embodiments, when it is necessary to associate the target image 141 of the pedestrian type with target images of other target types in close proximity to the periphery thereof, when there is no inclusion relationship between a target frame corresponding to the operated target image and a target frame of a target image of another other target map type, and the target frames intersect or the shortest distance between the target frames is smaller than a predetermined threshold, then the two target images are associated. The present invention can also be implemented in many different ways, which are not described herein.
In some embodiments of the present invention, the providing the video playing interface may further include the following steps: receiving a language type selected by a user (specifically, the user may select the language type by inputting, touch selecting, clicking selecting, scrolling selecting, hovering, and the invention is not limited thereto); searching a translation text corresponding to the selected language type from an array; and switching the characters of the video playing interface into the searched translation text.
Specifically, the video playing interface supports multiple languages to improve market export advantages. The technology realizes that interface translation texts need to be established, translation texts of multiple languages are recorded in the interface translation texts in an array mode, the translation arrays define the types of translation languages according to index values of subscripts, global variables LanguageArr are defined and assigned as the translation texts, and meanwhile, the global variable LanguaageType is defined as an array subscript to determine the language types of the current interface. The interface language type is uniformly controlled by way of reference to LanguageArr [ "array name" ] [ Language type ]. Changing the interface language simply changes the value of the global variable languagettype. The characters length of different language translations is different, and the same controlling part length can't compatible different language length, for solving above conflict, adopts the value according to languageType to set up the length of controlling part under the different language environment to satisfy the pleasing to the eye design demand of different language interfaces.
Therefore, on one hand, the method and the device support the function of displaying the corresponding analysis target frame in the video playing process, and improve the accuracy and efficiency of extracting information and positioning information by operators; on the other hand, the method supports the analysis of the target snapshot picture and the target analysis details related to the target snapshot picture, and the attributes in the target analysis details are sequentially output from high to low according to the confidence level. The information dynamic display can attract the attention of an observer better, the deep memory of the observer on the information is enhanced, and the capability of an operator for acquiring the information is improved; on the other hand, the method supports the regional display of analysis target images of different types (personnel, human faces, motor vehicles and non-motor vehicles), and supports the interactive modification of attributes in target details and storage, thereby improving the accuracy of the target analysis attributes of the video monitoring system; in another aspect, switching of multiple languages can be supported, interface display styles in different languages are compatible, and market advantages of products are improved.
The invention also provides a video playing device. Referring now to fig. 4, fig. 4 is a block diagram illustrating a video playback apparatus according to an embodiment of the present invention. The video playback device 200 includes an interface providing module 210.
The interface providing module 210 is configured to provide a video playing interface, which includes: the video playing area is used for playing videos; and a target image analysis display area for displaying a plurality of target images acquired from the video played in the video playing area and target attribute information of at least one target image in the plurality of target images.
The video playing device comprises a video playing interface of a video playing area and a target image analysis and display area, and a plurality of target images acquired from a video played in the video playing area and target attribute information of at least one target image in the plurality of target images are displayed while the video is played, so that a worker can be assisted in tracking the target and the target attribute and carrying out related operations in the video playing interface.
The present invention may implement the video playback device 200 in software, hardware, firmware, or any combination thereof. The splitting, merging and adding of modules are within the protection scope of the invention without departing from the concept of the invention.
In an exemplary embodiment of the present disclosure, a computer-readable storage medium is further provided, on which a computer program is stored, which when executed by, for example, a processor, can implement the steps of the video playing method in any of the above embodiments. In some possible embodiments, aspects of the present invention may also be implemented in the form of a program product including program code for causing a terminal device to perform the video playback method of the present specification, if the program product runs on the terminal device. Steps according to various exemplary embodiments of the present invention are described in the sections.
Referring to fig. 5, a program product 800 for implementing the above method according to an embodiment of the present invention is described, which may employ a portable compact disc read only memory (CD-ROM) and include program code, and may be run on a terminal device, such as a personal computer. However, the program product of the present invention is not limited in this regard and, in the present document, a readable storage medium may be any tangible medium that can contain, or store a program for use by or in connection with an instruction execution system, apparatus, or device.
The program product may employ any combination of one or more readable media. The readable medium may be a readable signal medium or a readable storage medium. A readable storage medium may be, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any combination of the foregoing. More specific examples (a non-exhaustive list) of the readable storage medium include: an electrical connection having one or more wires, a portable disk, a hard disk, a Random Access Memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or flash memory), an optical fiber, a portable compact disc read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing.
The computer readable storage medium may include a propagated data signal with readable program code embodied therein, for example, in baseband or as part of a carrier wave. Such a propagated data signal may take many forms, including, but not limited to, electro-magnetic, optical, or any suitable combination thereof. A readable storage medium may also be any readable medium that is not a readable storage medium and that can communicate, propagate, or transport a program for use by or in connection with an instruction execution system, apparatus, or device. Program code embodied on a readable storage medium may be transmitted using any appropriate medium, including but not limited to wireless, wireline, optical fiber cable, RF, etc., or any suitable combination of the foregoing.
Program code for carrying out operations for aspects of the present invention may be written in any combination of one or more programming languages, including an object oriented programming language such as Java, C + + or the like and conventional procedural programming languages, such as the "C" programming language or similar programming languages. The program code may execute entirely on the tenant computing device, partly on the tenant device, as a stand-alone software package, partly on the tenant computing device and partly on a remote computing device, or entirely on the remote computing device or server. In the case of remote computing devices, the remote computing devices may be connected to the tenant computing device through any kind of network, including a Local Area Network (LAN) or a Wide Area Network (WAN), or may be connected to an external computing device (e.g., through the internet using an internet service provider).
In an exemplary embodiment of the present disclosure, there is also provided an electronic device, which may include a processor, and a memory for storing executable instructions of the processor. Wherein the processor is configured to perform the steps of the video playing method in any of the above embodiments via execution of the executable instructions.
As will be appreciated by one skilled in the art, aspects of the present invention may be embodied as a system, method or program product. Thus, various aspects of the invention may be embodied in the form of: an entirely hardware embodiment, an entirely software embodiment (including firmware, microcode, etc.) or an embodiment combining hardware and software aspects that may all generally be referred to herein as a "circuit," module "or" system.
An electronic device 600 according to this embodiment of the invention is described below with reference to fig. 6. The electronic device 600 shown in fig. 6 is only an example, and should not bring any limitation to the functions and the scope of use of the embodiments of the present invention.
As shown in fig. 6, the electronic device 600 is embodied in the form of a general purpose computing device. The components of the electronic device 600 may include, but are not limited to: at least one processing unit 610, at least one storage unit 620, a bus 630 that connects the various system components (including the storage unit 620 and the processing unit 610), a display unit 640, and the like.
Wherein the storage unit stores program code executable by the processing unit 610, such that the processing unit 610 executes the video playback method of the present specification. Steps according to various exemplary embodiments of the present invention are described in the sections. For example, the processing unit 610 may execute the steps of the video playing method provided by the present invention.
The storage unit 620 may include readable media in the form of volatile memory units, such as a random access memory unit (RAM)6201 and/or a cache memory unit 6202, and may further include a read-only memory unit (ROM) 6203.
The memory unit 620 may also include a program/utility 6204 having a set (at least one) of program modules 6205, such program modules 6205 including, but not limited to: an operating system, one or more application programs, other program modules, and program data, each of which, or some combination thereof, may comprise an implementation of a network environment.
Bus 630 may be one or more of several types of bus structures, including a memory unit bus or memory unit controller, a peripheral bus, an accelerated graphics port, a processing unit, or a local bus using any of a variety of bus architectures.
The electronic device 600 may also communicate with one or more external devices 700 (e.g., keyboard, pointing device, bluetooth device, etc.), with one or more devices that enable a tenant to interact with the electronic device 600, and/or with any devices (e.g., router, modem, etc.) that enable the electronic device 600 to communicate with one or more other computing devices. Such communication may occur via an input/output (I/O) interface 650. Also, the electronic device 600 may communicate with one or more networks (e.g., a Local Area Network (LAN), a Wide Area Network (WAN), and/or a public network such as the Internet) via the network adapter 660. The network adapter 660 may communicate with other modules of the electronic device 600 via the bus 630. It should be appreciated that although not shown, other hardware and/or software modules may be used in conjunction with the electronic device 600, including but not limited to: microcode, device drivers, redundant processing units, external disk drive arrays, RAID systems, tape drives, and data backup storage systems, among others.
Through the above description of the embodiments, those skilled in the art will readily understand that the exemplary embodiments described herein may be implemented by software, or by software in combination with necessary hardware. Therefore, the technical solution according to the embodiments of the present disclosure may be embodied in the form of a software product, which may be stored in a non-volatile storage medium (which may be a CD-ROM, a usb disk, a removable hard disk, etc.) or on a network, and includes several instructions to enable a computing device (which may be a personal computer, a server, or a network device, etc.) to execute the video playing method according to the embodiments of the present disclosure.
Therefore, compared with the prior art, the scheme provided by the invention has the following advantages:
according to the invention, through the video playing interface comprising the video playing area and the target image analysis display area, a plurality of target images acquired from the video played in the video playing area and the target attribute information of at least one target image in the plurality of target images are displayed while the video is played, so that the video playing interface assists workers to track the target and the target attribute and carry out related operations.
Other embodiments of the disclosure will be apparent to those skilled in the art from consideration of the specification and practice of the disclosure disclosed herein. This application is intended to cover any variations, uses, or adaptations of the disclosure following, in general, the principles of the disclosure and including such departures from the present disclosure as come within known or customary practice within the art to which the disclosure pertains. It is intended that the specification and examples be considered as exemplary only, with a true scope and spirit of the disclosure being indicated by the following claims.

Claims (11)

1. A video playback method, comprising:
providing a video playing interface, wherein the video playing interface comprises:
the video playing area is used for playing videos; and
and the target image analysis display area is used for displaying a plurality of target images acquired from the video played in the video playing area and target attribute information of at least one target image in the plurality of target images.
2. The video playing method according to claim 1, wherein a target frame moving with a target is displayed in the video played in the video playing area, and the target image is captured according to a preset capturing rule according to the target frame.
3. The video playback method according to claim 2, wherein when the video playback area plays back a real-time video, the target image is updated as the target in the target frame changes, and the target attribute information of the target image is updated as the target image is updated.
4. The video playing method of claim 2, wherein when the video playing area plays the off-line video, the target image displayed in the target image analysis display area and the target attribute information of the target image are stored in a configuration file in association with the video frame number,
and reading a target image associated with the video frame number currently played by the offline video and target attribute information of the target image from the configuration file to be displayed in the target image analysis display area.
5. The video playback method according to claim 1, wherein the target attribute information of the target image is displayed in order of attribute confidence from high to low.
6. The video playback method of claim 1, wherein the video playback interface further comprises:
the target image display area is used for displaying a plurality of target images in different areas according to target types, and the number of the target images displayed in the target image display area is greater than or equal to the number of the target images displayed in the target image analysis display area.
7. The video playback method of any one of claims 1 to 6, wherein the video playback interface further comprises:
and the target detail display window is triggered and displayed through the operation on the target image, the target detail display window is used for displaying the target attribute information of the operated target image, and the number of the target attribute information of the target detail display window is more than or equal to the number of the target attribute information displayed in the target image analysis display area.
8. The video playback method according to claim 7, wherein the target attribute information of the target detail display window is edited and then updated to the corresponding target attribute information displayed in the target image analysis display area.
9. The video playback method according to claim 7, wherein the target detail display window displays the operated target image and target images of other target types associated with the operated target image.
10. The video playback method of claim 7, wherein providing the video playback interface further comprises:
receiving a language category selected by a user;
searching a translation text corresponding to the selected language type from an array;
and switching the characters of the video playing interface into the searched translation text.
11. A video playback apparatus, comprising:
an interface providing module configured to provide a video playing interface, the video playing interface comprising:
the video playing area is used for playing videos; and
and the target image analysis display area is used for displaying a plurality of target images acquired from the video played in the video playing area and target attribute information of at least one target image in the plurality of target images.
CN202010953413.7A 2020-09-11 2020-09-11 Video playing method and device Pending CN112055179A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010953413.7A CN112055179A (en) 2020-09-11 2020-09-11 Video playing method and device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010953413.7A CN112055179A (en) 2020-09-11 2020-09-11 Video playing method and device

Publications (1)

Publication Number Publication Date
CN112055179A true CN112055179A (en) 2020-12-08

Family

ID=73611381

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010953413.7A Pending CN112055179A (en) 2020-09-11 2020-09-11 Video playing method and device

Country Status (1)

Country Link
CN (1) CN112055179A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115334346A (en) * 2022-08-08 2022-11-11 北京达佳互联信息技术有限公司 Interface display method, video publishing method, video editing method and device

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101820530A (en) * 2009-02-27 2010-09-01 索尼公司 Image processing equipment, system, method and program and camera apparatus
CN108062510A (en) * 2017-11-17 2018-05-22 维库(厦门)信息技术有限公司 Dynamic display method and computer equipment during a kind of multiple target tracking fructufy
CN110830756A (en) * 2018-08-07 2020-02-21 华为技术有限公司 Monitoring method and device
CN111131902A (en) * 2019-12-13 2020-05-08 华为技术有限公司 Method for determining target object information and video playing equipment

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101820530A (en) * 2009-02-27 2010-09-01 索尼公司 Image processing equipment, system, method and program and camera apparatus
CN108062510A (en) * 2017-11-17 2018-05-22 维库(厦门)信息技术有限公司 Dynamic display method and computer equipment during a kind of multiple target tracking fructufy
CN110830756A (en) * 2018-08-07 2020-02-21 华为技术有限公司 Monitoring method and device
CN111131902A (en) * 2019-12-13 2020-05-08 华为技术有限公司 Method for determining target object information and video playing equipment

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115334346A (en) * 2022-08-08 2022-11-11 北京达佳互联信息技术有限公司 Interface display method, video publishing method, video editing method and device

Similar Documents

Publication Publication Date Title
AU2017254848B2 (en) Image matting using deep learning
US9912874B2 (en) Real-time visual effects for a live camera view
US8269773B2 (en) System and method for drag-and-drop graph building with live feedback
US5010500A (en) Gesture-modified diagram for retrieval of image resembling diagram, with parts selectable for further interactive retrieval
US20090278848A1 (en) Drawing familiar graphs while system determines suitable form
US20240126826A1 (en) System and method for integrating user feedback into website building system services
US20040183800A1 (en) Method and system for dynamic visualization of multi-dimensional data
US10963734B1 (en) Perception visualization tool
US20080044084A1 (en) Integrated human-computer interface for image recognition
CN110598095B (en) Method, device and storage medium for identifying article containing specified information
CN110428003B (en) Sample class label correction method and device and electronic equipment
US11755191B2 (en) Synthetic image data generation incorporating user-provided objects and relationships
Afzal et al. Visualization and visual analytics approaches for image and video datasets: A survey
CN113688290A (en) Interactive electronic maintenance system for vehicle chassis
CN112162672A (en) Information flow display processing method and device, electronic equipment and storage medium
CN115294308A (en) Augmented reality auxiliary assembly operation guiding system based on deep learning
CN112055179A (en) Video playing method and device
Höferlin et al. Interactive schematic summaries for exploration of surveillance video
US20220308742A1 (en) User interface with metadata content elements for video navigation
CN109815977B (en) Cutting and marking integrated method for mass production of machine learning samples
US20230126242A1 (en) Cell image processing system and cell image processing method
US20200082586A1 (en) Generating and providing composition effect tutorials for creating and editing digital content
Wigness Superlabel: A superpixel labeling interface for semantic image annotation
US11704141B2 (en) Real-time context preserving visual guidance
KR102198322B1 (en) Intelligent data visualization system using machine learning

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20201208

RJ01 Rejection of invention patent application after publication