WO2019219065A1 - 视频分析的方法和装置 - Google Patents
视频分析的方法和装置 Download PDFInfo
- Publication number
- WO2019219065A1 WO2019219065A1 PCT/CN2019/087288 CN2019087288W WO2019219065A1 WO 2019219065 A1 WO2019219065 A1 WO 2019219065A1 CN 2019087288 W CN2019087288 W CN 2019087288W WO 2019219065 A1 WO2019219065 A1 WO 2019219065A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- image
- target
- resolution
- image frame
- original resolution
- Prior art date
Links
- 238000004458 analytical method Methods 0.000 title claims abstract description 73
- 238000012545 processing Methods 0.000 claims abstract description 96
- 238000000034 method Methods 0.000 claims abstract description 70
- 238000005070 sampling Methods 0.000 abstract description 3
- 230000008569 process Effects 0.000 description 24
- 230000001133 acceleration Effects 0.000 description 9
- 230000002093 peripheral effect Effects 0.000 description 7
- 238000004891 communication Methods 0.000 description 6
- 238000010586 diagram Methods 0.000 description 6
- 230000003287 optical effect Effects 0.000 description 6
- 230000006870 function Effects 0.000 description 4
- 230000005540 biological transmission Effects 0.000 description 3
- 238000004364 calculation method Methods 0.000 description 3
- 238000001514 detection method Methods 0.000 description 3
- 238000005516 engineering process Methods 0.000 description 3
- 210000000056 organ Anatomy 0.000 description 3
- 238000013473 artificial intelligence Methods 0.000 description 2
- 238000006243 chemical reaction Methods 0.000 description 2
- 210000005069 ears Anatomy 0.000 description 2
- 230000005484 gravity Effects 0.000 description 2
- 238000003672 processing method Methods 0.000 description 2
- 238000009877 rendering Methods 0.000 description 2
- 241000282412 Homo Species 0.000 description 1
- 230000009471 action Effects 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 239000000919 ceramic Substances 0.000 description 1
- 238000013500 data storage Methods 0.000 description 1
- 230000003247 decreasing effect Effects 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 230000001788 irregular Effects 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 238000010801 machine learning Methods 0.000 description 1
- 238000010295 mobile communication Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000012544 monitoring process Methods 0.000 description 1
- 230000009467 reduction Effects 0.000 description 1
- 230000006641 stabilisation Effects 0.000 description 1
- 238000011105 stabilization Methods 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/40—Scenes; Scene-specific elements in video content
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/40—Scenes; Scene-specific elements in video content
- G06V20/49—Segmenting video sequences, i.e. computational techniques such as parsing or cutting the sequence, low-level clustering or determining units such as shots or scenes
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/42—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by implementation details or hardware specially adapted for video compression or decompression, e.g. dedicated software implementation
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/44—Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
- H04N21/4402—Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving reformatting operations of video signals for household redistribution, storage or real-time display
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/44—Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
- H04N21/4402—Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving reformatting operations of video signals for household redistribution, storage or real-time display
- H04N21/440281—Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving reformatting operations of video signals for household redistribution, storage or real-time display by altering the temporal resolution, e.g. by frame skipping
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02D—CLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
- Y02D10/00—Energy efficient computing, e.g. low power processors, power management or thermal management
Definitions
- the present application relates to the field of video surveillance technologies, and in particular, to a method and apparatus for video analysis.
- the intelligent analysis method of the video may be: hard decoding the original video to obtain a YUV (a color image data encoding method) format frame of the original resolution, and then encoding the original resolution YUV format frame into a low resolution.
- the video is then soft decoded by a CPU (Central Processing Unit) to obtain a low resolution YUV format frame, and then the video analysis can be performed based on the low resolution YUV format frame.
- a CPU Central Processing Unit
- Embodiments of the present application provide a method and apparatus for video analysis.
- the technical solution is as follows:
- a method of video analysis comprising:
- Video analysis processing is performed based on the image frame of the target resolution.
- performing video analysis processing on the image frame based on the target resolution including:
- Video analysis processing is performed based on the image frame of the target resolution and the image frame of the original resolution.
- performing video analysis processing on the image frame based on the target resolution and the image frame in the original resolution including:
- a second region image that matches the target image is intercepted and displayed.
- the determining, according to the target resolution and the original resolution, determining a second of the image frames of the original resolution corresponding to the first location information in the image frame of the target resolution Location information including:
- a second region image that matches the target image is determined based on the relative position.
- the intercepting and displaying the second area image that matches the target image in the image frame of the original resolution based on the second location information includes:
- the second region image is intercepted and displayed.
- the determining an information integrity score of the second area image includes:
- an apparatus for video analysis comprising:
- a first processing module configured to perform hardware decoding on the video data to obtain an image frame of an original resolution
- a second processing module configured to perform hardware downsampling processing on the image frame of the original resolution to obtain an image frame of a preset target resolution, where the target resolution is smaller than the original resolution;
- an analysis module configured to perform video analysis processing on the image frame based on the target resolution.
- the analyzing module is configured to:
- Video analysis processing is performed based on the image frame of the target resolution and the image frame of the original resolution.
- the analyzing module is configured to:
- a second region image that matches the target image is intercepted and displayed.
- the analyzing module is configured to:
- a second region image that matches the target image is determined based on the relative position.
- the analyzing module is configured to:
- the second region image is intercepted and displayed.
- the analyzing module is configured to:
- an electronic device comprising a processor and a memory, wherein the memory stores at least one instruction, at least one program, a code set or a set of instructions, the at least one instruction, the at least one A program, the set of codes, or a set of instructions loaded by the processor and executed to implement the method of video analysis as described in the first aspect above.
- a fourth aspect provides a computer readable storage medium, where the storage medium stores at least one instruction, at least one program, a code set, or a set of instructions, the at least one instruction, the at least one program, and the code
- a set or set of instructions is loaded by the processor and executed to implement the method of video analysis as described in the first aspect above.
- the video data to be analyzed is obtained; the video data is hardware decoded to obtain an image frame of a original resolution; and the image frame of the original resolution is subjected to hardware downsampling to obtain a preset target resolution image. a frame in which the target resolution is smaller than the original resolution; the image frame based on the target resolution is subjected to video analysis processing.
- the processing steps can be simplified, thereby improving the processing efficiency of intelligent analysis of the video.
- FIG. 1 is a flowchart of a method for video analysis provided by an embodiment of the present application.
- FIG. 2 is a flowchart of a method for video analysis provided by an embodiment of the present application.
- FIG. 3 is a flowchart of a method for video analysis provided by an embodiment of the present application.
- FIG. 4 is a schematic diagram of an interface of a video analysis method according to an embodiment of the present application.
- FIG. 5 is a schematic structural diagram of an apparatus for video analysis according to an embodiment of the present application.
- FIG. 6 is a schematic structural diagram of a terminal according to an embodiment of the present disclosure.
- FIG. 7 is a schematic structural diagram of a server provided by an embodiment of the present application.
- the embodiment of the present application provides a method for video analysis, which may be implemented by an electronic device.
- the electronic device can be a terminal or a server.
- the terminal can include components such as a processor, a memory, and the like.
- the processor can be a CPU (Central Processing Unit) and a GPU (Graphics Processing Unit), and can be used for hardware decoding of video data and hardware downsampling of image frames of original resolution.
- the image frame based on the target resolution is subjected to video analysis processing and the like.
- the memory may be a RAM (Random Access Memory), a Flash (flash memory), etc., and may be used to store received data, data required for processing, data generated during processing, and the like, such as video data, Image frames of the original resolution, image frames of the target resolution, and the like.
- the terminal may also include a screen, a transceiver, an image detecting component, an audio output component, an audio input component, and the like.
- the screen can be used to display the captured area image and so on.
- the transceiver can be used for data transmission with other devices, and can include an antenna, a matching circuit, a modem, and the like.
- the image detecting unit may be a camera or the like.
- the audio output unit can be a speaker, a headphone, or the like.
- the audio input component can be a microphone or the like.
- the server can include components such as a processor, a memory, a transceiver, and the like.
- the processor can be a CPU (Central Processing Unit) and a GPU (Graphics Processing Unit), and can be used for hardware decoding of video data and hardware downsampling of image frames of original resolution.
- the image frame based on the target resolution is subjected to video analysis processing and the like.
- the memory may be a RAM (Random Access Memory), a Flash (flash memory), etc., and may be used to store received data, data required for processing, data generated during processing, and the like, such as video data, Image frames of the original resolution, image frames of the target resolution, and the like.
- the transceiver can be used for data transmission with a terminal or other server (such as a positioning server), for example, sending a second area image to the terminal, and the transceiver can include an antenna, a matching circuit, a modem, and the like.
- a terminal or other server such as a positioning server
- the transceiver can include an antenna, a matching circuit, a modem, and the like.
- the processing flow of the method may include the following steps:
- step 101 video data to be analyzed is acquired.
- the video data can be acquired first.
- the acquired video data may be a piece of video data in the surveillance video, for example, the user wants to find a certain shooting target by monitoring the video, and the like.
- the acquired video data may also be a video clip in the movie video, for example, the user wants to perform special effects processing on a video clip in the movie, and the like.
- the obtained video data may also be other acquisition methods, which is not limited in this application.
- step 102 the video data is hardware decoded to obtain an image frame of the original resolution.
- decoding is a process of restoring a digital code to a content it represents or converting an electrical pulse signal into information, data, etc., represented by a specific method.
- Hardware decoding is a hardware, such as a GPU (Graphics Processing Unit, Graphics processor), a way of decoding a decoded video stream.
- the video data is hardware-decoded by the GPU to obtain an image frame of the original resolution.
- the method used for hardware decoding is an existing hardware decoding method, which is not described herein.
- step 103 the image frame of the original resolution is subjected to hardware downsampling processing to obtain an image frame of a preset target resolution, wherein the target resolution is smaller than the original resolution.
- downsampling is a processing method that reduces the number of sampling points.
- the quality of the image after downsampling processing is reduced, and image processing of the downsampled processed image can reduce the amount of processing. For example, for an image, if the downsampling coefficient is k, that is, one dot is taken every k points of each row and column in the original image to form an image.
- Hardware downsampling is a downsampling method implemented in hardware.
- the image frame of the original resolution may be downsampled, in order to improve processing efficiency and reduce
- the burden of the CPU can directly downsample the image frame of the original resolution on the hardware according to the preset downsampling rate, and obtain an image frame of a preset target resolution, and the target resolution must be smaller than the original resolution. In this way, it is not necessary to copy the original resolution image frame to the CPU, which can alleviate the problem of low efficiency caused by multiple copies of the data, and can reduce the processing load of the CPU.
- the downsampling process may be a proportional downsampling process, that is, the obtained image frame of the target resolution is the same as the aspect ratio of the image frame of the original resolution; the downsampling process may also be unequal.
- the downsampling process of the ratio that is, the obtained image frame of the target resolution is different from the aspect ratio of the image frame of the original resolution, which is not limited in this application.
- step 104 video analysis processing is performed based on the image frame of the target resolution.
- different video analysis processes may be performed based on the image frames of the target resolution and the image frames of the original resolution, respectively.
- the algorithm module with higher speed and efficiency can perform algorithm operation based on the image frame of the target resolution; the algorithm module with higher image quality can perform algorithm operation based on the image frame of the original resolution, the flow of the solution Can be as shown in Figure 2.
- the algorithm for recognizing and capturing the face is taken as an example, and the human face can be recognized according to the image frame of the target resolution and the image frame of the original resolution, and the recognized face image is intercepted and displayed, as shown in FIG. 3 .
- the processing of the above steps may be: determining, based on the acquired target image, first position information of the first area image that matches the target image in the image frame of the target resolution; based on the target resolution and the original resolution Determining second position information in the image frame of the original resolution corresponding to the first position information in the image frame of the target resolution; based on the second position information, intercepting and displaying in the image frame of the original resolution A second area image that matches the target image.
- the location information may be coordinate information of four vertex angles of the area image.
- the image of the captured target can be input to the electronic device, and the target is The image is matched with the image frame of the target resolution, the area image of the target image is matched with the target image (ie, the first area image), and the position information of the first area image is determined (ie, the first position) information).
- the second location information in the image frame of the original resolution corresponding to the first location information is obtained by scaling.
- a scheme of equal-scale downsampling processing can be adopted. If the ratio is 1:1, that is, the size of the image frame of the original resolution is the same as the size of the image frame of the target resolution, the first position information and the first position information The two position information is the same. Therefore, the area image corresponding to the second position information (ie, the second area image) can be determined in the image frame of the original resolution, and the second area image is the area image matching the target image. The second area image is intercepted and displayed to the user. If it is a scheme of equal-scale downsampling processing, and the ratio is not 1:1, that is, the size of the image frame of the original resolution is different from the size of the image frame of the target resolution, the first position information may be calculated first.
- the relative position of the first position information in the image frame of the target resolution may be determined first, and then, in the image frame of the original resolution, the second area image matching the target image is determined according to the determined relative position.
- the size of the image frame of the target resolution is 180 ⁇ 240 pixels
- the size of the image frame of the original resolution is 720 ⁇ 960 pixels
- the obtained first position information is (30, 40), (120, 40, respectively).
- the ratio information of the aspect ratio of the image frame of the first position information to the target resolution is (1/6, 1/6), (2). /3, 1/6), (1/6, 3/4), (2/3, 3/4)
- the scale information is calculated with the size of the image frame of the original resolution to obtain the image of the second region
- the second location information is (120, 160), (480, 160), (120, 720), (480, 720). Based on the obtained second location information, the second region image is determined, and then the second region image is intercepted and displayed to the user.
- the above conversion process may also adopt a scheme of unequal proportional down-sampling processing, and the method for determining the second region image according to the position information is the same as the equal-scale downsampling processing method in which the ratio is not 1:1, that is, first A relative position of the first position information in the image frame of the target resolution is determined, and then, in the image frame of the original resolution, the second area image matching the target image is determined according to the determined relative position.
- the size of the image frame of the target resolution is 180 ⁇ 240 pixels
- the size of the image frame of the original resolution is 720 ⁇ 720 pixels
- the first position information of the obtained first area image is (30, respectively).
- the ratio information (ie, relative position information) of the aspect ratio of the image frame of the first position information to the target resolution can be obtained (1/6, 1/6), (2/3, 1/6), (1/6, 3/4), (2/3, 3/4), the scale information is calculated with the size of the original resolution image frame
- second position information of the second area image which are (120, 120), (480, 120), (120, 540), (480, 540), respectively.
- the second region image is determined, and then the second region image is intercepted and displayed to the user.
- the above-exemplified manner of calculating the proportional information is only a feasible manner, and other calculation manners may be used, for example, first calculating the aspect ratio of the original resolution image frame and the target resolution.
- the ratio of the aspect ratio of the image frame is calculated according to the ratio, and the position information of the image of the second region is calculated according to the ratio.
- the matching method for matching the target image with the image frame of the target resolution may be: matching the target image with the preset first region image of the image frame of the target resolution, and calculating the target image. And a first matching degree of the preset first area image, and storing the first matching degree corresponding to the coordinates of the four top angles of the preset first area image, wherein the size of the first area image and the target image are preset The dimensions are exactly the same; then the abscissas of the four apex angles of the preset first area image are all increased by the same preset first increment to obtain a new preset first area image, and the new preset first area is The image is matched with the target image to obtain a second matching degree between the new preset first region image and the target image, and the second matching degree is compared with the stored first matching degree, and a relatively large matching degree between the two is obtained.
- the coordinates of the four apex angles of the preset first area image corresponding to the matching degree are stored, and the smaller matching degree and the coordinates of the four apex angles of the preset first area image corresponding to the matching degree are deleted.
- the abscissa of the four apex angles of the preset first area image is again increased by the same preset first increment to obtain a new preset first area image, and the new preset first area image is Matching with the target image to obtain a third matching degree between the new preset first area image and the target image, comparing the third matching degree with the stored matching degree, and the relatively large matching degree between the two and the matching
- the coordinates of the four apex angles of the preset first area image are stored, and the smaller matching degree and the coordinates of the four apex angles of the preset first area image corresponding to the matching degree are deleted...
- the abscissa of the two vertices of the preset first region image reaches the maximum value or the minimum value, that is, when the image of the first region image has reached the target resolution of the image frame.
- four The ordinate of the vertex is increased by a preset second increment to obtain a new preset first area image, and the new preset first area image is matched with the target image, and the target image and the preset first area image are calculated.
- the abscissas of the four apex angles of the preset first area image are all reduced by the same preset first increment to obtain a new preset first area image, and the new preset first area image and the target image are obtained.
- Performing matching obtaining a fifth matching degree between the new preset first area image and the target image, comparing the fifth matching degree with the stored matching degree, and comparing the relatively large matching degree of the two with the matching degree.
- the coordinates of the four apex angles of the preset first area image are stored, and the smaller matching degree and the coordinates of the four apex angles of the preset first area image corresponding to the matching degree are deleted.
- the maximum matching degree of the target image and all the preset first area images, and the coordinates of the four apex angles of the preset first area image corresponding to the matching degree are obtained, and the coordinates are the first Location information of the area image.
- the method of matching the target image with the image frame of the target resolution, determining the first region image in which the image frame of the target resolution matches the target image, and determining the position information of the image of the first region is further
- the image recognition model can be trained by the sample to obtain the trained image recognition model, and the image frame of the target image and the target resolution is input into the image recognition model, and the image frame of the target resolution and the target image can be obtained.
- the image recognition model is used to determine the position information of the image of the first region, and the determined speed is faster and the efficiency is higher.
- the method for matching the target image with the image frame of the target resolution, determining the first region image that matches the target image and the target image, and determining the location information of the first region image is further implemented.
- this application is not mentioned in this example, as long as the image of the target image and the target resolution is matched, the first region image of the image frame of the target resolution and the target image is determined, and the first The location information of an area image can be used. What method is used in this application is not limited in this application.
- the information integrity score processing may be performed on the second region image before the second region image is intercepted, as shown in FIG. 3, and corresponding processing is performed.
- the method may be as follows: determining, according to the second location information, a second region image that matches the target image in the image frame of the original resolution; determining an information integrity score of the second region image; when the information integrity score is greater than a preset rating At the threshold, the second area image is captured and displayed.
- the position information of the first area image is the coordinate information of the four vertices of the first area image, and then the four vertices of the second area image are determined at the original resolution image frame according to the four coordinate information, and then according to the four vertices Determine the second area image.
- the information integrity score processing is performed on the at least one second region image respectively, and the information integrity score of each second region image is obtained, and the information integrity score is higher, indicating the information integrity score.
- the display information of the corresponding second area image is more comprehensive. Therefore, the information integrity score is compared with the preset score threshold. If the information integrity score of the image of the second area is greater than the preset score threshold, the location information is intercepted according to the location information.
- the second area image is displayed to the user, as shown in FIG. In this way, the image of the second area displayed to the user is more complete and contains more useful information, so that the information acquired by the user is more comprehensive.
- the step of the information integrity scoring process may be based on: determining an information integrity score of the second region image based on the resolution of the region image corresponding to the location information, the photographic target integrity, and the photographic target capturing angle.
- one or more of the sharpness of the regional image, the completeness of the shooting target, and the shooting angle of the shooting target may be used as the basis of the information integrity scoring process when performing the information integrity scoring process on the regional image.
- the definition refers to the clarity of each detail and its boundary on the image. The higher the definition, the higher the resolution score of the image quality score, the lower the definition, and the highest the clarity score of the information integrity score.
- the completeness of the shooting target refers to the completeness of each component of the area image as the shooting target. For example, if the shooting target is a dog, it is determined whether the image of the area contains all the body parts of the dog, such as the head, the ears, the limbs, and the tail. The more the body parts are included, the more complete the shooting target of the image of the area; for example, if the shooting target is a human face, it is determined whether the image of the area contains all the organ parts of the face, such as hair and ears. The organ parts of the two eyes, the mouth, the chin, and the like, and the more the organ parts are included, the higher the degree of completeness of the image of the area. The higher the completeness of the target, the higher the score of the completeness of the target of the information integrity score, the lower the completeness of the target, and the lower the score of the completeness of the target of the information integrity score.
- the shooting angle of the shooting target is a basis for the shooting target to be a human face.
- the shooting face is a positive face
- the information that the user can obtain is more comprehensive, and the angle of the face of the shooting is larger. The less information the user can get. Therefore, it can be set that when the photographed face is a positive face, the photographing target photographing angle is 0 degree, and the photographing target photographing angle score of the information integrity score is the highest; the angle of the side turn is larger, the photographing target photographing angle is larger, the information The lower the score of the shooting target angle of the integrity score.
- the above-mentioned definition, the completeness of the shooting target, and the shooting angle of the shooting target are only examples of the rating basis of the present application. According to the actual application, other information can be used to perform information integrity scoring processing on the regional image. For example, the contrast of the area image, etc., is not limited in this application.
- the information integrity score processing is performed on the regional image according to the preset scoring basis, and finally the information integrity score of the regional image is obtained.
- the video data to be analyzed is obtained; the video data is hardware decoded to obtain an image frame of a original resolution; and the image frame of the original resolution is subjected to hardware downsampling to obtain a preset target. a resolution image frame, wherein the target resolution is smaller than the original resolution; a video analysis process is performed based on the image frame of the target resolution.
- the processing steps can be simplified, thereby improving the processing efficiency of intelligent analysis of the video.
- the embodiment of the present application further provides a device for video analysis, which may be the electronic device in the above embodiment.
- the device includes: an obtaining module 510, and a first processing module. 520.
- the obtaining module 510 is configured to acquire video data to be analyzed.
- the first processing module 520 is configured to perform hardware decoding on the video data to obtain an image frame of original resolution
- the second processing module 530 is configured to perform hardware downsampling processing on the image frame of the original resolution to obtain an image frame of a preset target resolution, where the target resolution is smaller than the original resolution;
- the analysis module 540 is configured to perform video analysis processing based on the image frame of the target resolution.
- the analyzing module 540 is configured to:
- Video analysis processing is performed based on the image frame of the target resolution and the image frame of the original resolution.
- the analyzing module 540 is configured to:
- a second region image that matches the target image is intercepted and displayed.
- the analyzing module 540 is configured to:
- the analyzing module 540 is configured to:
- the second region image is intercepted and displayed.
- the analyzing module 540 is configured to:
- the video data to be analyzed is obtained; the video data is hardware decoded to obtain an image frame of a original resolution; and the image frame of the original resolution is subjected to hardware downsampling to obtain a preset target.
- the device for video analysis provided by the foregoing embodiment is only illustrated by the division of each functional module. In actual applications, the function distribution may be completed by different functional modules as needed. The internal structure of the electronic device is divided into different functional modules to complete all or part of the functions described above.
- the apparatus for video analysis provided by the foregoing embodiment is the same as the method embodiment of the video analysis, and the specific implementation process is described in detail in the method embodiment, and details are not described herein again.
- FIG. 6 is a structural block diagram of a terminal according to an embodiment of the present application.
- the terminal 600 can be a portable mobile terminal, such as a smart phone or a tablet computer.
- Terminal 600 may also be referred to as a user device, a portable terminal, or the like.
- the terminal 600 includes a processor 601 and a memory 602.
- Processor 601 can include one or more processing cores, such as a 4-core processor, a 6-core processor, and the like.
- the processor 601 may be configured by at least one of a DSP (Digital Signal Processing), an FPGA (Field-Programmable Gate Array), and a PLA (Programmable Logic Array). achieve.
- the processor 601 may also include a main processor and a coprocessor.
- the main processor is a processor for processing data in an awake state, which is also called a CPU (Central Processing Unit); the coprocessor is A low-power processor for processing data in standby.
- the processor 601 may be integrated with a GPU (Graphics Processing Unit), which is responsible for rendering and rendering of the content that the display needs to display.
- the processor 601 may also include an AI (Artificial Intelligence) processor for processing computational operations related to machine learning.
- AI Artificial Intelligence
- Memory 602 can include one or more computer readable storage media that can be tangible and non-transitory. Memory 602 can also include high speed random access memory, as well as non-volatile memory, such as one or more disk storage devices, flash storage devices. In some embodiments, the non-transitory computer readable storage medium in memory 602 is for storing at least one instruction for a method performed by processor 601 to implement the video analysis provided in this application.
- the terminal 600 optionally further includes: a peripheral device interface 603 and at least one peripheral device.
- the peripheral device includes at least one of a radio frequency circuit 604, a touch display screen 605, a camera 606, an audio circuit 607, a positioning component 608, and a power source 609.
- the peripheral device interface 603 can be used to connect at least one peripheral device associated with an I/O (Input/Output) to the processor 601 and the memory 602.
- processor 601, memory 602, and peripheral interface 603 are integrated on the same chip or circuit board; in some other embodiments, any of processor 601, memory 602, and peripheral interface 603 or The two can be implemented on a separate chip or circuit board, which is not limited in this embodiment.
- the RF circuit 604 is configured to receive and transmit an RF (Radio Frequency) signal, also referred to as an electromagnetic signal.
- Radio frequency circuit 604 communicates with the communication network and other communication devices via electromagnetic signals.
- the RF circuit 604 converts the electrical signal into an electromagnetic signal for transmission, or converts the received electromagnetic signal into an electrical signal.
- the radio frequency circuit 604 includes an antenna system, an RF transceiver, one or more amplifiers, a tuner, an oscillator, a digital signal processor, a codec chipset, a subscriber identity module card, and the like.
- Radio frequency circuitry 604 can communicate with other terminals via at least one wireless communication protocol.
- the wireless communication protocols include, but are not limited to, the World Wide Web, a metropolitan area network, an intranet, generations of mobile communication networks (2G, 3G, 4G, and 5G), wireless local area networks, and/or WiFi (Wireless Fidelity) networks.
- the radio frequency circuit 604 may also include a circuit related to NFC (Near Field Communication), which is not limited in this application.
- the touch display 605 is used to display a UI (User Interface).
- the UI can include graphics, text, icons, video, and any combination thereof.
- Touch display 605 also has the ability to capture touch signals over the surface or surface of touch display 605.
- the touch signal can be input to the processor 601 as a control signal for processing.
- Touch display 605 is used to provide virtual buttons and/or virtual keyboards, also referred to as soft buttons and/or soft keyboards.
- the touch display screen 605 may be one, and the front panel of the terminal 600 is set; in other embodiments, the touch display screen 605 may be at least two, respectively disposed on different surfaces of the terminal 600 or in a folded design.
- the touch display 605 can be a flexible display disposed on a curved surface or a folded surface of the terminal 600. Even the touch display screen 605 can be set to a non-rectangular irregular pattern, that is, a profiled screen.
- the touch display screen 605 can be prepared by using an LCD (Liquid Crystal Display) or an OLED (Organic Light-Emitting Diode).
- Camera component 606 is used to capture images or video.
- camera assembly 606 includes a front camera and a rear camera.
- the front camera is used for video calls or self-timer
- the rear camera is used for photo or video capture.
- the rear camera is at least two, which are respectively a main camera, a depth of field camera, and a wide-angle camera, so that the main camera and the depth of field camera are combined to realize the background blur function, and the main camera and the wide-angle camera are integrated.
- Panoramic shooting and VR (Virtual Reality) shooting can also include a flash.
- the flash can be a monochrome temperature flash or a two-color temperature flash.
- the two-color temperature flash is a combination of a warm flash and a cool flash that can be used for light compensation at different color temperatures.
- Audio circuit 607 is used to provide an audio interface between the user and terminal 600.
- the audio circuit 607 can include a microphone and a speaker.
- the microphone is used to collect sound waves of the user and the environment, and convert the sound waves into electrical signals for processing to the processor 601 for processing, or input to the radio frequency circuit 604 for voice communication.
- the microphones may be multiple, and are respectively disposed at different parts of the terminal 600.
- the microphone can also be an array microphone or an omnidirectional acquisition microphone.
- the speaker is then used to convert electrical signals from the processor 601 or the RF circuit 604 into sound waves.
- the speaker can be either a conventional film speaker or a piezoceramic speaker.
- audio circuit 607 can also include a headphone jack.
- the location component 608 is used to locate the current geographic location of the terminal 600 to implement navigation or LBS (Location Based Service).
- the positioning component 608 can be a positioning component based on a US-based GPS (Global Positioning System), a Chinese Beidou system, or a Russian Galileo system.
- Power source 609 is used to power various components in terminal 600.
- the power source 609 can be an alternating current, a direct current, a disposable battery, or a rechargeable battery.
- the rechargeable battery may be a wired rechargeable battery or a wireless rechargeable battery.
- a wired rechargeable battery is a battery that is charged by a wired line
- a wireless rechargeable battery is a battery that is charged by a wireless coil.
- the rechargeable battery can also be used to support fast charging technology.
- terminal 600 also includes one or more sensors 610.
- the one or more sensors 610 include, but are not limited to, an acceleration sensor 611, a gyro sensor 612, a pressure sensor 613, a fingerprint sensor 614, an optical sensor 615, and a proximity sensor 616.
- the acceleration sensor 611 can detect the magnitude of the acceleration on the three coordinate axes of the coordinate system established by the terminal 600.
- the acceleration sensor 611 can be used to detect components of gravity acceleration on three coordinate axes.
- the processor 601 can control the touch display screen 605 to display the user interface in a landscape view or a portrait view according to the gravity acceleration signal collected by the acceleration sensor 611.
- the acceleration sensor 611 can also be used for the acquisition of game or user motion data.
- the gyro sensor 612 can detect the body direction and the rotation angle of the terminal 600, and the gyro sensor 612 can cooperate with the acceleration sensor 611 to collect the 3D motion of the user to the terminal 600. Based on the data collected by the gyro sensor 612, the processor 601 can implement functions such as motion sensing (such as changing the UI according to the user's tilting operation), image stabilization at the time of shooting, game control, and inertial navigation.
- functions such as motion sensing (such as changing the UI according to the user's tilting operation), image stabilization at the time of shooting, game control, and inertial navigation.
- the pressure sensor 613 may be disposed at a side border of the terminal 600 and/or a lower layer of the touch display screen 605.
- the pressure sensor 613 When the pressure sensor 613 is disposed at the side frame of the terminal 600, the user's holding signal to the terminal 600 can be detected, and the left and right hand recognition or shortcut operation is performed according to the holding signal.
- the operability control on the UI interface can be controlled according to the user's pressure operation on the touch display screen 605.
- the operability control includes at least one of a button control, a scroll bar control, an icon control, and a menu control.
- the fingerprint sensor 614 is configured to collect a fingerprint of the user to identify the identity of the user according to the collected fingerprint. Upon identifying that the user's identity is a trusted identity, the processor 601 authorizes the user to perform related sensitive operations including unlocking the screen, viewing encrypted information, downloading software, paying and changing settings, and the like.
- the fingerprint sensor 614 can be disposed on the front, back, or side of the terminal 600. When the physical button or vendor logo is provided on the terminal 600, the fingerprint sensor 614 can be integrated with the physical button or the manufacturer logo.
- Optical sensor 615 is used to collect ambient light intensity.
- the processor 601 can control the display brightness of the touch display 605 according to the ambient light intensity acquired by the optical sensor 615. Specifically, when the ambient light intensity is high, the display brightness of the touch display screen 605 is raised; when the ambient light intensity is low, the display brightness of the touch display screen 605 is lowered.
- the processor 601 can also dynamically adjust the shooting parameters of the camera assembly 606 according to the ambient light intensity collected by the optical sensor 615.
- Proximity sensor 616 also referred to as a distance sensor, is typically disposed on the front side of terminal 600. Proximity sensor 616 is used to capture the distance between the user and the front of terminal 600. In one embodiment, when the proximity sensor 616 detects that the distance between the user and the front side of the terminal 600 is gradually decreasing, the touch screen display 605 is controlled by the processor 601 to switch from the bright screen state to the screen state; when the proximity sensor 616 detects When the distance between the user and the front side of the terminal 600 gradually becomes larger, the processor 601 controls the touch display screen 605 to switch from the screen state to the bright screen state.
- FIG. 6 does not constitute a limitation to the terminal 600, and may include more or less components than those illustrated, or may combine some components or adopt different component arrangements.
- a computer readable storage medium having stored therein at least one instruction, at least one program, code set or instruction set, at least one instruction, at least one program, code set or instruction set A method of being loaded and executed by a processor to implement the recognition action category in the above embodiments.
- the computer readable storage medium can be a ROM, a random access memory (RAM), a CD-ROM, a magnetic tape, a floppy disk, and an optical data storage device.
- the video data to be analyzed is obtained; the video data is hardware decoded to obtain an image frame of a original resolution; and the image frame of the original resolution is subjected to hardware downsampling to obtain a preset target.
- FIG. 7 is a schematic structural diagram of a server according to an embodiment of the present application.
- the server 700 may generate a large difference due to different configurations or performances, and may include one or more central processing units (CPUs) 701 and one. Or more than one memory 702, wherein the memory 702 stores at least one instruction that is loaded and executed by the processor 701 to implement the method steps of video analysis described below:
- CPUs central processing units
- memory 702 stores at least one instruction that is loaded and executed by the processor 701 to implement the method steps of video analysis described below:
- Video analysis processing is performed based on the image frame of the target resolution.
- the at least one instruction is loaded and executed by the processor 701 to implement the following method steps:
- Video analysis processing is performed based on the image frame of the target resolution and the image frame of the original resolution.
- the at least one instruction is loaded and executed by the processor 701 to implement the following method steps:
- a second region image that matches the target image is intercepted and displayed.
- the at least one instruction is loaded and executed by the processor 701 to implement the following method steps:
- a second region image that matches the target image is determined based on the relative position.
- the at least one instruction is loaded and executed by the processor 701 to implement the following method steps:
- the second region image is intercepted and displayed.
- the at least one instruction is loaded and executed by the processor 701 to implement the following method steps:
- the video data to be analyzed is obtained; the video data is hardware decoded to obtain an image frame of a original resolution; and the image frame of the original resolution is subjected to hardware downsampling to obtain a preset target.
- a person skilled in the art may understand that all or part of the steps of implementing the above embodiments may be completed by hardware, or may be instructed by a program to execute related hardware, and the program may be stored in a computer readable storage medium.
- the storage medium mentioned may be a read only memory, a magnetic disk or an optical disk or the like.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Computing Systems (AREA)
- Image Analysis (AREA)
- Image Processing (AREA)
Abstract
Description
Claims (14)
- 一种视频分析的方法,所述方法包括:获取待分析的视频数据;对所述视频数据进行硬件解码,得到原始分辨率的图像帧;对所述原始分辨率的图像帧进行硬件降采样处理,得到预设的目标分辨率的图像帧,其中,所述目标分辨率小于所述原始分辨率;基于所述目标分辨率的图像帧进行视频分析处理。
- 根据权利要求1所述的方法,所述基于所述目标分辨率的图像帧进行视频分析处理,包括:基于所述目标分辨率的图像帧和所述原始分辨率的图像帧进行视频分析处理。
- 根据权利要求2所述的方法,所述基于所述目标分辨率的图像帧和所述原始分辨率的图像帧进行视频分析处理,包括:基于获取的目标图像,在所述目标分辨率的图像帧中,确定与所述目标图像相匹配的第一区域图像的第一位置信息;基于所述目标分辨率和所述原始分辨率,确定与所述目标分辨率的图像帧中的第一位置信息相对应的所述原始分辨率的图像帧中的第二位置信息;基于所述第二位置信息,在所述原始分辨率的图像帧中,截取并显示与所述目标图像相匹配的第二区域图像。
- 根据权利要求3所述的方法,所述基于所述目标分辨率和所述原始分辨率,确定与所述目标分辨率的图像帧中的第一位置信息相对应的所述原始分辨率的图像帧中的第二位置信息,包括:确定所述第一位置信息在所述目标分辨率的图像帧中的相对位置;在所述原始分辨率的图像帧中,根据所述相对位置,确定与所述目标图像相匹配的第二区域图像。
- 根据权利要求3所述的方法,所述基于所述第二位置信息,在所述原始分辨率的图像帧中,截取并显示与所述目标图像相匹配的第二区域图像,包括:基于所述第二位置信息,在所述原始分辨率的图像帧中,确定与所述目标图像相匹配的第二区域图像;确定所述第二区域图像的信息完整性评分;当所述信息完整性评分大于预设评分阈值时,截取并显示所述第二区域图像。
- 根据权利要求5所述的方法,所述确定所述第二区域图像的信息完整性评分,包括:基于所述位置信息对应的区域图像的清晰度、拍摄目标完整度和拍摄目标拍摄角度,确定所述第二区域图像的信息完整性评分。
- 一种视频分析的装置,所述装置包括:获取模块,用于获取待分析的视频数据;第一处理模块,用于对所述视频数据进行硬件解码,得到原始分辨率的图像帧;第二处理模块,用于对所述原始分辨率的图像帧进行硬件降采样处理,得到预设的目标分辨率的图像帧,其中,所述目标分辨率小于所述原始分辨率;分析模块,用于基于所述目标分辨率的图像帧进行视频分析处理。
- 根据权利要求7所述的装置,所述分析模块,用于:基于所述目标分辨率的图像帧和所述原始分辨率的图像帧进行视频分析处理。
- 根据权利要求8所述的装置,所述分析模块,用于:基于获取的目标图像,在所述目标分辨率的图像帧中,确定与所述目标图像相匹配的第一区域图像的第一位置信息;基于所述目标分辨率和所述原始分辨率,确定与所述目标分辨率的图像帧中的第一位置信息相对应的所述原始分辨率的图像帧中的第二位置信息;基于所述第二位置信息,在所述原始分辨率的图像帧中,截取并显示与所述目标图像相匹配的第二区域图像。
- 根据权利要求9所述的装置,所述分析模块,用于:确定所述第一位置信息在所述目标分辨率的图像帧中的相对位置;在所述原始分辨率的图像帧中,根据所述相对位置,确定与所述目标图像相匹配的第二区域图像。
- 根据权利要求9所述的装置,所述分析模块,用于:基于所述第二位置信息,在所述原始分辨率的图像帧中,确定与所述目标图像相匹配的第二区域图像;确定所述第二区域图像的信息完整性评分;当所述信息完整性评分大于预设评分阈值时,截取并显示所述第二区域图像。
- 根据权利要求11所述的装置,所述分析模块,用于:基于所述位置信息对应的区域图像的清晰度、拍摄目标完整度和拍摄目标拍摄角度,确定所述第二区域图像的信息完整性评分。
- 一种电子设备,所述电子设备包括处理器和存储器,所述存储器中存储有至少一条指令、至少一段程序、代码集或指令集,所述至少一条指令、所述至少一段程序、所述代码集或指令集由所述处理器加载并执行以实现如权利要求1至6任一所述的视频分析的方法。
- 一种计算机可读存储介质,所述存储介质中存储有至少一条指令、至少一段程序、代码集或指令集,所述至少一条指令、所述至少一段程序、所述代码集或指令集由所述处理器加载并执行以实现如权利要求1至6任一所述的视频分析的方法。
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201810473779.7 | 2018-05-17 | ||
CN201810473779.7A CN110502954B (zh) | 2018-05-17 | 2018-05-17 | 视频分析的方法和装置 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2019219065A1 true WO2019219065A1 (zh) | 2019-11-21 |
Family
ID=68539492
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2019/087288 WO2019219065A1 (zh) | 2018-05-17 | 2019-05-16 | 视频分析的方法和装置 |
Country Status (2)
Country | Link |
---|---|
CN (1) | CN110502954B (zh) |
WO (1) | WO2019219065A1 (zh) |
Cited By (18)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111726536A (zh) * | 2020-07-03 | 2020-09-29 | 腾讯科技(深圳)有限公司 | 视频生成方法、装置、存储介质及计算机设备 |
CN111753784A (zh) * | 2020-06-30 | 2020-10-09 | 广州酷狗计算机科技有限公司 | 视频的特效处理方法、装置、终端及存储介质 |
CN111859549A (zh) * | 2020-07-28 | 2020-10-30 | 奇瑞汽车股份有限公司 | 单一配置整车重量与重心信息的确定方法及相关设备 |
CN112135190A (zh) * | 2020-09-23 | 2020-12-25 | 成都市喜爱科技有限公司 | 视频处理方法、装置、系统、服务器及存储介质 |
CN112528760A (zh) * | 2020-11-24 | 2021-03-19 | 腾讯科技(深圳)有限公司 | 图像处理方法、装置、计算机设备及介质 |
CN112541429A (zh) * | 2020-12-08 | 2021-03-23 | 浙江大华技术股份有限公司 | 一种智能抓图方法、装置、电子设备以及存储介质 |
CN112804573A (zh) * | 2021-01-08 | 2021-05-14 | Oppo广东移动通信有限公司 | 电视信号的处理方法、装置、电子设备和存储介质 |
CN112817768A (zh) * | 2021-02-26 | 2021-05-18 | 北京梧桐车联科技有限责任公司 | 动画处理方法、装置、设备及计算机可读存储介质 |
CN113032590A (zh) * | 2021-03-29 | 2021-06-25 | 广州繁星互娱信息科技有限公司 | 特效展示方法、装置、计算机设备及计算机可读存储介质 |
CN113365027A (zh) * | 2021-05-28 | 2021-09-07 | 上海商汤智能科技有限公司 | 视频处理方法及装置、电子设备和存储介质 |
CN113392267A (zh) * | 2020-03-12 | 2021-09-14 | 平湖莱顿光学仪器制造有限公司 | 一种用于生成目标对象的二维显微视频信息的方法与设备 |
CN113422967A (zh) * | 2021-06-07 | 2021-09-21 | 深圳康佳电子科技有限公司 | 一种投屏显示控制方法、装置、终端设备及存储介质 |
CN113435530A (zh) * | 2021-07-07 | 2021-09-24 | 腾讯科技(深圳)有限公司 | 图像识别方法、装置、计算机设备及计算机可读存储介质 |
CN113561916A (zh) * | 2021-08-31 | 2021-10-29 | 长沙德壹科技有限公司 | 一种车载显示系统、车辆及车载摄像头图像显示方法 |
CN113992880A (zh) * | 2021-10-15 | 2022-01-28 | 上海佰贝科技发展股份有限公司 | 4k视频识别方法、系统、设备及计算机可读存储介质 |
CN114187349A (zh) * | 2021-11-03 | 2022-03-15 | 深圳市正运动技术有限公司 | 产品加工方法、装置、终端设备以及存储介质 |
CN114827567A (zh) * | 2022-03-23 | 2022-07-29 | 阿里巴巴(中国)有限公司 | 视频质量分析方法、设备和可读介质 |
CN116684626A (zh) * | 2023-08-04 | 2023-09-01 | 广东星云开物科技股份有限公司 | 视频压缩方法和共享售卖柜 |
Families Citing this family (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112068771A (zh) * | 2020-08-17 | 2020-12-11 | Oppo广东移动通信有限公司 | 视频处理方法、视频处理装置、终端设备及存储介质 |
CN112330541A (zh) * | 2020-11-11 | 2021-02-05 | 广州博冠信息科技有限公司 | 直播视频处理方法、装置、电子设备和存储介质 |
CN113852757B (zh) * | 2021-09-03 | 2023-05-26 | 维沃移动通信(杭州)有限公司 | 视频处理方法、装置、设备和存储介质 |
CN114710649B (zh) * | 2022-06-01 | 2023-01-24 | 广东中浦科技有限公司 | 污染源视频监控方法及系统 |
CN117495854B (zh) * | 2023-12-28 | 2024-05-03 | 淘宝(中国)软件有限公司 | 视频数据处理方法、设备及存储介质 |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2000031981A1 (en) * | 1998-11-20 | 2000-06-02 | Koninklijke Philips Electronics N.V. | Extraction of foreground information for stereoscopic video coding |
CN1960479A (zh) * | 2005-11-03 | 2007-05-09 | 中国科学院自动化研究所 | 利用单个摄像机进行主从视频跟踪的方法 |
CN101826157A (zh) * | 2010-04-28 | 2010-09-08 | 华中科技大学 | 一种地面静止目标实时识别跟踪方法 |
CN104067310A (zh) * | 2011-12-23 | 2014-09-24 | 超威半导体公司 | 显示图像的改进 |
CN105163127A (zh) * | 2015-09-07 | 2015-12-16 | 浙江宇视科技有限公司 | 视频分析方法及装置 |
CN106817608A (zh) * | 2015-11-27 | 2017-06-09 | 小米科技有限责任公司 | 实现局部显示的方法及装置 |
Family Cites Families (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7116833B2 (en) * | 2002-12-23 | 2006-10-03 | Eastman Kodak Company | Method of transmitting selected regions of interest of digital video data at selected resolutions |
CN105989802B (zh) * | 2015-03-05 | 2018-11-30 | 西安诺瓦电子科技有限公司 | 可编程逻辑器件及其亚像素下采样方法和相关应用 |
CN105472272A (zh) * | 2015-11-25 | 2016-04-06 | 浙江工业大学 | 基于fpga的多路视频拼接方法及装置 |
CN105718929B (zh) * | 2016-01-21 | 2019-04-30 | 成都信息工程大学 | 全天候未知环境下高精度快速圆形目标定位方法和系统 |
CN106791710B (zh) * | 2017-02-10 | 2020-12-04 | 北京地平线信息技术有限公司 | 目标检测方法、装置和电子设备 |
CN108012157B (zh) * | 2017-11-27 | 2020-02-04 | 上海交通大学 | 用于视频编码分数像素插值的卷积神经网络的构建方法 |
-
2018
- 2018-05-17 CN CN201810473779.7A patent/CN110502954B/zh active Active
-
2019
- 2019-05-16 WO PCT/CN2019/087288 patent/WO2019219065A1/zh active Application Filing
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2000031981A1 (en) * | 1998-11-20 | 2000-06-02 | Koninklijke Philips Electronics N.V. | Extraction of foreground information for stereoscopic video coding |
CN1960479A (zh) * | 2005-11-03 | 2007-05-09 | 中国科学院自动化研究所 | 利用单个摄像机进行主从视频跟踪的方法 |
CN101826157A (zh) * | 2010-04-28 | 2010-09-08 | 华中科技大学 | 一种地面静止目标实时识别跟踪方法 |
CN104067310A (zh) * | 2011-12-23 | 2014-09-24 | 超威半导体公司 | 显示图像的改进 |
CN105163127A (zh) * | 2015-09-07 | 2015-12-16 | 浙江宇视科技有限公司 | 视频分析方法及装置 |
CN106817608A (zh) * | 2015-11-27 | 2017-06-09 | 小米科技有限责任公司 | 实现局部显示的方法及装置 |
Cited By (30)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113392267B (zh) * | 2020-03-12 | 2024-01-16 | 平湖莱顿光学仪器制造有限公司 | 一种用于生成目标对象的二维显微视频信息的方法与设备 |
CN113392267A (zh) * | 2020-03-12 | 2021-09-14 | 平湖莱顿光学仪器制造有限公司 | 一种用于生成目标对象的二维显微视频信息的方法与设备 |
CN111753784A (zh) * | 2020-06-30 | 2020-10-09 | 广州酷狗计算机科技有限公司 | 视频的特效处理方法、装置、终端及存储介质 |
CN111726536A (zh) * | 2020-07-03 | 2020-09-29 | 腾讯科技(深圳)有限公司 | 视频生成方法、装置、存储介质及计算机设备 |
CN111726536B (zh) * | 2020-07-03 | 2024-01-05 | 腾讯科技(深圳)有限公司 | 视频生成方法、装置、存储介质及计算机设备 |
CN111859549A (zh) * | 2020-07-28 | 2020-10-30 | 奇瑞汽车股份有限公司 | 单一配置整车重量与重心信息的确定方法及相关设备 |
CN111859549B (zh) * | 2020-07-28 | 2024-05-14 | 奇瑞汽车股份有限公司 | 单一配置整车重量与重心信息的确定方法及相关设备 |
CN112135190A (zh) * | 2020-09-23 | 2020-12-25 | 成都市喜爱科技有限公司 | 视频处理方法、装置、系统、服务器及存储介质 |
CN112135190B (zh) * | 2020-09-23 | 2022-08-16 | 成都市喜爱科技有限公司 | 视频处理方法、装置、系统、服务器及存储介质 |
CN112528760A (zh) * | 2020-11-24 | 2021-03-19 | 腾讯科技(深圳)有限公司 | 图像处理方法、装置、计算机设备及介质 |
CN112528760B (zh) * | 2020-11-24 | 2024-01-09 | 腾讯科技(深圳)有限公司 | 图像处理方法、装置、计算机设备及介质 |
CN112541429B (zh) * | 2020-12-08 | 2024-05-31 | 浙江大华技术股份有限公司 | 一种智能抓图方法、装置、电子设备以及存储介质 |
CN112541429A (zh) * | 2020-12-08 | 2021-03-23 | 浙江大华技术股份有限公司 | 一种智能抓图方法、装置、电子设备以及存储介质 |
CN112804573A (zh) * | 2021-01-08 | 2021-05-14 | Oppo广东移动通信有限公司 | 电视信号的处理方法、装置、电子设备和存储介质 |
CN112817768A (zh) * | 2021-02-26 | 2021-05-18 | 北京梧桐车联科技有限责任公司 | 动画处理方法、装置、设备及计算机可读存储介质 |
CN112817768B (zh) * | 2021-02-26 | 2024-05-03 | 北京梧桐车联科技有限责任公司 | 动画处理方法、装置、设备及计算机可读存储介质 |
CN113032590A (zh) * | 2021-03-29 | 2021-06-25 | 广州繁星互娱信息科技有限公司 | 特效展示方法、装置、计算机设备及计算机可读存储介质 |
CN113032590B (zh) * | 2021-03-29 | 2024-05-03 | 广州繁星互娱信息科技有限公司 | 特效展示方法、装置、计算机设备及计算机可读存储介质 |
CN113365027A (zh) * | 2021-05-28 | 2021-09-07 | 上海商汤智能科技有限公司 | 视频处理方法及装置、电子设备和存储介质 |
CN113422967A (zh) * | 2021-06-07 | 2021-09-21 | 深圳康佳电子科技有限公司 | 一种投屏显示控制方法、装置、终端设备及存储介质 |
CN113435530B (zh) * | 2021-07-07 | 2023-10-10 | 腾讯科技(深圳)有限公司 | 图像识别方法、装置、计算机设备及计算机可读存储介质 |
CN113435530A (zh) * | 2021-07-07 | 2021-09-24 | 腾讯科技(深圳)有限公司 | 图像识别方法、装置、计算机设备及计算机可读存储介质 |
CN113561916A (zh) * | 2021-08-31 | 2021-10-29 | 长沙德壹科技有限公司 | 一种车载显示系统、车辆及车载摄像头图像显示方法 |
CN113992880B (zh) * | 2021-10-15 | 2024-04-12 | 上海佰贝科技发展股份有限公司 | 4k视频识别方法、系统、设备及计算机可读存储介质 |
CN113992880A (zh) * | 2021-10-15 | 2022-01-28 | 上海佰贝科技发展股份有限公司 | 4k视频识别方法、系统、设备及计算机可读存储介质 |
CN114187349A (zh) * | 2021-11-03 | 2022-03-15 | 深圳市正运动技术有限公司 | 产品加工方法、装置、终端设备以及存储介质 |
CN114827567B (zh) * | 2022-03-23 | 2024-05-28 | 阿里巴巴(中国)有限公司 | 视频质量分析方法、设备和可读介质 |
CN114827567A (zh) * | 2022-03-23 | 2022-07-29 | 阿里巴巴(中国)有限公司 | 视频质量分析方法、设备和可读介质 |
CN116684626B (zh) * | 2023-08-04 | 2023-11-24 | 广东星云开物科技股份有限公司 | 视频压缩方法和共享售卖柜 |
CN116684626A (zh) * | 2023-08-04 | 2023-09-01 | 广东星云开物科技股份有限公司 | 视频压缩方法和共享售卖柜 |
Also Published As
Publication number | Publication date |
---|---|
CN110502954B (zh) | 2023-06-16 |
CN110502954A (zh) | 2019-11-26 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2019219065A1 (zh) | 视频分析的方法和装置 | |
US11205282B2 (en) | Relocalization method and apparatus in camera pose tracking process and storage medium | |
CN110189340B (zh) | 图像分割方法、装置、电子设备及存储介质 | |
WO2021008456A1 (zh) | 图像处理方法、装置、电子设备及存储介质 | |
CN108594997B (zh) | 手势骨架构建方法、装置、设备及存储介质 | |
WO2019101021A1 (zh) | 图像识别方法、装置及电子设备 | |
JP7058760B2 (ja) | 画像処理方法およびその、装置、端末並びにコンピュータプログラム | |
CN110647865A (zh) | 人脸姿态的识别方法、装置、设备及存储介质 | |
WO2020221012A1 (zh) | 图像特征点的运动信息确定方法、任务执行方法和设备 | |
CN112907725B (zh) | 图像生成、图像处理模型的训练、图像处理方法和装置 | |
KR20140104753A (ko) | 신체 부위 검출을 이용한 이미지 프리뷰 | |
CN109886208B (zh) | 物体检测的方法、装置、计算机设备及存储介质 | |
CN110839128B (zh) | 拍照行为检测方法、装置及存储介质 | |
CN109360222B (zh) | 图像分割方法、装置及存储介质 | |
CN111723803B (zh) | 图像处理方法、装置、设备及存储介质 | |
CN111027490A (zh) | 人脸属性识别方法及装置、存储介质 | |
CN111754386A (zh) | 图像区域屏蔽方法、装置、设备及存储介质 | |
CN112135191A (zh) | 视频编辑方法、装置、终端及存储介质 | |
CN110675473B (zh) | 生成gif动态图的方法、装置、电子设备及介质 | |
WO2022199102A1 (zh) | 图像处理方法及装置 | |
CN112235650A (zh) | 视频处理方法、装置、终端及存储介质 | |
CN110232417B (zh) | 图像识别方法、装置、计算机设备及计算机可读存储介质 | |
CN111931712A (zh) | 人脸识别方法、装置、抓拍机及系统 | |
CN110853124A (zh) | 生成gif动态图的方法、装置、电子设备及介质 | |
CN111860064A (zh) | 基于视频的目标检测方法、装置、设备及存储介质 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 19803393 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 19803393 Country of ref document: EP Kind code of ref document: A1 |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 19803393 Country of ref document: EP Kind code of ref document: A1 |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS (EPO FORM 1205A DATED 01.06.2021) |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 19803393 Country of ref document: EP Kind code of ref document: A1 |