CN114549697A - Image processing method, device, equipment and storage medium - Google Patents

Image processing method, device, equipment and storage medium Download PDF

Info

Publication number
CN114549697A
CN114549697A CN202210152880.9A CN202210152880A CN114549697A CN 114549697 A CN114549697 A CN 114549697A CN 202210152880 A CN202210152880 A CN 202210152880A CN 114549697 A CN114549697 A CN 114549697A
Authority
CN
China
Prior art keywords
information
image frame
coordinate information
transparent canvas
target
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202210152880.9A
Other languages
Chinese (zh)
Inventor
王林芳
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Baidu Netcom Science and Technology Co Ltd
Original Assignee
Beijing Baidu Netcom Science and Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Baidu Netcom Science and Technology Co Ltd filed Critical Beijing Baidu Netcom Science and Technology Co Ltd
Priority to CN202210152880.9A priority Critical patent/CN114549697A/en
Publication of CN114549697A publication Critical patent/CN114549697A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T11/002D [Two Dimensional] image generation
    • G06T11/60Editing figures and text; Combining figures or text
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/70Information retrieval; Database structures therefor; File system structures therefor of video data
    • G06F16/74Browsing; Visualisation therefor
    • G06F16/745Browsing; Visualisation therefor the internal structure of a single video sequence
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/70Information retrieval; Database structures therefor; File system structures therefor of video data
    • G06F16/78Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/7867Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using information manually generated, e.g. tags, keywords, comments, title and artist information, manually generated time, location and usage information, user ratings
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformations in the plane of the image
    • G06T3/40Scaling of whole images or parts thereof, e.g. expanding or contracting

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Multimedia (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Databases & Information Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Human Computer Interaction (AREA)
  • Library & Information Science (AREA)
  • Processing Or Creating Images (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

The disclosure provides an image processing method, an image processing device, image processing equipment and a storage medium, and relates to the technical field of artificial intelligence, in particular to computer vision and intelligent cloud technology. The specific implementation scheme is as follows: in the process of playing a target image frame, acquiring first coordinate information of a target object in an original image frame corresponding to the target image frame; obtaining second coordinate information of the target object on the transparent canvas according to the first coordinate information, the acquisition size information of the original image frame and the display size information of the transparent canvas; the transparent canvas is overlaid on the target image frame in an equal ratio; and adding marking information for the target object on the transparent canvas according to the second coordinate information. By the technical scheme, the label information associated with the target image frame can be displayed in real time.

Description

Image processing method, device, equipment and storage medium
Technical Field
The present disclosure relates to the field of artificial intelligence technologies, and in particular, to computer vision and smart cloud technologies, and more particularly, to an image processing method, apparatus, device, and storage medium.
Background
With the continuous development of internet technology, videos gradually become main contents browsed by users through terminal devices. At present, for video image frames transmitted by a capturing device, a video player generally plays the video image frames directly in a single mode, and improvement is urgently needed.
Disclosure of Invention
The disclosure provides an image processing method, an apparatus, a device and a storage medium.
According to an aspect of the present disclosure, there is provided an image processing method including:
in the process of playing a target image frame, acquiring first coordinate information of a target object in an original image frame corresponding to the target image frame;
obtaining second coordinate information of the target object on the transparent canvas according to the first coordinate information, the acquisition size information of the original image frame and the display size information of the transparent canvas; the transparent canvas is overlaid on the target image frame in an equal ratio;
and adding marking information for the target object on the transparent canvas according to the second coordinate information.
According to another aspect of the present disclosure, there is provided an electronic device including:
at least one processor; and
a memory communicatively coupled to the at least one processor; wherein the content of the first and second substances,
the memory stores instructions executable by the at least one processor to enable the at least one processor to perform the image processing method of any of the embodiments of the present disclosure.
According to another aspect of the present disclosure, there is provided a non-transitory computer-readable storage medium storing computer instructions for causing a computer to perform the image processing method according to any one of the embodiments of the present disclosure.
According to the technology disclosed by the invention, the annotation information associated with the target image frame can be displayed in real time.
It should be understood that the statements in this section do not necessarily identify key or critical features of the embodiments of the present disclosure, nor do they limit the scope of the present disclosure. Other features of the present disclosure will become apparent from the following description.
Drawings
The drawings are included to provide a better understanding of the present solution and are not to be construed as limiting the present disclosure. Wherein:
fig. 1 is a flowchart of an image processing method provided according to an embodiment of the present disclosure;
FIG. 2 is a flow chart of another image processing method provided according to an embodiment of the present disclosure;
FIG. 3 is a flow chart of yet another image processing method provided in accordance with an embodiment of the present disclosure;
fig. 4 is a schematic structural diagram of an image processing apparatus provided according to an embodiment of the present disclosure;
fig. 5 is a block diagram of an electronic device to implement the image processing method of the embodiment of the present disclosure.
Detailed Description
Exemplary embodiments of the present disclosure are described below with reference to the accompanying drawings, in which various details of the embodiments of the disclosure are included to assist understanding, and which are to be considered as merely exemplary. Accordingly, those of ordinary skill in the art will recognize that various changes and modifications of the embodiments described herein can be made without departing from the scope and spirit of the present disclosure. Also, descriptions of well-known functions and constructions are omitted in the following description for clarity and conciseness.
Fig. 1 is a flowchart of an image processing method provided according to an embodiment of the present disclosure, which is suitable for a situation how to process an image in a scenario where a video player is used to play a video image frame; the method is particularly suitable for processing the images in the scene that the browser plays the video image frames through the video player.
The method can be executed by an image processing device, the device can be realized in a software and/or hardware mode, and can be integrated into an electronic device with an image processing function; preferably, the method can be configured in a browser in the electronic device. As shown in fig. 1, the image processing method of the present embodiment may include:
s101, in the process of playing the target image frame, acquiring first coordinate information of the target object in an original image frame corresponding to the target image frame.
In this embodiment, the target image frame may be a video image frame played in a browser. The target object is an object in the target image frame, that is, an object of interest to the user, and may be, for example, a person, a specific item (such as a dangerous item), or the like; illustratively, the number of target objects may be one or more. The original image frame is an original video image frame acquired by a video acquisition device. Optionally, the target image frame is obtained by essentially scaling the original image frame.
The first coordinate information is coordinate information of the target object in the original image frame corresponding to the target image frame, and may be, for example, vertex coordinate information of a labeling frame in which the target object is located. Optionally, after the video capture device captures the original image frame, the video capture device may use a target detection model to detect an object in the target image frame, so as to obtain position information of the target object, that is, the first coordinate information. For example, a face detection model may be used to detect a face in an original image frame to obtain position information of the face, i.e. coordinate information of the labeling frame.
Specifically, in the process of playing the target image frame through the video player, the first coordinate information of the target object in the original image frame corresponding to the target image frame may be obtained, and meanwhile, the acquisition size information of the original image frame may also be obtained.
Illustratively, if the browser acquires the first coordinate information of the target object in the original image frame corresponding to the target image frame and acquires the acquisition size information of the original image in the process of playing the target image frame through the video player, the following operations S102-S103 are performed, that is, the transparent canvas is added with the annotation information for the target object.
For example, in the process that the browser plays the target image frame by using flvjs, if the first coordinate information of the target object in the original image frame corresponding to the target image frame and the acquisition size information of the original image are acquired through the flvjs API, it is determined that the target image frame has associated annotation information, and then the following operations S102-S103 are performed, that is, the annotation information is added to the target object on the transparent canvas.
For another example, in the process that the browser uses flvjs to play the target image frame, acquiring the acquisition size information of the original image through the flvjs API, determining that the target image frame is not identified to have associated annotation information, and further not performing subsequent operations S102 to S103, that is, not adding annotation information to the target object on the transparent canvas.
And S102, obtaining second coordinate information of the target object on the transparent canvas according to the first coordinate information, the acquisition size information of the original image frame and the display size information of the transparent canvas.
In this embodiment, the transparent Canvas may be implemented by a Canvas tag; furthermore, the transparent canvas is overlaid on the target image frame in an equal ratio, that is, the display size information of the transparent picture is the same as the size information of the target image frame. The size information of the target image frame is substantially the width and height of a video tag corresponding to the video player; i.e., the width and height of the Canvas label corresponding to the transparent Canvas are the same as the width and height of the video label corresponding to the video player.
The acquisition size information is size information of an acquisition device that acquires an original image frame, and includes an original width and an original height. The presentation size information is size information to be presented on the browser page, and includes a presentation width and a presentation height. The second coordinate information is coordinate information of the target object on the transparent canvas.
In an optional mode, second coordinate information of the target object on the transparent canvas can be obtained based on a coordinate conversion model according to the first coordinate information, the acquisition size information of the original image frame and the display size information of the transparent canvas. The coordinate transformation model can be obtained based on a machine learning algorithm.
And S103, adding marking information to the target object on the transparent canvas according to the second coordinate information.
In this embodiment, the labeling information may include a labeling box of the area where the target object is located, and may also include other related information of the target object, such as an object name or an object category.
Specifically, the labeling information may be added to the target object on the transparent canvas according to the second coordinate information. For example, if the target object is a face, connecting vertices in sequence according to second coordinate information, that is, vertex coordinate information of a labeling frame of the face on a transparent canvas, and drawing the labeling frame for the face on the transparent canvas.
According to the technical scheme provided by the embodiment of the disclosure, in the process of playing the target image frame, first coordinate information of a target object in an original image frame corresponding to the target image frame is obtained, and then second coordinate information of the target object on a transparent canvas is obtained according to the first coordinate information, the acquisition size information of the original image frame and the display size information of the transparent canvas; and the transparent canvas is covered on the target image frame in an equal ratio, and then the labeling information is added to the target object on the transparent canvas according to the second coordinate information. According to the technical scheme, when the target image frame is played, the browser can display the labeling information associated with the target image frame in real time, so that a user can intuitively know the related information of the target object in the target image frame.
On the basis of the above embodiment, as an optional mode of the present disclosure, the annotation information on the transparent canvas may also be emptied after the target image frame is played.
Specifically, after the target image frame is played, the annotation information on the transparent Canvas may be deleted based on the deletion operation of the Canvas, so that the annotation information may be added to other video image frames in the following.
It can be understood that the annotation information on the transparent canvas is deleted after the target image frame is played, so that the same-frequency display of the annotation information and the target image frame is ensured, and the playing of the subsequent video image frame is not influenced.
Fig. 2 is a flowchart of another image processing method provided according to an embodiment of the present disclosure. On the basis of the above embodiment, the second coordinate information of the target object on the transparent canvas is further optimized according to the first coordinate information, the acquisition size information of the original image frame and the display size information of the transparent canvas, and an optional implementation scheme is provided. As shown in fig. 2, the image processing method of the present embodiment may include:
s201, in the process of playing the target image frame, acquiring first coordinate information of the target object in an original image frame corresponding to the target image frame.
S202, determining the scaling according to the acquisition size information of the original image frame and the display size information of the transparent canvas.
In this embodiment, the scaling includes a width scaling and a height scaling.
For example, the width scaling may be determined based on the original width of the original image frame and the display width of the transparent canvas. Specifically, the original width is divided by the display width to obtain the result of the width scaling.
Illustratively, the height scaling may be determined based on the original height of the original image frame and the presentation height of the transparent canvas. Specifically, the original height may be divided by the display height to obtain the height scaling.
For example, if the collection size information of the original image frame is 1800px, the original height is 1200 px; the display size information of the transparent canvas is display width 1200px and display height 600 px; the width scaling in the scaling is original width 1800/presentation width 1200-3/2 and the height scaling is original height 1200/presentation height 600-2.
And S203, obtaining second coordinate information of the target object on the transparent canvas according to the scaling and the first coordinate information.
Specifically, the result obtained by dividing the first coordinate information by the scaling is used as the second coordinate information of the target object on the transparent canvas. For example, the result of dividing the abscissa in the first coordinate information by the width scaling is taken as the abscissa in the second coordinate information; and dividing the vertical coordinate in the first coordinate information by the height scaling result to be used as the vertical coordinate in the second coordinate information.
And S204, adding marking information to the target object on the transparent canvas according to the second coordinate information.
According to the technical scheme provided by the embodiment of the disclosure, in the process of playing the target image frame, first coordinate information of the target object in the original image frame corresponding to the target image frame is obtained, then the scaling ratio is determined according to the acquisition size information of the original image frame and the display size information of the transparent canvas, second coordinate information of the target object on the transparent canvas is obtained according to the scaling ratio and the first coordinate information, and then labeling information is added to the target object on the transparent canvas according to the second coordinate information. According to the technical scheme, the second coordinate information of the target object on the transparent canvas is obtained by determining the scaling, so that the second coordinate information is more accurate, and the accuracy of the labeling information is ensured.
On the basis of the above embodiment, as an optional mode of the present disclosure, the second coordinate information of the target object on the transparent canvas is obtained according to the scaling ratio and the first coordinate information, or the scaling ratio is updated according to a ratio between the original size information and the display size information of the transparent canvas, and the second coordinate information of the target object on the transparent canvas is obtained according to the first coordinate information and the updated scaling ratio.
The original size information of the transparent canvas is the default size information of the transparent canvas, namely the size of the coordinate system of the transparent canvas, and comprises a default width and a default height.
Illustratively, the width scaling ratio of the scaling ratio can be updated according to the width ratio between the default width and the display width of the transparent canvas, the height scaling ratio of the scaling ratio is updated according to the height ratio between the default height and the display height of the transparent canvas, and then the second coordinate information of the target object on the transparent canvas is obtained according to the first coordinate information and the updated scaling ratio.
Specifically, the result of dividing the display width by the default width is used as a width ratio, and the result of multiplying the width ratio by the width scaling ratio is used as an updated width scaling ratio; and dividing the display height by the default height to obtain a height ratio, and multiplying the height ratio by the height scaling to obtain an updated height scaling. For example, the default size information of the transparent canvas is 300px in default width, 150px in default height, the display size of the transparent canvas is 1200px in display width, 600px in display height, the width scaling ratio shown in the above embodiment is 3/2, the height scaling ratio is 2, the width ratio is 1200 in display width/300 in default width is 4, the height ratio is 600 in display height/150 in default height is 4, the updated width scaling ratio is 3/2 in default height is 6, and the updated height scaling ratio is 2 in default height is 8.
And then according to the first coordinate information and the updated scaling, second coordinate information of the target object on the transparent canvas is obtained.
It can be appreciated that updating the scaling further improves the accuracy of the second coordinate information.
Fig. 3 is a flowchart of still another image processing method provided according to an embodiment of the present disclosure. On the basis of the above embodiment, further optimization is carried out, and an alternative implementation scheme is provided. As shown in fig. 3, the image processing method of the present embodiment may include:
s301, in the process of playing the target image frame, acquiring first coordinate information of the target object in an original image frame corresponding to the target image frame.
S302, obtaining second coordinate information of the target object on the transparent canvas according to the first coordinate information, the acquisition size information of the original image frame and the display size information of the transparent canvas.
And S303, adding marking information to the target object on the transparent canvas according to the second coordinate information.
S304, responding to the active annotation request of the target image frame, and determining third coordinate information of the target object on the transparent canvas according to the annotation operation acted on the transparent canvas.
In this embodiment, the active annotation request may be an annotation request triggered and generated when the user is unsatisfied with the displayed annotation information of the target image frame, or may also be an annotation request triggered and generated when the user wants to annotate the target image frame without the annotation information on the target image frame, that is, without acquiring the first coordinate information.
In an optional mode, when the user is not satisfied with the displayed labeling information of the target image frame, the second coordinate information is adjusted according to the labeling operation acting on the transparent canvas in response to the active labeling request of the user on the target image frame, and the third labeling information of the target object on the transparent canvas is determined.
Alternatively, in the case that the user is not satisfied with the displayed annotation information of the target image frame, in response to an active annotation request of the user for the target image frame, new annotation information is added to the target object according to the annotation operation acting on the transparent canvas, and third annotation information of the target object on the transparent canvas is determined.
In another optional mode, in the case that the first coordinate information is not acquired, in response to an active annotation request of a user for the target image frame, according to an annotation operation acting on the transparent canvas, annotation information is added to the unmarked target object, and third annotation information of the target object on the transparent canvas is determined.
And S305, updating the marking information according to the third coordinate information.
In an optional mode, under the condition that the user is not satisfied with the annotation information of the displayed target image frame, the corresponding annotation information is drawn on the transparent canvas according to the third annotation information so as to cover the annotation information corresponding to the second coordinate information.
In another optional mode, under the condition that the first coordinate information is not acquired, drawing corresponding annotation information on the transparent canvas according to the third annotation information.
According to the technical scheme of the embodiment, in the process of playing the target image frame, first coordinate information of a target object in an original image frame corresponding to the target image frame is obtained, and then second coordinate information of the target object on a transparent canvas is obtained according to the first coordinate information, the acquisition size information of the original image frame and the display size information of the transparent canvas; the transparent canvas is covered on the target image frame in an equal ratio, labeling information is added to the target object on the transparent canvas according to the second coordinate information, further, in response to an active labeling request for the target image frame, third coordinate information of the target object on the transparent canvas is determined according to labeling operation acting on the transparent canvas, and the labeling information is updated according to the third coordinate information. According to the technical scheme, the third coordinate information is introduced, the labeling information is updated, the labeling information is more accurate, and the user satisfaction is improved.
On the basis of the above embodiment, as an optional mode of the present disclosure, the target image frame and the updated annotation information may also be adjusted according to the adaptation size information of the target device; and associating and storing the adjusted target image frame and the annotation information in the target equipment.
The target device is a device for storing the target image frame and the labeling information. The fitting size information is size information of the target device, and includes a fitting width and a fitting height.
For example, a storage scaling ratio may be determined according to the display size information of the transparent canvas and the adaptation size information of the target device, and then, the target image frame and the updated annotation information are adjusted according to the storage scaling ratio, and the adjusted target image frame and the updated annotation information are stored in the target device in an associated manner. Wherein the storage scaling comprises a storage width ratio and a storage height ratio.
Specifically, the storage width ratio may be determined according to the display width of the transparent canvas and the adaptation width of the target device, the storage height ratio may be determined according to the display height of the transparent canvas and the adaptation height of the target device, then the target image frame and the updated annotation information may be adjusted according to the storage width ratio and the storage height ratio, and the adjusted target image frame and the updated annotation information may be stored in the target device in an associated manner.
For example, the result of dividing the display width of the transparent canvas by the adaptation width of the target device is used as a storage width ratio, the result of dividing the display height of the transparent canvas by the adaptation height of the target device is used as a storage height ratio, then, the width of the target image frame is divided by the storage width ratio, and the height of the target image frame is divided by the storage height ratio to obtain an adjusted target image frame; and dividing the abscissa in the labeling information by the storage width ratio, and dividing the ordinate in the labeling information by the storage height ratio to obtain the adjusted labeling information. And further associating and storing the adjusted target image frame and the annotation information in the target equipment.
It can be understood that the target image frame and the marking information are adjusted and then stored, and the original image frame and the original marking information are restored, so that the target image frame and the marking information are convenient to display without distortion.
Fig. 4 is a schematic structural diagram of a data processing apparatus according to an embodiment of the present disclosure. The embodiment of the disclosure is suitable for the situation of processing the image in the scene of playing the video image frame by adopting the video player; the method is particularly suitable for processing the images in the scene that the browser plays the video image frames through the video player.
The device can be realized by adopting a software and/or hardware mode, and can be integrated into an electronic device with an image processing function; preferably, the method can be configured in a browser in the electronic device. As shown in fig. 4, the image processing apparatus 400 includes:
a first coordinate information obtaining module 401, configured to obtain, in a process of playing a target image frame, first coordinate information of a target object in an original image frame corresponding to the target image frame;
a second coordinate information determining module 402, configured to obtain second coordinate information of the target object on the transparent canvas according to the first coordinate information, the acquisition size information of the original image frame, and the display size information of the transparent canvas; transparent canvas is covered on the target image frame in an equal ratio;
and a labeling information adding module 403, configured to add labeling information to the target object on the transparent canvas according to the second coordinate information.
According to the technical scheme provided by the embodiment of the disclosure, in the process of playing the target image frame, first coordinate information of a target object in an original image frame corresponding to the target image frame is obtained, and then second coordinate information of the target object on a transparent canvas is obtained according to the first coordinate information, the acquisition size information of the original image frame and the display size information of the transparent canvas; and the transparent canvas is covered on the target image frame in an equal ratio, and then the labeling information is added to the target object on the transparent canvas according to the second coordinate information. According to the technical scheme, when the target image frame is played, the browser can display the labeling information associated with the target image frame in real time, so that a user can intuitively know the related information of the target object in the target image frame.
Further, the second coordinate information determination module 403 includes:
the scaling determining unit is used for determining the scaling according to the acquisition size information of the original image frame and the display size information of the transparent canvas;
and the second coordinate information determining unit is used for obtaining second coordinate information of the target object on the transparent canvas according to the scaling and the first coordinate information.
Further, the second coordinate information determination unit is specifically configured to:
updating the scaling according to the ratio of the original size information and the display size information of the transparent canvas;
and obtaining second coordinate information of the target object on the transparent canvas according to the first coordinate information and the updated scaling.
Further, the apparatus further comprises:
and the annotation information emptying module is used for emptying the annotation information on the transparent canvas after the target image frame is played.
Further, the apparatus further comprises:
the third coordinate information determination module is used for responding to an active labeling request for the target image frame and determining third coordinate information of the target object on the transparent canvas according to labeling operation acting on the transparent canvas;
and the annotation information updating module is used for updating the annotation information according to the third coordinate information.
Further, the apparatus further comprises:
the adjusting module is used for adjusting the target image frame and the updated marking information according to the adaptive size information of the target equipment;
and the storage module is used for associating and storing the adjusted target image frame and the annotation information in the target equipment.
In the technical scheme of the disclosure, the acquisition, storage, application and the like of the related image frame data and the like all accord with the regulations of related laws and regulations, and do not violate the good customs of the public order.
The present disclosure also provides an electronic device, a readable storage medium, and a computer program product according to embodiments of the present disclosure.
FIG. 5 illustrates a schematic block diagram of an example electronic device 500 that can be used to implement embodiments of the present disclosure. Electronic devices are intended to represent various forms of digital computers, such as laptops, desktops, workstations, personal digital assistants, servers, blade servers, mainframes, and other appropriate computers. The electronic device may also represent various forms of mobile devices, such as personal digital processing, cellular phones, smart phones, wearable devices, and other similar computing devices. The components shown herein, their connections and relationships, and their functions, are meant to be examples only, and are not meant to limit implementations of the disclosure described and/or claimed herein.
As shown in fig. 5, the electronic device 500 includes a computing unit 501, which can perform various appropriate actions and processes according to a computer program stored in a Read Only Memory (ROM)502 or a computer program loaded from a storage unit 508 into a Random Access Memory (RAM) 503. In the RAM 503, various programs and data required for the operation of the electronic apparatus 500 can also be stored. The calculation unit 501, the ROM 502, and the RAM 503 are connected to each other by a bus 504. An input/output (I/O) interface 505 is also connected to bus 504.
A number of components in the electronic device 500 are connected to the I/O interface 505, including: an input unit 506 such as a keyboard, a mouse, or the like; an output unit 507 such as various types of displays, speakers, and the like; a storage unit 508, such as a magnetic disk, optical disk, or the like; and a communication unit 509 such as a network card, modem, wireless communication transceiver, etc. The communication unit 509 allows the electronic device 500 to exchange information/data with other devices through a computer network such as the internet and/or various telecommunication networks.
The computing unit 501 may be a variety of general-purpose and/or special-purpose processing components having processing and computing capabilities. Some examples of the computing unit 501 include, but are not limited to, a Central Processing Unit (CPU), a Graphics Processing Unit (GPU), various dedicated Artificial Intelligence (AI) computing chips, various computing units running machine learning model algorithms, a Digital Signal Processor (DSP), and any suitable processor, controller, microcontroller, and so forth. The calculation unit 501 executes the respective methods and processes described above, such as the image processing method. For example, in some embodiments, the image processing method may be implemented as a computer software program tangibly embodied in a machine-readable medium, such as storage unit 508. In some embodiments, part or all of the computer program may be loaded and/or installed onto the electronic device 500 via the ROM 502 and/or the communication unit 509. When the computer program is loaded into the RAM 503 and executed by the computing unit 501, one or more steps of the image processing method described above may be performed. Alternatively, in other embodiments, the computing unit 501 may be configured to perform the image processing method by any other suitable means (e.g., by means of firmware).
Various implementations of the systems and techniques described here above may be implemented in digital electronic circuitry, integrated circuitry, Field Programmable Gate Arrays (FPGAs), Application Specific Integrated Circuits (ASICs), Application Specific Standard Products (ASSPs), system on a chip (SOCs), Complex Programmable Logic Devices (CPLDs), computer hardware, firmware, software, and/or combinations thereof. These various embodiments may include: implemented in one or more computer programs that are executable and/or interpretable on a programmable system including at least one programmable processor, which may be special or general purpose, receiving data and instructions from, and transmitting data and instructions to, a storage system, at least one input device, and at least one output device.
Program code for implementing the methods of the present disclosure may be written in any combination of one or more programming languages. These program codes may be provided to a processor or controller of a general purpose computer, special purpose computer, or other programmable data processing apparatus, such that the program codes, when executed by the processor or controller, cause the functions/operations specified in the flowchart and/or block diagram to be performed. The program code may execute entirely on the machine, partly on the machine, as a stand-alone software package partly on the machine and partly on a remote machine or entirely on the remote machine or server.
In the context of this disclosure, a machine-readable medium may be a tangible medium that can contain, or store a program for use by or in connection with an instruction execution system, apparatus, or device. The machine-readable medium may be a machine-readable signal medium or a machine-readable storage medium. A machine-readable medium may include, but is not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any suitable combination of the foregoing. More specific examples of a machine-readable storage medium would include an electrical connection based on one or more wires, a portable computer diskette, a hard disk, a Random Access Memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or flash memory), an optical fiber, a portable compact disc read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing.
To provide for interaction with a user, the systems and techniques described here can be implemented on a computer having: a display device (e.g., a CRT (cathode ray tube) or LCD (liquid crystal display) monitor) for displaying information to a user; and a keyboard and a pointing device (e.g., a mouse or a trackball) by which a user can provide input to the computer. Other kinds of devices may also be used to provide for interaction with a user; for example, feedback provided to the user can be any form of sensory feedback (e.g., visual feedback, auditory feedback, or tactile feedback); and input from the user may be received in any form, including acoustic, speech, or tactile input.
The systems and techniques described here can be implemented in a computing system that includes a back-end component (e.g., as a data server), or that includes a middleware component (e.g., an application server), or that includes a front-end component (e.g., a user computer having a graphical user interface or a web browser through which a user can interact with an implementation of the systems and techniques described here), or any combination of such back-end, middleware, or front-end components. The components of the system can be interconnected by any form or medium of digital data communication (e.g., a communication network). Examples of communication networks include: local Area Networks (LANs), Wide Area Networks (WANs), and the Internet.
The computer system may include clients and servers. A client and server are generally remote from each other and typically interact through a communication network. The relationship of client and server arises by virtue of computer programs running on the respective computers and having a client-server relationship to each other. The server may be a cloud server, a server of a distributed system, or a server with a combined blockchain.
Artificial intelligence is the subject of research that makes computers simulate some human mental processes and intelligent behaviors (such as learning, reasoning, thinking, planning, etc.), both at the hardware level and at the software level. Artificial intelligence hardware technologies generally include technologies such as sensors, dedicated artificial intelligence chips, cloud computing, distributed storage, big data processing, and the like; the artificial intelligence software technology mainly comprises a computer vision technology, a voice recognition technology, a natural language processing technology, a machine learning/deep learning technology, a big data processing technology, a knowledge map technology and the like.
Cloud computing (cloud computing) refers to a technology system that accesses a flexibly extensible shared physical or virtual resource pool through a network, where resources may include servers, operating systems, networks, software, applications, storage devices, and the like, and may be deployed and managed in a self-service manner as needed. Through the cloud computing technology, high-efficiency and strong data processing capacity can be provided for technical application and model training of artificial intelligence, block chains and the like.
It should be understood that various forms of the flows shown above may be used, with steps reordered, added, or deleted. For example, the steps described in the present disclosure may be executed in parallel, sequentially, or in different orders, as long as the desired results of the technical solutions disclosed in the present disclosure can be achieved, and the present disclosure is not limited herein.
The above detailed description should not be construed as limiting the scope of the disclosure. It should be understood by those skilled in the art that various modifications, combinations, sub-combinations and substitutions may be made in accordance with design requirements and other factors. Any modification, equivalent replacement, and improvement made within the spirit and principle of the present disclosure should be included in the scope of protection of the present disclosure.

Claims (15)

1. An image processing method comprising:
in the process of playing a target image frame, acquiring first coordinate information of a target object in an original image frame corresponding to the target image frame;
obtaining second coordinate information of the target object on the transparent canvas according to the first coordinate information, the acquisition size information of the original image frame and the display size information of the transparent canvas; the transparent canvas is overlaid on the target image frame in an equal ratio;
and adding marking information for the target object on the transparent canvas according to the second coordinate information.
2. The method of claim 1, wherein the obtaining second coordinate information of the target object on the transparent canvas according to the first coordinate information, the acquisition size information of the original image frame and the presentation size information of the transparent canvas comprises:
determining a scaling ratio according to the acquisition size information of the original image frame and the display size information of the transparent canvas;
and obtaining second coordinate information of the target object on the transparent canvas according to the scaling and the first coordinate information.
3. The method of claim 2, wherein the deriving second coordinate information of the target object on the transparent canvas according to the scaling and the first coordinate information comprises:
updating the scaling according to the ratio between the original size information and the display size information of the transparent canvas;
and obtaining second coordinate information of the target object on the transparent canvas according to the first coordinate information and the updated scaling.
4. The method of claim 1, further comprising:
and after the target image frame is played, emptying the labeling information on the transparent canvas.
5. The method of claim 1, further comprising:
responding to an active labeling request of the target image frame, and determining third coordinate information of the target object on the transparent canvas according to a labeling operation acted on the transparent canvas;
and updating the labeling information according to the third coordinate information.
6. The method of claim 5, further comprising:
adjusting the target image frame and the updated marking information according to the adaptive size information of the target equipment;
and associating and storing the adjusted target image frame and the marking information in the target equipment.
7. An image processing apparatus comprising:
the first coordinate information acquisition module is used for acquiring first coordinate information of a target object in an original image frame corresponding to a target image frame in the process of playing the target image frame;
the second coordinate information determining module is used for obtaining second coordinate information of the target object on the transparent canvas according to the first coordinate information, the acquisition size information of the original image frame and the display size information of the transparent canvas; the transparent canvas is overlaid on the target image frame in an equal ratio;
and the marking information adding module is used for adding marking information for the target object on the transparent canvas according to the second coordinate information.
8. The apparatus of claim 7, wherein the second coordinate information determination module comprises:
the scaling determining unit is used for determining scaling according to the acquisition size information of the original image frame and the display size information of the transparent canvas;
and the second coordinate information determining unit is used for obtaining second coordinate information of the target object on the transparent canvas according to the scaling and the first coordinate information.
9. The apparatus according to claim 8, wherein the second coordinate information determination unit is specifically configured to:
updating the scaling according to the ratio between the original size information and the display size information of the transparent canvas;
and obtaining second coordinate information of the target object on the transparent canvas according to the first coordinate information and the updated scaling.
10. The apparatus of claim 7, further comprising:
and the marking information emptying module is used for emptying the marking information on the transparent canvas after the target image frame is played.
11. The apparatus of claim 7, further comprising:
the third coordinate information determination module is used for responding to an active labeling request of the target image frame and determining third coordinate information of the target object on the transparent canvas according to a labeling operation acted on the transparent canvas;
and the marking information updating module is used for updating the marking information according to the third coordinate information.
12. The apparatus of claim 11, further comprising:
the adjusting module is used for adjusting the target image frame and the updated marking information according to the adaptive size information of the target equipment;
and the storage module is used for associating and storing the adjusted target image frame and the marking information in the target equipment.
13. An electronic device, comprising:
at least one processor; and
a memory communicatively coupled to the at least one processor; wherein the content of the first and second substances,
the memory stores instructions executable by the at least one processor to enable the at least one processor to perform the image processing method of any one of claims 1-6.
14. A non-transitory computer-readable storage medium storing computer instructions for causing a computer to perform the image processing method according to any one of claims 1 to 6.
15. A computer program product comprising a computer program which, when executed by a processor, implements an image processing method according to any one of claims 1-6.
CN202210152880.9A 2022-02-18 2022-02-18 Image processing method, device, equipment and storage medium Pending CN114549697A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202210152880.9A CN114549697A (en) 2022-02-18 2022-02-18 Image processing method, device, equipment and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210152880.9A CN114549697A (en) 2022-02-18 2022-02-18 Image processing method, device, equipment and storage medium

Publications (1)

Publication Number Publication Date
CN114549697A true CN114549697A (en) 2022-05-27

Family

ID=81674612

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210152880.9A Pending CN114549697A (en) 2022-02-18 2022-02-18 Image processing method, device, equipment and storage medium

Country Status (1)

Country Link
CN (1) CN114549697A (en)

Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20160259766A1 (en) * 2015-03-08 2016-09-08 Microsoft Technology Licensing, Llc Ink experience for images
CN108760766A (en) * 2018-05-25 2018-11-06 哈尔滨工业大学 A kind of image split-joint method of large-aperture optical plane of crystal microdefect detection
US20190304150A1 (en) * 2018-03-27 2019-10-03 Faro Technologies, Inc. System and method of automatic room segmentation for two-dimensional floorplan annotation
CN110675465A (en) * 2019-09-23 2020-01-10 京东数字科技控股有限公司 Method and apparatus for generating image
CN111209721A (en) * 2018-11-16 2020-05-29 北京京东尚科信息技术有限公司 Bitmap font realization method and device, electronic equipment and storage medium
CN111367445A (en) * 2020-03-31 2020-07-03 中国建设银行股份有限公司 Image annotation method and device
CN111767490A (en) * 2020-06-30 2020-10-13 北京百度网讯科技有限公司 Method, device, equipment and storage medium for displaying image
CN112948735A (en) * 2021-02-26 2021-06-11 北京达佳互联信息技术有限公司 Image processing method, image processing device, electronic equipment and computer readable storage medium
CN114049631A (en) * 2021-11-06 2022-02-15 企查查科技有限公司 Data labeling method and device, computer equipment and storage medium

Patent Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20160259766A1 (en) * 2015-03-08 2016-09-08 Microsoft Technology Licensing, Llc Ink experience for images
US20190304150A1 (en) * 2018-03-27 2019-10-03 Faro Technologies, Inc. System and method of automatic room segmentation for two-dimensional floorplan annotation
CN108760766A (en) * 2018-05-25 2018-11-06 哈尔滨工业大学 A kind of image split-joint method of large-aperture optical plane of crystal microdefect detection
CN111209721A (en) * 2018-11-16 2020-05-29 北京京东尚科信息技术有限公司 Bitmap font realization method and device, electronic equipment and storage medium
CN110675465A (en) * 2019-09-23 2020-01-10 京东数字科技控股有限公司 Method and apparatus for generating image
CN111367445A (en) * 2020-03-31 2020-07-03 中国建设银行股份有限公司 Image annotation method and device
CN111767490A (en) * 2020-06-30 2020-10-13 北京百度网讯科技有限公司 Method, device, equipment and storage medium for displaying image
CN112948735A (en) * 2021-02-26 2021-06-11 北京达佳互联信息技术有限公司 Image processing method, image processing device, electronic equipment and computer readable storage medium
CN114049631A (en) * 2021-11-06 2022-02-15 企查查科技有限公司 Data labeling method and device, computer equipment and storage medium

Similar Documents

Publication Publication Date Title
WO2022227768A1 (en) Dynamic gesture recognition method and apparatus, and device and storage medium
CN113177469B (en) Training method and device of human attribute detection model, electronic equipment and medium
CN114494784A (en) Deep learning model training method, image processing method and object recognition method
US20220036111A1 (en) Method and device for associating panoramic image with point of interest, electronic device and storage medium
CN114511661A (en) Image rendering method and device, electronic equipment and storage medium
CN112861885A (en) Image recognition method and device, electronic equipment and storage medium
CN113205041A (en) Structured information extraction method, device, equipment and storage medium
CN112580666A (en) Image feature extraction method, training method, device, electronic equipment and medium
CN115861462A (en) Training method and device for image generation model, electronic equipment and storage medium
CN113705362A (en) Training method and device of image detection model, electronic equipment and storage medium
CN113885960B (en) Method and device for processing applet page, electronic equipment and storage medium
CN113837194A (en) Image processing method, image processing apparatus, electronic device, and storage medium
CN113221846A (en) Image recognition method, apparatus, device, storage medium, and program product
CN117312140A (en) Method and device for generating test case, electronic equipment and storage medium
CN114092608B (en) Expression processing method and device, computer readable storage medium and electronic equipment
CN113361519B (en) Target processing method, training method of target processing model and device thereof
CN114549697A (en) Image processing method, device, equipment and storage medium
CN114842476A (en) Watermark detection method and device and model training method and device
CN114093006A (en) Training method, device and equipment of living human face detection model and storage medium
CN114445668A (en) Image recognition method and device, electronic equipment and storage medium
CN112925942A (en) Data searching method, device, equipment and storage medium
CN113239889A (en) Image recognition method, device, equipment, storage medium and computer program product
CN113378958A (en) Automatic labeling method, device, equipment, storage medium and computer program product
CN114546199B (en) Image processing method, device, electronic equipment and storage medium
CN113420176B (en) Question searching method, question frame drawing device, question searching equipment and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination