CN113947566A - Visual restoration effect detection method, device and equipment and readable storage medium - Google Patents

Visual restoration effect detection method, device and equipment and readable storage medium Download PDF

Info

Publication number
CN113947566A
CN113947566A CN202111068492.4A CN202111068492A CN113947566A CN 113947566 A CN113947566 A CN 113947566A CN 202111068492 A CN202111068492 A CN 202111068492A CN 113947566 A CN113947566 A CN 113947566A
Authority
CN
China
Prior art keywords
information
visual
detection result
visual reduction
reduction effect
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202111068492.4A
Other languages
Chinese (zh)
Inventor
王钰
谢艳芬
潘文杰
李思琦
曾俊杰
周高景
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Sankuai Online Technology Co Ltd
Original Assignee
Beijing Sankuai Online Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Sankuai Online Technology Co Ltd filed Critical Beijing Sankuai Online Technology Co Ltd
Priority to CN202111068492.4A priority Critical patent/CN113947566A/en
Publication of CN113947566A publication Critical patent/CN113947566A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • G06T7/0004Industrial image inspection
    • G06T7/001Industrial image inspection using an image reference approach
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/44Arrangements for executing specific programs
    • G06F9/451Execution arrangements for user interfaces
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10004Still image; Photographic image

Landscapes

  • Engineering & Computer Science (AREA)
  • Software Systems (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Human Computer Interaction (AREA)
  • General Engineering & Computer Science (AREA)
  • Quality & Reliability (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

The application discloses a method, a device and equipment for detecting visual reduction effect and a readable storage medium, and belongs to the technical field of computers. Inputting a visual reference image and a visual restoration image of a page into an element detection model, and acquiring at least one first element, at least one second element, element information of the first element and element information of the second element, which are included in the visual reference image output by the element detection model; for any second element in at least one second element, obtaining a detection result of the visual reduction effect of the any second element based on element information of the first element corresponding to the any second element and element information of the any second element; and obtaining a detection result of the visual reduction effect of the visual reduction image based on the detection result of the visual reduction effect of the at least one second element. The method is high in detection efficiency, and good in detection effect when any second element is detected in the visual reduction effect.

Description

Visual restoration effect detection method, device and equipment and readable storage medium
Technical Field
The embodiment of the application relates to the technical field of computers, in particular to a method, a device and equipment for detecting visual reduction effect and a readable storage medium.
Background
During the development of an application program, a display page of the application program needs to be designed visually. Typically, a visual reference map of a page is designed by a designer, code is edited by a programmer, and the page is displayed on an electronic device based on the code. Wherein, the process of displaying the page on the electronic device is called a visual restoration process of the visual reference map. Because the visual reference image of the page may be different from the actually displayed page, the visual restoration effect needs to be detected to ensure that the visual restoration effect can meet the design requirements.
In the related art, when the visual reduction effect is detected, the visual reduction effect is detected in a manual mode. However, the detection of the visual restoration effect is carried out manually, so that the labor cost is high and the detection efficiency is low. Moreover, the visual reference map may include more elements, and the manual detection mode may have a missing detection condition, so that the detection effect is poor.
Disclosure of Invention
The embodiment of the application provides a method, a device and equipment for detecting visual reduction effect and a readable storage medium, which can be used for solving the problems in the related art. The technical scheme is as follows.
In a first aspect, an embodiment of the present application provides a method for detecting a visual reduction effect, where the method includes:
inputting a visual reference diagram and a visual reduction diagram of a page into an element detection model, and acquiring element information of at least one first element, at least one second element and at least one second element included in the visual reference diagram and output by the element detection model, wherein one second element corresponds to one first element;
for any second element in at least one second element, obtaining a detection result of the visual reduction effect of the any second element based on element information of the first element corresponding to the any second element and element information of the any second element;
and obtaining a detection result of the visual reduction effect of the visual reduction image based on the detection result of the visual reduction effect of the at least one second element.
In one possible implementation, the element information includes at least one of position information, size information, color information, or word weight information;
obtaining a detection result of the visual reduction effect of any second element based on the element information of the first element corresponding to the any second element and the element information of the any second element, including:
in response to the element information including position information, obtaining a position reduction detection result of any second element based on the position information of the first element corresponding to the any second element and the position information of the any second element;
obtaining a size reduction detection result of any second element based on the size information of the first element corresponding to any second element and the size information of any second element in response to the element information including the size information;
in response to the element information including color information, obtaining a color reduction detection result of any second element based on the color information of the first element corresponding to the any second element and the color information of the any second element;
and obtaining a reducibility detection result of any second element based on the word weight information of the first element corresponding to any second element and the word weight information of any second element in response to the element information including the word weight information.
In a possible implementation manner, obtaining the detection result of the visual reduction effect of the visual reduction image based on the detection result of the visual reduction effect of the at least one second element includes:
and responding to the detection result of the visual reduction effect of all the second elements that the visual reduction is abnormal, wherein the detection result of the visual reduction effect of the visual reduction image is that the visual reduction is abnormal.
In a possible implementation manner, obtaining the detection result of the visual reduction effect of the visual reduction image based on the detection result of the visual reduction effect of the at least one second element includes:
for any second element in the at least one second element, in response to the detection result of the visual reduction effect of the any second element being a visual reduction anomaly, the detection result of the visual reduction effect of the visual reduction image is a visual reduction anomaly.
In a possible implementation manner, after obtaining a detection result of the visual reduction effect of any one of the second elements based on the element information of the first element corresponding to the any one of the second elements and the element information of the any one of the second elements, the method further includes:
and for any second element in the at least one second element, responding to the visual reduction abnormity of any second element, and generating abnormity indication information which is used for indicating the visual reduction abnormity of any second element.
In a possible implementation manner, before inputting the visual reference map and the visual reduction map of the page into the element detection model, the method further includes:
displaying the page based on the uniform resource locator URL of the page;
and generating a screenshot of the page, wherein the screenshot of the page is used as the visual restoration image of the page.
In a second aspect, an embodiment of the present application provides an apparatus for detecting a visual reduction effect, where the apparatus includes:
the system comprises a first acquisition module, a second acquisition module and a third acquisition module, wherein the first acquisition module is used for inputting a visual reference map and a visual reduction map of a page into an element detection model, and acquiring at least one first element included in the visual reference map output by the element detection model, element information of the at least one first element, at least one second element included in the visual reduction map and element information of the at least one second element, wherein one second element corresponds to one first element;
the detection module is used for obtaining a detection result of the visual reduction effect of any one second element based on the element information of the first element corresponding to the any one second element and the element information of the any one second element for the any one second element in at least one second element;
and the second obtaining module is used for obtaining the detection result of the visual reduction effect of the visual reduction image based on the detection result of the visual reduction effect of the at least one second element.
In one possible implementation, the element information includes at least one of position information, size information, color information, or word weight information; the detection module is used for responding to the element information including position information, and obtaining a position reduction detection result of any second element based on the position information of the first element corresponding to any second element and the position information of any second element; obtaining a size reduction detection result of any second element based on the size information of the first element corresponding to any second element and the size information of any second element in response to the element information including the size information; in response to the element information including color information, obtaining a color reduction detection result of any second element based on the color information of the first element corresponding to the any second element and the color information of the any second element; and obtaining a reducibility detection result of any second element based on the word weight information of the first element corresponding to any second element and the word weight information of any second element in response to the element information including the word weight information.
In a possible implementation manner, the second obtaining module is configured to respond that the detection result of the visual reduction effect of all the second elements is that the visual reduction is not abnormal, and the detection result of the visual reduction effect of the visual reduction image is that the visual reduction is not abnormal.
In a possible implementation manner, the second obtaining module is configured to, for any second element in the at least one second element, respond that a detection result of the visual reduction effect of the any second element is a visual reduction exception, and determine that a detection result of the visual reduction effect of the visual reduction image is a visual reduction exception.
In one possible implementation, the apparatus further includes: and the first generation module is used for responding to the visual reduction abnormity of any second element in the at least one second element and generating abnormity indication information, and the abnormity indication information is used for indicating the visual reduction abnormity of any second element.
In one possible implementation, the apparatus further includes: the second generation module is used for displaying the page based on the URL of the page; and generating a screenshot of the page, wherein the screenshot of the page is used as the visual restoration image of the page.
In a third aspect, an embodiment of the present application provides an electronic device, where the electronic device includes a processor and a memory, where the memory stores at least one program code or instruction, and the at least one program code or instruction is loaded and executed by the processor, so as to enable the electronic device to implement the method for detecting a visual reduction effect in the first aspect or any one of the possible implementations of the first aspect.
In a fourth aspect, a computer-readable storage medium is further provided, where at least one program code or instruction is stored in the computer-readable storage medium, and the at least one program code or instruction is loaded and executed by a processor to enable a computer to implement the method for detecting a visual reduction effect in the foregoing first aspect or any one of the possible implementations of the first aspect.
In a fifth aspect, a computer program or a computer program product is further provided, where at least one computer instruction is stored, and the at least one computer instruction is loaded and executed by a processor, so as to enable a computer to implement the method for detecting a visual reduction effect in the first aspect or any one of the possible implementation manners of the first aspect.
The technical scheme provided by the embodiment of the application at least has the following beneficial effects.
According to the technical scheme, the first element included in the visual reference diagram of the page and the second element included in the visual restoration diagram of the page are acquired based on the element detection model, so that the first element and the second element can be automatically acquired, and the acquisition efficiency of the first element and the second element is high. For the corresponding first element and the corresponding second element, the detection result of the visual reduction effect of the second element can be obtained based on the element information of the first element and the element information of the second element, and then the detection result of the visual reduction effect of the visual reduction image can be obtained. In addition, the method detects the visual reduction effect of any second element, so that the method has a good detection effect on the visual reduction effect.
Drawings
In order to more clearly illustrate the technical solutions in the embodiments of the present application, the drawings needed to be used in the description of the embodiments are briefly introduced below, and it is obvious that the drawings in the following description are only some embodiments of the present application, and it is obvious for those skilled in the art to obtain other drawings based on these drawings without creative efforts.
Fig. 1 is a schematic implementation environment diagram of a method for detecting a visual reduction effect according to an embodiment of the present application;
fig. 2 is a flowchart of a method for detecting visual reduction effect according to an embodiment of the present application;
FIG. 3 is a schematic diagram of a first element included in a visual reference provided by an embodiment of the present application;
FIG. 4 is a schematic diagram of a second element included in a visual restoration map provided by an embodiment of the present application;
fig. 5 is a schematic diagram of abnormality indication information provided in an embodiment of the present application;
fig. 6 is a schematic structural diagram of an apparatus for detecting visual reduction effect according to an embodiment of the present application;
fig. 7 is a schematic structural diagram of an electronic device according to an embodiment of the present application.
Detailed Description
To make the objects, technical solutions and advantages of the present application more clear, embodiments of the present application will be described in further detail below with reference to the accompanying drawings.
Fig. 1 is a schematic implementation environment of a method for detecting a visual reduction effect according to an embodiment of the present application, and as shown in fig. 1, the implementation environment includes: an electronic device 101. The method for detecting the visual restoration effect in the embodiment of the present application may be executed by the electronic device 101. Illustratively, the electronic device 101 may comprise at least one of a terminal device or a server.
The terminal device may be at least one of a smart phone, a game console, a desktop computer, a tablet computer, an e-book reader, an MP3(Moving Picture Experts Group Audio Layer III, motion Picture Experts compression standard Audio Layer 3) player, an MP4(Moving Picture Experts Group Audio Layer IV, motion Picture Experts compression standard Audio Layer 4) player, or a laptop computer.
The server may be one server, or a server cluster formed by multiple servers, or any one of a cloud computing platform and a virtualization center, which is not limited in this embodiment of the present application. The server can be in communication connection with the terminal device through a wired network or a wireless network. The server may have functions of data processing, data storage, data transceiving, and the like, and is not limited in the embodiment of the present application.
Based on the implementation environment, the embodiment of the application provides a method for detecting a visual reduction effect. Taking the flowchart of the method for detecting the visual reduction effect provided in the embodiment of the present application shown in fig. 2 as an example, the method may be executed by the electronic device 101 in fig. 1. As shown in fig. 2, the method includes, but is not limited to, the following steps 201 to 203.
Step 201, inputting a visual reference map and a visual reduction map of a page into an element detection model, and obtaining at least one first element included in the visual reference map output by the element detection model, element information of the at least one first element, at least one second element included in the visual reduction map, and element information of the at least one second element, wherein one second element corresponds to one first element.
Illustratively, the visual reference map of a page is a visual reference map of the page entered by the user; the visual restoration of the page is a screenshot of the page as displayed on the electronic device. In one possible implementation, before inputting the visual reference map and the visual reduction map of the page into the element detection model, the method further includes: displaying a page based on a URL (Uniform Resource Locator) of the page; and generating a screenshot of the page, wherein the screenshot of the page is used as the visual restoration image of the page.
It should be noted that, for the way of displaying the page and generating the screenshot of the page based on the URL of the page, the embodiment of the present application is not limited. Illustratively, the page is a UI (User Interface).
Illustratively, after acquiring the visual reference image and the visual restoration image of the page, the electronic device performs the operation of step 201.
In One possible implementation, the element Detection model is any One of a FCOS (full volumetric One-Stage Object Detection), RCNN (Region-Based Convolutional Neural network) model, Fast RCNN (Fast Region-Based Convolutional Neural network) model, ultrafast Convolutional Neural network (ultrafast Region-Based Convolutional Neural network) model, YOLO (youonly local Single Convolutional Neural network) model, or Single SSD (Single multi box Detection) model.
The obtaining manner of the element detection model includes, but is not limited to, training the initial detection model based on the training sample. Illustratively, the training sample includes a visual reference image and a visual reduction image of the page; wherein, the visual reference map comprises at least one first element, element information of the at least one first element, at least one second element comprised by the visual reduction map, element information of the at least one second element, and a corresponding relation between the second element and the first element. Illustratively, meeting the requirement refers to reaching a threshold accuracy, or training a predetermined number of times, or model convergence, etc.
Illustratively, as shown in FIG. 3, the visual reference map of the page includes a plurality of first elements. As shown in FIG. 4, the visual restoration of the page includes a plurality of second elements. In addition, in the plurality of second elements shown in fig. 4, one second element corresponds to one first element shown in fig. 3.
Illustratively, the element information includes at least one of position information, size information, color information, or word weight information. That is, regardless of element information of the first element or element information of the second element, the element information of the first element includes at least one of position information, size information, color information, or word weight information, and the element information of the second element includes at least one of position information, size information, color information, or word weight information.
It should be noted that, for the corresponding first element and second element, the information type included in the element information of the first element is the same as the information type included in the element information of the second element. For example, the element information of a first element includes position information and size information, and the element information of a second element corresponding to the first element includes position information and size information.
Step 202, for any second element in at least one second element, obtaining a detection result of the visual reduction effect of the any second element based on the element information of the first element corresponding to the any second element and the element information of the any second element.
Illustratively, according to the information type included in the element information, the detection result of the visual reduction effect of any second element is obtained based on the element information of the first element corresponding to the any second element and the element information of the any second element, including but not limited to the following four cases.
In case 1, the element information includes location information.
In one possible implementation manner, for case 1, in response to that the element information includes position information, a position reduction detection result of any one second element is obtained based on the position information of the first element corresponding to the any one second element and the position information of the any one second element.
In one possible implementation, the location information includes first location information and/or second location information; the first position information is used for indicating the distance between the boundary of the element and the picture boundary, and the second position information is used for indicating the distance between the boundary of the element and the boundary of the element adjacent to the element.
For example, the position information of the first element includes first position information indicating a distance of a boundary of the first element from a boundary of the visual reference map and/or second position information indicating a distance of a boundary of the first element to a boundary of a first element adjacent to the first element. For another example, the position information of the second element includes first position information indicating a distance between a boundary of the second element and a boundary of the visual restoration image and/or second position information indicating a distance between a boundary of the second element and a boundary of an adjacent second element of the second element.
It should be noted that, for the corresponding first element and second element, the information type included in the position information of the first element is the same as the information type included in the position information of the second element.
Illustratively, according to the condition of the element information of the first element corresponding to the any second element and the position information included in the element information of the any second element, the position reduction detection result of the any second element is obtained based on the position information of the first element corresponding to the any second element and the position information of the any second element, including but not limited to the following conditions a to C.
In case a, the element information of the first element corresponding to the any one second element and the element information of the any one second element each include only the first position information.
Illustratively, for the case a, the position reduction detection result of any second element is obtained based on the first position information of the first element corresponding to the any second element and the first position information of the any second element.
In case B, the element information of the first element corresponding to the any one second element and the element information of the any one second element each include only the second position information.
Illustratively, for the case B, a position reduction detection result of any one second element is obtained based on the second position information of the first element corresponding to the any one second element and the second position information of the any one second element.
In case C, the element information of the first element and the element information of the second element corresponding to the second element each include first position information and second position information.
Illustratively, for the case C, the position reduction detection result of any one second element is obtained based on the first position information and the second position information of the first element corresponding to the any one second element, and the first position information and the second position information of the any one second element.
In a possible implementation manner, obtaining a position reduction detection result of any one second element based on the position information of the first element corresponding to the any one second element and the position information of the any one second element includes: acquiring the difference value between the position information of the first element corresponding to any one second element and the position information of any one second element; and obtaining a position reduction detection result of any second element based on the difference.
For example, for the case a, obtaining a position reduction detection result of any second element based on the first position information of the first element corresponding to the any second element and the first position information of the any second element includes: acquiring a first difference value between first position information of a first element corresponding to any one second element and first position information of any one second element; in response to the first difference not being greater than the first threshold, the position reduction detection result of any one second element is that the visual reduction is not abnormal; in response to the first difference being greater than a first threshold, the position restoration detection result of any one of the second elements is a visual restoration anomaly.
For example, for the above case B, obtaining a position reduction detection result of any one second element based on the second position information of the first element corresponding to the any one second element and the second position information of the any one second element includes: acquiring a second difference value between second position information of the first element corresponding to any one second element and second position information of any one second element; in response to the second difference being not greater than a second threshold, the position reduction detection result of any one of the second elements is that the visual reduction is not abnormal; and in response to the second difference being larger than a second threshold, the position reduction detection result of any one second element is a visual reduction abnormity.
For example, for the case C, obtaining a position reduction detection result of any second element based on the first position information and the second position information of the first element corresponding to the any second element and the first position information and the second position information of the any second element includes: acquiring a first difference value between first position information of a first element corresponding to any one second element and first position information of any one second element; acquiring a second difference value between second position information of the first element corresponding to any one second element and second position information of any one second element; in response to the first difference being not greater than a first threshold and the second difference being not greater than a second threshold, the position reduction detection result of any one of the second elements is that the visual reduction is abnormal-free; in response to the first difference being greater than a first threshold and/or the second difference being greater than a second threshold, the position restoration detection result of any one of the second elements is a visual restoration anomaly.
It should be noted that, both the first threshold and the second threshold may be set according to experience or actual requirements, which is not limited in the embodiments of the present application.
In case 2, the element information includes size information.
In one possible implementation manner, for case 2, in response to that the element information includes size information, a size reduction detection result of any second element is obtained based on the size information of the first element corresponding to the any second element and the size information of the any second element.
Illustratively, obtaining a size reduction detection result of any second element based on the size information of the first element corresponding to the any second element and the size information of the any second element includes: acquiring a third difference value between the size information of the first element corresponding to any one second element and the size information of any one second element; in response to the third difference being not greater than a third threshold, the size reduction detection result of any one of the second elements is that the visual reduction is abnormal-free; in response to the third difference being greater than a third threshold, the size reduction detection result of any of the second elements is a visual reduction anomaly.
It should be noted that the third threshold may be set according to experience or actual requirements, which is not limited in the embodiment of the present application.
In case 3, the element information includes color information.
In one possible implementation manner, for case 3, in response to that the element information includes color information, a color reduction detection result of any one second element is obtained based on the color information of the first element corresponding to the any one second element and the color information of the any one second element.
In one possible implementation, the color information is represented by HSV (Hue, Saturation, lightness) values. Illustratively, the color information is derived based on RGB (Red, Green, Blue, Red, Green, Blue) values of the elements.
For example, for any element, the V value of the element is the maximum value among the RGB values of the element, the S value of the element is obtained based on the following formula 1, and the H value of the element is obtained based on the following formula 2.
S ═ (Max-Min)/Max (equation 1)
Figure BDA0003259513030000101
Max is the maximum value of the RGB values, and Min is the minimum value of the RGB values. When the H value calculated based on formula 2 is smaller than 0, the value obtained by adding 360 to the H value is defined as the H value of the element.
Illustratively, obtaining a color reduction detection result of any second element based on the color information of the first element corresponding to the any second element and the color information of the any second element includes: acquiring a fourth difference value between the color information of the first element corresponding to any one second element and the color information of any one second element; in response to the fourth difference being not greater than a fourth threshold, the color reduction detection result of any one of the second elements is that the visual reduction is not abnormal; and responding to the fourth difference value being larger than a fourth threshold value, and the color reduction detection result of any second element is a visual reduction abnormity.
It should be noted that the fourth threshold may be set according to experience or actual requirements, which is not limited in the embodiment of the present application.
In case 4, the element information includes word repetition information.
In a possible implementation manner, for case 4, in response to that the element information includes word repetition information, a word repetition detection result of any second element is obtained based on the word repetition information of the first element corresponding to the any second element and the word repetition information of the any second element. Illustratively, the re-word information is a re-word level of the font of the element.
Exemplarily, obtaining a reducibility detection result of any second element based on the word weight information of the first element corresponding to the any second element and the word weight information of the any second element includes: acquiring a fifth difference value between the word weight information of the first element corresponding to any one second element and the word weight information of any one second element; responding to the fifth difference value not larger than a fifth threshold value, wherein the word reducibility detection result of any one second element is that the visual reduction is not abnormal; and responding to the fifth difference value being larger than a fifth threshold value, wherein the word reducibility detection result of any second element is visual reduction abnormity.
It should be noted that the fifth threshold may be set according to experience or actual requirements, which is not limited in the embodiment of the present application.
In one possible implementation manner, for any second element, the element information of the any second element includes a plurality of information types; responding to the detection results of all information types that the visual restoration is abnormal, wherein the detection result of the visual restoration effect of the second element is visual restoration abnormal; in response to the detection result of the presence of either information type being a visual reduction anomaly, the detection result of the visual reduction effect of the second element is a visual reduction anomaly.
It should be noted that, when the element information includes multiple information types, the detection of the multiple information types may be performed simultaneously, or the detection of one information type may be performed first, and then the detection of other information types may be performed, and the execution sequence is not limited in the embodiment of the present application.
For example, for a corresponding first element and second element, the element information of the first element and the element information of the second element each include size information and color information. The detection of the size information and the detection of the color information may be performed simultaneously, or the detection of the size information may be performed first and then the detection of the color information may be performed, or the detection of the color information may be performed first and then the detection of the size information may be performed. The method has flexible detection sequence of executing various information types. In addition, the method can perform detection of various information types, so that the detection effect of the detection of the visual restoration effect is better.
In a possible implementation manner, after obtaining a detection result of the visual reduction effect of any one of the second elements based on the element information of the first element corresponding to the any one of the second elements and the element information of the any one of the second elements, the method further includes: and for any second element in the at least one second element, responding to the visual reduction abnormity of any second element, and generating abnormity indication information which is used for indicating the visual reduction abnormity of any second element.
Illustratively, the anomaly indication information includes an anomaly type, wherein the anomaly type corresponds to an information type of the element information. For example, the abnormality indication information includes at least one of a position abnormality, a size abnormality, a color abnormality, or a word weight abnormality; wherein the position anomaly corresponds to position information, the size anomaly corresponds to size information, the color anomaly corresponds to color information, and the word weight anomaly corresponds to word weight information.
Illustratively, the abnormality indication information further includes an abnormality value for each abnormality type. For example, if the result of the position reduction detection of the second element is a visual reduction abnormality, the abnormality indication information further includes an abnormal value of the position abnormality. Illustratively, the outlier of the positional anomaly comprises the first difference and/or the second difference. It should be noted that the abnormal values of other abnormal types are the same as the abnormal values of the position abnormality in principle, and are not described herein again. Illustratively, referring to fig. 5, the result of the size reduction detection of the presence of the second element in the visual reduction image is a visual reduction exception, and the exception indication information includes: the anomaly type is a size anomaly, and the anomaly values are 10 pixels in length and 2 pixels in width.
Next, with continuing reference to fig. 2, after obtaining a detection result of the visual reduction effect of any one of the second elements, the electronic device executes step 203.
Step 203, obtaining a detection result of the visual reduction effect of the visual reduction image based on the detection result of the visual reduction effect of the at least one second element.
In one possible implementation manner, the detection result of the visual reduction effect of the visual reduction image is obtained based on the detection result of the visual reduction effect of the at least one second element according to the condition of the detection result of the visual reduction effect of the second element, which includes but is not limited to the following two conditions.
In case one, in response to that the detection results of the visual restoration effects of all the second elements are visual restoration non-abnormal, the detection result of the visual restoration effect of the visual restoration image is visual restoration non-abnormal.
For example, a visual reference map of a page includes 10 first elements and a visual restoration map of the page includes 10 second elements, where one second element corresponds to one first element. And responding to that the visual reduction effects of the 10 second elements are all visual reduction without abnormity, and the detection result of the visual reduction effect of the visual reduction image is that the visual reduction is not abnormal.
In case two, for any second element in the at least one second element, in response to the detection result of the visual reduction effect of the any second element being the visual reduction abnormality, the detection result of the visual reduction effect of the visual reduction image is the visual reduction abnormality.
Illustratively, the description will be given by taking the visual reference image of the page shown in fig. 3 and the visual restoration image of the page shown in fig. 4 as an example. The visual reference map includes 12 first elements and the visual restoration map includes 12 second elements, wherein one second element corresponds to one first element. In response to the detection result of the visual reduction effect of the second element "second shop" being the visual reduction abnormality, the detection result of the visual reduction effect of the visual reduction figure is the visual reduction abnormality.
According to the method, the first element included in the visual reference image of the page and the second element included in the visual reduction image of the page are obtained based on the element detection model, so that the first element and the second element can be automatically obtained, and the first element and the second element are obtained with high efficiency. For the corresponding first element and the corresponding second element, the detection result of the visual reduction effect of the second element can be obtained based on the element information of the first element and the element information of the second element, and then the detection result of the visual reduction effect of the visual reduction image can be obtained. In addition, the method detects the visual reduction effect of any second element, so that the method has a good detection effect on the visual reduction effect.
Fig. 6 is a schematic structural diagram of an apparatus for detecting visual reduction effect according to an embodiment of the present application, where the apparatus is applied to an electronic device, and as shown in fig. 6, the apparatus includes the following modules:
a first obtaining module 601, configured to input a visual reference diagram and a visual reduction diagram of a page into an element detection model, and obtain at least one first element included in the visual reference diagram output by the element detection model, element information of the at least one first element, at least one second element included in the visual reduction diagram, and element information of the at least one second element, where one second element corresponds to one first element;
a detecting module 602, configured to, for any second element in at least one second element, obtain a detection result of a visual reduction effect of the any second element based on element information of a first element corresponding to the any second element and the element information of the any second element;
the second obtaining module 603 is configured to obtain a detection result of the visual reduction effect of the visual reduction image based on the detection result of the visual reduction effect of the at least one second element.
In one possible implementation, the element information includes at least one of position information, size information, color information, or word weight information; a detection module 602, configured to, in response to that the element information includes position information, obtain a position reduction detection result of any one second element based on the position information of the first element corresponding to the any one second element and the position information of the any one second element; obtaining a size reduction detection result of any second element based on the size information of the first element corresponding to any second element and the size information of any second element in response to the element information including the size information; in response to the element information including color information, obtaining a color reduction detection result of any second element based on the color information of the first element corresponding to the any second element and the color information of the any second element; and obtaining a reducibility detection result of any second element based on the word weight information of the first element corresponding to any second element and the word weight information of any second element in response to the element information including the word weight information.
In a possible implementation manner, the second obtaining module 603 is configured to respond that the detection result of the visual reduction effect of all the second elements is that the visual reduction is not abnormal, and the detection result of the visual reduction effect of the visual reduction image is that the visual reduction is not abnormal.
In a possible implementation manner, the second obtaining module 603 is configured to, for any second element in the at least one second element, respond that a detection result of the visual reduction effect of the any second element is a visual reduction exception, and determine that a detection result of the visual reduction effect of the visual reduction image is a visual reduction exception.
In one possible implementation, the apparatus further includes: and the first generation module is used for responding to the visual reduction abnormity of any second element in the at least one second element and generating abnormity indication information, and the abnormity indication information is used for indicating the visual reduction abnormity of any second element.
In one possible implementation, the apparatus further includes: the second generation module is used for displaying the page based on the URL of the page; and generating a screenshot of the page, wherein the screenshot of the page is used as the visual restoration image of the page.
According to the device, the first element included in the visual reference image of the page and the second element included in the visual restoration image of the page are acquired based on the element detection model, so that the first element and the second element can be automatically acquired, and the acquisition efficiency of the first element and the second element is high. For the corresponding first element and the corresponding second element, the detection result of the visual reduction effect of the second element can be obtained based on the element information of the first element and the element information of the second element, and then the detection result of the visual reduction effect of the visual reduction image is obtained, so that the detection efficiency of detecting the visual reduction effect is high. In addition, as any second element is detected by the visual reduction effect, the detection effect of detecting the visual reduction effect is better.
It should be understood that, when the apparatus provided in fig. 6 implements its functions, it is only illustrated by the division of the functional modules, and in practical applications, the above functions may be distributed by different functional modules according to needs, that is, the internal structure of the apparatus is divided into different functional modules to implement all or part of the functions described above. In addition, the apparatus provided in the above embodiments and the corresponding method embodiments belong to the same concept, and specific implementation processes thereof are described in the method embodiments for details, which are not described herein again.
Fig. 7 shows a block diagram of an electronic device 700 according to an exemplary embodiment of the present application. The electronic device 700 may be a portable mobile terminal, such as: a smart phone, a tablet computer, an MP3 player (Moving Picture Experts Group Audio Layer III, motion video Experts compression standard Audio Layer 3), an MP4 player (Moving Picture Experts Group Audio Layer IV, motion video Experts compression standard Audio Layer 4), a notebook computer, or a desktop computer. The electronic device 700 may also be referred to by other names such as user equipment, portable terminal, laptop terminal, desktop terminal, and so forth.
In general, the electronic device 700 includes: a processor 701 and a memory 702.
The processor 701 may include one or more processing cores, such as a 4-core processor, an 8-core processor, and so on. The processor 701 may be implemented in at least one hardware form of a DSP (Digital Signal Processing), an FPGA (Field-Programmable Gate Array), and a PLA (Programmable Logic Array). The processor 701 may also include a main processor and a coprocessor, where the main processor is a processor for Processing data in an awake state, and is also called a Central Processing Unit (CPU); a coprocessor is a low power processor for processing data in a standby state. In some embodiments, the processor 701 may be integrated with a GPU (Graphics Processing Unit) which is responsible for rendering and drawing the content required to be displayed by the display screen. In some embodiments, the processor 701 may further include an AI (Artificial Intelligence) processor for processing computing operations related to machine learning.
Memory 702 may include one or more computer-readable storage media, which may be non-transitory. Memory 702 may also include high-speed random access memory, as well as non-volatile memory, such as one or more magnetic disk storage devices, flash memory storage devices. In some embodiments, a non-transitory computer readable storage medium in memory 702 is used to store at least one instruction for execution by processor 701 to implement a method of detecting a visual reduction effect provided by method embodiments herein.
In some embodiments, the electronic device 700 may further optionally include: a peripheral interface 703 and at least one peripheral. The processor 701, the memory 702, and the peripheral interface 703 may be connected by buses or signal lines. Various peripheral devices may be connected to peripheral interface 703 via a bus, signal line, or circuit board. Specifically, the peripheral device includes: at least one of a radio frequency circuit 704, a display screen 705, a camera assembly 706, an audio circuit 707, a positioning component 708, and a power source 709.
The peripheral interface 703 may be used to connect at least one peripheral related to I/O (Input/Output) to the processor 701 and the memory 702. In some embodiments, processor 701, memory 702, and peripheral interface 703 are integrated on the same chip or circuit board; in some other embodiments, any one or two of the processor 701, the memory 702, and the peripheral interface 703 may be implemented on a separate chip or circuit board, which is not limited in this embodiment.
The Radio Frequency circuit 704 is used for receiving and transmitting RF (Radio Frequency) signals, also called electromagnetic signals. The radio frequency circuitry 704 communicates with communication networks and other communication devices via electromagnetic signals. The rf circuit 704 converts an electrical signal into an electromagnetic signal to transmit, or converts a received electromagnetic signal into an electrical signal. Optionally, the radio frequency circuit 704 includes: an antenna system, an RF transceiver, one or more amplifiers, a tuner, an oscillator, a digital signal processor, a codec chipset, a subscriber identity module card, and so forth. The radio frequency circuitry 704 may communicate with other terminals via at least one wireless communication protocol. The wireless communication protocols include, but are not limited to: the world wide web, metropolitan area networks, intranets, generations of mobile communication networks (2G, 3G, 4G, and 5G), Wireless local area networks, and/or WiFi (Wireless Fidelity) networks. In some embodiments, the radio frequency circuit 704 may also include NFC (Near Field Communication) related circuits, which are not limited in this application.
The display screen 705 is used to display a UI (User Interface). The UI may include graphics, text, icons, video, and any combination thereof. When the display screen 705 is a touch display screen, the display screen 705 also has the ability to capture touch signals on or over the surface of the display screen 705. The touch signal may be input to the processor 701 as a control signal for processing. At this point, the display 705 may also be used to provide virtual buttons and/or a virtual keyboard, also referred to as soft buttons and/or a soft keyboard. In some embodiments, the display 705 may be one, disposed on the front panel of the electronic device 700; in other embodiments, the number of the display screens 705 may be at least two, and the at least two display screens are respectively disposed on different surfaces of the electronic device 700 or are in a folding design; in other embodiments, the display 705 may be a flexible display disposed on a curved surface or on a folded surface of the electronic device 700. Even more, the display 705 may be arranged in a non-rectangular irregular pattern, i.e. a shaped screen. The Display 705 may be made of LCD (Liquid Crystal Display), OLED (Organic Light-Emitting Diode), or the like.
The camera assembly 706 is used to capture images or video. Optionally, camera assembly 706 includes a front camera and a rear camera. Generally, a front camera is disposed at a front panel of the terminal, and a rear camera is disposed at a rear surface of the terminal. In some embodiments, the number of the rear cameras is at least two, and each rear camera is any one of a main camera, a depth-of-field camera, a wide-angle camera and a telephoto camera, so that the main camera and the depth-of-field camera are fused to realize a background blurring function, and the main camera and the wide-angle camera are fused to realize panoramic shooting and VR (Virtual Reality) shooting functions or other fusion shooting functions. In some embodiments, camera assembly 706 may also include a flash. The flash lamp can be a monochrome temperature flash lamp or a bicolor temperature flash lamp. The double-color-temperature flash lamp is a combination of a warm-light flash lamp and a cold-light flash lamp, and can be used for light compensation at different color temperatures.
The audio circuitry 707 may include a microphone and a speaker. The microphone is used for collecting sound waves of a user and the environment, converting the sound waves into electric signals, and inputting the electric signals to the processor 701 for processing or inputting the electric signals to the radio frequency circuit 704 to realize voice communication. For stereo capture or noise reduction purposes, the microphones may be multiple and disposed at different locations of the electronic device 700. The microphone may also be an array microphone or an omni-directional pick-up microphone. The speaker is used to convert electrical signals from the processor 701 or the radio frequency circuit 704 into sound waves. The loudspeaker can be a traditional film loudspeaker or a piezoelectric ceramic loudspeaker. When the speaker is a piezoelectric ceramic speaker, the speaker can be used for purposes such as converting an electric signal into a sound wave audible to a human being, or converting an electric signal into a sound wave inaudible to a human being to measure a distance. In some embodiments, the audio circuitry 707 may also include a headphone jack.
The positioning component 708 is operable to locate a current geographic Location of the electronic device 700 to implement a navigation or LBS (Location Based Service). The Positioning component 708 can be a Positioning component based on the Global Positioning System (GPS) in the united states, the beidou System in china, or the galileo System in russia.
The power supply 709 is used to supply power to various components in the electronic device 700. The power source 709 may be alternating current, direct current, disposable batteries, or rechargeable batteries. When the power source 709 includes a rechargeable battery, the rechargeable battery may be a wired rechargeable battery or a wireless rechargeable battery. The wired rechargeable battery is a battery charged through a wired line, and the wireless rechargeable battery is a battery charged through a wireless coil. The rechargeable battery may also be used to support fast charge technology.
In some embodiments, the electronic device 700 also includes one or more sensors 710. The one or more sensors 710 include, but are not limited to: acceleration sensor 711, gyro sensor 712, pressure sensor 713, fingerprint sensor 714, optical sensor 715, and proximity sensor 716.
The acceleration sensor 711 may detect the magnitude of acceleration in three coordinate axes of a coordinate system established with the electronic device 700. For example, the acceleration sensor 711 may be used to detect components of the gravitational acceleration in three coordinate axes. The processor 701 may control the display screen 705 to display the user interface in a landscape view or a portrait view according to the gravitational acceleration signal collected by the acceleration sensor 711. The acceleration sensor 711 may also be used for acquisition of motion data of a game or a user.
The gyro sensor 712 may detect a body direction and a rotation angle of the electronic device 700, and the gyro sensor 712 may cooperate with the acceleration sensor 711 to acquire a 3D motion of the user with respect to the electronic device 700. From the data collected by the gyro sensor 712, the processor 701 may implement the following functions: motion sensing (such as changing the UI according to a user's tilting operation), image stabilization at the time of photographing, game control, and inertial navigation.
Pressure sensors 713 may be disposed on a side bezel of electronic device 700 and/or underlying display screen 705. When the pressure sensor 713 is disposed on a side frame of the electronic device 700, a user holding signal of the electronic device 700 may be detected, and the processor 701 may perform left-right hand recognition or shortcut operation according to the holding signal collected by the pressure sensor 713. When the pressure sensor 713 is disposed at a lower layer of the display screen 705, the processor 701 controls the operability control on the UI interface according to the pressure operation of the user on the display screen 705. The operability control comprises at least one of a button control, a scroll bar control, an icon control and a menu control.
The fingerprint sensor 714 is used for collecting a fingerprint of a user, and the processor 701 identifies the identity of the user according to the fingerprint collected by the fingerprint sensor 714, or the fingerprint sensor 714 identifies the identity of the user according to the collected fingerprint. When the user identity is identified as a trusted identity, the processor 701 authorizes the user to perform relevant sensitive operations, including unlocking a screen, viewing encrypted information, downloading software, paying, changing settings, and the like. The fingerprint sensor 714 may be disposed on the front, back, or side of the electronic device 700. When a physical button or vendor Logo is provided on the electronic device 700, the fingerprint sensor 714 may be integrated with the physical button or vendor Logo.
The optical sensor 715 is used to collect the ambient light intensity. In one embodiment, the processor 701 may control the display brightness of the display screen 705 based on the ambient light intensity collected by the optical sensor 715. Specifically, when the ambient light intensity is high, the display brightness of the display screen 705 is increased; when the ambient light intensity is low, the display brightness of the display screen 705 is adjusted down. In another embodiment, processor 701 may also dynamically adjust the shooting parameters of camera assembly 706 based on the ambient light intensity collected by optical sensor 715.
A proximity sensor 716, also referred to as a distance sensor, is typically disposed on the front panel of the electronic device 700. The proximity sensor 716 is used to capture the distance between the user and the front of the electronic device 700. In one embodiment, the processor 701 controls the display screen 705 to switch from the bright screen state to the dark screen state when the proximity sensor 716 detects that the distance between the user and the front surface of the electronic device 700 is gradually decreased; when the proximity sensor 716 detects that the distance between the user and the front surface of the electronic device 700 is gradually increased, the processor 701 controls the display screen 705 to switch from the breath screen state to the bright screen state.
Those skilled in the art will appreciate that the configuration shown in fig. 7 does not constitute a limitation of the electronic device 700 and may include more or fewer components than those shown, or combine certain components, or employ a different arrangement of components.
In an exemplary embodiment, there is also provided a computer-readable storage medium having at least one program code stored therein, the at least one program code being loaded and executed by a processor to cause a computer to implement any one of the above-mentioned methods for detecting visual restoration effects.
Alternatively, the computer-readable storage medium may be a Read-Only Memory (ROM), a Random Access Memory (RAM), a Compact Disc Read-Only Memory (CD-ROM), a magnetic tape, a floppy disk, an optical data storage device, and the like.
In an exemplary embodiment, a computer program or a computer program product is further provided, in which at least one computer instruction is stored, and the at least one computer instruction is loaded and executed by a processor, so as to enable a computer to implement any one of the above-mentioned methods for detecting visual reduction effects.
It should be understood that reference to "a plurality" herein means two or more. "and/or" describes the association relationship of the associated objects, meaning that there may be three relationships, e.g., a and/or B, which may mean: a exists alone, A and B exist simultaneously, and B exists alone. The character "/" generally indicates that the former and latter associated objects are in an "or" relationship.
The above-mentioned serial numbers of the embodiments of the present application are merely for description and do not represent the merits of the embodiments.
The above description is only exemplary of the present application and should not be taken as limiting the present application, and any modifications, equivalents, improvements and the like that are made within the spirit and principle of the present application should be included in the protection scope of the present application.

Claims (10)

1. A method for detecting visual reduction effects, the method comprising:
inputting a visual reference diagram and a visual reduction diagram of a page into an element detection model, and acquiring at least one first element included in the visual reference diagram, element information of the at least one first element, at least one second element included in the visual reduction diagram and element information of the at least one second element, wherein one second element corresponds to one first element, which are output by the element detection model;
for any second element in the at least one second element, obtaining a detection result of the visual reduction effect of the any second element based on element information of a first element corresponding to the any second element and the element information of the any second element;
and obtaining a detection result of the visual reduction effect of the visual reduction image based on the detection result of the visual reduction effect of the at least one second element.
2. The method of claim 1, wherein the element information includes at least one of position information, size information, color information, or word weight information;
the obtaining a detection result of the visual reduction effect of any second element based on the element information of the first element corresponding to any second element and the element information of any second element includes:
in response to the element information including position information, obtaining a position reduction detection result of any second element based on the position information of the first element corresponding to the any second element and the position information of the any second element;
obtaining a size reduction detection result of any second element based on the size information of the first element corresponding to any second element and the size information of any second element in response to the element information including the size information;
in response to the element information including color information, obtaining a color reduction detection result of any second element based on the color information of the first element corresponding to the any second element and the color information of the any second element;
and in response to the element information including word repetition information, obtaining a word repetition reduction detection result of any second element based on the word repetition information of the first element corresponding to the any second element and the word repetition information of the any second element.
3. The method according to claim 1 or 2, wherein the obtaining the detection result of the visual reduction effect of the visual reduction image based on the detection result of the visual reduction effect of the at least one second element comprises:
and responding to the detection result of the visual reduction effect of all the second elements that the visual reduction is abnormal, wherein the detection result of the visual reduction effect of the visual reduction image is that the visual reduction is abnormal.
4. The method according to claim 1 or 2, wherein the obtaining the detection result of the visual reduction effect of the visual reduction image based on the detection result of the visual reduction effect of the at least one second element comprises:
for any second element in the at least one second element, in response to the detection result of the visual reduction effect of the any second element being a visual reduction anomaly, the detection result of the visual reduction effect of the visual reduction image being a visual reduction anomaly.
5. The method according to claim 1 or 2, wherein after obtaining the detection result of the visual reduction effect of any second element based on the element information of the first element corresponding to any second element and the element information of any second element, the method further comprises:
for any second element in the at least one second element, in response to the visual reduction abnormality of the any second element, generating abnormality indication information, wherein the abnormality indication information is used for indicating the visual reduction abnormality of the any second element.
6. The method according to claim 1 or 2, wherein before inputting the visual reference map and the visual reduction map of the page into the element detection model, the method further comprises:
displaying the page based on a Uniform Resource Locator (URL) of the page;
and generating a screenshot of the page, wherein the screenshot of the page is used as the visual restoration image of the page.
7. An apparatus for detecting visual reduction effects, the apparatus comprising:
the system comprises a first acquisition module, a second acquisition module and a third acquisition module, wherein the first acquisition module is used for inputting a visual reference map and a visual reduction map of a page into an element detection model, and acquiring at least one first element included in the visual reference map output by the element detection model, element information of the at least one first element, at least one second element included in the visual reduction map and element information of the at least one second element, wherein one second element corresponds to one first element;
the detection module is used for obtaining a detection result of the visual reduction effect of any second element based on the element information of the first element corresponding to the any second element and the element information of the any second element for any second element in the at least one second element;
and the second obtaining module is used for obtaining a detection result of the visual reduction effect of the visual reduction image based on the detection result of the visual reduction effect of the at least one second element.
8. An electronic device, comprising a processor and a memory, wherein at least one program code or instruction is stored in the memory, and the at least one program code or instruction is loaded and executed by the processor to cause the electronic device to implement the method for detecting visual reduction effect according to any one of claims 1-6.
9. A computer-readable storage medium having at least one program code stored therein, the at least one program code being loaded and executed by a processor to cause a computer to implement the method for detecting visual restoration effect according to any one of claims 1 to 6.
10. A computer program product having at least one computer instruction stored therein, the at least one computer instruction being loaded and executed by a processor to cause a computer to implement the method for detecting visual reduction effect according to any one of claims 1 to 6.
CN202111068492.4A 2021-09-13 2021-09-13 Visual restoration effect detection method, device and equipment and readable storage medium Pending CN113947566A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202111068492.4A CN113947566A (en) 2021-09-13 2021-09-13 Visual restoration effect detection method, device and equipment and readable storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202111068492.4A CN113947566A (en) 2021-09-13 2021-09-13 Visual restoration effect detection method, device and equipment and readable storage medium

Publications (1)

Publication Number Publication Date
CN113947566A true CN113947566A (en) 2022-01-18

Family

ID=79328130

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202111068492.4A Pending CN113947566A (en) 2021-09-13 2021-09-13 Visual restoration effect detection method, device and equipment and readable storage medium

Country Status (1)

Country Link
CN (1) CN113947566A (en)

Similar Documents

Publication Publication Date Title
CN110839128B (en) Photographing behavior detection method and device and storage medium
CN110321126B (en) Method and device for generating page code
CN109886208B (en) Object detection method and device, computer equipment and storage medium
CN110442521B (en) Control unit detection method and device
WO2022134632A1 (en) Work processing method and apparatus
CN112581358B (en) Training method of image processing model, image processing method and device
WO2023142915A1 (en) Image processing method, apparatus and device, and storage medium
CN114494469A (en) Image processing method, device, equipment and storage medium
CN110647881A (en) Method, device, equipment and storage medium for determining card type corresponding to image
CN111754386A (en) Image area shielding method, device, equipment and storage medium
CN111105474B (en) Font drawing method, font drawing device, computer device and computer readable storage medium
CN111857793B (en) Training method, device, equipment and storage medium of network model
CN109754439B (en) Calibration method, calibration device, electronic equipment and medium
CN112989198B (en) Push content determination method, device, equipment and computer-readable storage medium
CN112612405B (en) Window display method, device, equipment and computer readable storage medium
CN111127541A (en) Vehicle size determination method and device and storage medium
CN111179628B (en) Positioning method and device for automatic driving vehicle, electronic equipment and storage medium
CN111370096A (en) Interactive interface display method, device, equipment and storage medium
CN111860064A (en) Target detection method, device and equipment based on video and storage medium
CN111757146B (en) Method, system and storage medium for video splicing
CN110555443B (en) Color classification method, device and storage medium
CN112560903A (en) Method, device and equipment for determining image aesthetic information and storage medium
CN113947566A (en) Visual restoration effect detection method, device and equipment and readable storage medium
CN111523876A (en) Payment mode display method, device and system and storage medium
CN112990424A (en) Method and device for training neural network model

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination