CN114510187B - Image display method and device, electronic equipment and medium - Google Patents

Image display method and device, electronic equipment and medium Download PDF

Info

Publication number
CN114510187B
CN114510187B CN202210107799.9A CN202210107799A CN114510187B CN 114510187 B CN114510187 B CN 114510187B CN 202210107799 A CN202210107799 A CN 202210107799A CN 114510187 B CN114510187 B CN 114510187B
Authority
CN
China
Prior art keywords
image
area
resolution
magnification
display window
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202210107799.9A
Other languages
Chinese (zh)
Other versions
CN114510187A (en
Inventor
刘俊启
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Baidu Netcom Science and Technology Co Ltd
Original Assignee
Beijing Baidu Netcom Science and Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Baidu Netcom Science and Technology Co Ltd filed Critical Beijing Baidu Netcom Science and Technology Co Ltd
Priority to CN202210107799.9A priority Critical patent/CN114510187B/en
Publication of CN114510187A publication Critical patent/CN114510187A/en
Application granted granted Critical
Publication of CN114510187B publication Critical patent/CN114510187B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0484Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
    • G06F3/04845Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range for image manipulation, e.g. dragging, rotation, expansion or change of colour
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0481Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformation in the plane of the image
    • G06T3/40Scaling the whole image or part thereof
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformation in the plane of the image
    • G06T3/40Scaling the whole image or part thereof
    • G06T3/4053Super resolution, i.e. output image resolution higher than sensor resolution
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2203/00Indexing scheme relating to G06F3/00 - G06F3/048
    • G06F2203/048Indexing scheme relating to G06F3/048
    • G06F2203/04806Zoom, i.e. interaction techniques or interactors for controlling the zooming operation

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • General Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Human Computer Interaction (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

The disclosure provides an image display method and device, electronic equipment and medium, relates to the technical field of computers, and particularly relates to the technical field of image processing. The implementation scheme is as follows: responding to the amplifying operation of a user on a first image, judging whether the current first scaling of the first image is larger than or equal to a scaling threshold value; in response to determining that the first scale is greater than or equal to the scale threshold, performing enhancement processing on the first image to obtain a second image; and displaying the second image.

Description

Image display method and device, electronic equipment and medium
Technical Field
The present disclosure relates to the field of computer technology, and in particular, to the field of image processing technology, and in particular, to an image display method and apparatus, an electronic device, a computer readable storage medium, and a computer program product.
Background
Users are accustomed to obtaining information over the internet. The internet includes a large number of web pages. Images are widely available in a variety of web pages as a vivid, efficient information carrier.
The approaches described in this section are not necessarily approaches that have been previously conceived or pursued. Unless otherwise indicated, it should not be assumed that any of the approaches described in this section qualify as prior art merely by virtue of their inclusion in this section. Similarly, the problems mentioned in this section should not be considered as having been recognized in any prior art unless otherwise indicated.
Disclosure of Invention
The present disclosure provides an image display method and apparatus, an electronic device, a computer-readable storage medium, and a computer program product.
According to an aspect of the present disclosure, there is provided an image display method including: responding to the amplifying operation of a user on a first image, judging whether the current first scaling of the first image is larger than or equal to a scaling threshold value; in response to determining that the first scale is greater than or equal to the scale threshold, performing enhancement processing on the first image to obtain a second image; and displaying the second image.
According to an aspect of the present disclosure, there is provided an image display apparatus including: a judging module configured to judge whether a current first scaling of a first image is greater than or equal to a scaling threshold value in response to an amplifying operation of the first image by a user; an enhancement module configured to enhance the first image to obtain a second image in response to determining that the first scale is greater than or equal to the scale threshold; and a display module configured to display the second image.
According to an aspect of the present disclosure, there is provided an electronic apparatus including: at least one processor; and a memory communicatively coupled to the at least one processor, the memory storing instructions executable by the at least one processor to enable the at least one processor to perform the method.
According to an aspect of the present disclosure, there is provided a non-transitory computer-readable storage medium storing computer instructions for causing a computer to perform the above-described method.
According to an aspect of the present disclosure, there is provided a computer program product comprising a computer program which, when executed by a processor, implements the above method.
According to one or more embodiments of the present disclosure, an image can be enhanced based on a user's requirement, so that automatic optimization of image quality is achieved, a user can obtain interesting information from the image, and user experience is improved.
It should be understood that the description in this section is not intended to identify key or critical features of the embodiments of the disclosure, nor is it intended to be used to limit the scope of the disclosure. Other features of the present disclosure will become apparent from the following specification.
Drawings
The accompanying drawings illustrate exemplary embodiments and, together with the description, serve to explain exemplary implementations of the embodiments. The illustrated embodiments are for exemplary purposes only and do not limit the scope of the claims. Throughout the drawings, identical reference numerals designate similar, but not necessarily identical, elements.
FIG. 1 illustrates a schematic diagram of an exemplary system in which various methods described herein may be implemented, in accordance with an embodiment of the present disclosure;
FIG. 2 shows a flowchart of an image display method according to an embodiment of the present disclosure;
3A-3D illustrate schematic diagrams of adjusting image scaling according to embodiments of the present disclosure;
4A-4B show schematic diagrams of a first image, a second image, according to an embodiment of the present disclosure;
fig. 5 shows a block diagram of an image display apparatus according to an embodiment of the present disclosure; and
fig. 6 illustrates a block diagram of an exemplary electronic device that can be used to implement embodiments of the present disclosure.
Detailed Description
Exemplary embodiments of the present disclosure are described below in conjunction with the accompanying drawings, which include various details of the embodiments of the present disclosure to facilitate understanding, and should be considered as merely exemplary. Accordingly, one of ordinary skill in the art will recognize that various changes and modifications of the embodiments described herein can be made without departing from the scope of the present disclosure. Also, descriptions of well-known functions and constructions are omitted in the following description for clarity and conciseness.
In the present disclosure, the use of the terms "first," "second," and the like to describe various elements is not intended to limit the positional relationship, timing relationship, or importance relationship of the elements, unless otherwise indicated, and such terms are merely used to distinguish one element from another. In some examples, a first element and a second element may refer to the same instance of the element, and in some cases, they may also refer to different instances based on the description of the context.
The terminology used in the description of the various illustrated examples in this disclosure is for the purpose of describing particular examples only and is not intended to be limiting. Unless the context clearly indicates otherwise, the elements may be one or more if the number of the elements is not specifically limited. Furthermore, the term "and/or" as used in this disclosure encompasses any and all possible combinations of the listed items.
There are a large number of web pages in the internet that contain images. In the related art, a web page is generally capable of providing an image scaling function, and an image in the web page can be scaled and displayed based on user interaction. For example, when a user clicks on an image in a web page, the display window of the image will pop up in the current web page. The user can adjust the scaling (e.g., 80%, 100%, 150%, etc.) of the image by performing interactive operations (e.g., clicking, double clicking, dragging a control in the window, etc.) in the window, and zoom the image for display, thereby facilitating the user to view the image and obtain information of interest therefrom.
However, the quality of images in different web pages varies. For the image with poor quality, even if the image is displayed in an enlarged manner, the user cannot see the details of the image and cannot acquire the required information from the image.
In view of the above problems, an embodiment of the present disclosure provides an image display method, which can enhance an image based on a user's requirement, thereby implementing automatic optimization of image quality, facilitating a user to obtain interesting information from the image, and improving user experience.
Embodiments of the present disclosure will be described in detail below with reference to the accompanying drawings.
Fig. 1 illustrates a schematic diagram of an exemplary system 100 in which various methods and apparatus described herein may be implemented, in accordance with an embodiment of the present disclosure. Referring to fig. 1, the system 100 includes one or more client devices 101, 102, 103, 104, 105, and 106, a server 120, and one or more communication networks 110 coupling the one or more client devices to the server 120. Client devices 101, 102, 103, 104, 105, and 106 may be configured to execute one or more applications.
In an embodiment of the present disclosure, the server 120 may run one or more services or software applications that enable execution of the image display method.
In some embodiments, server 120 may also provide other services or software applications that may include non-virtual environments and virtual environments. In some embodiments, these services may be provided as web-based services or cloud services, for example, provided to users of client devices 101, 102, 103, 104, 105, and/or 106 under a software as a service (SaaS) model.
In the configuration shown in fig. 1, server 120 may include one or more components that implement the functions performed by server 120. These components may include software components, hardware components, or a combination thereof that are executable by one or more processors. A user operating client devices 101, 102, 103, 104, 105, and/or 106 may in turn utilize one or more client applications to interact with server 120 to utilize the services provided by these components. It should be appreciated that a variety of different system configurations are possible, which may differ from system 100. Accordingly, FIG. 1 is one example of a system for implementing the various methods described herein and is not intended to be limiting.
The user may navigate using client devices 101, 102, 103, 104, 105, and/or 106. The client device may provide an interface that enables a user of the client device to interact with the client device. The client device may also output information to the user via the interface. Although fig. 1 depicts only six client devices, those skilled in the art will appreciate that the present disclosure may support any number of client devices.
Client devices 101, 102, 103, 104, 105, and/or 106 may include various types of computer devices, such as portable handheld devices, general purpose computers (such as personal computers and laptop computers), workstation computers, wearable devices, smart screen devices, self-service terminal devices, service robots, gaming systems, thin clients, various messaging devices, sensors or other sensing devices, and the like. These computer devices may run various types and versions of software applications and operating systems, such as MICROSOFT Windows, APPLE iOS, UNIX-like operating systems, linux, or Linux-like operating systems (e.g., GOOGLE Chrome OS); or include various mobile operating systems such as MICROSOFT Windows Mobile OS, iOS, windows Phone, android. Portable handheld devices may include cellular telephones, smart phones, tablet computers, personal Digital Assistants (PDAs), and the like. Wearable devices may include head mounted displays (such as smart glasses) and other devices. The gaming system may include various handheld gaming devices, internet-enabled gaming devices, and the like. The client device is capable of executing a variety of different applications, such as various Internet-related applications, communication applications (e.g., email applications), short Message Service (SMS) applications, and may use a variety of communication protocols.
Network 110 may be any type of network known to those skilled in the art that may support data communications using any of a number of available protocols, including but not limited to TCP/IP, SNA, IPX, etc. By way of example only, the one or more networks 110 may be a Local Area Network (LAN), an ethernet-based network, a token ring, a Wide Area Network (WAN), the internet, a virtual network, a Virtual Private Network (VPN), an intranet, an extranet, a Public Switched Telephone Network (PSTN), an infrared network, a wireless network (e.g., bluetooth, wi-Fi), and/or any combination of these and/or other networks.
The server 120 may include one or more general purpose computers, special purpose server computers (e.g., PC (personal computer) servers, UNIX servers, mid-end servers), blade servers, mainframe computers, server clusters, or any other suitable arrangement and/or combination. The server 120 may include one or more virtual machines running a virtual operating system, or other computing architecture that involves virtualization (e.g., one or more flexible pools of logical storage devices that may be virtualized to maintain virtual storage devices of the server). In various embodiments, server 120 may run one or more services or software applications that provide the functionality described below.
The computing units in server 120 may run one or more operating systems including any of the operating systems described above as well as any commercially available server operating systems. Server 120 may also run any of a variety of additional server applications and/or middle tier applications, including HTTP servers, FTP servers, CGI servers, JAVA servers, database servers, etc.
In some implementations, server 120 may include one or more applications to analyze and consolidate data feeds and/or event updates received from users of client devices 101, 102, 103, 104, 105, and 106. Server 120 may also include one or more applications to display data feeds and/or real-time events via one or more display devices of client devices 101, 102, 103, 104, 105, and 106.
In some implementations, the server 120 may be a server of a distributed system or a server that incorporates a blockchain. The server 120 may also be a cloud server, or an intelligent cloud computing server or intelligent cloud host with artificial intelligence technology. The cloud server is a host product in a cloud computing service system, so as to solve the defects of large management difficulty and weak service expansibility in the traditional physical host and virtual private server (VPS, virtual Private Server) service.
The system 100 may also include one or more databases 130. In some embodiments, these databases may be used to store data and other information. For example, one or more of the databases 130 may be used to store information such as music files. Database 130 may reside in various locations. For example, the database used by the server 120 may be local to the server 120, or may be remote from the server 120 and may communicate with the server 120 via a network-based or dedicated connection. Database 130 may be of different types. In some embodiments, the database used by server 120 may be, for example, a relational database. One or more of these databases may store, update, and retrieve the databases and data from the databases in response to the commands.
In some embodiments, one or more of databases 130 may also be used by applications to store application data. The databases used by the application may be different types of databases, such as key value stores, object stores, or conventional stores supported by the file system.
The system 100 of fig. 1 may be configured and operated in various ways to enable application of the various methods and apparatus described in accordance with the present disclosure.
For purposes of embodiments of the present disclosure, in the example of FIG. 1, client devices 101, 102, 103, 104, 105, and 106 may include a client application for browsing web pages, which may be, for example, a browser, or an application containing WebView web page view controls, or the like.
The server 120 may be a server corresponding to a client application for web browsing in the client device, accordingly. The server 120 may include a service program that provides web browsing services to users based on web resources (including HTML files, images, audio, video, etc.) stored in the database 130. For example, a user may initiate a search request in a client application to search for a web page containing particular information (typically, the user may enter a search term in the client application to search for a web page containing the search term). Accordingly, the server 120 performs a web search in response to the search request, obtains web pages containing specific information that the user desires to search, and returns the web pages to the client application for presentation to the user.
The client devices 101, 102, 103, 104, 105 and 106 can execute the image display method of the embodiment of the disclosure, and enhance the image in the webpage currently browsed by the user based on the user requirement, so that the automatic optimization of the image quality is realized, and the user can acquire the interested information from the image conveniently.
Fig. 2 shows a flowchart of an image display method 200 according to an embodiment of the present disclosure. As described above, the method 200 is generally performed at a client device (e.g., client devices 101, 102, 103, 104, 105, and 106 shown in fig. 2). That is, the subject of execution of the various steps of method 200 may be client devices 101, 102, 103, 104, 105, and 106 shown in fig. 2.
As shown in fig. 2, the method 200 includes:
step 210, in response to the user's amplifying operation on the first image, determining whether the current first scaling of the first image is greater than or equal to a scaling threshold;
step 220, in response to determining that the first scaling is greater than or equal to the scaling threshold, performing enhancement processing on the first image to obtain a second image; and
step 230, displaying the second image.
According to the embodiment of the present disclosure, when the first image has been enlarged to a certain scale for display (i.e., the first scaling reaches the scaling threshold), if the user has an enlarging operation, it indicates that the quality of the current first image is poor, the user cannot obtain the information of interest from the current first image, and there is a need to improve the image quality. In this case, by performing enhancement processing on the first image, a high-quality second image can be obtained, and automatic optimization of image quality is achieved. The high quality second image is presented to the user so that the user can observe the image details and obtain the information of interest therefrom.
The various steps of method 200 are described in detail below.
According to some embodiments, the first image is displayed through an image display window. The image display window may be generated, for example, in response to a user interaction with a first image in the web page. For example, when a user browses a web page containing images in a client application, the user can select an image of interest (i.e., a first image) thereof by an interactive operation of clicking, long pressing, mouse-over, or the like, and in response to the interactive operation of the user, an image display window for displaying the first image will pop up in the web page. The user can zoom in and out the first image by performing an interactive operation in the image display window. Further, the user may drag the first image in the image display window, so that when the first image is enlarged to a certain extent (the image display window cannot completely display the whole first image, but only a local area of the first image is displayed), the local area of interest is selected for observation.
The scale is used to indicate the degree of reduction and enlargement of an image with respect to an original image at the time of display. For example, the image has a scaling ratio of 100%, and the image is displayed according to the size of the original image; the image is scaled to 80%, which means that the image is reduced to 80% of the original image and displayed; the scaling of the image is 150%, which means that the original image is enlarged by 1.5 times and displayed.
The scaling of the first image will change with the user's zoom-in and zoom-out operations in the image display window. It will be appreciated that the scaling only affects the display size of the first image on the screen, and does not change the size of the first image itself. In embodiments of the present disclosure, the size of an image refers to the resolution (resolution) of the image, which is generally expressed as the product of the number of pixels that the image includes horizontally and the number of pixels that the image includes vertically, e.g., 720×1280.
It should be noted that, for convenience of description and understanding, in the embodiment of the present disclosure, the scale of the first image is represented by the first scale, and the scale of the second image is represented by the second scale. The first image refers to an image before enhancement processing, and the second image refers to an image after enhancement processing.
Fig. 3A-3D illustrate schematic diagrams of adjusting a first scale in an image display window of a first image according to embodiments of the present disclosure.
As shown in fig. 3A, a web page 310 is currently displayed on the screen of the client device, the web page 310 including an image 320.
The user clicks on the image 320 in fig. 3A (accordingly, the image 320 is the first image), and the image display window 330 shown in fig. 3B is popped up. The image display window 330 is displayed full screen. Web page 310 is overlaid and therefore not visible in fig. 3B. The image 320 in fig. 3B is the same image as the image 320 in fig. 3A, except that the display sizes of the two are different, i.e., the scale (i.e., the first scale) is different. The first scale in fig. 3B is greater than the first scale in fig. 3A.
As shown in fig. 3B, the user may perform a zoom-in operation (e.g., a gesture operation of double-clicking, double-finger expansion, or a scroll operation of a mouse wheel, etc.) on the image 320 to further zoom in the display size of the image 320, i.e., to increase the first zoom scale. The enlarged image 320 is shown in fig. 3C.
In fig. 3C, the display size of the image 320 is larger than the size of the image display window 330, and the image display window 330 cannot display all of the image 320, but only a partial region 322 (as shown by the shaded portion) of the image 320. The user may drag the image 320 in the image display window 330 to move its local area of interest 324 into the image display window 330, as shown in fig. 3D.
In response to the user's zoom-in operation of the first image, it is determined whether the current first zoom scale of the first image is greater than or equal to a scale threshold in step 210 according to an embodiment of the present disclosure. The ratio threshold may be a predetermined value, such as 100%, 200%, etc.
If the current first scaling of the first image is smaller than the scaling threshold, the current user's zoom-in operation may be due to the first scaling being too small, i.e. the user cannot see the details of the image due to the first scaling being too small, instead of the quality problem of the image. Therefore, in this case, the first scaling of the first image may be increased in accordance with the user's enlargement operation, thereby further enlarging and displaying the first image.
If the current first scaling of the first image is greater than or equal to the scaling threshold, that is, if the user has already scaled the first image to a certain scale (i.e., the scaling threshold), there is still an image scaling requirement, which indicates that the quality of the first image is poor, and even if the first image is displayed in a scaled manner, the user cannot obtain the information of interest from the first image. In this case, therefore, enhancement processing is required for the first image to optimize its quality.
According to some embodiments, in step 220, performing enhancement processing on the first image includes: determining a target area to be processed of the first image; and performing enhancement processing on the target area.
According to some embodiments, the target area may be the whole of the first image, i.e. the first image is subjected to enhancement processing.
According to other embodiments, the target area may also be a local area of interest to the user in the first image, i.e. only the local area of the first image is subjected to the enhancement process. Compared with the whole enhancement of the first image, the enhancement processing of the local area in the first image reduces the calculated amount and can improve the image processing efficiency and the display efficiency.
In the case that the target area is a local area in the first image, in particular, according to some embodiments, the target area may be an area of the first image located in the image display window. Therefore, the region in the image display window can be directly used as the target region, extra calculation is not needed, and the image processing efficiency is further improved.
According to other embodiments, the target area may also be determined according to the following steps: acquiring a first area, wherein the first area is an area of a first image, which is positioned in an image display window; performing object recognition on the first image to determine the position areas corresponding to at least one object included in the first image; the position area overlapping with the first area is set as the target area. Therefore, the area (first area) displayed in the current image display window can be expanded, objects possibly interested by the user are identified, the area where the objects are located is used as a target area, and enhancement processing is carried out, so that the user can acquire interested information conveniently.
According to some embodiments, the enhancement processing of the target region includes: extracting characteristics of a target area; judging whether the target area is blurred or not based on the characteristics; and performing enhancement processing on the target area based on the judgment result.
According to some embodiments, a neural network model (e.g., a convolutional neural network) may be employed to determine whether the target region is blurred. Specifically, the target region may be input into a neural network model, the neural network model extracts characteristics of the target region, and a determination result of whether the target region is blurred is output.
According to other embodiments, it may be determined whether the target area is blurred by edge detection. For example, edge detection may be performed on the target area (e.g., the second derivative is performed on the target area to obtain edges therein), so as to obtain the total number of edge points of the target area, and if the total number of edge points is smaller than the threshold value, the target area is determined to be blurred. For another example, edge points in the target region may be acquired by edge detection, and then the variance of pixel values of the edge points is calculated, and if the variance is smaller than a threshold value, the target region is determined to be blurred.
Based on different fuzzy judgment results, different targeted enhancement processes can be performed on the target area.
Image blur is typically due to the fact that the image contains too little information (lower resolution). In this case, by improving the image resolution, the image quality can be optimized. Thus, according to some embodiments, in response to determining that the target region is blurred, the target region is subjected to a first enhancement process to cause the resolution of the second image to be greater than the resolution of the first image. In an embodiment of the present disclosure, the first enhancement process is used to increase the Resolution of the first image, i.e., the first enhancement process is a Super-Resolution (Super-Resolution) process. Specifically, the super-resolution processing may be implemented by an algorithm such as a deep learning model, interpolation, or the like.
The image itself is clear (i.e., not blurred), but it is difficult for the user to obtain the information of interest from it, typically due to poor visual performance of the image. In this case, the resolution can be kept unchanged, and the quality thereof can be optimized by enhancing the visual effect of the image (e.g., adjusting brightness, contrast, etc.). Thus, according to some embodiments, in response to determining that the target region is not blurred, the target region is subjected to a second enhancement process, and the resolution of the second image is equal to the resolution of the first image. In the embodiment of the present disclosure, the second enhancement processing is enhancement processing that keeps the resolution unchanged, such as adjusting brightness, adjusting contrast, removing noise points, and the like.
According to some embodiments, specifically, performing the first enhancement processing on the target area includes: performing super-resolution processing on the target area according to a preset magnification step length to obtain an intermediate image; judging whether the intermediate image is blurred; and in response to determining that the intermediate image is blurred and that the magnification of the intermediate image is less than the magnification threshold, continuing to perform super-resolution processing; or in response to the intermediate image not blurring or the magnification of the intermediate image reaching a magnification threshold, the intermediate image is taken as the second image. By circularly executing the above-mentioned steps of super resolution processing-blur detection, the resolution of the image can be automatically amplified to a suitable multiple, and the self-adaptive improvement of the resolution of the image can be realized.
The magnification step size is used to represent the interval of magnification of two super-resolution processes. For example, if the magnification step is 1, the interval (difference) representing the magnification of two times of super-resolution processing is 1, that is, the target area is sequentially super-resolution processed in the order of 2 times, 3 times, 4 times, …. For another example, if the magnification step is 2, the interval (difference) indicating the magnification of two times of super-resolution processing is 2, that is, the target area is sequentially super-resolution processed in the order of 2 times, 4 times, 8 times, …. And when the intermediate image obtained by a certain super-resolution processing is sufficiently clear (i.e. not blurred), or the magnification reaches a preset maximum magnification (i.e. a magnification threshold), the cycle is ended, and the current intermediate image is used as a second image after the enhancement processing.
It will be appreciated that resolution is typically expressed by the product of the number of pixels in both the lateral and longitudinal directions. Note that, in the embodiment of the present disclosure, the magnification of the resolution refers to the magnification of the number of pixels in a single direction. For example, the magnification of the resolution of the image p (resolution c×d) is 2 times, which means that the number of pixels in the horizontal direction and the number of pixels in the vertical direction of the image a are each enlarged to 2 times that of the original image, that is, the resolution of the enlarged image p' is 2c×2d.
After the high quality second image is obtained through step 220, step 230 is performed to display the second image to the user.
It will be appreciated that if the second image is obtained by super-resolution processing of the first image, then the second image is larger in size than the first image. After the super resolution processing is performed on the first image, the image content (first area) displayed in the previous image display window will overflow the window, the content displayed in the window having been not the content (content of interest) previously browsed by the user.
In order for the content in the image display window to remain interesting to the user after reloading the second image, step 230 further comprises, according to some embodiments: determining a second scaling of the second image based on the first scaling and a magnification, the magnification being a ratio of a resolution of the second image to a resolution of the first image; determining a second area in the second image, wherein the second area corresponds to the first area, and the first area is an area of the first image, which is positioned in the image display window; scaling the second area based on the second scaling scale to obtain an area to be displayed; and displaying the region to be displayed in the image display window, wherein the center point of the region to be displayed coincides with the center point of the image display window.
By redetermining the scaling (second scaling) based on the magnification from the first image to the second image and moving the scaled image (area to be displayed) to the center of the window for display, it can be ensured that the content displayed in the window is still the content of interest to the user, thereby improving the user experience.
Fig. 4A-4B show schematic diagrams of a first image, a second image, according to an embodiment of the present disclosure.
As shown in fig. 4A, the resolution of the first image 420 is w×h, and the point O is the origin of the coordinate system of the first image 420 (i.e., the pixel in the upper left corner of the first image 420). The image display window 410 in the client device is currently displaying a local area 422 (i.e., a first area) of the first image 420. Local region 422 includes pixels in first image 420 whose abscissas lie within the interval a through (a+w) and whose abscissas lie within the interval b through (b+h).
In fig. 4A, the first scale of the first image 420 is 100%. The preset ratio threshold is also 100%. The user performs a zoom-in operation on the first image 420 in the image display window 410, and since the current first scaling reaches the scaling threshold (100%), the first image 420 is subjected to enhancement processing. Specifically, by performing blur detection on the first image 420, it is determined that the first image 420 is blurred, and then subjected to first enhancement processing (i.e., super resolution processing), resulting in a second image 430 as shown in fig. 4B.
As shown in fig. 4B, the resolution of the second image is 2w×2h, i.e., the magnification of the resolution is 2 times. The first region 422 in fig. 4A is enlarged to the second region 432 in fig. 4B through super resolution processing. In order to enable the image display window 410 in fig. 4B to still display the same image content as that in fig. 4A, it is necessary to reduce the display size of the second image 430, to make the display size of the reduced second area 432 the same as that of the image display window 410, and to move the reduced second area 432 into the image display window 410 so that the center points of the two coincide. The second scale of the second area 432 is the quotient of the first scale and the magnification, i.e., 100%/2=50%, i.e., the second image 430 is reduced to 50% of the original image for display.
According to some embodiments, the method 200 further comprises: after the second image is displayed, the second image is scaled and displayed in response to a user's scaling operation of the second image.
After the second image is displayed in the image display window, a zoom operation of the second image by the user in the image display window may be received. And scaling and displaying the second image correspondingly based on the scaling operation of the user. Thus, the user can zoom the enhanced high quality second image to further obtain information of interest thereto.
According to an embodiment of the present disclosure, there is also provided an image display apparatus. Fig. 5 shows a block diagram of an image display apparatus 500 according to an embodiment of the present disclosure. As shown in fig. 5, the apparatus 500 includes:
a judging module 510 configured to judge whether a current first scaling of the first image is greater than or equal to a scaling threshold in response to a user's enlarging operation of the first image;
an enhancement module 520 configured to enhance the first image to obtain a second image in response to determining that the first scale is greater than or equal to the scale threshold; and
a display module 530 configured to display the second image.
According to the embodiment of the present disclosure, when the first image has been enlarged to a certain scale for display (i.e., the first scaling reaches the scaling threshold), if the user has an enlarging operation, it indicates that the quality of the current first image is poor, the user cannot obtain the information of interest from the current first image, and there is a need to improve the image quality. In this case, by performing enhancement processing on the first image, a high-quality second image can be obtained, and automatic optimization of image quality is achieved. The high quality second image is presented to the user so that the user can observe the image details and obtain the information of interest therefrom.
According to some embodiments, the enhancement module 520 includes: a determination unit configured to determine a target area to be processed of the first image; and an enhancement unit configured to perform enhancement processing on the target area.
According to some embodiments, the first image is displayed through an image display window, and the target area is an area of the first image located in the image display window.
According to some embodiments, the first image is displayed through an image display window, and the determining unit includes: an acquisition subunit configured to acquire a first area, wherein the first area is an area of the first image located in the image display window; an identification subunit configured to perform object identification on the first image, so as to determine a location area corresponding to at least one object included in the first image; and a determination subunit configured to take a location area overlapping the first area as the target area.
According to some embodiments, the enhancement unit comprises: an extraction subunit configured to extract features of the target region; a judging subunit configured to judge whether the target area is blurred based on the feature; and an enhancement subunit configured to perform enhancement processing on the target area based on a result of the determination.
According to some embodiments, the enhancer unit is further configured to: in response to determining that the target region is blurred, a first enhancement process is performed on the target region, wherein a resolution of the second image is greater than a resolution of the first image.
According to some embodiments, the enhancement subunit is further configured to perform a first enhancement process on the target area according to the following steps: performing super-resolution processing on the target area according to a preset magnification step length to obtain an intermediate image; judging whether the intermediate image is blurred; and in response to determining that the intermediate image is blurred and that the magnification of the intermediate image is less than a magnification threshold, continuing to perform the super-resolution processing; or in response to the intermediate image not blurring or the magnification of the intermediate image reaching the magnification threshold, regarding the intermediate image as the second image.
According to some embodiments, the enhancer unit is further configured to: in response to determining that the target region is not blurred, performing a second enhancement process on the target region, wherein a resolution of the second image is equal to a resolution of the first image.
According to some embodiments, the display module 530 includes: a first determination unit configured to determine a second scale of the second image based on the first scale and a magnification, the magnification being a ratio of a resolution of the second image to a resolution of the first image; a second determining unit configured to determine a second region in the second image, wherein the second region corresponds to a first region, which is a region of the first image located in an image display window; a scaling unit configured to scale the second area based on the second scaling ratio to obtain an area to be displayed; and a display unit configured to display the region to be displayed in the image display window, wherein a center point of the region to be displayed coincides with a center point of the image display window.
According to some embodiments, the apparatus 500 further comprises a scaling module configured to scale and display the second image in response to a scaling operation of the second image by a user.
It should be appreciated that the various modules or units of the apparatus 500 shown in fig. 5 may correspond to the various steps in the method 200 described with reference to fig. 2. Thus, the operations, features and advantages described above with respect to method 200 apply equally to apparatus 500 and the modules and units comprised thereof. For brevity, certain operations, features and advantages are not described in detail herein.
Although specific functions are discussed above with reference to specific modules, it should be noted that the functions of the various modules discussed herein may be divided into multiple modules and/or at least some of the functions of the multiple modules may be combined into a single module. For example, the determination module 510 and the enhancement module 520 described above may be combined into a single module in some embodiments.
It should also be appreciated that various techniques may be described herein in the general context of software hardware elements or program modules. The various modules described above with respect to fig. 5 may be implemented in hardware or in hardware in combination with software and/or firmware. For example, the modules may be implemented as computer program code/instructions configured to be executed in one or more processors and stored in a non-transitory computer-readable storage medium. Alternatively, these modules may be implemented as hardware logic/circuitry. For example, in some embodiments, one or more of the modules 510-530 may be implemented together in a System on Chip (SoC). The SoC may include an integrated circuit chip including one or more components of a processor (e.g., a central processing unit (Central Processing Unit, CPU), microcontroller, microprocessor, digital signal processor (Digital Signal Processor, DSP), etc.), memory, one or more communication interfaces, and/or other circuitry, and may optionally execute received program code and/or include embedded firmware to perform functions.
There is also provided, in accordance with an embodiment of the present disclosure, an electronic device including: at least one processor; and a memory communicatively coupled to the at least one processor, the memory storing instructions executable by the at least one processor to enable the at least one processor to perform the image display method 200 of an embodiment of the present disclosure.
According to an embodiment of the present disclosure, there is provided a non-transitory computer-readable storage medium storing computer instructions for causing a computer to perform the image display method 200 of the embodiment of the present disclosure.
According to an embodiment of the present disclosure, there is provided a computer program product comprising a computer program which, when executed by a processor, implements the image display method 200 of the embodiments of the present disclosure.
Referring to fig. 6, a block diagram of an electronic device 600 that may be a server or a client of the present disclosure, which is an example of a hardware device that may be applied to aspects of the present disclosure, will now be described. Electronic devices are intended to represent various forms of digital electronic computer devices, such as laptops, desktops, workstations, personal digital assistants, servers, blade servers, mainframes, and other suitable computers. The electronic device may also represent various forms of mobile devices, such as personal digital processing, cellular telephones, smartphones, wearable devices, and other similar computing devices. The components shown herein, their connections and relationships, and their functions, are meant to be exemplary only, and are not meant to limit implementations of the disclosure described and/or claimed herein.
As shown in fig. 6, the electronic device 600 includes a computing unit 601 that can perform various appropriate actions and processes according to a computer program stored in a Read Only Memory (ROM) 602 or a computer program loaded from a storage unit 608 into a Random Access Memory (RAM) 603. In the RAM603, various programs and data required for the operation of the device 600 may also be stored. The computing unit 601, ROM602, and RAM603 are connected to each other by a bus 604. An input/output (I/O) interface 605 is also connected to bus 604.
A number of components in the electronic device 600 are connected to the I/O interface 605, including: an input unit 606,An output unit 607, a storage unit 608, and a communication unit 609. The input unit 606 may be any type of device capable of inputting information to the device 600, the input unit 606 may receive input numeric or character information and generate key signal inputs related to user settings and/or function control of the electronic device, and may include, but is not limited to, a mouse, a keyboard, a touch screen, a trackpad, a trackball, a joystick, a microphone, and/or a remote control. The output unit 607 may be any type of device capable of presenting information and may include, but is not limited to, a display, speakers, video/audio output terminals, vibrators, and/or printers. Storage unit 608 may include, but is not limited to, magnetic disks, optical disks. The communication unit 609 allows the device 600 to exchange information/data with other devices via a computer network, such as the internet, and/or various telecommunications networks, and may include, but is not limited to, a modem, a network card, an infrared communication device, a wireless communication transceiver, and/or a chipset, such as bluetooth TM Devices, 802.11 devices, wi-Fi devices, wiMAX devices, cellular communication devices, and/or the like.
The computing unit 601 may be a variety of general and/or special purpose processing components having processing and computing capabilities. Some examples of computing unit 601 include, but are not limited to, a Central Processing Unit (CPU), a Graphics Processing Unit (GPU), various specialized Artificial Intelligence (AI) computing chips, various computing units running machine learning model algorithms, a Digital Signal Processor (DSP), and any suitable processor, controller, microcontroller, etc. The computing unit 601 performs the various methods and processes described above, such as method 200. For example, in some embodiments, the method 200 may be implemented as a computer software program tangibly embodied on a machine-readable medium, such as the storage unit 608. In some embodiments, part or all of the computer program may be loaded and/or installed onto the device 600 via the ROM 602 and/or the communication unit 609. One or more of the steps of the method 200 described above may be performed when a computer program is loaded into RAM 603 and executed by the computing unit 601. Alternatively, in other embodiments, computing unit 601 may be configured to perform method 200 by any other suitable means (e.g., by means of firmware).
Various implementations of the systems and techniques described here above may be implemented in digital electronic circuitry, integrated circuit systems, field Programmable Gate Arrays (FPGAs), application Specific Integrated Circuits (ASICs), application Specific Standard Products (ASSPs), systems On Chip (SOCs), complex Programmable Logic Devices (CPLDs), computer hardware, firmware, software, and/or combinations thereof. These various embodiments may include: implemented in one or more computer programs, the one or more computer programs may be executed and/or interpreted on a programmable system including at least one programmable processor, which may be a special purpose or general-purpose programmable processor, that may receive data and instructions from, and transmit data and instructions to, a storage system, at least one input device, and at least one output device.
Program code for carrying out methods of the present disclosure may be written in any combination of one or more programming languages. These program code may be provided to a processor or controller of a general purpose computer, special purpose computer, or other programmable data processing apparatus such that the program code, when executed by the processor or controller, causes the functions/operations specified in the flowchart and/or block diagram to be implemented. The program code may execute entirely on the machine, partly on the machine, as a stand-alone software package, partly on the machine and partly on a remote machine or entirely on the remote machine or server.
In the context of this disclosure, a machine-readable medium may be a tangible medium that can contain, or store a program for use by or in connection with an instruction execution system, apparatus, or device. The machine-readable medium may be a machine-readable signal medium or a machine-readable storage medium. The machine-readable medium may include, but is not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any suitable combination of the foregoing. More specific examples of a machine-readable storage medium would include an electrical connection based on one or more wires, a portable computer diskette, a hard disk, a Random Access Memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or flash memory), an optical fiber, a portable compact disc read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing.
To provide for interaction with a user, the systems and techniques described here can be implemented on a computer having: a display device (e.g., a CRT (cathode ray tube) or LCD (liquid crystal display) monitor) for displaying information to a user; and a keyboard and pointing device (e.g., a mouse or trackball) by which a user can provide input to the computer. Other kinds of devices may also be used to provide for interaction with a user; for example, feedback provided to the user may be any form of sensory feedback (e.g., visual feedback, auditory feedback, or tactile feedback); and input from the user may be received in any form, including acoustic input, speech input, or tactile input.
The systems and techniques described here can be implemented in a computing system that includes a background component (e.g., as a data server), or that includes a middleware component (e.g., an application server), or that includes a front-end component (e.g., a user computer having a graphical user interface or a web browser through which a user can interact with an implementation of the systems and techniques described here), or any combination of such background, middleware, or front-end components. The components of the system can be interconnected by any form or medium of digital data communication (e.g., a communication network). Examples of communication networks include: local Area Networks (LANs), wide Area Networks (WANs), and the internet.
The computer system may include a client and a server. The client and server are typically remote from each other and typically interact through a communication network. The relationship of client and server arises by virtue of computer programs running on the respective computers and having a client-server relationship to each other. The server may be a cloud server, a server of a distributed system, or a server incorporating a blockchain.
It should be appreciated that various forms of the flows shown above may be used to reorder, add, or delete steps. For example, the steps recited in the present disclosure may be performed in parallel, sequentially or in a different order, provided that the desired results of the disclosed aspects are achieved, and are not limited herein.
Although embodiments or examples of the present disclosure have been described with reference to the accompanying drawings, it is to be understood that the foregoing methods, systems, and apparatus are merely exemplary embodiments or examples, and that the scope of the present invention is not limited by these embodiments or examples but only by the claims following the grant and their equivalents. Various elements of the embodiments or examples may be omitted or replaced with equivalent elements thereof. Furthermore, the steps may be performed in a different order than described in the present disclosure. Further, various elements of the embodiments or examples may be combined in various ways. It is important that as technology evolves, many of the elements described herein may be replaced by equivalent elements that appear after the disclosure.

Claims (11)

1. An image display method, comprising:
responding to the amplifying operation of a user on a first image, judging whether the current first scaling of the first image is larger than or equal to a scaling threshold value;
In response to determining that the first scale is greater than or equal to the scale threshold, performing enhancement processing on the first image to obtain a second image, wherein the performing enhancement processing on the first image includes:
determining a target area to be processed of the first image;
extracting characteristics of the target area;
judging whether the target area is blurred or not based on the characteristics;
in response to determining that the target region is blurred, performing a first enhancement process on the target region, wherein the resolution of the second image is greater than the resolution of the first image, the performing the first enhancement process on the target region comprising:
performing super-resolution processing on the target area according to a preset magnification step length to obtain an intermediate image;
judging whether the intermediate image is blurred;
in response to determining that the intermediate image is blurred and that the magnification of the intermediate image is less than a magnification threshold, continuing to perform the step of super-resolution processing; or alternatively
In response to the intermediate image not blurring or the magnification of the intermediate image reaching the magnification threshold, taking the intermediate image as the second image;
Determining a second scale of the second image based on the first scale and a magnification of the second image, the magnification of the second image being a ratio of a resolution of the second image to a resolution of the first image;
determining a second area in the second image, wherein the second area corresponds to a first area, and the first area is an area of the first image, which is positioned in an image display window;
scaling the second area based on the second scaling to obtain an area to be displayed; and
and displaying the region to be displayed in the image display window, wherein the center point of the region to be displayed coincides with the center point of the image display window.
2. The method of claim 1, wherein the first image is displayed through the image display window and the target area is an area of the first image that is located in the image display window.
3. The method of claim 1, wherein the first image is displayed through the image display window, and wherein determining a target area of the first image to be processed comprises:
Acquiring a first area, wherein the first area is an area of the first image, which is positioned in the image display window;
performing object recognition on the first image to determine a position area corresponding to at least one object included in the first image;
and taking a position area overlapped with the first area as the target area.
4. The method of claim 1, further comprising:
in response to determining that the target region is not blurred, performing a second enhancement process on the target region, wherein a resolution of the second image is equal to a resolution of the first image.
5. The method of any of claims 1-4, further comprising:
and zooming and displaying the second image in response to a zooming operation of the user on the second image.
6. An image display device comprising:
a judging module configured to judge whether a current first scaling of a first image is greater than or equal to a scaling threshold value in response to an amplifying operation of the first image by a user;
an enhancement module configured to enhance the first image to obtain a second image in response to determining that the first scale is greater than or equal to the scale threshold, wherein the enhancement module comprises:
A determination unit configured to determine a target area to be processed of the first image;
an enhancement unit configured to perform enhancement processing on the target area, including:
an extraction subunit configured to extract features of the target region;
a judging subunit configured to judge whether the target area is blurred based on the feature;
an enhancement subunit configured to perform, in response to determining that the target region is blurred, a first enhancement process on the target region, wherein a resolution of the second image is greater than a resolution of the first image, the enhancement process on the target region comprising:
performing super-resolution processing on the target area according to a preset magnification step length to obtain an intermediate image;
judging whether the intermediate image is blurred;
in response to determining that the intermediate image is blurred and that the magnification of the intermediate image is less than a magnification threshold, continuing to perform the step of super-resolution processing; or alternatively
In response to the intermediate image not blurring or the magnification of the intermediate image reaching the magnification threshold, taking the intermediate image as the second image;
A display module configured to display the second image, the displaying the second image comprising:
determining a second scale of the second image based on the first scale and a magnification of the second image, the magnification of the second image being a ratio of a resolution of the second image to a resolution of the first image;
determining a second area in the second image, wherein the second area corresponds to a first area, and the first area is an area of the first image, which is positioned in an image display window;
scaling the second area based on the second scaling to obtain an area to be displayed; and
and displaying the region to be displayed in the image display window, wherein the center point of the region to be displayed coincides with the center point of the image display window.
7. The apparatus of claim 6, wherein the first image is displayed through the image display window and the target area is an area of the first image that is located in the image display window.
8. The apparatus according to claim 6, wherein the first image is displayed through the image display window, and wherein the determining unit includes:
An acquisition subunit configured to acquire a first area, wherein the first area is an area of the first image located in the image display window;
an identification subunit configured to perform object identification on the first image, so as to determine a location area corresponding to at least one object included in the first image; and
and a determination subunit configured to take a location area overlapping with the first area as the target area.
9. The apparatus of claim 6, wherein the enhancer unit is further configured to:
in response to determining that the target region is not blurred, performing a second enhancement process on the target region, wherein a resolution of the second image is equal to a resolution of the first image.
10. An electronic device, comprising:
at least one processor; and
a memory communicatively coupled to the at least one processor; wherein the method comprises the steps of
The memory stores instructions executable by the at least one processor to enable the at least one processor to perform the method of any one of claims 1-5.
11. A non-transitory computer readable storage medium storing computer instructions for causing a computer to perform the method of any one of claims 1-5.
CN202210107799.9A 2022-01-28 2022-01-28 Image display method and device, electronic equipment and medium Active CN114510187B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202210107799.9A CN114510187B (en) 2022-01-28 2022-01-28 Image display method and device, electronic equipment and medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210107799.9A CN114510187B (en) 2022-01-28 2022-01-28 Image display method and device, electronic equipment and medium

Publications (2)

Publication Number Publication Date
CN114510187A CN114510187A (en) 2022-05-17
CN114510187B true CN114510187B (en) 2023-06-23

Family

ID=81552284

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210107799.9A Active CN114510187B (en) 2022-01-28 2022-01-28 Image display method and device, electronic equipment and medium

Country Status (1)

Country Link
CN (1) CN114510187B (en)

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110944163A (en) * 2019-11-21 2020-03-31 维沃移动通信有限公司 Image processing method and electronic equipment
CN110969580A (en) * 2019-12-10 2020-04-07 北京猎豹移动科技有限公司 Image display method and device
CN112738625A (en) * 2020-12-24 2021-04-30 广东九联科技股份有限公司 Video image enhancement method and device based on set top box
CN113256495A (en) * 2021-06-07 2021-08-13 北京百度网讯科技有限公司 Image processing method, apparatus, device, storage medium, and computer program product
KR20210108027A (en) * 2020-02-25 2021-09-02 삼성전자주식회사 Electronic apparatus and control method thereof

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110944163A (en) * 2019-11-21 2020-03-31 维沃移动通信有限公司 Image processing method and electronic equipment
CN110969580A (en) * 2019-12-10 2020-04-07 北京猎豹移动科技有限公司 Image display method and device
KR20210108027A (en) * 2020-02-25 2021-09-02 삼성전자주식회사 Electronic apparatus and control method thereof
CN112738625A (en) * 2020-12-24 2021-04-30 广东九联科技股份有限公司 Video image enhancement method and device based on set top box
CN113256495A (en) * 2021-06-07 2021-08-13 北京百度网讯科技有限公司 Image processing method, apparatus, device, storage medium, and computer program product

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
The Error-Amended Sharp Edge (EASE) Scheme for Image Zooming;Youngjoon Cha等;《 IEEE Transactions on Image Processing 》;全文 *
图像处理中的模糊算法及实现分析;倪迎花;;信息与电脑(理论版)(08);全文 *

Also Published As

Publication number Publication date
CN114510187A (en) 2022-05-17

Similar Documents

Publication Publication Date Title
US11321593B2 (en) Method and apparatus for detecting object, method and apparatus for training neural network, and electronic device
CN114511758A (en) Image recognition method and device, electronic device and medium
US8887059B2 (en) Method and apparatus of locally controlling display content of a remote system
US20220301108A1 (en) Image quality enhancing
KR20160003683A (en) Automatically manipulating visualized data based on interactivity
CN112967196A (en) Image restoration method and device, electronic device and medium
CN108008876B (en) Display method, device and equipment of floating window and storage medium
CN113657518B (en) Training method, target image detection method, device, electronic device, and medium
CN110727383A (en) Touch interaction method and device based on small program, electronic equipment and storage medium
CN114510187B (en) Image display method and device, electronic equipment and medium
CN109739403B (en) Method and apparatus for processing information
CN114510308B (en) Method, device, equipment and medium for storing application page by mobile terminal
CN116030185A (en) Three-dimensional hairline generating method and model training method
CN114092556A (en) Method, apparatus, electronic device, medium for determining human body posture
CN114327718A (en) Interface display method and device, equipment and medium
CN114071024A (en) Image shooting method, neural network training method, device, equipment and medium
CN113627534A (en) Method and device for identifying type of dynamic image and electronic equipment
CN115170536B (en) Image detection method, training method and device of model
CN114491345A (en) Webpage image processing method, webpage image loading method and device
CN114140852B (en) Image detection method and device
CN113378001B (en) Video playing progress adjusting method and device, electronic equipment and medium
CN115331077B (en) Training method of feature extraction model, target classification method, device and equipment
CN116205819B (en) Character image generation method, training method and device of deep learning model
CN113744414B (en) Image processing method, device, equipment and storage medium
CN113793290B (en) Parallax determining method, device, equipment and medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant