WO2022166624A1 - 一种屏幕显示方法及相关装置 - Google Patents

一种屏幕显示方法及相关装置 Download PDF

Info

Publication number
WO2022166624A1
WO2022166624A1 PCT/CN2022/073339 CN2022073339W WO2022166624A1 WO 2022166624 A1 WO2022166624 A1 WO 2022166624A1 CN 2022073339 W CN2022073339 W CN 2022073339W WO 2022166624 A1 WO2022166624 A1 WO 2022166624A1
Authority
WO
WIPO (PCT)
Prior art keywords
pixel
electronic device
sub
bitmap data
display
Prior art date
Application number
PCT/CN2022/073339
Other languages
English (en)
French (fr)
Inventor
陈健
张朋
周越海
李煜
王亮
Original Assignee
华为技术有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 华为技术有限公司 filed Critical 华为技术有限公司
Priority to EP22748903.6A priority Critical patent/EP4261675A4/en
Publication of WO2022166624A1 publication Critical patent/WO2022166624A1/zh

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/14Digital output to display device ; Cooperation and interconnection of the display device with other functional units
    • G06F3/147Digital output to display device ; Cooperation and interconnection of the display device with other functional units using display panels
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/14Digital output to display device ; Cooperation and interconnection of the display device with other functional units
    • G06F3/1454Digital output to display device ; Cooperation and interconnection of the display device with other functional units involving copying of the display data of a local workstation or window to a remote workstation or window so that an actual copy of the data is displayed simultaneously on two or more displays, e.g. teledisplay
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G3/00Control arrangements or circuits, of interest only in connection with visual indicators other than cathode-ray tubes
    • G09G3/20Control arrangements or circuits, of interest only in connection with visual indicators other than cathode-ray tubes for presentation of an assembly of a number of characters, e.g. a page, by composing the assembly by combination of individual elements arranged in a matrix no fixed position being assigned to or needed to be assigned to the individual characters or partial characters
    • G09G3/2007Display of intermediate tones
    • G09G3/2074Display of intermediate tones using sub-pixels
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G3/00Control arrangements or circuits, of interest only in connection with visual indicators other than cathode-ray tubes
    • G09G3/20Control arrangements or circuits, of interest only in connection with visual indicators other than cathode-ray tubes for presentation of an assembly of a number of characters, e.g. a page, by composing the assembly by combination of individual elements arranged in a matrix no fixed position being assigned to or needed to be assigned to the individual characters or partial characters
    • G09G3/22Control arrangements or circuits, of interest only in connection with visual indicators other than cathode-ray tubes for presentation of an assembly of a number of characters, e.g. a page, by composing the assembly by combination of individual elements arranged in a matrix no fixed position being assigned to or needed to be assigned to the individual characters or partial characters using controlled light sources
    • G09G3/30Control arrangements or circuits, of interest only in connection with visual indicators other than cathode-ray tubes for presentation of an assembly of a number of characters, e.g. a page, by composing the assembly by combination of individual elements arranged in a matrix no fixed position being assigned to or needed to be assigned to the individual characters or partial characters using controlled light sources using electroluminescent panels
    • G09G3/32Control arrangements or circuits, of interest only in connection with visual indicators other than cathode-ray tubes for presentation of an assembly of a number of characters, e.g. a page, by composing the assembly by combination of individual elements arranged in a matrix no fixed position being assigned to or needed to be assigned to the individual characters or partial characters using controlled light sources using electroluminescent panels semiconductive, e.g. using light-emitting diodes [LED]
    • G09G3/3208Control arrangements or circuits, of interest only in connection with visual indicators other than cathode-ray tubes for presentation of an assembly of a number of characters, e.g. a page, by composing the assembly by combination of individual elements arranged in a matrix no fixed position being assigned to or needed to be assigned to the individual characters or partial characters using controlled light sources using electroluminescent panels semiconductive, e.g. using light-emitting diodes [LED] organic, e.g. using organic light-emitting diodes [OLED]
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G5/00Control arrangements or circuits for visual indicators common to cathode-ray tube indicators and other visual indicators
    • G09G5/003Details of a display terminal, the details relating to the control arrangement of the display terminal and to the interfaces thereto
    • G09G5/005Adapting incoming signals to the display format of the display terminal
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G5/00Control arrangements or circuits for visual indicators common to cathode-ray tube indicators and other visual indicators
    • G09G5/22Control arrangements or circuits for visual indicators common to cathode-ray tube indicators and other visual indicators characterised by the display of characters or indicia using display control signals derived from coded signals representing the characters or indicia, e.g. with a character-code memory
    • G09G5/24Generation of individual character patterns
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G5/00Control arrangements or circuits for visual indicators common to cathode-ray tube indicators and other visual indicators
    • G09G5/22Control arrangements or circuits for visual indicators common to cathode-ray tube indicators and other visual indicators characterised by the display of characters or indicia using display control signals derived from coded signals representing the characters or indicia, e.g. with a character-code memory
    • G09G5/24Generation of individual character patterns
    • G09G5/28Generation of individual character patterns for enhancement of character form, e.g. smoothing
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G5/00Control arrangements or circuits for visual indicators common to cathode-ray tube indicators and other visual indicators
    • G09G5/36Control arrangements or circuits for visual indicators common to cathode-ray tube indicators and other visual indicators characterised by the display of a graphic pattern, e.g. using an all-points-addressable [APA] memory
    • G09G5/39Control of the bit-mapped memory
    • G09G5/393Arrangements for updating the contents of the bit-mapped memory
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G5/00Control arrangements or circuits for visual indicators common to cathode-ray tube indicators and other visual indicators
    • G09G5/36Control arrangements or circuits for visual indicators common to cathode-ray tube indicators and other visual indicators characterised by the display of a graphic pattern, e.g. using an all-points-addressable [APA] memory
    • G09G5/39Control of the bit-mapped memory
    • G09G5/395Arrangements specially adapted for transferring the contents of the bit-mapped memory to the screen
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G2300/00Aspects of the constitution of display devices
    • G09G2300/04Structural and physical details of display devices
    • G09G2300/0439Pixel structures
    • G09G2300/0443Pixel structures with several sub-pixels for the same colour in a pixel, not specifically used to display gradations
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G2300/00Aspects of the constitution of display devices
    • G09G2300/04Structural and physical details of display devices
    • G09G2300/0439Pixel structures
    • G09G2300/0452Details of colour pixel setup, e.g. pixel composed of a red, a blue and two green components
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G2340/00Aspects of display data processing
    • G09G2340/04Changes in size, position or resolution of an image
    • G09G2340/0457Improvement of perceived resolution by subpixel rendering
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G2340/00Aspects of display data processing
    • G09G2340/10Mixing of images, i.e. displayed pixel being the result of an operation, e.g. adding, on the corresponding input pixels
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G2360/00Aspects of the architecture of display systems
    • G09G2360/16Calculation or use of calculated indices related to luminance levels in display data
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G2360/00Aspects of the architecture of display systems
    • G09G2360/18Use of a frame buffer in a display terminal, inclusive of the display panel
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G2370/00Aspects of data communication
    • G09G2370/04Exchange of auxiliary data, i.e. other than image data, between monitor and graphics controller

Definitions

  • the present application relates to the field of electronic technology, and in particular, to a screen display method and related devices.
  • Sub-pixel rendering (SPR) technology is often used in electronic devices with organic light-emitting diode (organic light-emitting diode, OLED) display screens.
  • the sub-pixel rendering technology uses the difference in the recognition of different colors by the human eye, changing the conventional mode of defining a pixel by red, green and blue sub-pixels, and using a relatively small number of sub-pixels by sharing some color sub-pixels between different pixels. It can simulate the ability to achieve the same pixel resolution performance, thereby reducing the manufacturing process and manufacturing cost.
  • the sub-pixel rendering technology has two key elements: one is the arrangement of the sub-pixels, and the other is the sub-pixel rendering algorithm.
  • the arrangement of sub-pixels is determined by the OLED display panel, which defines the distribution rules of sub-pixels of different colors on the display panel; the sub-pixel rendering algorithm is determined by the display controller, which defines how the display device lights up sub-pixels of different colors according to the input data.
  • the sub-pixel arrangement and sub-pixel rendering algorithm jointly affect the display effect of the display device. A suitable combination will make the display screen clear, delicate, soft and accurate in color, while an inappropriate combination may result in graininess, color cast, color fringing, etc. happening.
  • the sub-pixel arrangement depends on the display panel device, and the sub-pixel rendering algorithm is often integrated in the display controller, both of which are determined by hardware, and have the characteristics of long update cycle, poor customizability, and inability to update through software.
  • the sub-pixel arrangement and sub-pixel rendering algorithm of the display screen cannot be changed by means of software update, and it is difficult to optimize the display effect. How to optimize the display effect of the display screen is a problem being studied by those skilled in the art.
  • the embodiments of the present application provide a screen display method and a related device, which can improve the screen display effect.
  • the present application provides a screen display method, which is applied to an electronic device, and the electronic device includes an OLED display panel and a display controller integrated with a sub-pixel rendering algorithm.
  • the bitmap data of the content the electronic device performs software filtering on the bitmap data, wherein the software filtering is determined based on the sub-pixel rendering algorithm, and is used to optimize the display edge of the content to be displayed on the OLED display panel;
  • Sub-pixel rendering is performed on the bitmap data after that to determine the RGB grayscale value of each pixel;
  • the display controller drives the OLED light-emitting points on the OLED display panel according to the RGB grayscale value to display the content to be displayed.
  • the content to be displayed may include text, characters, images, and the like.
  • the subpixel rendering algorithm is a hardware algorithm and difficult to update via software. After acquiring the bitmap data of the content to be displayed, the electronic device performs software filtering processing on the bitmap data before sending the bitmap data of the content to be displayed to the display screen; The sub-pixel rendering is performed by the sub-pixel rendering algorithm, and the RGB gray value of each pixel point (including the gray value of the red pixel, the gray value of the green pixel, and the gray value of the blue pixel) is determined.
  • the filtering parameters of the software filtering are based on As determined by the sub-pixel rendering algorithm, the software filtering and the sub-pixel rendering algorithm work together to optimize the display edge of the content to be displayed on the OLED display panel, so as to achieve the effect of changing the display effect. In this way, the difficulty that the sub-pixel rendering algorithm cannot be updated online is solved.
  • the electronic device performs software filtering on the bitmap data including: the bitmap data includes RGB three-channel bitmap data and single-channel grayscale bitmap data; when the bitmap data is a single-channel When grayscale bitmap data is used, the bitmap data is converted into RGB three-channel bitmap data; the electronic device performs single-pixel channel software filtering on the red pixel gray value and blue pixel gray value of the RGB three-channel bitmap data respectively. ; The electronic device keeps the green pixel gray value of the bitmap data unchanged.
  • Bitmap data includes RGB three-channel bitmap data, that is, gray value data including three RGB channels of each pixel (red pixel gray value, blue pixel gray value and green pixel gray value of each pixel point). degree value), in which the red pixel gray value of each pixel constitutes the red channel in the bitmap data, the blue pixel gray value of each pixel constitutes the blue channel in the bitmap data, and each pixel The green pixel gray value of a point constitutes the green channel in the bitmap data.
  • Single-pixel channel software filtering processing refers to performing software filtering processing on one channel in the RGB three-channel bitmap data.
  • the electronic device When the bitmap data is RGB three-channel bitmap data, the electronic device performs single-pixel channel software filtering processing on the red pixel gray value and blue pixel gray value in the bitmap data respectively, and determines the filtered red pixel gray value. and the blue pixel gray value; where the green pixel gray value remains unchanged.
  • the bitmap data also includes single-channel grayscale bitmap data, that is, the grayscale value data of one channel including each pixel point (that is, the red pixel grayscale value, blue pixel grayscale value and green pixel grayscale value of each pixel point). same gray value).
  • the electronic device first converts the single-channel grayscale bitmap data into RGB three-channel bitmap data, that is, if the size of the original bitmap (single-channel grayscale bitmap data) If it is [w, h], the electronic device allocates a memory of size [w, h, 3] to convert single-channel grayscale bitmap data into RGB three-channel bitmap data.
  • the electronic device performs single-pixel channel software filtering processing on the red pixel gray value and blue pixel gray value in the RGB three-channel bitmap data respectively, and determines the filtered red pixel gray value and blue pixel gray value. ; where the gray value of the green pixel remains unchanged.
  • the single-pixel channel software filtering processing includes: the single-pixel channel software filtering processing is perpendicular to the one-dimensional filtering processing direction in the sub-pixel rendering algorithm. That is, the filtering parameters in the single-pixel channel software filtering process are determined based on the sub-pixel rendering algorithm.
  • the filter in the sub-pixel rendering algorithm is a one-dimensional filter in the horizontal direction
  • the filtering parameters in the single-pixel channel software filtering process are determined. is a one-dimensional filter parameter in the vertical direction
  • the filter parameter in the single-pixel channel software filtering process is a one-dimensional filter in the horizontal direction. parameter.
  • the formula for one-dimensional filtering includes:
  • dst(x, y) is the pixel gray value of the output pixel point (x, y)
  • src(x, y) is the pixel gray value of the input pixel point (x, y).
  • the kernel is the convolution kernel.
  • the filter kernel of one-dimensional filtering in the horizontal direction is (0.5, 0.5)
  • the value of the kernel is (0.5, 0.5)
  • the filter kernel is a 1-by-2 matrix
  • x ⁇ The value of ' is 0 and 1, that is, kernel.cols is 2, and the value of y ⁇ ' is 0, that is, kernel.cols is 1.
  • anchor is the anchor point
  • anchor.x is the abscissa of the anchor point
  • anchor.y is the ordinate of the anchor point.
  • the anchor point of one-dimensional filtering in the horizontal direction can be (0,0).
  • Each value in the output matrix is equal to the sum of the multiplication results of the input matrix and the corresponding position elements of the convolution kernel.
  • the method before the electronic device performs software filtering on the bitmap data, the method further includes: the electronic device determines a sub-pixel rendering algorithm based on the model of the display screen.
  • the models of different display screens may be different.
  • the electronic device can determine the sub-pixel rendering algorithm integrated in the display screen through the model of the display screen, so that the electronic device can determine the filtering parameters of the software filtering or single-pixel channel software filtering processing of the electronic device through the filtering method adopted in the sub-pixel rendering algorithm. filter parameters in .
  • the electronic device performing software filtering on the bitmap data includes: when confirming that the electronic device is not in a scene of shooting, sharing or screencasting, the electronic device performs software filtering on the bitmap data. That is, the electronic device needs to distinguish different scenarios to determine whether to perform software filtering on the bitmap data.
  • the electronic device receives the operation of shooting, sharing and screencasting, the electronic device captures the image and sends it to other electronic devices through the application.
  • the image data sent by the electronic device to other electronic devices is a bitmap. If the electronic device performs a preset filtering operation on the collected image, the sent image data will be a bitmap filtered by software.
  • the electronic device performs software filtering on the bitmap data of the content to be displayed, and further includes: the electronic device saves the bitmap data of the content to be displayed after software filtering.
  • the electronic device saves the bitmap of the display content for the next call.
  • the method further includes: the electronic device calls the saved bitmap data of the content to be displayed after software filtering; the display controller performs sub-pixel rendering on the bitmap data after software filtering, and determines each The RGB grayscale value of the pixel point, and the OLED light-emitting point on the OLED display panel is driven according to the RGB grayscale value to display the content to be displayed. Since the electronic device saves the bitmap data of the content to be displayed after software filtering, the electronic device can directly call the bitmap data filtered by the software for the same content to be displayed next time, that is, the electronic device only needs to understand the newly appeared text. Or characters or images perform a rendering and software filtering operation, and then the software filtering results can be reused, which greatly reduces the amount of calculation.
  • obtaining the bitmap data of the content to be displayed by the electronic device includes: the electronic device determines the vector image of the content to be displayed; the electronic device rasterizes the vector image of the content to be displayed, and obtains Bitmap data of the content to be displayed.
  • the data type of the content to be displayed includes text, characters, and images.
  • the present application provides an electronic device, comprising: one or more processors, one or more memories, a display screen, and a display screen controller integrating a sub-pixel rendering algorithm; the one or more memories are associated with a or more processors, a display screen, and a display screen controller are coupled; the one or more memories are used to store computer program code, the computer program code comprising computer instructions; when the computer instructions are executed on the processor, cause the Electronic equipment performs:
  • Software filtering is performed on the bitmap data by the processor, wherein the software filtering is determined based on a sub-pixel rendering algorithm, and is used to optimize the display edge of the content to be displayed on the OLED display panel;
  • the sub-pixel rendering is performed on the bitmap data filtered by the software through the display controller, and the RGB gray value of each pixel is determined;
  • the OLED light-emitting points on the OLED display panel are driven according to the RGB grayscale values by the display controller to display the content to be displayed.
  • the content to be displayed may include text, characters, images, and the like.
  • the subpixel rendering algorithm is a hardware algorithm and difficult to update via software. After acquiring the bitmap data of the content to be displayed, the electronic device performs software filtering processing on the bitmap data before sending the bitmap data of the content to be displayed to the display screen; The sub-pixel rendering is performed by the sub-pixel rendering algorithm, in which the filtering parameters of the software filtering are determined based on the sub-pixel rendering algorithm.
  • the software filtering and the sub-pixel rendering algorithm work together to optimize the display edge of the content to be displayed on the OLED display panel. Show what the effect does. In this way, the difficulty that the sub-pixel rendering algorithm cannot be updated online is solved.
  • performing software filtering on the bitmap data by the processor includes: the bitmap data includes RGB three-channel bitmap data and single-channel grayscale bitmap data; when the bitmap data is a single channel When channel grayscale bitmap data, the processor converts the bitmap data into RGB three-channel bitmap data; Pixel channel software filtering processing; the green pixel gray value of the bitmap data remains unchanged.
  • the single-pixel channel software filtering processing includes: the single-pixel channel software filtering processing is perpendicular to the one-dimensional filtering processing direction in the sub-pixel rendering algorithm. That is, the filtering parameters in the single-pixel channel software filtering process are determined based on the sub-pixel rendering algorithm.
  • the filter in the sub-pixel rendering algorithm is a one-dimensional filter in the horizontal direction
  • the filtering parameters in the single-pixel channel software filtering process are determined. is a one-dimensional filter parameter in the vertical direction
  • the filter parameter in the single-pixel channel software filtering process is a one-dimensional filter in the horizontal direction. parameter.
  • the formula for one-dimensional filtering includes:
  • dst(x, y) is the pixel gray value of the output pixel point (x, y)
  • src(x, y) is the pixel gray value of the input pixel point (x, y).
  • the kernel is the convolution kernel.
  • the filter kernel of one-dimensional filtering in the horizontal direction is (0.5, 0.5)
  • the value of the kernel is (0.5, 0.5)
  • the filter kernel is a 1-by-2 matrix
  • x ⁇ The value of ' is 0 and 1, that is, kernel.cols is 2, and the value of y ⁇ ' is 0, that is, kernel.cols is 1.
  • anchor is the anchor point
  • anchor.x is the abscissa of the anchor point
  • anchor.y is the ordinate of the anchor point.
  • the anchor point of one-dimensional filtering in the horizontal direction can be (0,0).
  • Each value in the output matrix is equal to the sum of the multiplication results of the input matrix and the corresponding position elements of the convolution kernel.
  • the method before performing software filtering on the bitmap data by the processor, the method further includes: determining, by the processor, a sub-pixel rendering algorithm based on the model of the display screen.
  • the models of different display screens may be different.
  • the electronic device can determine the sub-pixel rendering algorithm integrated in the display screen through the model of the display screen, so that the electronic device can determine the filtering parameters of the software filtering or single-pixel channel software filtering processing of the electronic device through the filtering method adopted in the sub-pixel rendering algorithm. filter parameters in .
  • performing software filtering on the bitmap data by the processor includes: performing software filtering on the bitmap data by the processor when recognizing that the electronic device is not in the scene of shooting, sharing and screen projection. That is, the electronic device needs to distinguish different scenarios to determine whether to perform software filtering on the bitmap data.
  • the electronic device receives the operation of shooting, sharing and screencasting, the electronic device captures the image and sends it to other electronic devices through the application.
  • the image data sent by the electronic device to other electronic devices is a bitmap. If the electronic device performs a preset filtering operation on the collected image, the sent image data will be a bitmap filtered by software.
  • the processor performs software filtering on the bitmap data of the content to be displayed, and then further includes: storing the bitmap data of the content to be displayed after software filtering in a memory.
  • the electronic device saves a bitmap of the display content for the next recall.
  • the electronic device further executes: calling the saved bitmap data of the content to be displayed after software filtering through the processor; Perform sub-pixel rendering, determine the RGB grayscale value of each pixel, and drive the OLED light-emitting points on the OLED display panel according to the RGB grayscale value to display the content to be displayed. Since the electronic device saves the bitmap data of the content to be displayed after software filtering, the electronic device can directly call the bitmap data filtered by the software for the same content to be displayed next time, that is, the electronic device only needs to understand the newly appeared text. Or characters or images perform a rendering and software filtering operation, and then the software filtering results can be reused, which greatly reduces the amount of calculation.
  • obtaining the bitmap data of the content to be displayed by the processor includes: determining, by the processor, a vector diagram of the content to be displayed; to obtain the bitmap data of the content to be displayed.
  • the data type of the content to be displayed includes text, characters, and images.
  • an embodiment of the present application provides a computer storage medium, including computer instructions, when the computer instructions are executed on an electronic device, the communication apparatus is made to execute the screen display method in any of the possible implementations of any of the above aspects .
  • an embodiment of the present application provides a computer program product that, when the computer program product runs on a computer, enables the computer to execute the screen display method in any of the possible implementations of any of the foregoing aspects.
  • FIG. 1 is a schematic structural diagram of an electronic device according to an embodiment of the present application.
  • FIGS. 2a and 2b are schematic diagrams of an arrangement of sub-pixels according to an embodiment of the present application.
  • 3a and 3b are schematic scene diagrams of an interface display principle provided by an embodiment of the present application.
  • FIG. 4 is a schematic diagram of a display effect of a sub-pixel rendering algorithm provided by an embodiment of the present application.
  • 5a and 5b are schematic diagrams of filtering of a sub-pixel rendering algorithm provided by an embodiment of the present application.
  • 6a and 6b are schematic diagrams of filtering of another sub-pixel rendering algorithm provided by an embodiment of the present application.
  • FIG. 7 is a schematic structural diagram of an electronic device according to an embodiment of the present application.
  • 8a and 8b are schematic diagrams of display effects of a sub-pixel rendering algorithm provided by an embodiment of the present application.
  • FIG. 9 is a method flowchart of a screen display method provided by an embodiment of the present application.
  • 10a to 10c are schematic diagrams of display effects of a screen display method provided by an embodiment of the present application.
  • 11a to 11c are schematic diagrams of filtering of a screen display method provided by an embodiment of the present application.
  • FIG. 12 is a schematic diagram of another sub-pixel arrangement method provided by an embodiment of the present application.
  • FIG. 13 is a schematic diagram of a display effect of another sub-pixel rendering algorithm provided by an embodiment of the present application.
  • FIG. 14 is a schematic scene diagram of a screen display method provided by an embodiment of the present application.
  • first and second are only used for descriptive purposes, and should not be construed as implying or implying relative importance or implying the number of indicated technical features. Therefore, the features defined as “first” and “second” may explicitly or implicitly include one or more of the features. In the description of the embodiments of the present application, unless otherwise specified, the “multiple” The meaning is two or more. The orientation or positional relationship indicated by the terms “middle”, “left”, “right”, “upper”, “lower”, etc.
  • FIG. 1 shows a schematic structural diagram of an exemplary electronic device 100 provided by an embodiment of the present application.
  • the electronic device 100 may be a cell phone, tablet computer, desktop computer, laptop computer, handheld computer, notebook computer, ultra-mobile personal computer (UMPC), netbook, as well as cellular telephones, personal digital assistants (personal digital assistants) digital assistant (PDA), augmented reality (AR) devices, virtual reality (VR) devices, artificial intelligence (AI) devices, wearable devices, in-vehicle devices, smart home devices and/or Smart city equipment, the embodiment of the present application does not limit the specific type of the electronic equipment.
  • PDA personal digital assistants
  • AR augmented reality
  • VR virtual reality
  • AI artificial intelligence
  • the electronic device 100 may include a processor 110, an external memory interface 120, an internal memory 121, a universal serial bus (USB) interface 130, a charge management module 140, a power management module 141, a battery 142, an antenna 1, an antenna 2 , mobile communication module 150, wireless communication module 160, audio module 170, speaker 170A, receiver 170B, microphone 170C, headphone jack 170D, sensor module 180, buttons 190, motor 191, indicator 192, camera 193, display screen 194, and Subscriber identification module (subscriber identification module, SIM) card interface 195 and so on.
  • SIM Subscriber identification module
  • the sensor module 180 may include a pressure sensor 180A, a gyroscope sensor 180B, an air pressure sensor 180C, a magnetic sensor 180D, an acceleration sensor 180E, a distance sensor 180F, a proximity light sensor 180G, a fingerprint sensor 180H, a temperature sensor 180J, a touch sensor 180K, and ambient light. Sensor 180L, bone conduction sensor 180M, etc.
  • the structures illustrated in the embodiments of the present invention do not constitute a specific limitation on the electronic device 100 .
  • the electronic device 100 may include more or less components than shown, or combine some components, or separate some components, or arrange different components.
  • the illustrated components may be implemented in hardware, software, or a combination of software and hardware.
  • the processor 110 may include one or more processing units, for example, the processor 110 may include an application processor (application processor, AP), a modem processor, a graphics processor (graphics processing unit, GPU), an image signal processor (image signal processor, ISP), controller, video codec, digital signal processor (digital signal processor, DSP), baseband processor, and/or neural-network processing unit (neural-network processing unit, NPU), etc. Wherein, different processing units may be independent devices, or may be integrated in one or more processors.
  • application processor application processor, AP
  • modem processor graphics processor
  • ISP image signal processor
  • controller video codec
  • digital signal processor digital signal processor
  • baseband processor baseband processor
  • neural-network processing unit neural-network processing unit
  • the controller can generate an operation control signal according to the instruction operation code and timing signal, and complete the control of fetching and executing instructions.
  • a memory may also be provided in the processor 110 for storing instructions and data.
  • the memory in processor 110 is cache memory. This memory may hold instructions or data that have just been used or recycled by the processor 110 . If the processor 110 needs to use the instruction or data again, it can be called directly from the memory. Repeated accesses are avoided and the latency of the processor 110 is reduced, thereby increasing the efficiency of the system.
  • the processor 110 may include one or more interfaces.
  • the interface may include an integrated circuit (inter-integrated circuit, I2C) interface, an integrated circuit built-in audio (inter-integrated circuit sound, I2S) interface, a pulse code modulation (pulse code modulation, PCM) interface, a universal asynchronous transceiver (universal asynchronous transmitter) receiver/transmitter, UART) interface, mobile industry processor interface (MIPI), general-purpose input/output (GPIO) interface, subscriber identity module (SIM) interface, and / or universal serial bus (universal serial bus, USB) interface, etc.
  • I2C integrated circuit
  • I2S integrated circuit built-in audio
  • PCM pulse code modulation
  • PCM pulse code modulation
  • UART universal asynchronous transceiver
  • MIPI mobile industry processor interface
  • GPIO general-purpose input/output
  • SIM subscriber identity module
  • USB universal serial bus
  • the I2C interface is a bidirectional synchronous serial bus that includes a serial data line (SDA) and a serial clock line (SCL).
  • the processor 110 may contain multiple sets of I2C buses.
  • the processor 110 can be respectively coupled to the touch sensor 180K, the charger, the flash, the camera 193 and the like through different I2C bus interfaces.
  • the processor 110 may couple the touch sensor 180K through the I2C interface, so that the processor 110 and the touch sensor 180K communicate with each other through the I2C bus interface, so as to realize the touch function of the electronic device 100 .
  • the I2S interface can be used for audio communication.
  • the processor 110 may contain multiple sets of I2S buses.
  • the processor 110 may be coupled with the audio module 170 through an I2S bus to implement communication between the processor 110 and the audio module 170 .
  • the audio module 170 can transmit audio signals to the wireless communication module 160 through the I2S interface, so as to realize the function of answering calls through a Bluetooth headset.
  • the PCM interface can also be used for audio communications, sampling, quantizing and encoding analog signals.
  • the audio module 170 and the wireless communication module 160 may be coupled through a PCM bus interface.
  • the audio module 170 can also transmit audio signals to the wireless communication module 160 through the PCM interface, so as to realize the function of answering calls through the Bluetooth headset. Both the I2S interface and the PCM interface can be used for audio communication.
  • the UART interface is a universal serial data bus used for asynchronous communication.
  • the bus may be a bidirectional communication bus. It converts the data to be transmitted between serial communication and parallel communication.
  • a UART interface is typically used to connect the processor 110 with the wireless communication module 160 .
  • the processor 110 communicates with the Bluetooth module in the wireless communication module 160 through the UART interface to implement the Bluetooth function.
  • the audio module 170 can transmit audio signals to the wireless communication module 160 through the UART interface, so as to realize the function of playing music through the Bluetooth headset.
  • the MIPI interface can be used to connect the processor 110 with peripheral devices such as the display screen 194 and the camera 193 .
  • MIPI interfaces include camera serial interface (CSI), display serial interface (DSI), etc.
  • the processor 110 communicates with the camera 193 through a CSI interface, so as to realize the photographing function of the electronic device 100 .
  • the processor 110 communicates with the display screen 194 through the DSI interface to implement the display function of the electronic device 100 .
  • the GPIO interface can be configured by software.
  • the GPIO interface can be configured as a control signal or as a data signal.
  • the GPIO interface may be used to connect the processor 110 with the camera 193, the display screen 194, the wireless communication module 160, the audio module 170, the sensor module 180, and the like.
  • the GPIO interface can also be configured as I2C interface, I2S interface, UART interface, MIPI interface, etc.
  • the USB interface 130 is an interface that conforms to the USB standard specification, and may specifically be a Mini USB interface, a Micro USB interface, a USB Type C interface, and the like.
  • the USB interface 130 can be used to connect a charger to charge the electronic device 100, and can also be used to transmit data between the electronic device 100 and peripheral devices. It can also be used to connect headphones to play audio through the headphones.
  • the interface can also be used to connect other electronic devices, such as AR devices.
  • the interface connection relationship between the modules illustrated in the embodiment of the present invention is only a schematic illustration, and does not constitute a structural limitation of the electronic device 100 .
  • the electronic device 100 may also adopt different interface connection manners in the foregoing embodiments, or a combination of multiple interface connection manners.
  • the charging management module 140 is used to receive charging input from the charger.
  • the charger may be a wireless charger or a wired charger.
  • the charging management module 140 may receive charging input from the wired charger through the USB interface 130 .
  • the charging management module 140 may receive wireless charging input through a wireless charging coil of the electronic device 100 . While the charging management module 140 charges the battery 142, the power management module 141 can also supply power to the electronic device.
  • the power management module 141 is used for connecting the battery 142 , the charging management module 140 and the processor 110 .
  • the power management module 141 receives input from the battery 142 and/or the charging management module 140, and supplies power to the processor 110, the internal memory 121, the display screen 194, the camera 193, and the wireless communication module 160.
  • the power management module 141 can also be used to monitor parameters such as battery capacity, battery cycle times, battery health status (leakage, impedance).
  • the power management module 141 may also be provided in the processor 110 .
  • the power management module 141 and the charging management module 140 may also be provided in the same device.
  • the wireless communication function of the electronic device 100 may be implemented by the antenna 1, the antenna 2, the mobile communication module 150, the wireless communication module 160, the modulation and demodulation processor, the baseband processor, and the like.
  • Antenna 1 and Antenna 2 are used to transmit and receive electromagnetic wave signals.
  • Each antenna in electronic device 100 may be used to cover a single or multiple communication frequency bands. Different antennas can also be reused to improve antenna utilization.
  • the antenna 1 can be multiplexed into a diversity antenna of the wireless local area network. In other embodiments, the antenna may be used in conjunction with a tuning switch.
  • the mobile communication module 150 may provide wireless communication solutions including 2G/3G/4G/5G etc. applied on the electronic device 100 .
  • the mobile communication module 150 may include at least one filter, switch, power amplifier, low noise amplifier (LNA) and the like.
  • the mobile communication module 150 can receive electromagnetic waves from the antenna 1, filter and amplify the received electromagnetic waves, and transmit them to the modulation and demodulation processor for demodulation.
  • the mobile communication module 150 can also amplify the signal modulated by the modulation and demodulation processor, and then turn it into an electromagnetic wave for radiation through the antenna 1 .
  • at least part of the functional modules of the mobile communication module 150 may be provided in the processor 110 .
  • at least part of the functional modules of the mobile communication module 150 may be provided in the same device as at least part of the modules of the processor 110 .
  • the modem processor may include a modulator and a demodulator.
  • the modulator is used to modulate the low frequency baseband signal to be sent into a medium and high frequency signal.
  • the demodulator is used to demodulate the received electromagnetic wave signal into a low frequency baseband signal. Then the demodulator transmits the demodulated low-frequency baseband signal to the baseband processor for processing.
  • the low frequency baseband signal is processed by the baseband processor and passed to the application processor.
  • the application processor outputs sound signals through audio devices (not limited to the speaker 170A, the receiver 170B, etc.), or displays images or videos through the display screen 194 .
  • the modem processor may be a stand-alone device.
  • the modem processor may be independent of the processor 110, and may be provided in the same device as the mobile communication module 150 or other functional modules.
  • the wireless communication module 160 can provide applications on the electronic device 100 including wireless local area networks (WLAN) (such as wireless fidelity (Wi-Fi) networks), bluetooth (BT), global navigation satellites Wireless communication solutions such as global navigation satellite system (GNSS), frequency modulation (FM), near field communication (NFC), and infrared technology (IR).
  • WLAN wireless local area networks
  • BT Bluetooth
  • GNSS global navigation satellite system
  • FM frequency modulation
  • NFC near field communication
  • IR infrared technology
  • the wireless communication module 160 may be one or more devices integrating at least one communication processing module.
  • the wireless communication module 160 receives electromagnetic waves via the antenna 2 , frequency modulates and filters the electromagnetic wave signals, and sends the processed signals to the processor 110 .
  • the wireless communication module 160 can also receive the signal to be sent from the processor 110 , perform frequency modulation on it, amplify it, and convert it into electromagnetic waves for radiation through the antenna 2 .
  • the antenna 1 of the electronic device 100 is coupled with the mobile communication module 150, and the antenna 2 is coupled with the wireless communication module 160, so that the electronic device 100 can communicate with the network and other devices through wireless communication technology.
  • the wireless communication technology may include global system for mobile communications (GSM), general packet radio service (GPRS), code division multiple access (CDMA), broadband Code Division Multiple Access (WCDMA), Time Division Code Division Multiple Access (TD-SCDMA), Long Term Evolution (LTE), BT, GNSS, WLAN, NFC , FM, and/or IR technology, etc.
  • the GNSS may include global positioning system (global positioning system, GPS), global navigation satellite system (global navigation satellite system, GLONASS), Beidou navigation satellite system (beidou navigation satellite system, BDS), quasi-zenith satellite system (quasi -zenith satellite system, QZSS) and/or satellite based augmentation systems (SBAS).
  • global positioning system global positioning system, GPS
  • global navigation satellite system global navigation satellite system, GLONASS
  • Beidou navigation satellite system beidou navigation satellite system, BDS
  • quasi-zenith satellite system quadsi -zenith satellite system, QZSS
  • SBAS satellite based augmentation systems
  • the electronic device 100 implements a display function through a GPU, a display screen 194, an application processor, and the like.
  • the GPU is a microprocessor for image processing, and is connected to the display screen 194 and the application processor.
  • the GPU is used to perform mathematical and geometric calculations for graphics rendering.
  • Processor 110 may include one or more GPUs that execute program instructions to generate or alter display information.
  • Display screen 194 is used to display images, videos, and the like.
  • Display screen 194 includes a display panel.
  • the display panel can be a liquid crystal display (LCD), an organic light-emitting diode (OLED), an active-matrix organic light-emitting diode or an active-matrix organic light-emitting diode (active-matrix organic light).
  • LED diode AMOLED
  • flexible light-emitting diode flexible light-emitting diode (flex light-emitting diode, FLED), Miniled, MicroLed, Micro-oLed, quantum dot light-emitting diode (quantum dot light emitting diodes, QLED) and so on.
  • the electronic device 100 may include one or N display screens 194 , where N is a positive integer greater than one.
  • the electronic device 100 may implement a shooting function through an ISP, a camera 193, a video codec, a GPU, a display screen 194, an application processor, and the like.
  • the ISP is used to process the data fed back by the camera 193 .
  • the shutter is opened, the light is transmitted to the camera photosensitive element through the lens, the light signal is converted into an electrical signal, and the camera photosensitive element transmits the electrical signal to the ISP for processing, and converts it into an image visible to the naked eye.
  • ISP can also perform algorithm optimization on image noise, brightness, and skin tone.
  • ISP can also optimize the exposure, color temperature and other parameters of the shooting scene.
  • the ISP may be provided in the camera 193 .
  • Camera 193 is used to capture still images or video.
  • the object is projected through the lens to generate an optical image onto the photosensitive element.
  • the photosensitive element may be a charge coupled device (CCD) or a complementary metal-oxide-semiconductor (CMOS) phototransistor.
  • CMOS complementary metal-oxide-semiconductor
  • the photosensitive element converts the optical signal into an electrical signal, and then transmits the electrical signal to the ISP to convert it into a digital image signal.
  • the ISP outputs the digital image signal to the DSP for processing.
  • DSP converts digital image signals into standard RGB, YUV and other formats of image signals.
  • the electronic device 100 may include 1 or N cameras 193 , where N is a positive integer greater than 1.
  • a digital signal processor is used to process digital signals, in addition to processing digital image signals, it can also process other digital signals. For example, when the electronic device 100 selects a frequency point, the digital signal processor is used to perform Fourier transform on the frequency point energy and so on.
  • Video codecs are used to compress or decompress digital video.
  • the electronic device 100 may support one or more video codecs.
  • the electronic device 100 can play or record videos of various encoding formats, such as: Moving Picture Experts Group (moving picture experts group, MPEG) 1, MPEG2, MPEG3, MPEG4 and so on.
  • MPEG Moving Picture Experts Group
  • MPEG2 moving picture experts group
  • MPEG3 MPEG4
  • MPEG4 Moving Picture Experts Group
  • the NPU is a neural-network (NN) computing processor.
  • NN neural-network
  • Applications such as intelligent cognition of the electronic device 100 can be implemented through the NPU, such as image recognition, face recognition, speech recognition, text understanding, and the like.
  • the internal memory 121 may include one or more random access memories (RAM) and one or more non-volatile memories (NVM).
  • RAM random access memories
  • NVM non-volatile memories
  • Random access memory can include static random-access memory (SRAM), dynamic random access memory (DRAM), synchronous dynamic random access memory (SDRAM), double data rate synchronization Dynamic random access memory (double data rate synchronous dynamic random access memory, DDR SDRAM, such as the fifth generation DDR SDRAM is generally called DDR5 SDRAM), etc.;
  • SRAM static random-access memory
  • DRAM dynamic random access memory
  • SDRAM synchronous dynamic random access memory
  • DDR SDRAM double data rate synchronous dynamic random access memory
  • DDR SDRAM double data rate synchronous dynamic random access memory
  • DDR SDRAM double data rate synchronous dynamic random access memory
  • DDR5 SDRAM double data rate synchronous dynamic random access memory
  • Non-volatile memory may include magnetic disk storage devices, flash memory.
  • Flash memory can be divided into NOR FLASH, NAND FLASH, 3D NAND FLASH, etc. according to the operating principle, and can include single-level memory cell (SLC), multi-level memory cell (multi-level memory cell, SLC) according to the level of storage cell potential.
  • cell, MLC multi-level memory cell
  • TLC triple-level cell
  • QLC quad-level cell
  • UFS universal flash storage
  • eMMC embedded multimedia memory card
  • the random access memory can be directly read and written by the processor 110, and can be used to store executable programs (eg, machine instructions) of an operating system or other running programs, and can also be used to store data of users and application programs.
  • executable programs eg, machine instructions
  • the random access memory can be directly read and written by the processor 110, and can be used to store executable programs (eg, machine instructions) of an operating system or other running programs, and can also be used to store data of users and application programs.
  • the non-volatile memory can also store executable programs and store data of user and application programs, etc., and can be loaded into the random access memory in advance for the processor 110 to directly read and write.
  • the external memory interface 120 can be used to connect an external non-volatile memory, so as to expand the storage capacity of the electronic device 100 .
  • the external non-volatile memory communicates with the processor 110 through the external memory interface 120 to realize the data storage function. For example, save music, video, etc. files in external non-volatile memory.
  • the electronic device 100 may implement audio functions through an audio module 170, a speaker 170A, a receiver 170B, a microphone 170C, an earphone interface 170D, an application processor, and the like. Such as music playback, recording, etc.
  • the audio module 170 is used for converting digital audio information into analog audio signal output, and also for converting analog audio input into digital audio signal. Audio module 170 may also be used to encode and decode audio signals. In some embodiments, the audio module 170 may be provided in the processor 110 , or some functional modules of the audio module 170 may be provided in the processor 110 .
  • Speaker 170A also referred to as a "speaker" is used to convert audio electrical signals into sound signals.
  • the electronic device 100 can listen to music through the speaker 170A, or listen to a hands-free call.
  • the receiver 170B also referred to as "earpiece" is used to convert audio electrical signals into sound signals.
  • the voice can be answered by placing the receiver 170B close to the human ear.
  • the microphone 170C also called “microphone” or “microphone” is used to convert sound signals into electrical signals.
  • the user can make a sound by approaching the microphone 170C through a human mouth, and input the sound signal into the microphone 170C.
  • the electronic device 100 may be provided with at least one microphone 170C. In other embodiments, the electronic device 100 may be provided with two microphones 170C, which can implement a noise reduction function in addition to collecting sound signals. In other embodiments, the electronic device 100 may further be provided with three, four or more microphones 170C to collect sound signals, reduce noise, identify sound sources, and implement directional recording functions.
  • the earphone jack 170D is used to connect wired earphones.
  • the earphone interface 170D may be the USB interface 130, or may be a 3.5mm open mobile terminal platform (OMTP) standard interface, a cellular telecommunications industry association of the USA (CTIA) standard interface.
  • OMTP open mobile terminal platform
  • CTIA cellular telecommunications industry association of the USA
  • the pressure sensor 180A is used to sense pressure signals, and can convert the pressure signals into electrical signals.
  • the pressure sensor 180A may be provided on the display screen 194 .
  • the capacitive pressure sensor may be comprised of at least two parallel plates of conductive material. When a force is applied to the pressure sensor 180A, the capacitance between the electrodes changes.
  • the electronic device 100 determines the intensity of the pressure according to the change in capacitance. When a touch operation acts on the display screen 194, the electronic device 100 detects the intensity of the touch operation according to the pressure sensor 180A.
  • the electronic device 100 may also calculate the touched position according to the detection signal of the pressure sensor 180A.
  • touch operations acting on the same touch position but with different touch operation intensities may correspond to different operation instructions. For example, when a touch operation whose intensity is less than the first pressure threshold acts on the short message application icon, the instruction for viewing the short message is executed. When a touch operation whose intensity is greater than or equal to the first pressure threshold acts on the short message application icon, the instruction to create a new short message is executed.
  • the gyro sensor 180B may be used to determine the motion attitude of the electronic device 100 .
  • the angular velocity of electronic device 100 about three axes ie, x, y, and z axes
  • the gyro sensor 180B can be used for image stabilization.
  • the gyro sensor 180B detects the shaking angle of the electronic device 100, calculates the distance that the lens module needs to compensate according to the angle, and allows the lens to offset the shaking of the electronic device 100 through reverse motion to achieve anti-shake.
  • the gyro sensor 180B can also be used for navigation and somatosensory game scenarios.
  • the air pressure sensor 180C is used to measure air pressure.
  • the electronic device 100 calculates the altitude through the air pressure value measured by the air pressure sensor 180C to assist in positioning and navigation.
  • the magnetic sensor 180D includes a Hall sensor.
  • the electronic device 100 can detect the opening and closing of the flip holster using the magnetic sensor 180D.
  • the electronic device 100 can detect the opening and closing of the flip according to the magnetic sensor 180D. Further, according to the detected opening and closing state of the leather case or the opening and closing state of the flip cover, characteristics such as automatic unlocking of the flip cover are set.
  • the acceleration sensor 180E can detect the magnitude of the acceleration of the electronic device 100 in various directions (generally three axes).
  • the magnitude and direction of gravity can be detected when the electronic device 100 is stationary. It can also be used to identify the posture of electronic devices, and can be used in applications such as horizontal and vertical screen switching, pedometers, etc.
  • the electronic device 100 can measure the distance through infrared or laser. In some embodiments, when shooting a scene, the electronic device 100 can use the distance sensor 180F to measure the distance to achieve fast focusing.
  • Proximity light sensor 180G may include, for example, light emitting diodes (LEDs) and light detectors, such as photodiodes.
  • the light emitting diodes may be infrared light emitting diodes.
  • the electronic device 100 emits infrared light to the outside through the light emitting diode.
  • Electronic device 100 uses photodiodes to detect infrared reflected light from nearby objects. When sufficient reflected light is detected, it can be determined that there is an object near the electronic device 100 . When insufficient reflected light is detected, the electronic device 100 may determine that there is no object near the electronic device 100 .
  • the electronic device 100 can use the proximity light sensor 180G to detect that the user holds the electronic device 100 close to the ear to talk, so as to automatically turn off the screen to save power.
  • Proximity light sensor 180G can also be used in holster mode, pocket mode automatically unlocks and locks the screen.
  • the ambient light sensor 180L is used to sense ambient light brightness.
  • the electronic device 100 can adaptively adjust the brightness of the display screen 194 according to the perceived ambient light brightness.
  • the ambient light sensor 180L can also be used to automatically adjust the white balance when taking pictures.
  • the ambient light sensor 180L can also cooperate with the proximity light sensor 180G to detect whether the electronic device 100 is in a pocket, so as to prevent accidental touch.
  • the fingerprint sensor 180H is used to collect fingerprints.
  • the electronic device 100 can use the collected fingerprint characteristics to realize fingerprint unlocking, accessing application locks, taking photos with fingerprints, answering incoming calls with fingerprints, and the like.
  • the temperature sensor 180J is used to detect the temperature.
  • the electronic device 100 uses the temperature detected by the temperature sensor 180J to execute a temperature processing strategy. For example, when the temperature reported by the temperature sensor 180J exceeds a threshold value, the electronic device 100 reduces the performance of the processor located near the temperature sensor 180J in order to reduce power consumption and implement thermal protection.
  • the electronic device 100 when the temperature is lower than another threshold, the electronic device 100 heats the battery 142 to avoid abnormal shutdown of the electronic device 100 caused by the low temperature.
  • the electronic device 100 boosts the output voltage of the battery 142 to avoid abnormal shutdown caused by low temperature.
  • Touch sensor 180K also called “touch device”.
  • the touch sensor 180K may be disposed on the display screen 194 , and the touch sensor 180K and the display screen 194 form a touch screen, also called a “touch screen”.
  • the touch sensor 180K is used to detect a touch operation on or near it.
  • the touch sensor can pass the detected touch operation to the application processor to determine the type of touch event.
  • Visual output related to touch operations may be provided through display screen 194 .
  • the touch sensor 180K may also be disposed on the surface of the electronic device 100 , which is different from the location where the display screen 194 is located.
  • the bone conduction sensor 180M can acquire vibration signals. In some embodiments, the bone conduction sensor 180M can acquire vibration signals of the vibrating bone mass of the human voice. The bone conduction sensor 180M can also contact the pulse of the human body and receive the blood pressure beating signal. In some embodiments, the bone conduction sensor 180M can also be disposed in the earphone, combined with the bone conduction earphone.
  • the audio module 170 can analyze the voice signal based on the vibration signal of the vocal vibration bone block obtained by the bone conduction sensor 180M, so as to realize the voice function.
  • the application processor can analyze the heart rate information based on the blood pressure beat signal obtained by the bone conduction sensor 180M, and realize the function of heart rate detection.
  • the keys 190 include a power-on key, a volume key, and the like. Keys 190 may be mechanical keys. Touch buttons are also possible.
  • the electronic device 100 may receive key inputs and generate key signal inputs related to user settings and function control of the electronic device 100 .
  • Motor 191 can generate vibrating cues.
  • the motor 191 can be used for vibrating alerts for incoming calls, and can also be used for touch vibration feedback.
  • touch operations acting on different applications can correspond to different vibration feedback effects.
  • the motor 191 can also correspond to different vibration feedback effects for touch operations on different areas of the display screen 194 .
  • Different application scenarios for example: time reminder, receiving information, alarm clock, games, etc.
  • the touch vibration feedback effect can also support customization.
  • the indicator 192 can be an indicator light, which can be used to indicate the charging state, the change of the power, and can also be used to indicate a message, a missed call, a notification, and the like.
  • the SIM card interface 195 is used to connect a SIM card.
  • the SIM card can be contacted and separated from the electronic device 100 by inserting into the SIM card interface 195 or pulling out from the SIM card interface 195 .
  • the electronic device 100 may support 1 or N SIM card interfaces, where N is a positive integer greater than 1.
  • the SIM card interface 195 can support Nano SIM card, Micro SIM card, SIM card and so on. Multiple cards can be inserted into the same SIM card interface 195 at the same time. The types of the plurality of cards may be the same or different.
  • the SIM card interface 195 can also be compatible with different types of SIM cards.
  • the SIM card interface 195 is also compatible with external memory cards.
  • the electronic device 100 interacts with the network through the SIM card to realize functions such as call and data communication.
  • the electronic device 100 employs an eSIM, ie: an embedded SIM card.
  • the eSIM card can be embedded in the electronic device 100 and cannot be separated from the electronic device 100 .
  • the embodiment of the present application provides a screen display method, which can improve the screen display effect.
  • Each pixel in the display screen 194 can be composed of three sub-pixels of red, green and blue (RGB).
  • RGB red, green and blue
  • the principle is that the three primary colors of red, blue and green (RGB) can form any color.
  • the three sub-pixels emit light with different display brightness respectively, and can be mixed into the required color visually.
  • the arrangement of sub-pixels on different types of display screens is also different.
  • the light-emitting of the LCD screen mainly depends on its backlight layer.
  • the backlight of the backlight layer is only white, so it is necessary to add a layer of color filters to project the three primary colors.
  • the droplets in this liquid crystal layer are all contained in tiny cell structures, one or more of which make up a pixel on the screen.
  • FIG. 2a exemplarily shows the arrangement of sub-pixels in the LCD screen.
  • G represents a green sub-pixel
  • B represents a blue sub-pixel
  • R represents a red sub-pixel.
  • a green sub-pixel G, a blue sub-pixel B, and a red sub-pixel R are grouped together to form a pixel, and the arrangement order of the three-color sub-pixels is not limited.
  • the OLED screen is a display screen made of organic electric self-luminous diodes.
  • a minimum light-emitting unit can be regarded as a sub-pixel, and each sub-pixel can only emit light of a single color (red, blue or green).
  • the subpixels that emit red light are called red subpixels
  • the subpixels that emit blue light are called blue subpixels
  • the subpixels that emit green light are called green subpixels.
  • the sub-pixels of the OLED screen can be arranged in various ways. For example, one arrangement is shown in Figure 2b, in which a single pixel consists of only red, green or blue and green sub-pixels, that is, each pixel contains a green sub-pixel.
  • Sub-pixels and a red (or blue) sub-pixel the pixels including the red sub-pixel and the blue sub-pixel are arranged at intervals in the horizontal and vertical directions. In the screen, 9 sub-pixels are made in the horizontal direction, while only 6 sub-pixels are made in the horizontal direction in this OLED screen, and the number of sub-pixels is reduced by one-third.
  • the three primary colors of RGB can constitute all colors, two colors cannot constitute all colors, so when an image is actually displayed, each pixel in the OLED screen in the arrangement shown in Figure 2b will borrow the luminescence of its surrounding pixels. Points to form the RGB three primary colors, which can also be called borrowed colors.
  • the first pixel point P1 includes sub-pixels of all three colors, and there is no need to share any color sub-pixels from the adjacent second pixel point P2; while in FIG. 2b, the third pixel point P3 includes red color The sub-pixels and the green sub-pixels can share the blue sub-pixels from the adjacent fourth pixel point P4.
  • the blue sub-pixel in P4 is referred to as the compensation sub-pixel of the third pixel point P3.
  • One pixel can share sub-pixels with multiple pixels, the third pixel P3 can also share a blue sub-pixel with the fifth pixel P5, and the third pixel P3 can also share a red sub-pixel with the sixth pixel P6 pixel.
  • the interface content of the display interface of the electronic device 100 includes text, images, and the like.
  • the display of "HUAWEI" in the device 100 triggered by an application is taken as an example for introduction, wherein the application can be a reading app, a browsing app, a text editing application, etc., for example, when the user is in a file management application , when a file "Document 1.txt" containing "HUAWEI" character content is opened through a text editing application, the text editing application will display the characters on the user interface, so by calling the interface for displaying text in the system,
  • the character code corresponding to "HUAWEI” is sent to the character display module or frame of the electronic device for display, and the character display module or frame of the electronic device 100 performs character layout on the field "HUAWEI" according to the character code corresponding to "HUAWEI", Determine the display form (including size, color, font, word spacing, line spacing, etc.) and display position of each character (which can be a
  • the electronic device 100 queries the corresponding font file for the display data corresponding to each character (for example, the character "E", the electronic device 100 queries the font file of the example font A for the word "E” in the font of the example font A
  • the currently commonly used font file type is vector font, in which the display data of characters are stored in the form of vector graphics, specifically, the outline of characters is described by quadratic or cubic Bezier curves.
  • the electronic device 100 renders the vector graphics of the display data of each character, and the rendering process includes rasterization.
  • the vector graphics can be converted into bitmaps. Bitmaps are made up of multiple pixels that can be arranged and colored differently to form patterns.
  • the bitmap of the image to be displayed includes gray value data of three RGB channels of each pixel (the red gray value of each pixel constitutes a red vector matrix, The blue gray value constitutes a blue vector matrix and the green gray value constitutes a green vector matrix); if the image to be displayed is a grayscale image, since the grayscale values of the red, green and blue primary colors of the grayscale image are the same, then the to-be-displayed image is a grayscale image.
  • the bitmap of the display image includes gray value data of one channel of each pixel (a gray value vector matrix).
  • the pixel of a character bitmap usually only includes gray value data of one channel. Among them, the grayscale value ranges from 0 to 255.
  • the resolution of an image is 500*338, and the image is a color image, because the color of a pixel is represented by three RGB values, so the pixel matrix of the image corresponds to three grayscale values
  • the vector matrix is the red (R) vector matrix (500*338 size), the green (G) vector matrix (500*338 size), and the blue (B) vector matrix (500*338 size). If the values of the first row and first column of each matrix are: R: 240, G: 223, B: 204, so the color of this pixel is (240, 223, 204).
  • the resolution of an image is 500*338, and the image is a grayscale image, because the grayscale values of the three primary colors of red, green and blue are the same for each pixel in the bitmap, so the pixel matrix of the image corresponds to A vector matrix (500*338 size), if the value of the first row and first column of the matrix is 240, the color of this pixel is (240, 240, 240).
  • the electronic device 100 performs mixed overlay and layer overlay on the rendered bitmap.
  • Mixed overlay refers to pasting the bitmap on the corresponding position of the screen, and the size of the screen is the same as the size of the display screen of the application interface.
  • the way of mixing and overlaying includes CPU drawing and GPU drawing.
  • Layer overlay refers to combining the application interfaces of multiple applications and interface display elements such as status bars, navigation bars, and floating windows into an interface to be displayed.
  • the storage form of the interface to be displayed is a bitmap form with gray value data.
  • the electronic device 100 sends the superimposed interface to be displayed to the display screen, the display screen controller of the display screen lights the light-emitting points (sub-pixel points) on the display screen based on the sub-pixel rendering algorithm, and the display screen displays the to-be-displayed point interface.
  • the electronic device 100 obtains display data (a vector diagram) of the character "E”, and the electronic device 100 renders the vector diagram, Generate a bitmap composed of pixel points, superimpose the bitmap with other display contents to form the interface to be displayed, and finally send it to the display screen, and light the light-emitting points on the screen based on the sub-pixel rendering algorithm.
  • Subpixel rendering algorithms can be integrated into the display controller to define how the display device illuminates subpixels of different colors based on input data. Based on different sub-pixel rendering algorithms, the ways of lighting sub-pixels on the display device are also different. For example, for the sub-pixel arrangement on the display device of the electronic device 100 is the sub-pixel arrangement shown in FIG.
  • the three sub-pixels The display brightness of the point is determined based on the RGB gray value of the single-pixel color point; through the second sub-pixel rendering algorithm, the five sub-pixel points shown in Figure 4 are also lit on the display device, and the display brightness of the five sub-pixel points is is determined based on the RGB grayscale value of the single-pixel color point.
  • FIGS 5a and 5b illustrate the technical principle of the sub-pixel rendering algorithm 1 in detail.
  • the electronic device 100 sends the bitmap of the interface to be displayed to the display screen.
  • the display controller of the display screen renders the display based on the sub-pixels. Algorithms light up the glowing dots (subpixels) on the display.
  • the input bitmap of the interface to be displayed includes grayscale data of three RGB channels, while a pixel on the display screen has only two light-emitting points, corresponding to a green sub-pixel and a red/blue sub-pixel respectively .
  • pixel point (n-2, m), pixel point (n-1, m), and pixel point (n, m) as examples in Figure 5a
  • sub-pixel rendering algorithm 1 pixel point (n-1, m)
  • the blue sub-pixel can be shared with the adjacent pixel point (n-2, m)
  • the pixel point (n, m) can be shared with the adjacent pixel point (n-1, m) with the red sub-pixel.
  • Green sub-pixels correspond one-to-one and do not need to be shared with other pixels.
  • the pixel gray value corresponding to each light-emitting point on the display screen is determined, and the pixel point (n-1, m) and the adjacent pixel point (n-2, m) share the same value.
  • the shared parameter between sub-pixels in FIG. 5b is 0.5 is just an example, and does not constitute a limitation to the embodiments of the present application.
  • FIGS 6a and 6b illustrate the technical principle of the second sub-pixel rendering algorithm in detail.
  • the electronic device 100 sends the bitmap of the interface to be displayed to the display screen.
  • the display controller of the display screen renders the display based on the sub-pixels.
  • Algorithm 2 lights up the light-emitting points (sub-pixels) on the display.
  • the input bitmap of the interface to be displayed includes grayscale data of three RGB channels, and a pixel on the display screen has only two light-emitting points, corresponding to a green sub-pixel and a red/blue sub-pixel respectively .
  • the sub-pixel rendering algorithm exists in the form of hardware algorithm, which is solidified in DDIC or processing chip, and it is difficult to update through software.
  • the interface content of the display interface of the electronic device 100 includes characters and images.
  • the electronic The device 100 performs image decoding on the image to be displayed to obtain the image data to be displayed.
  • the image display module or framework of the electronic device 100 typesets the image to be displayed according to the image data, and determines the to-be-displayed image.
  • Display image display form including display size, color, etc.
  • display position can be a pixel coordinate, positioned at the upper left corner of the display position.
  • the image data is stored in the form of vector graphics or bitmaps. Then, the electronic device 100 renders the vector image of the to-be-displayed image to generate a bitmap.
  • the electronic device 100 performs mixed overlay and layer overlay on the rendered bitmap, pastes the bitmap on the corresponding position of the picture, the size of the picture is the same as the size of the display picture of the application, and then the application of the multiple applications
  • the interface and interface display elements such as the status bar and the navigation bar are combined into the interface to be displayed.
  • the electronic device 100 sends the superimposed interface to be displayed to the display screen, and the display screen controller of the display screen lights up the display screen based on the sub-pixel rendering algorithm and the arrangement of the light-emitting points (sub-pixel points) on the display screen. , the interface to be displayed is displayed on the display screen.
  • the display of a single white pixel requires sub-pixels of three colors of red, green and blue to emit light.
  • one pixel includes sub-pixels of three colors, so one pixel can display a single white pixel; while for the diamond arrangement of the above-mentioned OLED screen, a pixel has only two sub-pixels of two colors. Therefore, it is necessary to borrow sub-pixels from the surrounding pixels to display a single white pixel.
  • the following introduces the hardware structure of an electronic device 101 with a display screen, and further introduces the sub-pixel rendering algorithm in the display principle of the OLED display screen.
  • the electronic device 101 may be mounted on or integrated on the electronic device 100 , and the electronic device 100 may include the electronic device 101 .
  • the electronic device 101 includes an application processor AP and a display module, wherein the display module includes a display screen and a display driver integrated circuit (display driver integrated circuits, DDIC).
  • the DDIC is one of the main components of the display controller, and the sub-pixel rendering algorithm is integrated in the DDIC.
  • DDIC is used to drive the display screen and light up the light-emitting points (sub-pixel points) on the display screen based on the sub-pixel rendering algorithm.
  • the display screen may be the above-mentioned OLED screen, and the arrangement of the sub-pixels may adopt the above-mentioned diamond arrangement.
  • the application processor AP typesets, renders, and superimposes the interface content (including text, images, etc.) into a bitmap, and sends it to the DDIC.
  • the DDIC filters and samples the received bitmap. After filtering and sampling, the DDIC determines the display brightness of the sub-pixels, and converts the display brightness of each sub-pixel into a voltage or current to drive the OLED screen with the corresponding voltage. Or the current lights up the light-emitting points (sub-pixels) on the OLED screen.
  • the DDIC and AP can be integrated in the same chip.
  • DDIC can also be integrated in the OLED display.
  • the bitmap sent by the application processor AP to the DDIC is a rasterized bitmap, the bitmap is composed of pixels, and the bitmap includes the gray value data of three RGB channels of each pixel, or Gray value data for one channel.
  • DDIC receives the bitmap sent by the application processor AP, DDIC divides the bitmap into three channels for processing respectively. The three channels are the green channel, the red channel and the blue channel.
  • DDIC converts the display brightness of each sub-pixel into voltage or current, etc. Glowing point.
  • DDIC divides the processing process of the bitmap into green channel, red channel and blue channel. degree value), the red vector matrix for processing the bitmap in the red channel (including the pixel gray value of the red sub-pixels of each pixel), and the blue vector matrix for processing the bitmap in the blue channel (including each pixel pixel gray value of the blue subpixel of the point). Since there is only one red sub-pixel and one blue sub-pixel for every two pixels in the diamond arrangement of the OLED screen, the red sub-pixel and blue sub-pixel need to be used as compensation pixels for the surrounding pixels, and the red sub-pixel and blue sub-pixel The pixel gray value of a pixel is related to its surrounding pixels.
  • the electronic device 100 needs to filter and sample the bitmap to determine the pixel grayscale values of the red sub-pixels and the blue sub-pixels on the OLED screen. Therefore, in the red channel and blue channel, DDIC needs to filter and sample the input bitmap, and then based on the filtered and sampled bitmap, DDIC transforms the pixel gray value of the filtered bitmap into voltage or
  • the OLED screen is driven in the form of electric current, and the red light-emitting point and the blue light-emitting point on the OLED screen are lighted with corresponding voltage or current.
  • each pixel in the diamond arrangement of the OLED screen has a green sub-pixel
  • the display brightness of the green sub-pixel has nothing to do with other pixels, so there is no need for filtering and sampling for the green channel, and the DDIC can directly input the bitmap to
  • the green channel converts the pixel gray value of the bitmap into the form of voltage or current, etc., drives the OLED screen, and lights the green light-emitting points on the OLED screen with the corresponding voltage or current.
  • One-dimensional filtering may be filtering in the horizontal direction or filtering in the vertical direction, which is hereinafter referred to as 1D-SPR for short.
  • the formula for one-dimensional filtering can be:
  • dst(x, y) is the pixel gray value of the output pixel point (x, y)
  • src(x, y) is the pixel gray value of the input pixel point (x, y).
  • the kernel is the convolution kernel.
  • the filter kernel of the one-dimensional filter in the horizontal direction is (0.5, 0.5)
  • the value of the kernel is (0.5, 0.5)
  • the filter kernel is a 1-by-2 matrix
  • x The value of ' is 0 and 1, that is, kernel.cols is 2, and the value of y' is 0, that is, kernel.cols is 1.
  • anchor is the anchor point
  • anchor.x is the abscissa of the anchor point
  • anchor.y is the ordinate of the anchor point.
  • the anchor point of the one-dimensional filtering in the above horizontal direction is (0,0).
  • Each value in the output matrix is equal to the sum of the multiplication results of the input matrix and the corresponding position elements of the convolution kernel.
  • the filter kernel can be described as (0.5, 0.5).
  • the electronic device 100 performs horizontal filtering on the input bitmap with a filter kernel of (0.5, 0.5), and the gray value of each pixel in the filtering result is the gray value of two adjacent pixels in the input bitmap.
  • the red pixel gray value of , I_r(x, y) is the red pixel gray value of the input pixel (x, y)
  • the abscissa of the anchor point is 0,
  • the ordinate is 0, and the value of the kernel convolution kernel is (0.5, 0.5)
  • the filtering method of the first sub-pixel rendering algorithm shown in FIGS. 5 a and 5 b may be referred to, that is, one-dimensional filtering in the horizontal direction, and the filter kernel is (0.5, 0.5).
  • DDIC Based on the gray value of each pixel in the red channel, combined with the arrangement of sub-pixels on the display screen, DDIC can determine the display brightness of each red sub-pixel on the display screen. That is, the display brightness of each red/blue sub-pixel is related to the gray value of two adjacent pixels in the bitmap in the input channel.
  • a white dot can be displayed through three sub-pixels of red, green, and blue. It can be seen that since the filtering is only performed in the horizontal direction, the horizontal line and the horizontal stroke edge of the text There may be green sub-pixels exposed at the edge, and when the electronic device 100 displays a black and white boundary line in a horizontal direction, a color fringing (green fringing) phenomenon may be caused visually. Since the human eye is most sensitive to green, green sub-pixels exposed at the edge will have a more obvious green fringe and will have a stronger graininess.
  • Two-dimensional filtering, filtering in both the horizontal direction and the vertical direction is hereinafter referred to as 2D-SPR.
  • the filter kernel for two-dimensional filtering can be described as
  • the filtering method of the second sub-pixel rendering algorithm shown in FIG. 6a and FIG. 6b can be referred to as two-dimensional filtering, and the filter kernel is
  • DDIC can determine the display brightness of each red sub-pixel on the display screen. That is, the display brightness of each red/blue sub-pixel is related to the grayscale values of the four adjacent pixels in the horizontal and vertical directions in the bitmap in the input channel.
  • two red sub-pixels, two blue sub-pixels, and one green sub-pixel can display a white point. It can be seen that all green sub-pixels are Surrounded by red and blue sub-pixels, and with low brightness at the edges (25% red display brightness and 25% blue display brightness), there is a certain brightness gradient effect. At this time, when the electronic device 100 displays the black and white boundary line in the horizontal direction, there is no color edge (green edge) phenomenon visually.
  • the sub-pixel rendering algorithm determines how the DDIC lights up the sub-pixels of different colors on the display according to the input data.
  • the sub-pixel rendering algorithm is a hardware algorithm and cannot be updated by software. Then, in the case of horizontal filtering, when a black and white boundary line in a horizontal direction (for example, a horizontal stroke edge of a character) is displayed on the OLED screen arranged with diamonds, the color fringing phenomenon will be caused visually.
  • the sub-pixel arrangement and sub-pixel rendering algorithm of the electronic device 101 cannot be changed through software update, and it is difficult to optimize the display effect.
  • the embodiment of the present application provides a screen display method, which can solve the color fringing phenomenon and improve the screen display effect of the display screen without changing the sub-pixel arrangement of the display screen and the sub-pixel rendering algorithm.
  • the step flow of a screen display method provided by the present application may include:
  • Step S101 the electronic device 100 determines a vector diagram of the content to be displayed.
  • the electronic device 100 acquires the content to be displayed, where the content to be displayed may be text, characters, images, or the like.
  • the electronic device 100 determines the display form (display position, size, color, etc.) of the content to be displayed, and forms a vector diagram of the content to be displayed.
  • the electronic device 100 determines the display form of the image in the display interface based on the image data to be displayed and the display position in the content to be displayed, and determines the text based on the text data to be displayed, the display position and the font file in the content to be displayed.
  • the display form in the display interface thereby forming a vector diagram of the content to be displayed.
  • Step S102 the electronic device 100 renders the vector diagram of the content to be displayed, and generates a bitmap of the content to be displayed.
  • the electronic device 100 After determining the vector image of the content to be displayed, the electronic device 100 renders the vector image of the content to be displayed, that is, the vector image is rasterized, and the image represented by the vector image is converted into a bitmap.
  • the bitmap is composed of multiple pixels. It is composed of points, and the bitmap includes the gray value data of three RGB channels of each pixel or the gray value data of one channel.
  • the electronic device 100 generates a bitmap of the content to be displayed.
  • the above steps S101 and S102 are optional, and the electronic device 100 may directly acquire a bitmap of the content to be displayed.
  • the electronic device 100 receives the bitmap of the content to be displayed sent by other devices; for example, the electronic device 100 calls the bitmap of the content to be displayed stored by itself; and so on.
  • Step S103 the electronic device 100 performs software filtering on the bitmap of the content to be displayed.
  • the electronic device 100 After the electronic device 100 generates the bitmap of the content to be displayed, the electronic device 100 performs software filtering processing on the bitmap to generate a software-filtered bitmap.
  • software filtering processing There are various manners of software filtering, which are not limited in this embodiment of the present application.
  • the filtering parameters of the software filtering herein are determined based on a sub-pixel rendering algorithm, and the software filtering is used to optimize the display edge of the content to be displayed on the display panel of the display screen.
  • the processing direction of the software filtering is perpendicular to the processing direction of the one-dimensional filtering in the sub-pixel rendering algorithm.
  • the electronic device 100 performs software filtering on the bitmap of the content to be displayed, and the filter kernel is Among them, software filtering Combined with the resampling filter (0.5, 0.5) as a virtual resampling filter
  • the electronic device 100 filters the bitmap of the content to be displayed, and the filter kernel is (0.5, 0.5).
  • the filter kernel is (0.5, 0.5).
  • software filtering (0.5, 0.5) and resampling filter Combined together as a virtual resampling filter It can be seen that the above two cases are both to make the final filtering result achieve the effect of two-dimensional filtering.
  • the bitmap of the content to be displayed includes RGB three-channel bitmap data and single-channel grayscale bitmap data.
  • the RGB three-channel bitmap data includes the gray value data of the RGB three channels of each pixel point (red pixel gray value, blue pixel gray value and green pixel gray value of each pixel point);
  • the red pixel gray value, blue pixel gray value and green pixel gray value of each pixel in the grayscale bitmap are the same, so the single-channel grayscale bitmap includes the grayscale value of one channel of each pixel. data.
  • the red pixel gray value of each pixel constitutes the red channel in the bitmap data
  • the blue pixel gray value of each pixel constitutes the blue channel in the bitmap data
  • the green pixel of each pixel Grayscale values make up the green channel in bitmap data.
  • the electronic device When the bitmap of the content to be displayed includes RGB three-channel bitmap data, the electronic device performs single-pixel channel software filtering on the red pixel grayscale value and the blue pixel grayscale value in the bitmap data, respectively, to determine the filtered red color. Pixel gray value and blue pixel gray value; where the green pixel gray value remains unchanged.
  • Single-pixel channel software filtering processing refers to performing software filtering processing on one channel in the RGB three-channel bitmap data. At this time, the electronic device 100 obtains the filtered bitmap.
  • the bitmap of the content to be displayed includes the first red pixel gray value, the first blue pixel gray value and the first green pixel gray value of each pixel.
  • the electronic device 100 respectively filters the red channel and the blue channel in the bitmap, and obtains a filtered bitmap, where the filtered bitmap includes the second red pixel gray value, the second blue pixel gray value of each pixel color pixel gray value and the first green pixel gray value.
  • the electronic device When the bitmap of the content to be displayed includes single-channel grayscale bitmap data, the electronic device first converts the single-channel grayscale bitmap data into RGB three-channel bitmap data, if the original bitmap (the bitmap of the content to be displayed) ) has a size of [w, h]; the electronic device 100 allocates a memory with a size of [w, h, 3] to convert the single-channel grayscale bitmap data into RGB three-channel bitmap data (that is, the red pixels of each pixel point The gray value, the gray value of the blue pixel and the gray value of the green pixel are the same), wherein the electronic device performs the single-pixel channel software respectively on the gray value of the red pixel and the gray value of the blue pixel in the RGB three-channel bitmap data.
  • the filtering process determines the gray value of the red pixel and the gray value of the blue pixel after filtering; the gray value of the green pixel remains unchanged.
  • the electronic device 100 obtains the filtered bitmap.
  • the filtered bitmap includes RGB three-channel bitmap data.
  • the bitmap of the content to be displayed includes the first pixel gray value of each pixel, that is, the red pixel gray value, blue pixel gray value and green pixel gray value of each pixel are all is the gray value of the first pixel.
  • the electronic device copies the grayscale bitmap to the green channel, and then the electronic device 100 filters the grayscale bitmap to obtain the second pixel grayscale value.
  • the electronic device 100 copies the filtered grayscale bitmap to Red channel and blue channel, get the filtered bitmap.
  • the filtered bitmap includes the red pixel gray value (the second pixel gray value), the blue pixel gray value (the second pixel gray value), and the green pixel gray value (the first pixel gray value) of each pixel. pixel gray value).
  • the following is an example of displaying the letter "E", using Siyuan bold font, the font size is 8, and the following bitmap is obtained after rasterization and gamma mapping.
  • the size of the bitmap is 7*11, and the character bitmap usually includes single-channel grayscale bitmap data.
  • the specific bitmap data is shown in Table 1.
  • Software filtering is performed on the bitmap data of the above letter "E", wherein, the electronic device performs single-pixel channel software filtering processing on the gray value of the red pixel and the gray value of the blue pixel, and the gray value of the green pixel of the bitmap data remains unchanged. change, and obtain the bitmap data after software filtering as shown in Table 2.
  • degamma mapping transformation from nonlinear domain to linear domain
  • gamma mapping is performed again after software filtering (from transform the linear domain back into the nonlinear domain). If the bitmap to be displayed is in the linear domain, filter it directly. In practical situations, degamma mapping and gamma mapping may lead to loss of accuracy of the data, resulting in errors.
  • the filtering parameters in the above software filtering manner are determined in combination with the information of the display screen of the electronic device 100 .
  • the information of the display screen includes the type and quantity of the display screen, and the display direction of the display screen. in,
  • the electronic device 100 can determine the sub-pixel rendering algorithm integrated in the display screen by the model of the display screen, and the electronic device 100 determines the software filtering mode of the electronic device 100 by the filtering mode adopted in the sub-pixel rendering algorithm.
  • the number of display screens in the electronic device 100 may be multiple.
  • the models of the multiple display screens may be different, and the software filtering methods of the electronic device 100 are also different for different models of display screens.
  • the electronic device 100 switches the display screen to display, the software filtering mode of the content to be displayed by the electronic device 100 can be switched.
  • the electronic device 100 determines a software filtering method for the content to be displayed based on the display screen on which the content to be displayed is currently displayed.
  • the display directions of the display screen of the electronic device 100 include landscape orientation (90 degrees and 270 degrees) and portrait orientation (0 degrees and 180 degrees).
  • the electronic device 100 determines the software filtering mode of the electronic device 100 based on the current display direction of the display screen. Since the sub-pixel rendering algorithm is a hardware algorithm, the filtering direction in the sub-pixel rendering algorithm is fixed. When the display direction of the display screen changes, the layout of the content to be displayed also changes. At this time, the filtering parameters of the software filtering should also be change accordingly.
  • the filtering parameters are not only filter kernels, the filtering parameters for horizontal screens with different angles are also different, and the filtering parameters for vertical screens with different angles are also different.
  • the filter kernels are both (0.5, 0.5), but one is the average of the current pixel and its left pixel, and the other is the current pixel.
  • the average of the pixels on the right produces a different filtering result.
  • the electronic device 100 changes the filtering parameters of the software filtering based on the model of the electronic device and changes in the state of the electronic device (display screen) (eg, switching between different display screens, changing the display direction of the display screen, etc.). In this way, the display effect of the display screen in different situations can be solved, and the practicability and flexibility of the present application can be improved.
  • the following takes the electronic device 100 to perform software filtering processing on a single-channel bitmap (each pixel in the bitmap includes grayscale information of one channel), and the filter kernel of the software filtering is (0.5, 0.5) as an example.
  • the code looks like this:
  • Fig. 10c exemplarily shows a schematic display effect of the bitmap data in the red channel or the blue channel extracted before and after filtering.
  • a filter kernel of (0.5, 0.5) the vertical strokes of the word " ⁇ " in the red and blue channels Widening, because of the filtering in the horizontal direction, the red sub-pixels and the blue sub-pixels in the adjacent pixels in the horizontal direction perform color compensation, thereby achieving the effect of eliminating the green edge in the vertical direction.
  • the horizontal stroke of the word " ⁇ " will become wider (thicker), because of the filtering in the vertical direction, the red sub-pixels and blue sub-pixels in the adjacent pixels in the vertical direction will be Perform color compensation to achieve the effect of eliminating green edges in the horizontal direction.
  • the electronic device 100 sends the filtered content to the display, and the filter core of the sub-pixel rendering algorithm in the display is After the secondary filtering and sampling process of the resampling filter of the sub-pixel rendering algorithm, the DDIC drives the display to light up the sub-pixels on the screen.
  • the above horizontal filtering (0.5, 0.5) is related to the resampling filter Combined together as a virtual resampling filter That is, the method can enable the 1D SPR electronic device to obtain a display effect similar to that of the 2D SPR electronic device.
  • Step S104 the electronic device 100 superimposes the filtered bitmap.
  • the electronic device 100 pastes the bitmap on the corresponding position of the screen, and the size of the screen is the same as the size of the display screen of the application, so that the bitmap can be displayed in the same size as the display screen of the display screen.
  • the size is shown on the display of the electronic device.
  • the way of mixing and overlaying includes CPU drawing and GPU drawing.
  • the electronic device 100 combines the application interfaces of multiple applications and interface display elements such as status bars and navigation bars into an interface to be displayed.
  • Step S105 Based on the superimposed bitmap, the electronic device 100 lights the light-emitting points on the display screen through a sub-pixel rendering algorithm to display the to-be-displayed content.
  • the electronic device 100 sends the superimposed bitmap to the display screen, and the display screen controller (DDIC) of the display screen uses the sub-pixel rendering algorithm to light up the light-emitting points (sub-pixel points) on the display screen based on the superimposed bitmap. , so that the content to be displayed is displayed on the display screen of the electronic device 100 .
  • DDIC display screen controller
  • the superimposed bitmap includes gray value information of three RGB channels (red gray value, blue gray value and green gray value of each pixel).
  • the electronic device 100 needs to filter the input bitmap (the superimposed bitmap) to determine the pixel grayscale values of the red sub-pixels and the blue sub-pixels. Since there is only one red sub-pixel and one blue sub-pixel for every two pixels in the diamond arrangement of the OLED screen, the red sub-pixel and blue sub-pixel need to be used as compensation pixels for the surrounding pixels, and the red sub-pixel and blue sub-pixel
  • the display brightness of a pixel is related to its surrounding pixels.
  • the electronic device 100 performs a software filter on the red channel and the blue channel.
  • the electronic device 100 performs secondary filtering on the red channel and the blue channel to determine the red sub-pixel and the blue sub-pixel. pixel gray value. Therefore, in the red channel and the blue channel, based on the filtered and sampled bitmap, DDIC converts the pixel gray value of the filtered bitmap into a voltage or current, etc., and drives the OLED screen with the corresponding voltage or current. The current drives the red light-emitting points and the blue light-emitting points on the OLED screen to emit light.
  • the filtering method here is different from the filtering method in the above step S103.
  • the filtering method here is determined by the sub-pixel rendering algorithm integrated in the DDIC in the electronic device 100, and it is difficult to change by software update.
  • DDIC can directly input the superimposed bitmap to the green channel, and convert the pixel gray value of the green sub-pixel on the superimposed bitmap into voltage or current. and other forms, drive the display screen to light the green light-emitting point on the display screen with the corresponding voltage or current.
  • the electronic device 100 displays the content to be displayed corresponding to the superimposed bitmap.
  • the electronic device 100 performs pre-filtering (software filtering) on the content to be displayed before being sent to the display screen, and works together with the resampling filter in the sub-pixel rendering algorithm to change the display effect.
  • pre-filtering software filtering
  • the embodiments of the present application solve the difficulty that the sub-pixel rendering algorithm cannot be updated online.
  • the filter kernel of software filtering is as follows
  • the resampling filter kernel used in the sub-pixel rendering algorithm in DDIC is (0.5, 0.5) as an example, when the sub-pixel arrangement on the display screen is the sub-pixel arrangement shown in Figure 2b, the software filtering and sub-pixel arrangement are described in detail.
  • software filtering Combined with the resampling filter (0.5, 0.5) as a virtual resampling filter
  • the electronic device 100 performs software filtering processing on the bitmap.
  • the bitmap of the content to be displayed includes grayscale data of three RGB channels
  • the electronic device 100 performs software filtering on the red channel and the blue channel
  • the filter kernel is Determines the grayscale value of each pixel in the bitmap.
  • the electronic device 100 sends the filtered bitmap to the display screen, and the display screen controller of the display screen lights the light-emitting points (sub-pixel points) on the display screen based on the sub-pixel rendering algorithm.
  • the filter kernel of this subpixel rendering algorithm is (0.5, 0.5).
  • the pixel point (n-1, m) can share the blue sub-pixel with the adjacent pixel point (n-2, m), and the pixel point (n, m) can share the blue sub-pixel with the adjacent pixel point.
  • (n-1, m) share red sub-pixels. Green sub-pixels correspond one-to-one and do not need to be shared with other pixels.
  • the above-mentioned preprocessing software filtering algorithm may occur at any stage of the rendering and display (send to the display screen) process.
  • the filtering algorithm of the preprocessing software can occur between any two steps of rasterization, mixed overlay, layer overlay, and display.
  • the preprocessing software filtering algorithm can occur between any two steps in image decoding, rasterization, hybrid overlay, layer overlay, and display.
  • step S103 is integrated after the rasterization flow (step S102 ) and before the mixing and superposition process (step S104 ). If step S103 is performed after the overlay process (step S104 ) and before the display (step S105 ), after the electronic device 100 obtains the bitmap of the content to be displayed, the mixed overlay and layer overlay process are performed, and then the superimposed bitmap is processed. After software filtering is performed, the electronic device 100 sends the filtered bitmap to the display screen, and uses the sub-pixel rendering algorithm to light up the light-emitting points on the display screen to display the to-be-displayed content. Here, before performing software filtering on the superimposed bitmap, the electronic device 100 needs to transform the superimposed bitmap into the linear domain, and then perform software filtering, and then re-transform the filtered bitmap back to the nonlinear domain .
  • the electronic device 100 adopts different processing methods according to different properties of the content to be displayed (eg, text, images, and vector graphics).
  • the electronic device 100 executes step S103 only on words or characters.
  • the edge of the text stroke has a higher contrast, and the color edge and graininess are the most obvious at the edge of the text stroke.
  • the electronic device 100 acquires the content to be displayed, and determines the display form of the content to be displayed. Based on the to-be-displayed content, the electronic device 100 queries whether the bitmap of the to-be-displayed content is stored in the electronic device 100. If the to-be-displayed content is text or font, size, etc.), query the electronic device 100 whether the bitmap of the content to be displayed under the display form is stored; if the content to be displayed is an image, the electronic device 100 based on the content to be displayed and the display form (display of the image size, etc.), inquire whether the electronic device 100 stores the bitmap of the content to be displayed in the display form.
  • the electronic device 100 calls the bitmap of the content to be displayed, and directly executes step S104.
  • the electronic device 100 executes steps S102 and S103, and after the electronic device renders and filters the content to be displayed, a bitmap of the content to be displayed is generated.
  • the electronic device 100 saves the displayed content. bitmap for the next call.
  • the electronic device 100 only needs to perform rendering and software filtering operations on the newly appeared characters or characters or images once, and then the software filtering results can be reused, which greatly reduces the amount of calculation.
  • the electronic device 100 determines whether to perform the step S103 based on the current state of the electronic device 100 .
  • the electronic device 100 executes step S103.
  • the preset condition includes that the electronic device 100 is not currently in a scene such as shooting and sharing, screen projection, or the like.
  • the electronic device 100 captures the image and sends it to other electronic devices through an application.
  • the image data sent by the electronic device 100 to other electronic devices is a bitmap. If the electronic device 100 performs a preset filtering operation on the captured image, the sent image data will be a bitmap filtered by software.
  • the preset filter kernel of software filtering is (0.5, 0.5)
  • the image filtered by the software is an image with a green edge (for example, as shown in FIG. 6a ).
  • Other electronic devices receive the image with the green border, which affects the display effect of the image on other electronic devices. Therefore, when the electronic device 100 is not currently in the shooting and sharing scene, the electronic device 100 executes step S103.
  • the electronic device 100 when the electronic device 100 receives the screen projection operation, the electronic device 100 sends the captured image to other electronic devices in real time for display. If the electronic device 100 performs a preset filtering operation on the captured image, it will cause The image data sent is a software filtered bitmap. Other electronic devices receive the software-filtered bitmap, which affects the display effect of the bitmap on other electronic devices. Therefore, when the electronic device 100 is not currently in a screen projection scenario, the electronic device 100 executes step S103. Optionally, when the electronic device 100 receives the operation of turning off the screen projection, the electronic device 100 resumes performing step S103.
  • the method for improving the display effect of the display screen described in the above embodiments can be applied to the case where the display screen of the electronic device 100 is an OLED screen and the arrangement of the sub-pixels is a diamond arrangement.
  • the arrangement of the sub-pixels is a diamond arrangement.
  • the screen display method provided by the embodiment of the present application can also be applied to display screens with other sub-pixel arrangements, improving the display of display screens with different arrangements. Effect.
  • FIG. 12 exemplarily shows another sub-pixel arrangement of the OLED display screen.
  • the arrangement shown in FIG. 12 includes two types of pixel points, one type of pixel point includes one red sub-pixel and one blue sub-pixel, and the other type of pixel point includes two green sub-pixels. and spaced vertically.
  • different sub-pixel rendering algorithms can achieve different display effects.
  • the following is an exemplary brief introduction of four sub-pixel rendering algorithms suitable for this arrangement.
  • sub-pixel rendering algorithm 3 when a single-pixel white point is input (the RGB gray value of the single-pixel white point is (255, 255, 255)), through sub-pixel rendering algorithm three, five sub-pixels are lit on the display device, It includes two red sub-pixels, two blue sub-pixels and one green sub-pixel; the two red sub-pixels have the same brightness, and the two blue sub-pixels have the same brightness.
  • the lighting method implemented by the sub-pixel rendering algorithm 3 can make the human eye perceive white, and the actual effect produced is high definition, but serious graininess and obvious color fringing.
  • the sub-pixel rendering algorithm four eight sub-pixels are lit on the display device, including two red sub-pixels, two blue sub-pixels and four green sub-pixels; the brightness of the two red sub-pixels is the same, and the two blue sub-pixels The sub-pixels have the same brightness, and the four green sub-pixels have different brightness.
  • the display brightness borne by two green sub-pixels is 37.5%
  • the brightness borne by the other two green sub-pixels is 12.5%.
  • the lighting method implemented by the sub-pixel rendering algorithm 4 can make the human eye perceive white, and the actual effect produced is that the definition is low and the color fringing is obvious.
  • sub-pixel rendering algorithm five Six sub-pixels are lit on the display device, including two red sub-pixels, two blue sub-pixels and two green sub-pixels; two red sub-pixels have the same brightness, and two blue sub-pixels have the same brightness.
  • the sub-pixels have the same brightness, and the two green sub-pixels have the same brightness.
  • the lighting method implemented by the sub-pixel rendering algorithm 5 can make the human eye perceive white, and the actual effect produced is that the definition is low and the color fringing is obvious.
  • sub-pixel rendering algorithm six Through sub-pixel rendering algorithm six, ten sub-pixels are lit on the display device, including three red sub-pixels, three blue sub-pixels and four green sub-pixels; three red sub-pixels have the same brightness, and three blue sub-pixels have the same brightness.
  • the sub-pixels have the same brightness, and the four green sub-pixels have different brightness.
  • the display brightness borne by two green sub-pixels is 37.5%
  • the brightness borne by the other two green sub-pixels is 12.5%.
  • the lighting method implemented by the sub-pixel rendering algorithm 6 can make the human eye perceive white, and the actual effect produced is that the definition is average, and the color fringing is generally obvious.
  • DDIC provides a certain ability to adjust the sub-pixel rendering algorithm, but the number and range of adjustable parameters are small.
  • the above four sub-pixel rendering algorithms all have different degrees of unclearness and obvious color edge problems. A more satisfactory result cannot be obtained. Therefore, to improve the display effect of the above-mentioned display screen, it cannot be achieved by changing the sub-pixel rendering algorithm.
  • the electronic device 100 can finally achieve high definition and no color fringing by pre-filtering the content to be displayed before being sent to the display screen, and working together with the sub-pixel rendering algorithm. effect, improve the display effect. Solved the difficulty that the subpixel rendering algorithm could not be updated online.
  • the pre-filtering process here is that after the electronic device 100 obtains the content to be displayed, and performs typesetting and rendering on the content to be displayed, the electronic device 100 performs software filtering processing on the bitmap of the content to be displayed.
  • the bitmap of the content to be displayed includes gray value data of three RGB channels of each pixel (red gray value, blue gray value and green gray value of each pixel).
  • the electronic device 100 filters the green channel in the bitmap, and determines the filtered pixel grayscale value of the bitmap in the green channel. At this time, the electronic device 100 obtains the filtered bitmap.
  • the bitmap of the content to be displayed includes a third red pixel gray value, a third blue pixel gray value, and a third green pixel gray value of each pixel.
  • the electronic device 100 filters the green channel in the bitmap, and obtains a filtered bitmap, where the filtered bitmap includes the gray value of the third red pixel and the gray value of the third blue pixel of each pixel. and the fourth green pixel gray value.
  • the electronic device 100 allocates a The memory is used to save the filtered bitmap, and the filtered bitmap includes grayscale value data of three RGB channels.
  • the electronic device 100 copies the original bitmap to two channels in the filtered bitmap, which are the red channel and the blue channel; due to the above-mentioned arrangement of sub-pixels, an ideal display effect is to be achieved.
  • the red and blue sub-pixels do not need to be adjusted, so no filtering is required for the red and blue channels, and the electronic device 100 can directly copy the original bitmap to the red and blue channels.
  • the electronic device 100 performs software filtering on the original bitmap to determine the pixel gray value of the green sub-pixel, and then copies the software-filtered grayscale bitmap to another channel, which is the green channel. At this time, the electronic device 100 obtains the filtered bitmap.
  • the bitmap of the content to be displayed includes the third pixel gray value of each pixel, that is, the red pixel gray value, blue pixel gray value and green pixel gray value of each pixel are all is the gray value of the third pixel.
  • the electronic device copies the grayscale bitmap to the red channel and the blue channel, and then the electronic device 100 filters the grayscale bitmap to obtain the fourth pixel grayscale value.
  • the electronic device 100 stores the filtered grayscale bitmap. Copy the image to the green channel to get the filtered bitmap.
  • the filtered bitmap includes the red pixel gray value (third pixel gray value), blue pixel gray value (third pixel gray value), and green pixel gray value (fourth pixel gray value) of each pixel. pixel gray value).
  • the filter kernel can be described as (0.125, 0.75, 0.125).
  • the brightness of one green sub-pixel in sub-pixel rendering algorithm 3 is 100%.
  • the brightness of one green sub-pixel can be divided into three green sub-pixels, so that the brightness of the three green sub-pixels is 12.5%, 75% and 12.5%.
  • the electronic device 100 sends the filtered content to the display, and after the filtering and sampling process of the resampling filter of the display screen control chip (DDIC), the sub-pixels on the screen are lit.
  • DDIC display screen control chip
  • the electronic device 100 searches the font file of the example font A for the display data of the word “E” in the font of the example font A, and the display data It is stored in the form of a vector diagram. Then, the electronic device 100 renders the vector image of the character "net” to generate a bitmap.
  • the bitmap is composed of multiple pixels, and the bitmap includes the gray value data of three RGB channels of each pixel (the red gray value of each pixel constitutes a red vector matrix, and the blue gray value constitutes a red vector matrix. The blue vector matrix and the green grayscale values form the green vector matrix).
  • the electronic device 100 performs software filtering on the green channel in the bitmap, and the filter kernel of the software filtering may be (0.125, 0.75, 0.125).
  • the electronic device 100 performs mixed overlay and layer overlay on the filtered bitmap. Finally, the electronic device 100 sends the superimposed bitmap to the display screen, and the display screen controller of the display screen performs secondary filtering based on the resampling filter in the sub-pixel rendering algorithm to determine the final grayscale values of the three RGB channels information.
  • the electronic device 100 converts the pixel gray value of the bitmap after the secondary filtering into the form of voltage or current, drives the OLED screen, lights the light-emitting points (sub-pixel points) on the display screen with the corresponding voltage or current, and displays the The character bitmap is displayed on the screen.
  • the filtering parameters in the above software filtering method can be adjusted as required, and the software filtering parameters for the sub-pixel rendering algorithm 4 are different from the software filtering parameters for the sub-pixel rendering algorithm 3.
  • the parameters of the above software filtering are determined based on the sub-pixel rendering algorithm of the electronic device 100 and the desired display effect.
  • the preprocessing software filtering algorithm described above may occur at any stage in the image rendering and display (send to the display) process.
  • the above-mentioned embodiments it may be implemented in whole or in part by software, hardware, firmware or any combination thereof.
  • software it can be implemented in whole or in part in the form of a computer program product.
  • the computer program product includes one or more computer instructions. When the computer program instructions are loaded and executed on a computer, all or part of the processes or functions described in the embodiments of the present application are generated.
  • the computer may be a general purpose computer, special purpose computer, computer network, or other programmable device.
  • the computer instructions may be stored in or transmitted from one computer-readable storage medium to another computer-readable storage medium, for example, the computer instructions may be downloaded from a website site, computer, server, or data center Transmission to another website site, computer, server, or data center by wire (eg, coaxial cable, optical fiber, digital subscriber line) or wireless (eg, infrared, wireless, microwave, etc.).
  • the computer-readable storage medium can be any available medium that can be accessed by a computer or a data storage device such as a server, a data center, or the like that includes an integration of one or more available media.
  • the usable media may be magnetic media (eg, floppy disks, hard disks, magnetic tapes), optical media (eg, DVDs), or semiconductor media (eg, solid state drives), and the like.
  • the process can be completed by instructing the relevant hardware by a computer program, and the program can be stored in a computer-readable storage medium.
  • the program When the program is executed , which may include the processes of the foregoing method embodiments.
  • the aforementioned storage medium includes: ROM or random storage memory RAM, magnetic disk or optical disk and other mediums that can store program codes.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Computer Hardware Design (AREA)
  • Human Computer Interaction (AREA)
  • General Engineering & Computer Science (AREA)
  • Control Of Indicators Other Than Cathode Ray Tubes (AREA)

Abstract

公开了一种屏幕显示方法,其特征在于,所述方法包括:电子设备在获取到待显示内容的位图数据后,在将该待显示内容的位图数据发送到显示屏之前,将该位图数据进行软件滤波处理;然后将经过了软件滤波的位图数据通过子像素渲染算法进行子像素渲染,进而根据子像素渲染后确定的RGB灰度值驱动显示屏的显示面板上的OLED发光点以显示待显示内容。其中软件滤波的滤波参数是基于子像素渲染算法确定的,软件滤波和子像素渲染算法共同作用,能够优化待显示内容在OLED显示面板上的显示边缘,达到改变显示效果的作用。这样,解决了子像素渲染算法无法在线更新的困难。

Description

一种屏幕显示方法及相关装置
本申请要求于2021年02月02日提交中国专利局、申请号为202110144623.6、申请名称为“一种改善屏幕显示效果的装置与方法”的中国专利申请的优先权,本申请要求于2021年02月02日提交中国专利局、申请号为202110143801.3、申请名称为“一种软硬联合调节子像素渲染的装置与方法”的中国专利申请的优先权,本申请要求于2021年6月30日提交中国专利局、申请号为202110742740.2、申请名称为“一种屏幕显示方法及相关装置”的中国专利申请的优先权,其全部内容通过引用结合在本申请中。
技术领域
本申请涉及电子技术领域,尤其涉及一种屏幕显示方法及相关装置。
背景技术
子像素渲染(sub-pixel rendering,SPR)技术常被用于具有有机发光二极管(organic light-emitting diode,OLED)显示屏幕的电子设备。子像素渲染技术利用人眼对不同色彩辨识度的差异,改变常规由红绿蓝三色子像素定义一个像素的模式,通过不同像素间分享部分颜色子像素的方式,用相对较少的子像素数,模拟实现相同像素分辨率表现能力,从而降低制作工艺和制作成本。
子像素渲染技术有两个关键要素:其一是子像素的排列,其二是子像素渲染算法。子像素的排列由OLED显示面板决定,定义显示面板上不同颜色子像素的分布规律;子像素渲染算法由显示屏控制器决定,定义显示器件如何根据输入数据点亮不同颜色的子像素。子像素排列和子像素渲染算法共同影响显示器件的显示效果,合适的组合会使显示画面清晰、细腻、柔和、颜色准确,而不合适的组合可能会导致画面出现颗粒感、偏色、彩边等情况。
因此,子像素排列取决于显示面板器件,子像素渲染算法常集成在显示屏控制器中,二者均由硬件决定,有更新周期长、可定制性差、无法通过软件更新的特点。搭载有显示屏幕的电子设备在选型、制造、生产、销售后,无法通过软件更新的方式改变显示屏幕的子像素排列和子像素渲染算法,难以优化显示效果。如何优化显示屏幕的显示效果,是本领域技术人员正在研究的问题。
发明内容
本申请实施例提供了一种屏幕显示方法及相关装置,可以改善屏幕显示效果。
第一方面,本申请提供了一种屏幕显示方法,应用于电子设备,电子设备包括OLED显示面板和集成子像素渲染算法的显示屏控制器,其特征在于,该方法包括:电子设备获取待显示内容的位图数据;电子设备对位图数据进行软件滤波,其中,软件滤波基于子像素渲染算法确定,用于优化待显示内容在OLED显示面板上的显示边缘;显示屏控制器对经过软件滤波后的位图数据进行子像素渲染,确定每个像素点的RGB灰度值;显示屏控制器根据RGB灰度值驱动OLED显示面板上的OLED发光点以显示待显示内容。
本申请实施例中,待显示内容可以包括文字、字符、图像等等。子像素渲染算法是一个硬件算法,难以通过软件更新。电子设备在获取到待显示内容的位图数据后,在将该待显示内容的位图数据发送到显示屏之前,将该位图数据进行软件滤波处理;然后将经过了软件滤 波的位图数据通过子像素渲染算法进行子像素渲染,确定每个像素点的RGB灰度值(包括红色像素灰度值、绿色像素灰度值、蓝色像素灰度值),其中软件滤波的滤波参数是基于子像素渲染算法确定的,软件滤波和子像素渲染算法共同作用,能够优化待显示内容在OLED显示面板上的显示边缘,达到改变显示效果的作用。这样,解决了子像素渲染算法无法在线更新的困难。
结合第一方面,在一些可能的实施方式中,电子设备对位图数据进行软件滤波包括:位图数据包括RGB三通道位图数据和单通道灰度位图数据;当位图数据为单通道灰度位图数据时,将位图数据转换为RGB三通道位图数据;电子设备对RGB三通道位图数据的红色像素灰度值和蓝色像素灰度值分别进行单像素通道软件滤波处理;电子设备对位图数据的绿色像素灰度值保持不变。
位图数据包括RGB三通道位图数据,即为包括每个像素点的RGB三个通道的灰度值数据(每个像素点的红色像素灰度值、蓝色像素灰度值和绿色像素灰度值),其中每个像素点的红色像素灰度值构成了位图数据中的红色通道,每个像素点的蓝色像素灰度值构成了位图数据中的蓝色通道,每个像素点的绿色像素灰度值构成了位图数据中的绿色通道。单像素通道软件滤波处理指的是对RGB三通道位图数据中的其中一个通道进行软件滤波处理。
当位图数据为RGB三通道位图数据,电子设备对该位图数据中红色像素灰度值和蓝色像素灰度值分别进行单像素通道软件滤波处理,确定滤波后的红色像素灰度值和蓝色像素灰度值;其中绿色像素灰度值保持不变。
位图数据还包括单通道灰度位图数据,即为包括每个像素点的一个通道的灰度值数据(即每个像素点的红色像素灰度值、蓝色像素灰度值和绿色像素灰度值相同)。当位图数据为单通道灰度位图数据,电子设备首先将该单通道灰度位图数据转化为RGB三通道位图数据,即若原始位图(单通道灰度位图数据)的尺寸为[w,h],则电子设备分配尺寸为[w,h,3]的内存将单通道灰度位图数据转化为RGB三通道位图数据。其中,电子设备对该RGB三通道位图数据中红色像素灰度值和蓝色像素灰度值分别进行单像素通道软件滤波处理,确定滤波后的红色像素灰度值和蓝色像素灰度值;其中绿色像素灰度值保持不变。
结合第一方面,在一些可能的实施方式中,单像素通道软件滤波处理包括:单像素通道软件滤波处理与子像素渲染算法中的一维滤波处理方向垂直。即单像素通道软件滤波处理中的滤波参数是基于子像素渲染算法确定的,当子像素渲染算法中的滤波器为水平方向上的一维滤波器,则单像素通道软件滤波处理中的滤波参数为竖直方向上的一维滤波参数;当子像素渲染算法中的滤波器为竖直方向上的一维滤波器,则单像素通道软件滤波处理中的滤波参数为水平方向上的一维滤波参数。
在一些可能的实施方式中,一维滤波的公式包括:
Figure PCTCN2022073339-appb-000001
其中,dst(x,y)为输出的像素点(x,y)的像素灰度值,src(x,y)为输入的像素点(x,y)的像素灰度值。kernel为卷积核,例如水平方向上的一维滤波的滤波器核为(0.5,0.5),则kernel的值为(0.5,0.5),该滤波器核为1乘2的矩阵,则x^'的取值为0和1,即kernel.cols为2,y^'的取值为0,即kernel.cols为1。anchor为锚点,anchor.x为锚点的横坐标,anchor.y为锚点的纵坐标,例如在水平方向上的一维滤波的锚点可以为(0,0)。输出矩阵中的每个数值等于输入矩阵与卷积核对应位置元素相乘结果的和。
结合第一方面,在一些可能的实施方式中,在电子设备对位图数据进行软件滤波之前,方法还包括:电子设备基于显示屏的型号确定子像素渲染算法。其中,不同的显示屏的型号可以不同。电子设备通过显示屏的型号可以确定该显示屏中集成的子像素渲染算法,从而电子设备可以通过子像素渲染算法中采用的滤波方式确定电子设备的软件滤波的滤波参数或单像素通道软件滤波处理中的滤波参数。
结合第一方面,在一些可能的实施方式中,电子设备对位图数据进行软件滤波包括:确认电子设备不处于拍摄分享、投屏的场景时,电子设备对位图数据进行软件滤波。即电子设备需要区分不同的场景来确定是否需要对位图数据进行软件滤波。当电子设备接收到拍摄分享、投屏操作,电子设备采集图像,并通过应用程序发送到其他电子设备。其中,电子设备向其他电子设备发送的图像数据为位图,电子设备若对采集到的图像进行预设的滤波操作,会导致发送的图像数据是经过软件滤波的位图。其他电子设备接收到该经过软件滤波的位图,影响该图像在其他电子设备上的显示效果。因此,电子设备需要确认电子设备不处于拍摄分享、投屏的场景时,电子设备对位图数据进行软件滤波。
结合第一方面,在一些可能的实施方式中,电子设备对待显示内容的位图数据进行软件滤波,之后还包括:电子设备保存软件滤波后的待显示内容的位图数据。这里,电子设备保存该显示内容的位图,以供下一次调用。
在一些可能的实施方式中,方法还包括:电子设备调用保存的软件滤波后的待显示内容的位图数据;显示屏控制器对经过软件滤波后的位图数据进行子像素渲染,确定每个像素点的RGB灰度值,并根据RGB灰度值驱动OLED显示面板上的OLED发光点以显示待显示内容。由于电子设备保存了软件滤波后的待显示内容的位图数据,则在下一次对相同的待显示内容,电子设备可以直接调用软件滤波后的位图数据,即电子设备只需要对新出现的文字或字符或图像执行一次渲染和软件滤波操作,之后可以复用软件滤波结果,极大程度上减少了计算量。
结合第一方面,在一些可能的实施方式中,电子设备获取待显示内容的位图数据,包括:电子设备确定待显示内容的矢量图;电子设备对待显示内容的矢量图进行栅格化,获取待显示内容的位图数据。
结合第一方面,在一些可能的实施方式中,待显示内容的数据类型包括文字、字符、图像。
第二方面,本申请提供了一种电子设备,包括:一个或多个处理器、一个或多个存储器、显示屏和集成子像素渲染算法的显示屏控制器;该一个或多个存储与一个或多个处理器、显示屏、显示屏控制器耦合;该一个或多个存储器用于存储计算机程序代码,该计算机程序代码包括计算机指令;当该计算机指令在该处理器上运行时,使得该电子设备执行:
通过处理器获取待显示内容的位图数据;
通过处理器对位图数据进行软件滤波,其中,软件滤波基于子像素渲染算法确定,用于优化待显示内容在OLED显示面板上的显示边缘;
通过显示屏控制器对经过软件滤波后的位图数据进行子像素渲染,确定每个像素点的RGB灰度值;
通过显示屏控制器根据RGB灰度值驱动OLED显示面板上的OLED发光点以显示待显示内容。
本申请实施例中,待显示内容可以包括文字、字符、图像等等。子像素渲染算法是一个硬件算法,难以通过软件更新。电子设备在获取到待显示内容的位图数据后,在将该待显示 内容的位图数据发送到显示屏之前,将该位图数据进行软件滤波处理;然后将经过了软件滤波的位图数据通过子像素渲染算法进行子像素渲染,其中软件滤波的滤波参数是基于子像素渲染算法确定的,软件滤波和子像素渲染算法共同作用,能够优化待显示内容在OLED显示面板上的显示边缘,达到改变显示效果的作用。这样,解决了子像素渲染算法无法在线更新的困难。
结合第二方面,在一些可能的实施方式中,通过处理器对位图数据进行软件滤波包括:位图数据包括RGB三通道位图数据和单通道灰度位图数据;当位图数据为单通道灰度位图数据时,通过处理器将位图数据转换为RGB三通道位图数据;通过处理器对RGB三通道位图数据的红色像素灰度值和蓝色像素灰度值分别进行单像素通道软件滤波处理;位图数据的绿色像素灰度值保持不变。
结合第二方面,在一些可能的实施方式中,单像素通道软件滤波处理包括:单像素通道软件滤波处理与子像素渲染算法中的一维滤波处理方向垂直。即单像素通道软件滤波处理中的滤波参数是基于子像素渲染算法确定的,当子像素渲染算法中的滤波器为水平方向上的一维滤波器,则单像素通道软件滤波处理中的滤波参数为竖直方向上的一维滤波参数;当子像素渲染算法中的滤波器为竖直方向上的一维滤波器,则单像素通道软件滤波处理中的滤波参数为水平方向上的一维滤波参数。
在一些可能的实施方式中,一维滤波的公式包括:
Figure PCTCN2022073339-appb-000002
其中,dst(x,y)为输出的像素点(x,y)的像素灰度值,src(x,y)为输入的像素点(x,y)的像素灰度值。kernel为卷积核,例如水平方向上的一维滤波的滤波器核为(0.5,0.5),则kernel的值为(0.5,0.5),该滤波器核为1乘2的矩阵,则x^'的取值为0和1,即kernel.cols为2,y^'的取值为0,即kernel.cols为1。anchor为锚点,anchor.x为锚点的横坐标,anchor.y为锚点的纵坐标,例如在水平方向上的一维滤波的锚点可以为(0,0)。输出矩阵中的每个数值等于输入矩阵与卷积核对应位置元素相乘结果的和。
结合第二方面,在一些可能的实施方式中,在通过处理器对位图数据进行软件滤波之前,方法还包括:通过处理器基于显示屏的型号确定子像素渲染算法。其中,不同的显示屏的型号可以不同。电子设备通过显示屏的型号可以确定该显示屏中集成的子像素渲染算法,从而电子设备可以通过子像素渲染算法中采用的滤波方式确定电子设备的软件滤波的滤波参数或单像素通道软件滤波处理中的滤波参数。
结合第二方面,在一些可能的实施方式中,通过处理器对位图数据进行软件滤波包括:认电子设备不处于拍摄分享、投屏的场景时,通过处理器对位图数据进行软件滤波。即电子设备需要区分不同的场景来确定是否需要对位图数据进行软件滤波。当电子设备接收到拍摄分享、投屏操作,电子设备采集图像,并通过应用程序发送到其他电子设备。其中,电子设备向其他电子设备发送的图像数据为位图,电子设备若对采集到的图像进行预设的滤波操作,会导致发送的图像数据是经过软件滤波的位图。其他电子设备接收到该经过软件滤波的位图,影响该图像在其他电子设备上的显示效果。因此,电子设备需要确认电子设备不处于拍摄分享、投屏的场景时,电子设备对位图数据进行软件滤波。
结合第二方面,在一些可能的实施方式中,通过处理器对待显示内容的位图数据进行软件滤波,之后还包括:通过存储器保存软件滤波后的待显示内容的位图数据。这里,电子设 备保存该显示内容的位图,以供下一次调用。
结合第二方面,在一些可能的实施方式中,电子设备还执行:通过处理器调用保存的软件滤波后的待显示内容的位图数据;通过显示屏控制器对经过软件滤波后的位图数据进行子像素渲染,确定每个像素点的RGB灰度值,并根据RGB灰度值驱动OLED显示面板上的OLED发光点以显示待显示内容。由于电子设备保存了软件滤波后的待显示内容的位图数据,则在下一次对相同的待显示内容,电子设备可以直接调用软件滤波后的位图数据,即电子设备只需要对新出现的文字或字符或图像执行一次渲染和软件滤波操作,之后可以复用软件滤波结果,极大程度上减少了计算量。
结合第二方面,在一些可能的实施方式中,通过处理器获取待显示内容的位图数据,包括:通过处理器确定待显示内容的矢量图;通过处理器对待显示内容的矢量图进行栅格化,获取待显示内容的位图数据。
结合第二方面,在一些可能的实施方式中,待显示内容的数据类型包括文字、字符、图像。
第三方面,本申请实施例提供了一种计算机存储介质,包括计算机指令,当计算机指令在电子设备上运行时,使得通信装置执行上述任一方面任一项可能的实现方式中的屏幕显示方法。
第四方面,本申请实施例提供了一种计算机程序产品,当计算机程序产品在计算机上运行时,使得计算机执行上述任一方面任一项可能的实现方式中的屏幕显示方法。
附图说明
图1为本申请实施例提供的一种电子设备的结构示意图;
图2a和图2b为本申请实施例提供的一种子像素排列方式的示意图;
图3a和图3b为本申请实施例提供的一种界面显示原理的场景示意图;
图4为本申请实施例提供的一种子像素渲染算法的显示效果示意图;
图5a和图5b为本申请实施例提供的一种子像素渲染算法的滤波原理图;
图6a和图6b为本申请实施例提供的又一种子像素渲染算法的滤波原理图;
图7为本申请实施例提供的一种电子设备的结构示意图;
图8a和图8b为本申请实施例提供的一种子像素渲染算法的显示效果示意图;
图9为本申请实施例提供的一种屏幕显示方法的方法流程图;
图10a~图10c为本申请实施例提供的一种屏幕显示方法的显示效果示意图;
图11a~图11c为本申请实施例提供的一种屏幕显示方法的滤波原理图;
图12为本申请实施例提供的又一种子像素排列方式的示意图;
图13为本申请实施例提供的又一种子像素渲染算法的显示效果示意图;
图14为本申请实施例提供的一种屏幕显示方法的场景示意图。
具体实施方式
下面将结合附图对本申请实施例中的技术方案进行地描述。其中,在本申请实施例的描述中,除非另有说明,“/”表示或的意思,例如,A/B可以表示A或B;文本中的“和/或”仅仅是一种描述关联对象的关联关系,表示可以存在三种关系,例如,A和/或B,可以表示:单独存在A,同时存在A和B,单独存在B这三种情况,另外,在本申请实施例的描述中,“多个”是指两个或多于两个。
以下,术语“第一”、“第二”仅用于描述目的,而不能理解为暗示或暗示相对重要性或者隐含指明所指示的技术特征的数量。由此,限定有“第一”、“第二”的特征可以明示或者隐含地包括一个或者更多个该特征,在本申请实施例的描述中,除非另有说明,“多个”的含义是两个或两个以上。术语“中间”、“左”、“右”、“上”、“下”等指示的方位或位置关系为基于附图所示的方位或位置关系,仅是为了便于描述本申请和简化描述,而不是指示或暗示所指的装置或元件必须具有特定的方位、以特定的方位构造和操作,因此不能理解为对本申请的限制。
下面首先介绍本申请实施例中涉及的电子设备100。
参见图1,图1示出了本申请实施例提供的示例性电子设备100的结构示意图。
电子设备100可以是手机、平板电脑、桌面型计算机、膝上型计算机、手持计算机、笔记本电脑、超级移动个人计算机(ultra-mobile personal computer,UMPC)、上网本,以及蜂窝电话、个人数字助理(personal digital assistant,PDA)、增强现实(augmented reality,AR)设备、虚拟现实(virtual reality,VR)设备、人工智能(artificial intelligence,AI)设备、可穿戴式设备、车载设备、智能家居设备和/或智慧城市设备,本申请实施例对该电子设备的具体类型不作特殊限制。
电子设备100可以包括处理器110,外部存储器接口120,内部存储器121,通用串行总线(universal serial bus,USB)接口130,充电管理模块140,电源管理模块141,电池142,天线1,天线2,移动通信模块150,无线通信模块160,音频模块170,扬声器170A,受话器170B,麦克风170C,耳机接口170D,传感器模块180,按键190,马达191,指示器192,摄像头193,显示屏194,以及用户标识模块(subscriber identification module,SIM)卡接口195等。其中传感器模块180可以包括压力传感器180A,陀螺仪传感器180B,气压传感器180C,磁传感器180D,加速度传感器180E,距离传感器180F,接近光传感器180G,指纹传感器180H,温度传感器180J,触摸传感器180K,环境光传感器180L,骨传导传感器180M等。
可以理解的是,本发明实施例示意的结构并不构成对电子设备100的具体限定。在本申请另一些实施例中,电子设备100可以包括比图示更多或更少的部件,或者组合某些部件,或者拆分某些部件,或者不同的部件布置。图示的部件可以以硬件,软件或软件和硬件的组合实现。
处理器110可以包括一个或多个处理单元,例如:处理器110可以包括应用处理器(application processor,AP),调制解调处理器,图形处理器(graphics processing unit,GPU),图像信号处理器(image signal processor,ISP),控制器,视频编解码器,数字信号处理器(digital signal processor,DSP),基带处理器,和/或神经网络处理器(neural-network processing unit,NPU)等。其中,不同的处理单元可以是独立的器件,也可以集成在一个或多个处理器中。
控制器可以根据指令操作码和时序信号,产生操作控制信号,完成取指令和执行指令的控制。
处理器110中还可以设置存储器,用于存储指令和数据。在一些实施例中,处理器110中的存储器为高速缓冲存储器。该存储器可以保存处理器110刚用过或循环使用的指令或数据。如果处理器110需要再次使用该指令或数据,可从所述存储器中直接调用。避免了重复存取,减少了处理器110的等待时间,因而提高了系统的效率。
在一些实施例中,处理器110可以包括一个或多个接口。接口可以包括集成电路(inter-integrated circuit,I2C)接口,集成电路内置音频(inter-integrated circuit sound,I2S)接口, 脉冲编码调制(pulse code modulation,PCM)接口,通用异步收发传输器(universal asynchronous receiver/transmitter,UART)接口,移动产业处理器接口(mobile industry processor interface,MIPI),通用输入输出(general-purpose input/output,GPIO)接口,用户标识模块(subscriber identity module,SIM)接口,和/或通用串行总线(universal serial bus,USB)接口等。
I2C接口是一种双向同步串行总线,包括一根串行数据线(serial data line,SDA)和一根串行时钟线(derail clock line,SCL)。在一些实施例中,处理器110可以包含多组I2C总线。处理器110可以通过不同的I2C总线接口分别耦合触摸传感器180K,充电器,闪光灯,摄像头193等。例如:处理器110可以通过I2C接口耦合触摸传感器180K,使处理器110与触摸传感器180K通过I2C总线接口通信,实现电子设备100的触摸功能。
I2S接口可以用于音频通信。在一些实施例中,处理器110可以包含多组I2S总线。处理器110可以通过I2S总线与音频模块170耦合,实现处理器110与音频模块170之间的通信。在一些实施例中,音频模块170可以通过I2S接口向无线通信模块160传递音频信号,实现通过蓝牙耳机接听电话的功能。
PCM接口也可以用于音频通信,将模拟信号抽样,量化和编码。在一些实施例中,音频模块170与无线通信模块160可以通过PCM总线接口耦合。在一些实施例中,音频模块170也可以通过PCM接口向无线通信模块160传递音频信号,实现通过蓝牙耳机接听电话的功能。所述I2S接口和所述PCM接口都可以用于音频通信。
UART接口是一种通用串行数据总线,用于异步通信。该总线可以为双向通信总线。它将要传输的数据在串行通信与并行通信之间转换。在一些实施例中,UART接口通常被用于连接处理器110与无线通信模块160。例如:处理器110通过UART接口与无线通信模块160中的蓝牙模块通信,实现蓝牙功能。在一些实施例中,音频模块170可以通过UART接口向无线通信模块160传递音频信号,实现通过蓝牙耳机播放音乐的功能。
MIPI接口可以被用于连接处理器110与显示屏194,摄像头193等外围器件。MIPI接口包括摄像头串行接口(camera serial interface,CSI),显示屏串行接口(display serial interface,DSI)等。在一些实施例中,处理器110和摄像头193通过CSI接口通信,实现电子设备100的拍摄功能。处理器110和显示屏194通过DSI接口通信,实现电子设备100的显示功能。
GPIO接口可以通过软件配置。GPIO接口可以被配置为控制信号,也可被配置为数据信号。在一些实施例中,GPIO接口可以用于连接处理器110与摄像头193,显示屏194,无线通信模块160,音频模块170,传感器模块180等。GPIO接口还可以被配置为I2C接口,I2S接口,UART接口,MIPI接口等。
USB接口130是符合USB标准规范的接口,具体可以是Mini USB接口,Micro USB接口,USB Type C接口等。USB接口130可以用于连接充电器为电子设备100充电,也可以用于电子设备100与外围设备之间传输数据。也可以用于连接耳机,通过耳机播放音频。该接口还可以用于连接其他电子设备,例如AR设备等。
可以理解的是,本发明实施例示意的各模块间的接口连接关系,只是示意性说明,并不构成对电子设备100的结构限定。在本申请另一些实施例中,电子设备100也可以采用上述实施例中不同的接口连接方式,或多种接口连接方式的组合。
充电管理模块140用于从充电器接收充电输入。其中,充电器可以是无线充电器,也可以是有线充电器。在一些有线充电的实施例中,充电管理模块140可以通过USB接口130接收有线充电器的充电输入。在一些无线充电的实施例中,充电管理模块140可以通过电子设备100的无线充电线圈接收无线充电输入。充电管理模块140为电池142充电的同时,还可 以通过电源管理模块141为电子设备供电。
电源管理模块141用于连接电池142,充电管理模块140与处理器110。电源管理模块141接收电池142和/或充电管理模块140的输入,为处理器110,内部存储器121,显示屏194,摄像头193,和无线通信模块160等供电。电源管理模块141还可以用于监测电池容量,电池循环次数,电池健康状态(漏电,阻抗)等参数。在其他一些实施例中,电源管理模块141也可以设置于处理器110中。在另一些实施例中,电源管理模块141和充电管理模块140也可以设置于同一个器件中。
电子设备100的无线通信功能可以通过天线1,天线2,移动通信模块150,无线通信模块160,调制解调处理器以及基带处理器等实现。
天线1和天线2用于发射和接收电磁波信号。电子设备100中的每个天线可用于覆盖单个或多个通信频带。不同的天线还可以复用,以提高天线的利用率。例如:可以将天线1复用为无线局域网的分集天线。在另外一些实施例中,天线可以和调谐开关结合使用。
移动通信模块150可以提供应用在电子设备100上的包括2G/3G/4G/5G等无线通信的解决方案。移动通信模块150可以包括至少一个滤波器,开关,功率放大器,低噪声放大器(low noise amplifier,LNA)等。移动通信模块150可以由天线1接收电磁波,并对接收的电磁波进行滤波,放大等处理,传送至调制解调处理器进行解调。移动通信模块150还可以对经调制解调处理器调制后的信号放大,经天线1转为电磁波辐射出去。在一些实施例中,移动通信模块150的至少部分功能模块可以被设置于处理器110中。在一些实施例中,移动通信模块150的至少部分功能模块可以与处理器110的至少部分模块被设置在同一个器件中。
调制解调处理器可以包括调制器和解调器。其中,调制器用于将待发送的低频基带信号调制成中高频信号。解调器用于将接收的电磁波信号解调为低频基带信号。随后解调器将解调得到的低频基带信号传送至基带处理器处理。低频基带信号经基带处理器处理后,被传递给应用处理器。应用处理器通过音频设备(不限于扬声器170A,受话器170B等)输出声音信号,或通过显示屏194显示图像或视频。在一些实施例中,调制解调处理器可以是独立的器件。在另一些实施例中,调制解调处理器可以独立于处理器110,与移动通信模块150或其他功能模块设置在同一个器件中。
无线通信模块160可以提供应用在电子设备100上的包括无线局域网(wireless local area networks,WLAN)(如无线保真(wireless fidelity,Wi-Fi)网络),蓝牙(bluetooth,BT),全球导航卫星系统(global navigation satellite system,GNSS),调频(frequency modulation,FM),近距离无线通信技术(near field communication,NFC),红外技术(infrared,IR)等无线通信的解决方案。无线通信模块160可以是集成至少一个通信处理模块的一个或多个器件。无线通信模块160经由天线2接收电磁波,将电磁波信号调频以及滤波处理,将处理后的信号发送到处理器110。无线通信模块160还可以从处理器110接收待发送的信号,对其进行调频,放大,经天线2转为电磁波辐射出去。
在一些实施例中,电子设备100的天线1和移动通信模块150耦合,天线2和无线通信模块160耦合,使得电子设备100可以通过无线通信技术与网络以及其他设备通信。所述无线通信技术可以包括全球移动通讯系统(global system for mobile communications,GSM),通用分组无线服务(general packet radio service,GPRS),码分多址接入(code division multiple access,CDMA),宽带码分多址(wideband code division multiple access,WCDMA),时分码分多址(time-division code division multiple access,TD-SCDMA),长期演进(long term evolution,LTE),BT,GNSS,WLAN,NFC,FM,和/或IR技术等。所述GNSS可以包括全球卫星定 位系统(global positioning system,GPS),全球导航卫星系统(global navigation satellite system,GLONASS),北斗卫星导航系统(beidou navigation satellite system,BDS),准天顶卫星系统(quasi-zenith satellite system,QZSS)和/或星基增强系统(satellite based augmentation systems,SBAS)。
电子设备100通过GPU,显示屏194,以及应用处理器等实现显示功能。GPU为图像处理的微处理器,连接显示屏194和应用处理器。GPU用于执行数学和几何计算,用于图形渲染。处理器110可包括一个或多个GPU,其执行程序指令以生成或改变显示信息。
显示屏194用于显示图像,视频等。显示屏194包括显示面板。显示面板可以采用液晶显示屏(liquid crystal display,LCD),有机发光二极管(organic light-emitting diode,OLED),有源矩阵有机发光二极体或主动矩阵有机发光二极体(active-matrix organic light emitting diode的,AMOLED),柔性发光二极管(flex light-emitting diode,FLED),Miniled,MicroLed,Micro-oLed,量子点发光二极管(quantum dot light emitting diodes,QLED)等。在一些实施例中,电子设备100可以包括1个或N个显示屏194,N为大于1的正整数。
电子设备100可以通过ISP,摄像头193,视频编解码器,GPU,显示屏194以及应用处理器等实现拍摄功能。
ISP用于处理摄像头193反馈的数据。例如,拍照时,打开快门,光线通过镜头被传递到摄像头感光元件上,光信号转换为电信号,摄像头感光元件将所述电信号传递给ISP处理,转化为肉眼可见的图像。ISP还可以对图像的噪点,亮度,肤色进行算法优化。ISP还可以对拍摄场景的曝光,色温等参数优化。在一些实施例中,ISP可以设置在摄像头193中。
摄像头193用于捕获静态图像或视频。物体通过镜头生成光学图像投射到感光元件。感光元件可以是电荷耦合器件(charge coupled device,CCD)或互补金属氧化物半导体(complementary metal-oxide-semiconductor,CMOS)光电晶体管。感光元件把光信号转换成电信号,之后将电信号传递给ISP转换成数字图像信号。ISP将数字图像信号输出到DSP加工处理。DSP将数字图像信号转换成标准的RGB,YUV等格式的图像信号。在一些实施例中,电子设备100可以包括1个或N个摄像头193,N为大于1的正整数。
数字信号处理器用于处理数字信号,除了可以处理数字图像信号,还可以处理其他数字信号。例如,当电子设备100在频点选择时,数字信号处理器用于对频点能量进行傅里叶变换等。
视频编解码器用于对数字视频压缩或解压缩。电子设备100可以支持一种或多种视频编解码器。这样,电子设备100可以播放或录制多种编码格式的视频,例如:动态图像专家组(moving picture experts group,MPEG)1,MPEG2,MPEG3,MPEG4等。
NPU为神经网络(neural-network,NN)计算处理器,通过借鉴生物神经网络结构,例如借鉴人脑神经元之间传递模式,对输入信息快速处理,还可以不断的自学习。通过NPU可以实现电子设备100的智能认知等应用,例如:图像识别,人脸识别,语音识别,文本理解等。
内部存储器121可以包括一个或多个随机存取存储器(random access memory,RAM)和一个或多个非易失性存储器(non-volatile memory,NVM)。
随机存取存储器可以包括静态随机存储器(static random-access memory,SRAM)、动态随机存储器(dynamic random access memory,DRAM)、同步动态随机存储器(synchronous dynamic random access memory,SDRAM)、双倍资料率同步动态随机存取存储器(double data rate synchronous dynamic random access memory,DDR SDRAM,例如第五代DDR SDRAM一般称为DDR5 SDRAM)等;
非易失性存储器可以包括磁盘存储器件、快闪存储器(flash memory)。
快闪存储器按照运作原理划分可以包括NOR FLASH、NAND FLASH、3D NAND FLASH等,按照存储单元电位阶数划分可以包括单阶存储单元(single-level cell,SLC)、多阶存储单元(multi-level cell,MLC)、三阶储存单元(triple-level cell,TLC)、四阶储存单元(quad-level cell,QLC)等,按照存储规范划分可以包括通用闪存存储(英文:universal flash storage,UFS)、嵌入式多媒体存储卡(embedded multi media Card,eMMC)等。
随机存取存储器可以由处理器110直接进行读写,可以用于存储操作系统或其他正在运行中的程序的可执行程序(例如机器指令),还可以用于存储用户及应用程序的数据等。
非易失性存储器也可以存储可执行程序和存储用户及应用程序的数据等,可以提前加载到随机存取存储器中,用于处理器110直接进行读写。
外部存储器接口120可以用于连接外部的非易失性存储器,实现扩展电子设备100的存储能力。外部的非易失性存储器通过外部存储器接口120与处理器110通信,实现数据存储功能。例如将音乐,视频等文件保存在外部的非易失性存储器中。
电子设备100可以通过音频模块170,扬声器170A,受话器170B,麦克风170C,耳机接口170D,以及应用处理器等实现音频功能。例如音乐播放,录音等。
音频模块170用于将数字音频信息转换成模拟音频信号输出,也用于将模拟音频输入转换为数字音频信号。音频模块170还可以用于对音频信号编码和解码。在一些实施例中,音频模块170可以设置于处理器110中,或将音频模块170的部分功能模块设置于处理器110中。
扬声器170A,也称“喇叭”,用于将音频电信号转换为声音信号。电子设备100可以通过扬声器170A收听音乐,或收听免提通话。
受话器170B,也称“听筒”,用于将音频电信号转换成声音信号。当电子设备100接听电话或语音信息时,可以通过将受话器170B靠近人耳接听语音。
麦克风170C,也称“话筒”,“传声器”,用于将声音信号转换为电信号。当拨打电话或发送语音信息时,用户可以通过人嘴靠近麦克风170C发声,将声音信号输入到麦克风170C。电子设备100可以设置至少一个麦克风170C。在另一些实施例中,电子设备100可以设置两个麦克风170C,除了采集声音信号,还可以实现降噪功能。在另一些实施例中,电子设备100还可以设置三个,四个或更多麦克风170C,实现采集声音信号,降噪,还可以识别声音来源,实现定向录音功能等。
耳机接口170D用于连接有线耳机。耳机接口170D可以是USB接口130,也可以是3.5mm的开放移动电子设备平台(open mobile terminal platform,OMTP)标准接口,美国蜂窝电信工业协会(cellular telecommunications industry association of the USA,CTIA)标准接口。
压力传感器180A用于感受压力信号,可以将压力信号转换成电信号。在一些实施例中,压力传感器180A可以设置于显示屏194。压力传感器180A的种类很多,如电阻式压力传感器,电感式压力传感器,电容式压力传感器等。电容式压力传感器可以是包括至少两个具有导电材料的平行板。当有力作用于压力传感器180A,电极之间的电容改变。电子设备100根据电容的变化确定压力的强度。当有触摸操作作用于显示屏194,电子设备100根据压力传感器180A检测所述触摸操作强度。电子设备100也可以根据压力传感器180A的检测信号计算触摸的位置。在一些实施例中,作用于相同触摸位置,但不同触摸操作强度的触摸操作,可以对应不同的操作指令。例如:当有触摸操作强度小于第一压力阈值的触摸操作作用于短消息应用图标时,执行查看短消息的指令。当有触摸操作强度大于或等于第一压力阈值的触 摸操作作用于短消息应用图标时,执行新建短消息的指令。
陀螺仪传感器180B可以用于确定电子设备100的运动姿态。在一些实施例中,可以通过陀螺仪传感器180B确定电子设备100围绕三个轴(即,x,y和z轴)的角速度。陀螺仪传感器180B可以用于拍摄防抖。示例性的,当按下快门,陀螺仪传感器180B检测电子设备100抖动的角度,根据角度计算出镜头模组需要补偿的距离,让镜头通过反向运动抵消电子设备100的抖动,实现防抖。陀螺仪传感器180B还可以用于导航,体感游戏场景。
气压传感器180C用于测量气压。在一些实施例中,电子设备100通过气压传感器180C测得的气压值计算海拔高度,辅助定位和导航。
磁传感器180D包括霍尔传感器。电子设备100可以利用磁传感器180D检测翻盖皮套的开合。在一些实施例中,当电子设备100是翻盖机时,电子设备100可以根据磁传感器180D检测翻盖的开合。进而根据检测到的皮套的开合状态或翻盖的开合状态,设置翻盖自动解锁等特性。
加速度传感器180E可检测电子设备100在各个方向上(一般为三轴)加速度的大小。当电子设备100静止时可检测出重力的大小及方向。还可以用于识别电子设备姿态,应用于横竖屏切换,计步器等应用。
距离传感器180F,用于测量距离。电子设备100可以通过红外或激光测量距离。在一些实施例中,拍摄场景,电子设备100可以利用距离传感器180F测距以实现快速对焦。
接近光传感器180G可以包括例如发光二极管(LED)和光检测器,例如光电二极管。发光二极管可以是红外发光二极管。电子设备100通过发光二极管向外发射红外光。电子设备100使用光电二极管检测来自附近物体的红外反射光。当检测到充分的反射光时,可以确定电子设备100附近有物体。当检测到不充分的反射光时,电子设备100可以确定电子设备100附近没有物体。电子设备100可以利用接近光传感器180G检测用户手持电子设备100贴近耳朵通话,以便自动熄灭屏幕达到省电的目的。接近光传感器180G也可用于皮套模式,口袋模式自动解锁与锁屏。
环境光传感器180L用于感知环境光亮度。电子设备100可以根据感知的环境光亮度自适应调节显示屏194亮度。环境光传感器180L也可用于拍照时自动调节白平衡。环境光传感器180L还可以与接近光传感器180G配合,检测电子设备100是否在口袋里,以防误触。
指纹传感器180H用于采集指纹。电子设备100可以利用采集的指纹特性实现指纹解锁,访问应用锁,指纹拍照,指纹接听来电等。
温度传感器180J用于检测温度。在一些实施例中,电子设备100利用温度传感器180J检测的温度,执行温度处理策略。例如,当温度传感器180J上报的温度超过阈值,电子设备100执行降低位于温度传感器180J附近的处理器的性能,以便降低功耗实施热保护。在另一些实施例中,当温度低于另一阈值时,电子设备100对电池142加热,以避免低温导致电子设备100异常关机。在其他一些实施例中,当温度低于又一阈值时,电子设备100对电池142的输出电压执行升压,以避免低温导致的异常关机。
触摸传感器180K,也称“触控器件”。触摸传感器180K可以设置于显示屏194,由触摸传感器180K与显示屏194组成触摸屏,也称“触控屏”。触摸传感器180K用于检测作用于其上或附近的触摸操作。触摸传感器可以将检测到的触摸操作传递给应用处理器,以确定触摸事件类型。可以通过显示屏194提供与触摸操作相关的视觉输出。在另一些实施例中,触摸传感器180K也可以设置于电子设备100的表面,与显示屏194所处的位置不同。
骨传导传感器180M可以获取振动信号。在一些实施例中,骨传导传感器180M可以获 取人体声部振动骨块的振动信号。骨传导传感器180M也可以接触人体脉搏,接收血压跳动信号。在一些实施例中,骨传导传感器180M也可以设置于耳机中,结合成骨传导耳机。音频模块170可以基于所述骨传导传感器180M获取的声部振动骨块的振动信号,解析出语音信号,实现语音功能。应用处理器可以基于所述骨传导传感器180M获取的血压跳动信号解析心率信息,实现心率检测功能。
按键190包括开机键,音量键等。按键190可以是机械按键。也可以是触摸式按键。电子设备100可以接收按键输入,产生与电子设备100的用户设置以及功能控制有关的键信号输入。
马达191可以产生振动提示。马达191可以用于来电振动提示,也可以用于触摸振动反馈。例如,作用于不同应用(例如拍照,音频播放等)的触摸操作,可以对应不同的振动反馈效果。作用于显示屏194不同区域的触摸操作,马达191也可对应不同的振动反馈效果。不同的应用场景(例如:时间提醒,接收信息,闹钟,游戏等)也可以对应不同的振动反馈效果。触摸振动反馈效果还可以支持自定义。
指示器192可以是指示灯,可以用于指示充电状态,电量变化,也可以用于指示消息,未接来电,通知等。
SIM卡接口195用于连接SIM卡。SIM卡可以通过插入SIM卡接口195,或从SIM卡接口195拔出,实现和电子设备100的接触和分离。电子设备100可以支持1个或N个SIM卡接口,N为大于1的正整数。SIM卡接口195可以支持Nano SIM卡,Micro SIM卡,SIM卡等。同一个SIM卡接口195可以同时插入多张卡。所述多张卡的类型可以相同,也可以不同。SIM卡接口195也可以兼容不同类型的SIM卡。SIM卡接口195也可以兼容外部存储卡。电子设备100通过SIM卡和网络交互,实现通话以及数据通信等功能。在一些实施例中,电子设备100采用eSIM,即:嵌入式SIM卡。eSIM卡可以嵌在电子设备100中,不能和电子设备100分离。
本申请实施例提供了一种屏幕显示方法,能够改善屏幕显示效果。
为了便于理解,结合上述电子设备100的硬件结构,下面先对本申请实施例涉及的相关原理和相关概念进行介绍。
(1)显示屏194的种类。
显示屏194中每一个像素可以由红绿蓝(red green blue,RGB)三种子像素构成,其原理是红蓝绿(RGB)三基色可以构成任何一种颜色,当需要显示不同的颜色的时候,三种子像素分别以不同的显示亮度发光,在视觉上就可以混合成所需要的颜色。不同种类的显示屏上的子像素的排列方式也不同。
a.LCD屏
LCD屏的发光主要依赖其背光层,背光层的背光灯只有白色一种,所以就需要增加一层彩色滤光片来投射三基色。同时,为了控制红绿蓝三色的比例,LCD屏的背光层和彩色滤光之间存在一个调节电压大小的液晶层。该液晶层中的液滴都被包含在细小的单元格结构中,其中一个或多个单元格构成屏幕上的一个像素。
在LCD屏中,它的单个像素由三个单元格组成,分别透过这三个单元格在屏幕上显示出红色、蓝色和绿色。因此将这个三个单元格称为每个像素的红色子像素、绿色子像素和蓝色子像素。如图2a所示,图2a示例性的示出了LCD屏中子像素的排列方式。其中,G表示一个绿色子像素,B表示一个蓝色子像素,R表示一个红色子像素。一个绿色子像素G、一个 蓝色子像素B以及一个红色子像素R为一组,组成一个像素,三种颜色子像素的排列顺序不做限制。
b.OLED屏
OLED屏是利用有机电自发光二极管制成的显示屏。在OLED屏中,一个最小发光单位可以看作是一个子像素,每个子像素只可以发出单色(红色、蓝色或绿色)的光。发出红色光的子像素称为红色子像素,发出蓝色光的子像素称为蓝色子像素,发出绿色光的子像素称为绿色子像素。OLED屏的子像素的排列可以有多种排列方式,例如,一种排列方式如图2b所示,其中的单个像素点只有红绿或者蓝绿两个子像素组成,即每个像素均包含一个绿色子像素以及一个红色(或蓝色)子像素,包含红色子像素和包含蓝色子像素的像素点在水平和垂直方向间隔排列,根据图2b可以看出,同样显示3×3个像素,LCD屏中在水平方向上做了9个子像素,而这种OLED屏中在水平方向上只做了6个子像素,子像素数量减少了三分之一。
由于RGB三基色才能构成所有的颜色,两种颜色是不可以构成所有的颜色,所以在实际显示图像时,以图2b所示的排列方式的OLED屏中每一个像素会借用其周边像素的发光点来构成RGB三基色,这也可称为借色。例如在图2a中,第一像素点P1包括全部三种颜色的子像素,无需从相邻的第二像素点P2共用任何颜色的子像素;而在图2b中,第三像素点P3包括红色子像素和绿色子像素,可以从相邻的第四像素点P4共用蓝色子像素,这里,把P4中的蓝色子像素称为第三像素点P3的补偿子像素。其中,一个像素点可以和多个像素点共用子像素,第三像素点P3还可以和第五像素点P5共用蓝色子像素,第三像素点P3还可以和第六像素点P6共用红色子像素。
OLED屏的子像素的排列还有其他排列方式,这里不再一一列举。
(2)界面显示原理
电子设备100的显示界面的界面内容包括文字、图像等。如图3a所示,以由应用程序触发在设备100中显示“HUAWEI”为例来介绍,其中应用程序可以为阅读app、浏览类app、文本编辑应用等,例如,当用户在文件管理应用中,通过文本编辑应用打开一个包含“HUAWEI”字符内容的文件”文档1.txt”时,所述文本编辑应用要将所述字符显示到用户界面上,所以会通过调用系统中显示文字的接口,把“HUAWEI”对应的字符编码发送给电子设备的文字显示模块或框架进行显示,所述电子设备100的文字显示模块或框架根据“HUAWEI”对应的字符编码对“HUAWEI”这个字段进行字符排版,确定每个字符的显示形式(包括大小、颜色、字体、字间距、行间距等)和显示位置(可以是一个像素坐标,定位在显示位置的左上角)。
随后电子设备100在相应的字体文件中查询每个字符对应的显示数据(例如“E”这个字符,电子设备100在示例字体A的字体文件中查询“E”这个字在示例字体A这个字体中的显示数据),当前常用的字体文件类型为矢量字体,其中字符的显示数据是以矢量图的形式进行存储的,具体而言为通过二次或三次贝塞尔曲线来描述字符轮廓。
然后,电子设备100对每个字符的显示数据的矢量图进行渲染,渲染的过程包括栅格化,通过对矢量图进行栅格化,可以将矢量图转化为位图。位图是由多个像素点组成的,这些像素点可以进行不同的排列和染色以构成图样。这里,若该待显示图像为彩色图像,则该待显示图像的位图中包括每个像素点的RGB三个通道的灰度值数据(每个像素点的红色灰度值构成红色向量矩阵、蓝色灰度值构成蓝色向量矩阵以及绿色灰度值构成绿色向量矩阵);若该待显示图像为灰度图像,由于灰度图像的红绿蓝三基色的灰度值相同,则该待显示图像的位 图中包括每个像素点的一个通道的灰度值数据(一个灰度值向量矩阵),例如字符位图的像素点通常都只包括一个通道的灰度值数据。其中,灰度值的范围在0~255之间。
举例来说,一个图像的分辨率为500*338,且该图像为彩色图像,因为一个像素点的颜色是由RGB三个值来表现的,所以该图像的像素点矩阵对应三个灰度值向量矩阵,分别是红色(R)向量矩阵(500*338大小),绿色(G)向量矩阵(500*338大小),蓝色(B)向量矩阵(500*338大小)。如果每个矩阵的第一行第一列的值分别为:R:240,G:223,B:204,所以这个像素点的颜色就是(240,223,204)。
又例如,一个图像的分辨率为500*338,且该图像为灰度图像,因为该位图中每个像素点的红绿蓝三基色的灰度值相同,所以该图像的像素点矩阵对应一个向量矩阵(500*338大小),如果该矩阵的第一行第一列的值为240,则这个像素点的颜色就是(240,240,240)。
接着,电子设备100将渲染后的位图进行混合叠加和图层叠加。混合叠加指的是将位图贴在画面的对应的位置,该画面的大小和应用界面的显示画面的大小相同。其中,混合叠加的方式包括CPU绘图和GPU绘图。图层叠加指的是把多个应用的应用界面以及状态栏、导航栏、悬浮窗口等界面显示元素合并成待显示界面。该待显示界面的存储形式为带有灰度值数据的位图形式。
最后,电子设备100将叠加后的待显示界面发送到显示屏,显示屏的显示屏控制器基于子像素渲染算法点亮显示屏上的发光点(子像素点),显示屏上显示该待显示界面。示例性的,如图3b所示,图3b中以“E”这个字符为例,电子设备100获取到“E”这个字符的显示数据(矢量图),电子设备100对该矢量图进行渲染,生成由一个个像素点组成的位图,将该位图与其他显示内容进行叠加为待显示界面,最终发送到显示屏,基于子像素渲染算法点亮屏幕上的发光点。
子像素渲染算法可以集成在显示屏控制器中,定义显示器件如何根据输入数据点亮不同颜色的子像素。基于不同的子像素渲染算法,显示器件上点亮子像素的方式也是不同的,例如对于电子设备100的显示器件上的子像素排列方式为图2b中示出的子像素排列方式来说,如图4所示,输入一个单像素白点(单像素白点的RGB灰度值为(255,255,255)),通过子像素渲染算法一,显示器件上点亮三个子像素点(一个红色子像素、一个蓝色子像素和一个绿色子像素),可以使人眼感知到白色;通过子像素渲染算法二,显示器件上点亮五个子像素点(两个红色子像素、两个蓝色子像素和一个绿色子像素),可以使人眼感知到白色。当输入同一个单像素彩色点,通过子像素渲染算法一,显示器件上同样点亮图4所示的三个子像素点(红色子像素、蓝色子像素和绿色子像素),该三个子像素点的显示亮度是基于该单像素彩色点的RGB灰度值确定的;通过子像素渲染算法二,显示器件上同样点亮图4所示的五个子像素点,该五个子像素点的显示亮度是基于该单像素彩色点的RGB灰度值确定的。
图5a和图5b详细说明了子像素渲染算法一的技术原理。
电子设备100将待显示界面的位图发送到显示屏,以该显示屏上的子像素排列方式为图2b中示出的子像素排列方式为例,显示屏的显示屏控制器基于子像素渲染算法一点亮显示屏上的发光点(子像素点)。
如图5a所示,输入的待显示界面的位图包括RGB三通道的灰度值数据,而显示屏上一个像素只有两个发光点,分别对应一个绿色子像素和一个红色/蓝色子像素。图5a中以像素点(n-2,m)、像素点(n-1,m)、像素点(n,m)为例,通过子像素渲染算法一,像素点(n-1,m)可以和相邻像素点(n-2,m)共用蓝色子像素,像素点(n,m)可以和相邻像素点(n-1,m)共用红色子像素。绿色子像素一一对应,无需和其他像素点共用。
如图5b所示,通过子像素渲染算法一,确定显示屏上每个发光点对应的像素灰度值,像素点(n-1,m)和相邻像素点(n-2,m)共用蓝色子像素,那么像素点(n-1,m)的蓝色子像素灰度值为B(n-1,m)=0.5b(n-2,m)+0.5b(n-1,m);像素点(n,m)和相邻像素点(n-1,m)共用红色子像素,那么像素点(n,m)的红色子像素灰度值为R(n,m)=0.5r(n-1,m)+0.5r(n,m);绿色子像素灰度值为G(n,m)=g(n,m)。其中,图5b中子像素之间的共用参数为0.5只是一个示例,不构成对本申请实施例的限制。
图6a和图6b详细说明了子像素渲染算法二的技术原理。
电子设备100将待显示界面的位图发送到显示屏,以该显示屏上的子像素排列方式为图2b中示出的子像素排列方式为例,显示屏的显示屏控制器基于子像素渲染算法二点亮显示屏上的发光点(子像素点)。
如图6a所示,输入的待显示界面的位图包括RGB三通道的灰度值数据,而显示屏上一个像素只有两个发光点,分别对应一个绿色子像素和一个红色/蓝色子像素。图6a中以像素点(n-2,m)、像素点(n-1,m)、像素点(n,m)、像素点(n-2,m-1)、像素点(n-1,m-1)、像素点(n,m-1)为例,通过子像素渲染算法二,像素点(n-1,m)可以和周边像素点(n-2,m-1)、像素点(n-1,m-1)、像素点(n-2,m)共用蓝色子像素,像素点(n,m)可以和周边像素点(n-1,m-1)、像素点(n,m-1)、像素点(n-1,m)共用红色子像素。绿色子像素一一对应,无需和其他像素点共用。
如图6b所示,通过子像素渲染算法二,确定显示屏上每个发光点对应的像素灰度值,像素点(n-1,m)和周边像素点(n-2,m-1)、像素点(n-1,m-1)、像素点(n-2,m)共用蓝色子像素,那么像素点(n-1,m)的蓝色子像素灰度值为B(n-1,m)=0.25b(n-2,m-1)+0.25b(n-1,m-1)+0.25b(n-2,m)+0.25b(n-1,m);像素点(n,m)和周边像素点(n-1,m-1)、像素点(n,m-1)、像素点(n-1,m)共用红色子像素,那么像素点(n,m)的红色子像素灰度值为R(n,m)=0.25r(n-1,m-1)+0.25r(n,m-1)+0.25r(n-1,m)+0.25r(n,m);绿色子像素灰度值为G(n,m)=g(n,m)。其中,图6b中子像素之间的共用参数为0.25只是一个示例,不构成对本申请实施例的限制。
目前子像素渲染算法是以硬件算法的形式存在,固化在DDIC或处理芯片中,难以通过软件更新。
上述对字符的显示原理进行了示例性的描述,电子设备100的显示界面的界面内容包括字符和图像,同理于电子设备100对字符的显示原理,对于图像的显示原理来说来说,电子设备100对待显示图像进行图像解码获取到待显示的图像数据,电子设备100基于显示界面中待显示的图像数据电子设备100的图像显示模块或框架根据该图像数据对待显示图像进行排版,确定该待显示图像的显示形式(包括显示大小、颜色等)和显示位置(可以是一个像素坐标,定位在显示位置的左上角)。其中该图像数据是以矢量图或位图的形式进行存储的。然后,电子设备100对该待显示图像的矢量图进行渲染,生成位图。
接着,电子设备100将渲染后的位图进行混合叠加和图层叠加,将位图贴在画面的对应的位置,该画面的大小和应用的显示画面的大小相同,之后将多个应用的应用界面以及状态栏、导航栏等界面显示元素合并成待显示界面。最后,电子设备100将叠加后的待显示界面发送到显示屏,显示屏的显示屏控制器基于子像素渲染算法和显示屏上的发光点(子像素点)的排列方式,点亮显示屏上的发光点,显示屏上显示该待显示界面。
本申请实施例,一个白色单像素的显示需要红绿蓝三种颜色的子像素发光。对于LCD屏 来说,一个像素点包括三个颜色的子像素,因此一个像素点即可以显示出一个白色单像素;而对于上述OLED屏的钻石排列来说,一个像素点只有两个颜色的子像素,因此需要向周边的像素点借用子像素来显示出一个白色单像素。
基于上述界面显示原理,下面介绍一种带有显示屏幕的电子设备101的硬件结构,对OLED显示屏的显示原理中的子像素渲染算法进行进一步介绍。在一些实施例中,该电子设备101可以搭载或集成在电子设备100上,电子设备100可以包括电子设备101。
图7示出了一种电子设备101的硬件结构图,电子设备101包括应用处理器AP和显示模块,其中显示模块包括显示屏和显示驱动器集成电路(display driver integrated circuits,DDIC)。DDIC是显示屏控制器的主要元件之一,DDIC中集成有子像素渲染算法。DDIC用于驱动显示屏,并基于子像素渲染算法点亮显示屏上的发光点(子像素点)。其中显示屏可以是上述OLED屏,子像素点的排列方式可以采用上述的钻石排列。应用处理器AP将界面内容(其中包括文字、图像等)排版、渲染、叠加成位图,发送给DDIC。DDIC对接收到的位图进行滤波和采样,经过滤波和采样后,DDIC确定子像素的显示亮度,并将每个子像素的显示亮度变换成电压或电流等形式,驱动OLED屏,以相应的电压或电流点亮OLED屏上的发光点(子像素点)。在一些实施例中,DDIC和AP可以集成在同一个芯片中。可选的,DDIC也可以集成在OLED显示屏中。
其中,应用处理器AP发送给DDIC的位图为经过栅格化的位图,该位图由一个个像素点构成,位图中包括每个像素点的RGB三个通道的灰度值数据或者一个通道的灰度值数据。DDIC接收到应用处理器AP发送的位图后,DDIC将该位图分为三个通道分别进行处理,该三个通道为绿色通道、红色通道和蓝色通道,结合子像素渲染算法(包括进行滤波采样等)和显示屏的子像素排列方式确定显示屏上子像素的显示亮度,DDIC将每一个子像素的显示亮度变换为电压或电流等形式,驱动显示屏并点亮显示屏上相应的发光点。
例如在图7中,DDIC将对位图的处理过程划分为绿色通道、红色通道和蓝色通道,绿色通道中处理该位图的绿色向量矩阵(包括每个像素点的绿色子像素的像素灰度值),红色通道中处理该位图的红色向量矩阵(包括每个像素点的红色子像素的像素灰度值),蓝色通道中处理该位图的蓝色向量矩阵(包括每个像素点的蓝色子像素的像素灰度值)。由于OLED屏的钻石排列中每两个像素点才有一个红色子像素和一个蓝色子像素,因此红色子像素和蓝色子像素需要作为周边像素的补偿像素点,红色子像素和蓝色子像素的像素灰度值与其周边的像素点有关。即针对于红色通道和蓝色通道,电子设备100需要将位图进行滤波和采样,确定出OLED屏上红色子像素和蓝色子像素的像素灰度值。因此,在红色通道和蓝色通道中,DDIC需要将输入的位图进行滤波和采样,然后基于滤波和采样后的位图,DDIC将该滤波后的位图的像素灰度值变换为电压或电流等形式,驱动OLED屏,以相应的电压或电流点亮OLED屏上的红色发光点和蓝色发光点。又由于OLED屏的钻石排列中每一个像素点都有绿色子像素,绿色子像素的显示亮度与其他像素点无关,因此对于绿色通道来说无需进行滤波和采样,DDIC可以直接将位图输入到绿色通道,将该位图的像素灰度值变换为电压或电流等形式,驱动OLED屏,以相应的电压或电流点亮OLED屏上的绿色发光点。
针对于上述滤波的方式,下面示例性的介绍两种滤波方式,包括一维滤波和二维滤波。其中,
一维滤波可以是水平方向上的滤波或垂直方向上的滤波,后续简称为1D-SPR。
一维滤波的公式可以为:
Figure PCTCN2022073339-appb-000003
其中,dst(x,y)为输出的像素点(x,y)的像素灰度值,src(x,y)为输入的像素点(x,y)的像素灰度值。kernel为卷积核,例如上述水平方向上的一维滤波的滤波器核为(0.5,0.5),则kernel的值为(0.5,0.5),该滤波器核为1乘2的矩阵,则x′的取值为0和1,即kernel.cols为2,y′的取值为0,即kernel.cols为1。anchor为锚点,anchor.x为锚点的横坐标,anchor.y为锚点的纵坐标,例如在上述水平方向上的一维滤波的锚点为(0,0)。输出矩阵中的每个数值等于输入矩阵与卷积核对应位置元素相乘结果的和。
以水平方向上的滤波为例,其滤波器核可以描述为(0.5,0.5)。例如对于红色通道来说,电子设备100对输入位图进行滤波器核为(0.5,0.5)的水平滤波,滤波结果中每个像素的灰度值是输入位图中相邻两个像素的灰度值的平均,即L_r(x,y)=(I_r(x,y)+I_r(x+1,y))/2,其中L_r(x,y)为输出的像素点(x,y)的红色像素灰度值,I_r(x,y)为输入的像素点(x,y)的红色像素灰度值,锚点anchor的横坐标为0,纵坐标为0,kernel卷积核的值为(0.5,0.5),则基于上述一维滤波的公式可以得出像素点(x,y)的红色像素灰度值L_r(x,y)=(I_r(x,y)+I_r(x+1,y))/2。
这里,可以参考上述图5a和图5b示出的子像素渲染算法一的滤波方式即为水平方向上的一维滤波,滤波器核为(0.5,0.5)。
将上述滤波结果进行采样、驱动等步骤,基于红色通道中每个像素的灰度值,并结合显示屏上子像素排列方式,DDIC可以确定出显示屏上每一个红色子像素的显示亮度。即每个红色/蓝色子像素的显示亮度与输入通道内位图中相邻两个像素的灰度值相关。
如图8a所示,经过水平滤波和采样后,通过红、绿、蓝三个子像素即可以显示出一个白点,可以看出,由于只在水平方向进行滤波,在水平线以及文字水平笔画边缘处会存在暴露在边缘的绿色子像素,当电子设备100显示水平方向的黑白交界线时,在视觉上会导致彩边(绿边)现象。由于人眼对绿色最为敏感,暴露在边缘的绿色子像素会产生较明显的绿边,并会有较强的颗粒感。
二维滤波,在水平方向和垂直方向均进行滤波,后续简称为2D-SPR。示例性的,二维滤波的滤波器核可以描述为
Figure PCTCN2022073339-appb-000004
例如对于红色通道来说,电子设备100对输入位图进行滤波器核为
Figure PCTCN2022073339-appb-000005
的二维滤波,滤波结果中每个像素的灰度值是输入位图中周边2x2区域内,四个像素的灰度值的平均,即L_r(x,y)=(I_r(x,y)+I_r(x+1,y)+I_r(x,y+1)+I_r(x+1,y+1))/4。这里,可以参考上述图6a和图6b示出的子像素渲染算法二的滤波方式即为二维滤波,滤波器核为
Figure PCTCN2022073339-appb-000006
将上述滤波结果进行采样、驱动等步骤,基于红色通道中每个像素的灰度值,并结合显示屏上子像素排列方式,DDIC可以确定出显示屏上每一个红色子像素的显示亮度。即每个红色/蓝色子像素的显示亮度与输入通道内位图中水平方向和垂直方向相邻的四个像素点的灰度值相关。
如图8b所示,经过二维滤波和采样后,通过两个红色子像素、两个蓝色子像素、一个绿色子像素即可以显示出一个白点,可以看出,所有的绿色子像素均被红、蓝子像素包围,且在边缘处亮度较低(25%的红色显示亮度和25%的蓝色显示亮度),有一定的亮度渐变效果。此时当电子设备100显示水平方向的黑白交界线,在视觉上没有彩边(绿边)现象。
综上所述,子像素渲染算法决定了DDIC如何根据输入数据点亮显示屏上不同颜色的子像素,然而子像素渲染算法是一个硬件算法,无法通过软件更新。那么在水平滤波的情况下, 当以钻石排列的OLED屏中显示水平方向的黑白交界线(例如文字水平笔画边缘)时,在视觉上会导致彩边现象。搭载有电子设备101的电子设备100在选型、制造、生产、销售后,无法通过软件更新的方式改变电子设备101的子像素排列和子像素渲染算法,难以优化显示效果。
本申请实施例,提供了一种屏幕显示方法,能够在不改变显示屏的子像素排列并且不更改子像素渲染算法的情况下,解决彩边现象,改善显示屏的屏幕显示效果。
接下来介绍本申请提供的一种屏幕显示方法的步骤流程,如图9所示,可包括:
步骤S101、电子设备100确定待显示内容的矢量图。
电子设备100获取待显示内容,其中该待显示内容可以是文字、字符、图像等。电子设备100确定该待显示内容的显示形式(显示位置、大小、颜色等),形成待显示内容的矢量图。
可选的,电子设备100基于该待显示内容中待显示的图像数据和显示位置确定图像在显示界面中的显示形式,基于该待显示内容中待显示的文字数据、显示位置以及字体文件确定文字在显示界面中的显示形式,从而形成该待显示内容的矢量图。
步骤S102、电子设备100对该待显示内容的矢量图进行渲染,生成该待显示内容的位图。
电子设备100确定待显示内容的矢量图后,对该待显示内容的矢量图进行渲染,即对矢量图进行栅格化,将矢量图表示的图像转化为位图,位图是由多个像素点组成的,位图中包括每个像素点的RGB三个通道的灰度值数据或者一个通道的灰度值数据。这里,电子设备100生成该待显示内容的位图。
在一些实施例中,上述步骤S101和步骤S102为可选的,电子设备100可以直接获取待显示内容的位图。例如电子设备100接收到其他设备发送的待显示内容的位图;又例如电子设备100调用自身存储的待显示内容的位图;等等。
步骤S103、电子设备100对该待显示内容的位图进行软件滤波。
电子设备100生成该待显示内容的位图后,电子设备100对该位图进行软件滤波处理,生成软件滤波后的位图。其中,软件滤波的方式有多种,本申请实施例不作限制。在一些实施例中,这里的软件滤波的滤波参数是基于子像素渲染算法确定的,该软件滤波用于优化待显示内容在显示屏的显示面板上的显示边缘。
可选的,软件滤波的处理方向和子像素渲染算法中的一维滤波处理方向垂直。
举例来说,当DDIC中的子像素渲染算法中采用的重采样滤波器核为(0.5,0.5),则电子设备100此时对该待显示内容的位图进行软件滤波,滤波器核为
Figure PCTCN2022073339-appb-000007
其中,软件滤波
Figure PCTCN2022073339-appb-000008
与重采样滤波器(0.5,0.5)共同组合为虚拟的重采样滤波器
Figure PCTCN2022073339-appb-000009
当子像素渲染算法中采用的重采样滤波器核为
Figure PCTCN2022073339-appb-000010
则电子设备100此时对该待显示内容的位图进行滤波,滤波器核为(0.5,0.5)。其中,软件滤波(0.5,0.5)与重采样滤波器
Figure PCTCN2022073339-appb-000011
共同组合为虚拟的重采样滤波器
Figure PCTCN2022073339-appb-000012
可以看出,上述两种情况均是为了让最终的滤波结果达到二维滤波的效果。
在一些实施例中,该待显示内容的位图包括RGB三通道位图数据和单通道灰度位图数据。其中,RGB三通道位图数据包括每个像素点的RGB三个通道的灰度值数据(每个像素点的红色像素灰度值、蓝色像素灰度值和绿色像素灰度值);由于灰度位图中每个像素点的红色 像素灰度值、蓝色像素灰度值和绿色像素灰度值相同,因此单通道灰度位图中包括每个像素点的一个通道的灰度值数据。其中每个像素点的红色像素灰度值构成了位图数据中的红色通道,每个像素点的蓝色像素灰度值构成了位图数据中的蓝色通道,每个像素点的绿色像素灰度值构成了位图数据中的绿色通道。
当该待显示内容的位图包括RGB三通道位图数据,电子设备对该位图数据中红色像素灰度值和蓝色像素灰度值分别进行单像素通道软件滤波处理,确定滤波后的红色像素灰度值和蓝色像素灰度值;其中绿色像素灰度值保持不变。单像素通道软件滤波处理指的是对RGB三通道位图数据中的其中一个通道进行软件滤波处理。此时,电子设备100获取到滤波后的位图。
示例性的,该待显示内容的位图中包括每个像素点的第一红色像素灰度值、第一蓝色像素灰度值和第一绿色像素灰度值。电子设备100对该位图中红色通道和蓝色通道分别进行滤波,获取到滤波后的位图,该滤波后的位图中包括每个像素点的第二红色像素灰度值、第二蓝色像素灰度值和第一绿色像素灰度值。
当该待显示内容的位图包括单通道灰度位图数据,电子设备首先将该单通道灰度位图数据转化为RGB三通道位图数据,若原始位图(该待显示内容的位图)的尺寸为[w,h];电子设备100分配尺寸为[w,h,3]的内存将单通道灰度位图数据转化为RGB三通道位图数据(即每个像素点的红色像素灰度值、蓝色像素灰度值和绿色像素灰度值相同),其中,电子设备对该RGB三通道位图数据中红色像素灰度值和蓝色像素灰度值分别进行单像素通道软件滤波处理,确定滤波后的红色像素灰度值和蓝色像素灰度值;其中绿色像素灰度值保持不变。此时,电子设备100获取到滤波后的位图。该滤波后的位图包括RGB三通道位图数据。
示例性的,该待显示内容的位图中包括每个像素点的第一像素灰度值,即每个像素点的红色像素灰度值、蓝色像素灰度值和绿色像素灰度值均为第一像素灰度值。首先电子设备将该灰度位图拷贝到绿色通道,然后电子设备100将该灰度位图进行滤波,获取到第二像素灰度值,电子设备100将滤波后的灰度位图分别拷贝到红色通道和蓝色通道,获取到滤波后的位图。该滤波后的位图中包括每个像素点的红色像素灰度值(第二像素灰度值)、蓝色像素灰度值(第二像素灰度值)、绿色像素灰度值(第一像素灰度值)。
示例性的,下面以显示字母“E”为例,使用思源黑体,字体大小为8号,经栅格化和gamma映射后得到如下位图。其中该位图大小为7*11,字符位图通常包括单通道灰度位图数据,具体的位图数据如表一所示。
表一
Figure PCTCN2022073339-appb-000013
Figure PCTCN2022073339-appb-000014
对上述字母“E”的位图数据进行软件滤波,其中,电子设备对红色像素灰度值和蓝色像素灰度值进行单像素通道软件滤波处理,位图数据的绿色像素灰度值保持不变,得到如表二所示的经过软件滤波后的位图数据。
表二
Figure PCTCN2022073339-appb-000015
Figure PCTCN2022073339-appb-000016
在一些实施例中,如果待显示的位图处于非线性域,则需要在进行软件滤波前进行degamma映射(从非线性域变换到线性域),并在经过软件滤波后重新进行gamma映射(从线性域变换回非线性域)。如果待显示位图处于线性域,则直接进行滤波。在实际情况中,degamma映射和gamma映射可能会导致数据的精度损失,产生误差。
在一些实施例中,上述软件滤波方式中的滤波参数是结合电子设备100的显示屏的信息来确定的。显示屏的信息包括显示屏的型号、数量、以及显示屏的显示方向等。其中,
不同的显示屏的型号可以不同。电子设备100通过显示屏的型号可以确定该显示屏中集成的子像素渲染算法,电子设备100通过子像素渲染算法中采用的滤波方式确定电子设备100的软件滤波方式。
电子设备100中显示屏的数量可以是多个。该多个显示屏的型号可以不同,针对不同型号的显示屏,电子设备100的软件滤波方式也不同。当电子设备100在切换显示屏进行显示时,可以切换电子设备100对待显示内容的软件滤波方式。电子设备100基于待显示内容当前显示的显示屏,确定对该待显示内容的软件滤波方式。
电子设备100的显示屏的显示方向包括横屏(90度和270度)和竖屏(0度和180度)。电子设备100基于当前显示屏的显示方向,确定电子设备100的软件滤波方式。由于子像素渲染算法是一个硬件算法,子像素渲染算法中的滤波方向是固定的,当显示屏的显示方向进行改变,待显示内容的排版也发生了改变,此时软件滤波的滤波参数也要相应的改变。举例来说,当子像素渲染算法中采用的重采样滤波器核为(0.5,0.5),若电子设备100当前的显示方向为竖屏(0度),则电子设备100的软件滤波的滤波器核为
Figure PCTCN2022073339-appb-000017
若电子设备100当前的显示方向为横屏(90度),则电子设备100的软件滤波的滤波器核为(0.5,0.5)。可选的,滤波参数不只是滤波器核,不同角度的横屏的滤波参数也不同,不同角度的竖屏的滤波参数也不同。例如两种竖屏形式,电子设备100的显示方向为0度和180度时的滤波器核都为(0.5,0.5),但一个是当前像素和其左侧像素的平均,另一个是当前像素和右侧像素的平均,产生的滤波结果不同。
这里,电子设备100基于电子设备的型号、电子设备(显示屏)状态的变化(如不同显示屏之间的切换、显示屏的显示方向改变等),改变软件滤波的滤波参数。这样能够解决不同情况下显示屏的显示效果,提高本申请的实用性和灵活性。
下面以电子设备100对单通道位图(位图中每个像素点包括一个通道的灰度信息)进行软件滤波处理,该软件滤波的滤波器核为(0.5,0.5)为例,示例性的代码如下所示:
Figure PCTCN2022073339-appb-000018
以显示“王”字为例,如图10c所示,图10c示例性的示出了在滤波前后提取出的红色通道或蓝色通道中的位图数据的示意性的显示效果。在滤波之前,“王”字的竖直笔画上存在绿边;在经过滤波器核为(0.5,0.5)的滤波器水平滤波后,红色通道和蓝色通道中“王”字的竖直笔画变宽了,因为在水平方向上的滤波,从而水平方向上相邻像素点中的红色子像素和蓝色子像素进行颜色补偿,从而达到了在竖直方向上消除绿边的效果。
同理,若将该位图经过滤波器核为
Figure PCTCN2022073339-appb-000019
的滤波器竖直滤波后,“王”字的水平笔画会变宽(粗),因为在竖直方向上的滤波,从而竖直方向上相邻像素点中的红色子像素和蓝色子像素进行颜色补偿,从而能达到在水平方向上消除绿边的效果。
随后电子设备100将滤波后的内容送交至显示器,显示器中子像素渲染算法的滤波器核为
Figure PCTCN2022073339-appb-000020
经过子像素渲染算法的重采样滤波器的二次滤波和采样过程后,DDIC驱动显示屏,点亮屏幕上子像素。可以看出,上述水平滤波(0.5,0.5)与重采样滤波器
Figure PCTCN2022073339-appb-000021
共同组合为虚拟的重采样滤波器
Figure PCTCN2022073339-appb-000022
即该方法可以使1D SPR电子设备取得与2D SPR电子设备相近的显示效果。
步骤S104、电子设备100将该滤波后的位图进行叠加。
电子设备100获取到滤波后的位图后,将该位图贴在画面的对应的位置,该画面的大小和应用的显示画面的大小相同,以使该位图能够以显示屏的显示画面的大小显示在电子设备的显示屏上。其中,混合叠加的方式包括CPU绘图和GPU绘图。之后电子设备100将多个应用的应用界面以及状态栏、导航栏等界面显示元素合并成待显示界面。
步骤S105、电子设备100基于该叠加后的位图,通过子像素渲染算法点亮显示屏上的发 光点,显示该待显示内容。
电子设备100将叠加后的位图发送到显示屏,显示屏的显示屏控制器(DDIC)基于该叠加后的位图,通过子像素渲染算法点亮显示屏上的发光点(子像素点),从而电子设备100的显示屏上显示该待显示内容。
下面具体说明电子设备100对叠加后的位图通过子像素渲染算法进行处理的过程。
叠加后的位图包括了RGB三个通道的灰度值信息(每个像素点的红色灰度值、蓝色灰度值和绿色灰度值)。针对于红色通道和蓝色通道,电子设备100需要对输入位图(叠加后的位图)进行滤波,确定出红色子像素和蓝色子像素的像素灰度值。由于OLED屏的钻石排列中每两个像素点才有一个红色子像素和一个蓝色子像素,因此红色子像素和蓝色子像素需要作为周边像素的补偿像素点,红色子像素和蓝色子像素的显示亮度与其周边的像素点有关。步骤S103中电子设备100对红色通道和蓝色通道进行了一次软件滤波,这里基于子像素渲染算法电子设备100对红色通道和蓝色通道进行二次滤波,确定出红色子像素和蓝色子像素的像素灰度值。因此,在红色通道和蓝色通道中,基于滤波和采样后的位图,DDIC将该滤波后的位图的像素灰度值变换为电压或电流等形式,驱动OLED屏,以相应的电压或电流驱动OLED屏上的红色发光点和蓝色发光点发光。需要注意的是,这里的滤波方式和上述步骤S103中的滤波方式不同,这里的滤波方式为电子设备100中集成在DDIC的子像素渲染算法决定的,难以通过软件更新的方式进行改变。
对于绿色通道来说无需进行上述重采样滤波的过程,DDIC可以直接将叠加后的位图输入到绿色通道,将该叠加后的位图上的绿色子像素的像素灰度值变换为电压或电流等形式,驱动显示屏,以相应的电压或电流点亮显示屏上的绿色发光点。电子设备100显示该叠加后的位图对应的待显示内容。
本申请实施例中,电子设备100对发送到显示屏之前的待显示内容进行预先滤波处理(软件滤波),和子像素渲染算法中的重采样滤波器共同作用,达到改变显示效果的作用。本申请实施例解决了子像素渲染算法无法在线更新的困难。
下面以软件滤波的滤波器核为
Figure PCTCN2022073339-appb-000023
DDIC中的子像素渲染算法中采用的重采样滤波器核为(0.5,0.5)为例,当显示屏上的子像素排列方式为图2b中示出的子像素排列方式,详细说明软件滤波和子像素渲染算法共同组合实现的滤波效果。其中,软件滤波
Figure PCTCN2022073339-appb-000024
与重采样滤波器(0.5,0.5)共同组合为虚拟的重采样滤波器
Figure PCTCN2022073339-appb-000025
如图11a所示,电子设备100生成该待显示内容的位图后,电子设备100对该位图进行软件滤波处理。待显示内容的位图包括RGB三通道的灰度值数据,电子设备100对红色通道和蓝色通道进行软件滤波,滤波器核为
Figure PCTCN2022073339-appb-000026
确定位图中每个像素的灰度值。其中,滤波后像素点(n,m-1)的蓝色子像素灰度值为B(n,m-1)=0.5b(n,m-2)+0.5b(n,m-1);滤波后像素点(n,m)的红色子像素灰度值为R(n,m)=0.5r(n,m-1)+0.5r(n,m);绿色子像素灰度值为G(n,m)=g(n,m)。
如图11b所示,电子设备100将滤波后的位图发送到显示屏,显示屏的显示屏控制器基于子像素渲染算法点亮显示屏上的发光点(子像素点)。该子像素渲染算法的滤波器核为(0.5,0.5)。其中,通过该子像素渲染算法,像素点(n-1,m)可以和相邻像素点(n-2,m)共用蓝色子像素,像素点(n,m)可以和相邻像素点(n-1,m)共用红色子像素。绿色子像素一一对应,无需和其他像素点共用。图11b中,通过该子像素渲染算法,像素点(n-1,m)的蓝色子像素灰度值为B1(n-1,m)=0.5B(n-2,m)+0.5B(n-1,m);像素点(n,m)的红色子像素 灰度值为R1(n,m)=0.5R(n-1,m)+0.5R(n,m);绿色子像素灰度值为G(n,m)=g(n,m)。
其中,如图11c中可以看出,软件滤波
Figure PCTCN2022073339-appb-000027
与重采样滤波器(0.5,0.5)共同组合为虚拟的重采样滤波器
Figure PCTCN2022073339-appb-000028
即该方法可以使1D SPR电子设备取得与2D SPR电子设备相近的显示效果,在不改变子像素渲染算法的情况下,解决了子像素渲染算法为一维滤波时会产生的彩边现象。
在一些实施例中,上述预处理软件滤波算法(步骤S103)可以发生在渲染及送显(发送到显示屏)过程中的任一阶段。对待显示内容为文字来说,预处理软件滤波算法可以发生在栅格化、混合叠加、图层叠加、送显中任意两个步骤之间。对待显示内容为图像、UI图形、3D画面来说,预处理软件滤波算法可以发生在图像解码、栅格化、混合叠加、图层叠加、送显中任意两个步骤之间。
上述方法流程中步骤S103集成在栅格化流程(步骤S102)之后、混合叠加过程(步骤S104)之前。若步骤S103执行在叠加过程(步骤S104)之后、送显(步骤S105)之前,电子设备100获取到待显示内容的位图后,进行混合叠加和图层叠加过程,然后对叠加后的位图进行软件滤波,电子设备100将滤波后的位图发送到显示屏,通过子像素渲染算法点亮显示屏上的发光点,显示该待显示内容。这里,电子设备100在对叠加后的位图进行软件滤波之前,需要将该叠加后的位图变换为线性域上,然后进行软件滤波之后,再将滤波后的位图重新变换回非线性域。
在一些实施例中,电子设备100根据待显示内容的不同性质(如文字、图像、矢量图形),采用不同的处理方式。可选的,电子设备100只对文字或字符执行步骤S103。由于与显示界面中其他UI元素相比,文字笔画边缘具有较高对比度,彩边和颗粒感等情况在文字笔画边缘最为明显,我们可以仅对UI界面中的文字部分进行处理,可以在不太影响显示效果的情况下,节约电子设备100的计算资源。
在一些实施例中,在上述步骤S101中,电子设备100获取待显示内容,并确定该待显示内容的显示形式。电子设备100基于该待显示内容,查询电子设备100中是否存储了该待显示内容的位图,若待显示内容为文字或字符,电子设备100基于该待显示内容和显示形式(该待显示内容的字体、尺寸等),查询电子设备100中是否存储了该显示形式下的待显示内容的位图;若待显示内容为图像,电子设备100基于该待显示内容和显示形式(该图像的显示大小等),查询电子设备100中是否存储了该显示形式下的该待显示内容的位图。
若存储了,则电子设备100调用该待显示内容的位图,直接执行步骤S104。可选的,若没有存储,则电子设备100执行步骤S102和步骤S103,电子设备对该待显示内容进行渲染和滤波之后,生成该待显示内容的位图,这里,电子设备100保存该显示内容的位图,以供下一次调用。
这样,电子设备100只需要对新出现的文字或字符或图像执行一次渲染和软件滤波操作,之后可以复用软件滤波结果,极大程度上减少了计算量。
在一些实施例中,在电子设备100执行上述步骤S103之前,电子设备100基于电子设备100的当前状态,确定是否执行步骤S103。当电子设备100当前状态满足预设条件时,电子设备100执行步骤S103。
该预设条件包括电子设备100当前不处于拍摄分享、投屏等场景下。示例性的,当电子 设备100接收到拍摄分享操作,电子设备100采集图像,并通过应用程序发送到其他电子设备。其中,电子设备100向其他电子设备发送的图像数据为位图,电子设备100若对采集到的图像进行预设的滤波操作,会导致发送的图像数据是经过软件滤波的位图。例如预设的软件滤波的滤波器核为(0.5,0.5),则经过该软件滤波的图像为带有绿边的图像(例如图6a所示)。其他电子设备接收到该带有绿边的图像,影响该图像在其他电子设备上的显示效果。因此,当电子设备100当前不处于拍摄分享场景下,电子设备100执行步骤S103。
同理,当电子设备100接收到开启投屏操作,电子设备100将采集到的图像实时发送到其他电子设备上显示出来,电子设备100若对采集到的图像进行预设的滤波操作,会导致发送的图像数据是经过软件滤波的位图。其他电子设备接收到该经过软件滤波的位图,影响该位图在其他电子设备上的显示效果。因此,当电子设备100当前不处于投屏场景下,电子设备100执行步骤S103。可选的,当电子设备100接收到关闭投屏操作,电子设备100恢复执行步骤S103。
综上,上述实施例描述的改善显示屏的显示效果的方式,可以适用于电子设备100的显示屏为OLED屏,且子像素的排列方式为钻石排列的情况。其中OLED屏的子像素的排列方式有多种,不限于钻石排列,本申请实施例提供的屏幕显示方法,也能够适用于其他子像素排列方式的显示屏,改善不同排列方式的显示屏的显示效果。
如图12所示,图12示例性的示出了OLED显示屏的又一种子像素排列方式。图12中示出的排列方式中包括两种像素点,一种像素点包括一个红色子像素和一个蓝色子像素,另一种像素点包括两个绿色子像素,这两种像素点在水平和垂直方向间隔排列。在这种排列方式下,不同的子像素渲染算法能够实现的显示效果不同。下面示例性的简要介绍适用于这种排列方式的四种子像素渲染算法。
如图13所示,当输入一个单像素白点(单像素白点的RGB灰度值为(255,255,255)),通过子像素渲染算法三,显示器件上点亮五个子像素点,包括两个红色子像素、两个蓝色子像素和一个绿色子像素;两个红色子像素的亮度相同,两个蓝色子像素的亮度相同。其中,子像素渲染算法三实现的点亮方式可以使人眼感知到白色,产生的实际效果为清晰度高,但是颗粒感严重和彩边明显。
通过子像素渲染算法四,显示器件上点亮八个子像素点,包括两个红色子像素、两个蓝色子像素和四个绿色子像素;两个红色子像素的亮度相同,两个蓝色子像素的亮度相同,四个绿色子像素的亮度不同。例如其中两个绿色子像素承担的显示亮度为37.5%,另外两个绿色子像素承担的亮度为12.5%。其中,子像素渲染算法四实现的点亮方式可以使人眼感知到白色,产生的实际效果为清晰度低,且彩边明显。
通过子像素渲染算法五,显示器件上点亮六个子像素点,包括两个红色子像素、两个蓝色子像素和两个绿色子像素;两个红色子像素的亮度相同,两个蓝色子像素的亮度相同,两个绿色子像素的亮度相同。其中,子像素渲染算法五实现的点亮方式可以使人眼感知到白色,产生的实际效果为清晰度低,且彩边明显。
通过子像素渲染算法六,显示器件上点亮十个子像素点,包括三个红色子像素、三个蓝色子像素和四个绿色子像素;三个红色子像素的亮度相同,三个蓝色子像素的亮度相同,四个绿色子像素的亮度不同。例如其中两个绿色子像素承担的显示亮度为37.5%,另外两个绿色子像素承担的亮度为12.5%。其中,子像素渲染算法六实现的点亮方式可以使人眼感知到白色,产生的实际效果为清晰度一般,且彩边一般明显。
可以看出,DDIC提供了一定的对子像素渲染算法进行调节的能力,但是可调节参数数量和参数范围较小,上述四种子像素渲染算法均有不同程度的不清晰和彩边明显的问题,无法取得较为满意的效果。因此,想要改善上述显示屏的显示效果,无法通过更改子像素渲染算法的方式实现。
本申请实施例提供的一种屏幕显示方法,电子设备100通过对发送到显示屏之前的待显示内容进行预先滤波处理,和子像素渲染算法共同作用,最终能够实现清晰度高且无彩边现象的效果,改善了显示效果。解决了子像素渲染算法无法在线更新的困难。
这里的预先滤波处理是在电子设备100获取到待显示内容,对该待显示内容进行排版、渲染后,电子设备100对该待显示内容的位图进行软件滤波处理。
当该待显示内容的位图中包括每个像素点的RGB三个通道的灰度值数据(每个像素点的红色灰度值、蓝色灰度值和绿色灰度值)。电子设备100对该位图中绿色通道进行滤波,确定出绿色通道中的位图通过滤波后的像素灰度值。此时,电子设备100获取到滤波后的位图。
示例性的,该待显示内容的位图中包括每个像素点的第三红色像素灰度值、第三蓝色像素灰度值和第三绿色像素灰度值。电子设备100对该位图中绿色通道进行滤波,获取到滤波后的位图,该滤波后的位图中包括每个像素点的第三红色像素灰度值、第三蓝色像素灰度值和第四绿色像素灰度值。
当该待显示内容的位图中包括每个像素点的一个通道的灰度值数据,若原始位图的尺寸为[w,h];电子设备100分配尺寸为[w,h,3]的内存用于保存滤波后的位图,该滤波后的位图包括RGB三个通道的灰度值数据。其一,电子设备100将原始位图拷贝到滤波后的位图中的两个通道,这两个通道为红色通道和蓝色通道;由于针对上述子像素排列方式,想要达到理想的显示效果不需要对红色子像素和蓝色子像素进行调整,因此对于红色通道和蓝色通道来说无需进行滤波,电子设备100可以直接将原始位图拷贝到红色通道和蓝色通道。
其二,电子设备100将原始位图进行软件滤波,确定出绿色子像素的像素灰度值,然后将软件滤波后的灰度位图拷贝到另外一个通道,该通道为绿色通道。此时,电子设备100获取到滤波后的位图。
示例性的,该待显示内容的位图中包括每个像素点的第三像素灰度值,即每个像素点的红色像素灰度值、蓝色像素灰度值和绿色像素灰度值均为第三像素灰度值。首先电子设备将该灰度位图拷贝到红色通道和蓝色通道,然后电子设备100将该灰度位图进行滤波,获取到第四像素灰度值,电子设备100将滤波后的灰度位图拷贝到绿色通道,获取到滤波后的位图。该滤波后的位图中包括每个像素点的红色像素灰度值(第三像素灰度值)、蓝色像素灰度值(第三像素灰度值)、绿色像素灰度值(第四像素灰度值)。
举例来说,上述软件滤波方式中其滤波器核可以描述为(0.125,0.75,0.125)。子像素渲染算法三中一个绿色子像素的亮度为100%,通过该滤波处理可以将一个绿色子像素的亮度分到三个绿色子像素中,以使该三个绿色子像素承担的亮度分别为12.5%、75%和12.5%。
随后电子设备100将滤波后的内容送交至显示器,经过显示屏控制芯片(DDIC)的重采样滤波器的滤波和采样过程后,点亮屏幕上子像素。
示例性的,如图14所示,以“E”这个字符为例,电子设备100在示例字体A的字体文件中查询“E”这个字在示例字体A这个字体中的显示数据,该显示数据是以矢量图的形式进行存储的。然后,电子设备100对“网”这个字符的矢量图进行渲染,生成位图。位图是由多个像素点组成的,位图中包括每个像素点的RGB三个通道的灰度值数据(每个像素点的 红色灰度值构成红色向量矩阵、蓝色灰度值构成蓝色向量矩阵以及绿色灰度值构成绿色向量矩阵)。电子设备100对该位图中的绿色通道进行软件滤波,该软件滤波的滤波器核可以为(0.125,0.75,0.125)。
接着电子设备100对滤波后的位图进行混合叠加和图层叠加。最后,电子设备100将叠加后的位图发送到显示屏,显示屏的显示屏控制器基于子像素渲染算法中的重采样滤波器进行二次滤波,确定出RGB三个通道最终的灰度值信息。电子设备100将该二次滤波后的位图的像素灰度值变换为电压或电流等形式,驱动OLED屏,以相应的电压或电流点亮显示屏上的发光点(子像素点),显示屏上显示该字符位图。
可以看出,上述软件滤波(0.125,0.75,0.125)与子像素渲染算法三中的重采样滤波器共同作用,能够实现比子像素渲染算法三实现的显示效果更好的显示效果。
可以理解的,上述软件滤波方式中的滤波参数可以根据需要进行调节,针对子像素渲染算法四的软件滤波的参数和针对子像素渲染算法三的软件滤波的参数不同。上述软件滤波的参数是基于电子设备100子像素渲染算法和想要实现的显示效果确定的。
在一些实施例中,上述预处理软件滤波算法可以发生在图像渲染及送显(发送到显示屏)过程中的任一阶段。
在上述实施例中,可以全部或部分地通过软件、硬件、固件或者其任意组合来实现。当使用软件实现时,可以全部或部分地以计算机程序产品的形式实现。所述计算机程序产品包括一个或多个计算机指令。在计算机上加载和执行所述计算机程序指令时,全部或部分地产生按照本申请实施例所述的流程或功能。所述计算机可以是通用计算机、专用计算机、计算机网络、或者其他可编程装置。所述计算机指令可以存储在计算机可读存储介质中,或者从一个计算机可读存储介质向另一个计算机可读存储介质传输,例如,所述计算机指令可以从一个网站站点、计算机、服务器或数据中心通过有线(例如同轴电缆、光纤、数字用户线)或无线(例如红外、无线、微波等)方式向另一个网站站点、计算机、服务器或数据中心进行传输。所述计算机可读存储介质可以是计算机能够存取的任何可用介质或者是包含一个或多个可用介质集成的服务器、数据中心等数据存储设备。所述可用介质可以是磁性介质,(例如,软盘、硬盘、磁带)、光介质(例如DVD)、或者半导体介质(例如固态硬盘)等。
本领域普通技术人员可以理解实现上述实施例方法中的全部或部分流程,该流程可以由计算机程序来指令相关的硬件完成,该程序可存储于计算机可读取存储介质中,该程序在执行时,可包括如上述各方法实施例的流程。而前述的存储介质包括:ROM或随机存储记忆体RAM、磁碟或者光盘等各种可存储程序代码的介质。

Claims (21)

  1. 一种屏幕显示方法,应用于电子设备,所述电子设备包括OLED显示面板和集成子像素渲染算法的显示屏控制器,其特征在于,所述方法包括:
    所述电子设备获取待显示内容的位图数据;
    所述电子设备对所述位图数据进行软件滤波,其中,所述软件滤波基于所述子像素渲染算法确定,用于优化所述待显示内容在OLED显示面板上的显示边缘;
    所述显示屏控制器对经过所述软件滤波后的位图数据进行子像素渲染,确定每个像素点的RGB灰度值;
    所述显示屏控制器根据所述RGB灰度值驱动所述OLED显示面板上的OLED发光点以显示所述待显示内容。
  2. 根据权利要求1所述的方法,其特征在于,所述电子设备对所述位图数据进行软件滤波包括:
    所述位图数据包括RGB三通道位图数据和单通道灰度位图数据;
    当所述位图数据为单通道灰度位图数据时,将所述位图数据转换为RGB三通道位图数据;
    所述电子设备对所述RGB三通道位图数据的红色像素灰度值和蓝色像素灰度值分别进行单像素通道软件滤波处理;
    所述电子设备对所述位图数据的绿色像素灰度值保持不变。
  3. 根据权利要求2所述的方法,其特征在于,所述单像素通道软件滤波处理包括:
    所述单像素通道软件滤波处理与子像素渲染算法中的一维滤波处理方向垂直。
  4. 根据权利要求3所述的方法,其特征在于,所述一维滤波的公式包括:
    Figure PCTCN2022073339-appb-100001
    其中,dst(x,y)为输出的像素点(x,y)的像素灰度值,src(x,y)为输入的像素点(x,y)的像素灰度值,kernel为卷积核,anchor为锚点,anchor.x为锚点的横坐标,anchor.x为锚点的纵坐标。
  5. 根据权利要求1所述的方法,其特征在于,在所述电子设备对所述位图数据进行软件滤波之前,所述方法还包括:
    所述电子设备基于所述显示屏的型号确定所述子像素渲染算法。
  6. 根据权利要求1所述的方法,其特征在于,所述电子设备对所述位图数据进行软件滤波包括:
    确认所述电子设备不处于拍摄分享、投屏的场景时,所述电子设备对所述位图数据进行软件滤波。
  7. 根据权利要求1所述的方法,其特征在于,所述电子设备对所述待显示内容的位图数据进行软件滤波,之后还包括:
    所述电子设备保存所述软件滤波后的待显示内容的位图数据。
  8. 根据权利要求7所述的方法,其特征在于,所述方法还包括:
    所述电子设备调用所述保存的软件滤波后的待显示内容的位图数据;
    所述显示屏控制器对经过所述软件滤波后的位图数据进行子像素渲染,确定每个像素点的RGB灰度值,并根据所述RGB灰度值驱动所述OLED显示面板上的OLED发光点以显示所述待显示内容。
  9. 根据权利要求1所述的方法,其特征在于,所述电子设备获取待显示内容的位图数据,包括:
    所述电子设备确定待显示内容的矢量图;
    所述电子设备对所述待显示内容的矢量图进行栅格化,获取所述待显示内容的位图数据。
  10. 根据权利要求1所述的方法,其特征在于,所述待显示内容的数据类型包括文字、字符、图像。
  11. 一种电子设备,其特征在于,包括:一个或多个处理器、一个或多个存储器、显示屏和集成子像素渲染算法的显示屏控制器;所述一个或多个存储器分别与所述一个或多个处理、所述显示屏、所述显示屏控制器耦合;所述一个或多个存储器用于存储计算机程序代码,所述计算机程序代码包括计算机指令;当所述计算机指令在所述处理器上运行时,使得所述电子设备执行:
    通过所述处理器获取待显示内容的位图数据;
    通过所述处理器对所述位图数据进行软件滤波,其中,所述软件滤波基于所述子像素渲染算法确定,用于优化所述待显示内容在OLED显示面板上的显示边缘;
    通过所述显示屏控制器对经过所述软件滤波后的位图数据进行子像素渲染,确定每个像素点的RGB灰度值;
    通过所述显示屏控制器根据所述RGB灰度值驱动所述OLED显示面板上的OLED发光点以显示所述待显示内容。
  12. 根据权利要求11所述的电子设备,其特征在于,所述通过所述处理器对所述位图数据进行软件滤波包括:
    所述位图数据包括RGB三通道位图数据和单通道灰度位图数据;
    当所述位图数据为单通道灰度位图数据时,通过所述处理器将所述位图数据转换为RGB三通道位图数据;
    通过所述处理器对所述RGB三通道位图数据的红色像素灰度值和蓝色像素灰度值分别进行单像素通道软件滤波处理;所述位图数据的绿色像素灰度值保持不变。
  13. 根据权利要求12所述的电子设备,其特征在于,所述单像素通道软件滤波处理包括:
    所述单像素通道软件滤波处理与子像素渲染算法中的一维滤波处理方向垂直。
  14. 根据权利要求13所述的电子设备,其特征在于,所述一维滤波的公式包括:
    Figure PCTCN2022073339-appb-100002
    其中,dst(x,y)为输出的像素点(x,y)的像素灰度值,src(x,y)为输入的像素点(x,y)的像素灰度值,kernel为卷积核,anchor为锚点,anchor.x为锚点的横坐标,anchor.x为锚点的纵坐标。
  15. 根据权利要求11所述的电子设备,其特征在于,在通过所述处理器对所述位图数据进行软件滤波之前,所述方法还包括:
    通过所述处理器基于所述显示屏的型号确定所述子像素渲染算法。
  16. 根据权利要求11所述的电子设备,其特征在于,所述通过所述处理器对所述位图数据进行软件滤波包括:
    确认所述电子设备不处于拍摄分享、投屏的场景时,通过所述处理器对所述位图数据进行软件滤波。
  17. 根据权利要求11所述的电子设备,其特征在于,通过所述处理器对所述待显示内容的位图数据进行软件滤波,之后还包括:
    通过所述存储器保存所述软件滤波后的待显示内容的位图数据。
  18. 根据权利要求17所述的电子设备,其特征在于,所述电子设备还执行:
    通过所述处理器调用所述保存的软件滤波后的待显示内容的位图数据;
    通过所述显示屏控制器对经过所述软件滤波后的位图数据进行子像素渲染,确定每个像素点的RGB灰度值,并根据所述RGB灰度值驱动所述OLED显示面板上的OLED发光点以显示所述待显示内容。
  19. 根据权利要求11所述的电子设备,其特征在于,所述通过所述处理器获取待显示内容的位图数据,包括:
    通过所述处理器确定待显示内容的矢量图;
    通过所述处理器对所述待显示内容的矢量图进行栅格化,获取所述待显示内容的位图数据。
  20. 根据权利要求11所述的电子设备,其特征在于,所述待显示内容的数据类型包括文字、字符、图像。
  21. 一种计算机可读介质,用于存储一个或多个程序,其中所述一个或多个程序被配置为被所述一个或多个处理器执行,所述一个或多个程序包括指令,所述指令用于执行如权利要求1-10所述的方法。
PCT/CN2022/073339 2021-02-02 2022-01-22 一种屏幕显示方法及相关装置 WO2022166624A1 (zh)

Priority Applications (1)

Application Number Priority Date Filing Date Title
EP22748903.6A EP4261675A4 (en) 2021-02-02 2022-01-22 SCREEN DISPLAY METHOD AND RELATED APPARATUS

Applications Claiming Priority (6)

Application Number Priority Date Filing Date Title
CN202110143801.3 2021-02-02
CN202110143801 2021-02-02
CN202110144623 2021-02-02
CN202110144623.6 2021-02-02
CN202110742740.2A CN114840164A (zh) 2021-02-02 2021-06-30 一种屏幕显示方法及相关装置
CN202110742740.2 2021-06-30

Publications (1)

Publication Number Publication Date
WO2022166624A1 true WO2022166624A1 (zh) 2022-08-11

Family

ID=82562918

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2022/073339 WO2022166624A1 (zh) 2021-02-02 2022-01-22 一种屏幕显示方法及相关装置

Country Status (3)

Country Link
EP (1) EP4261675A4 (zh)
CN (1) CN114840164A (zh)
WO (1) WO2022166624A1 (zh)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115657992A (zh) * 2022-10-10 2023-01-31 荣耀终端有限公司 一种屏幕显示的方法、装置、设备及存储介质
CN116364023A (zh) * 2023-04-12 2023-06-30 广州文石信息科技有限公司 基于gpu的电子墨水屏驱动方法、装置、介质及设备

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100045695A1 (en) * 2007-04-20 2010-02-25 Candice Hellen Brown Elliott Subpixel rendering area resample functions for display device
US20100141670A1 (en) * 2008-12-10 2010-06-10 Microsoft Corporation Color Packing Glyph Textures with a Processor
CN109192084A (zh) * 2018-10-11 2019-01-11 惠科股份有限公司 显示面板及一种渲染方法
CN109983527A (zh) * 2016-11-23 2019-07-05 三星电子株式会社 显示装置和显示面板的驱动方法
CN110337685A (zh) * 2017-02-23 2019-10-15 辛纳普蒂克斯公司 编码demura校准信息

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB9815907D0 (en) * 1998-07-21 1998-09-16 British Broadcasting Corp Improvements in colour displays
EP1934970B1 (en) * 2005-10-14 2016-08-17 Samsung Display Co., Ltd. Improved memory structures for image processing

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100045695A1 (en) * 2007-04-20 2010-02-25 Candice Hellen Brown Elliott Subpixel rendering area resample functions for display device
US20100141670A1 (en) * 2008-12-10 2010-06-10 Microsoft Corporation Color Packing Glyph Textures with a Processor
CN109983527A (zh) * 2016-11-23 2019-07-05 三星电子株式会社 显示装置和显示面板的驱动方法
CN110337685A (zh) * 2017-02-23 2019-10-15 辛纳普蒂克斯公司 编码demura校准信息
CN109192084A (zh) * 2018-10-11 2019-01-11 惠科股份有限公司 显示面板及一种渲染方法

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
See also references of EP4261675A4

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115657992A (zh) * 2022-10-10 2023-01-31 荣耀终端有限公司 一种屏幕显示的方法、装置、设备及存储介质
CN115657992B (zh) * 2022-10-10 2024-03-19 荣耀终端有限公司 一种屏幕显示的方法、装置、设备及存储介质
CN116364023A (zh) * 2023-04-12 2023-06-30 广州文石信息科技有限公司 基于gpu的电子墨水屏驱动方法、装置、介质及设备
CN116364023B (zh) * 2023-04-12 2023-11-28 广州文石信息科技有限公司 基于gpu的电子墨水屏驱动方法、装置、介质及设备

Also Published As

Publication number Publication date
CN114840164A (zh) 2022-08-02
EP4261675A1 (en) 2023-10-18
EP4261675A4 (en) 2024-07-10

Similar Documents

Publication Publication Date Title
WO2021135730A1 (zh) 显示界面适配方法、显示界面适配设计方法和电子设备
CN115473957B (zh) 一种图像处理方法和电子设备
WO2020093988A1 (zh) 一种图像处理方法及电子设备
WO2021036715A1 (zh) 一种图文融合方法、装置及电子设备
WO2021052342A1 (zh) 电子设备调整画面色彩的方法和装置
WO2022166624A1 (zh) 一种屏幕显示方法及相关装置
WO2022100685A1 (zh) 一种绘制命令处理方法及其相关设备
CN114579016A (zh) 一种共享输入设备的方法、电子设备及系统
US20220245778A1 (en) Image bloom processing method and apparatus, and storage medium
WO2022095744A1 (zh) Vr显示控制方法、电子设备及计算机可读存储介质
CN113935898A (zh) 图像处理方法、系统、电子设备及计算机可读存储介质
WO2021057626A1 (zh) 图像处理方法、装置、设备及计算机存储介质
CN112684969A (zh) 始终显示方法及移动设备
CN113810603A (zh) 点光源图像检测方法和电子设备
CN114089932A (zh) 多屏显示方法、装置、终端设备及存储介质
CN115631250B (zh) 图像处理方法与电子设备
CN115150542B (zh) 一种视频防抖方法及相关设备
CN114708289A (zh) 一种图像帧预测的方法及电子设备
CN114527903A (zh) 一种按键映射方法、电子设备及系统
WO2023000745A1 (zh) 显示控制方法及相关装置
US20240144451A1 (en) Image Processing Method and Electronic Device
US20240045557A1 (en) Method for Sharing Input Device, Electronic Device, and System
CN113610943B (zh) 图标圆角化的处理方法及装置
CN115482143B (zh) 应用的图像数据调用方法、系统、电子设备及存储介质
CN114390195A (zh) 一种自动对焦的方法、装置、设备及存储介质

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 22748903

Country of ref document: EP

Kind code of ref document: A1

ENP Entry into the national phase

Ref document number: 2022748903

Country of ref document: EP

Effective date: 20230713

NENP Non-entry into the national phase

Ref country code: DE