WO2023230965A1 - 显示方法及装置 - Google Patents

显示方法及装置 Download PDF

Info

Publication number
WO2023230965A1
WO2023230965A1 PCT/CN2022/096698 CN2022096698W WO2023230965A1 WO 2023230965 A1 WO2023230965 A1 WO 2023230965A1 CN 2022096698 W CN2022096698 W CN 2022096698W WO 2023230965 A1 WO2023230965 A1 WO 2023230965A1
Authority
WO
WIPO (PCT)
Prior art keywords
image data
osd
display
sub
area
Prior art date
Application number
PCT/CN2022/096698
Other languages
English (en)
French (fr)
Inventor
邵继洋
孙建康
薛亚冲
Original Assignee
京东方科技集团股份有限公司
北京京东方技术开发有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 京东方科技集团股份有限公司, 北京京东方技术开发有限公司 filed Critical 京东方科技集团股份有限公司
Priority to PCT/CN2022/096698 priority Critical patent/WO2023230965A1/zh
Priority to CN202280001600.2A priority patent/CN117529911A/zh
Publication of WO2023230965A1 publication Critical patent/WO2023230965A1/zh

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/14Digital output to display device ; Cooperation and interconnection of the display device with other functional units
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/30Image reproducers

Definitions

  • the present disclosure relates to the field of signal recognition, and in particular, to a display method and device.
  • OSD On Screen Display
  • a display method is provided, which is applied to a display device.
  • the display screen of the display device includes multiple sub-display areas.
  • the display method includes: first, obtaining image data to be displayed.
  • the image data to be displayed includes OSD image data.
  • the OSD image data includes OSD sub-image data from n viewpoints, where n is an integer greater than or equal to 2. Then, the OSD image data is parsed, and the parsed OSD image data is rearranged.
  • the position information of the OSD image data determine the area to be displayed corresponding to the OSD image data on the display screen, and display the rearranged OSD image data in the area to be displayed; the area to be displayed includes at least one sub-display area, and the rearranged
  • the resulting OSD image data includes rearranged OSD sub-image data of some or all viewpoints.
  • the image data to be displayed also includes source image data
  • the method further includes: parsing the source image data, and performing data rearrangement on the parsed source image data.
  • displaying the rearranged OSD image data in the area to be displayed according to the area to be displayed corresponding to the OSD image data on the display screen includes: according to the area to be displayed corresponding to the OSD image data on the display screen, using The rearranged OSD image data replaces the rearranged source image data corresponding to the area to be displayed, or the rearranged OSD image data is superimposed on the rearranged source image data corresponding to the area to be displayed.
  • overlaying the rearranged OSD image data on the rearranged source image data corresponding to the area to be displayed includes: setting transparency for the rearranged OSD image data, and superimposing the rearranged source image data corresponding to the area to be displayed.
  • the rearranged OSD image data after setting the transparency is superimposed on the source image data.
  • the OSD sub-image data of each viewpoint includes A sub-pixels, and A is an integer greater than or equal to 2; performing data rearrangement on the parsed OSD image data includes: OSD sub-images of n viewpoints All sub-pixels in the data are rearranged; different sub-pixels in the rearranged OSD sub-image data corresponding to n viewpoints are located in different rows.
  • the A sub-pixels include a first sub-pixel, the first sub-pixel is any one of the A sub-pixels, the first sub-pixel includes t sub-pixels, and t is an integer greater than or equal to 2.
  • parsing the OSD image data and rearranging the parsed OSD image data includes: first, parsing the OSD image data to obtain the display parameters of the OSD image data and the display data of the OSD image data.
  • the display parameters of the OSD image data include the number of viewpoints and sub-display area identifiers corresponding to the OSD image data, and the capacity of the display data of the OSD image data. Then, data rearrangement is performed on the display data of the OSD image data according to the display parameters of the OSD image data.
  • a method of obtaining OSD sub-image data of n viewpoints includes: using n cameras to capture the first OSD model from different viewpoints to obtain OSD sub-image data of n viewpoints. Or use a three-dimensional modeling or image processor to obtain the second OSD model, and obtain OSD sub-image data of n viewpoints based on the second OSD model.
  • the position information of the OSD image data includes the starting position of the area to be displayed, and the length information and width information of the area to be displayed.
  • the display method further includes: first, obtaining a first command.
  • the first command is used to switch the display mode of the OSD image from 3D display to 2D display, or to switch the display mode of the OSD image from 2D display to 2D display. 3D display.
  • two-dimensional OSD image data is acquired.
  • the two-dimensional OSD image is displayed on the display screen according to the area to be displayed corresponding to the two-dimensional OSD image data.
  • the display method further includes: first, obtaining the user's face image or pupil image to determine the user's main observation area; and/or obtaining the user's gesture image to determine the user's main viewpoint area. Then perform high-definition display on the OSD image data and/or source image data in the main observation area and/or main viewpoint area; perform low-definition display on the OSD image data and/or source image data outside the main observation area and/or main viewpoint area.
  • a computer-readable storage medium stores computer program instructions.
  • the computer program instructions When the computer program instructions are run on a computer (for example, a display device), they cause the computer to perform the display method as described in any of the above embodiments.
  • a computer program product includes computer program instructions.
  • the computer program instructions When the computer program instructions are executed on a computer (eg, a display device), the computer program instructions cause the computer to perform the display method as described in any of the above embodiments.
  • a computer program is provided.
  • the computer program When the computer program is executed on a computer (e.g., a display device), the computer program causes the computer to perform the display method as described in any of the above embodiments.
  • a display device including a display screen, a data acquisition device and a data processing device.
  • the display screen includes multiple sub-display areas.
  • the data acquisition device is configured to acquire image data to be displayed.
  • the image data to be displayed includes OSD image data.
  • the OSD image data includes OSD sub-image data from n viewpoints, where n is an integer greater than or equal to 2.
  • the data processing device is configured to parse the OSD image data and perform data rearrangement on the parsed OSD image data.
  • the position information of the OSD image data determine the area to be displayed corresponding to the OSD image data on the display screen, and control the area to be displayed on the display screen to display the rearranged OSD image data; the area to be displayed includes at least one sub-display area,
  • the rearranged OSD image data includes rearranged OSD sub-image data of some or all viewpoints.
  • the image data to be displayed also includes source image data
  • the data processing device is further configured to: parse the source image data, and perform data rearrangement on the parsed source image data.
  • the data processing device is configured to replace the rearranged source image data corresponding to the area to be displayed with the rearranged OSD image data according to the area to be displayed corresponding to the OSD image data, or, in the area to be displayed
  • the rearranged OSD image data is superimposed on the corresponding rearranged source image data.
  • the data processing device is configured to set transparency on the rearranged OSD image data, and superimpose the rearranged OSD image data with the transparency set on the rearranged source image data corresponding to the area to be displayed.
  • the OSD sub-image data of each viewpoint includes A sub-pixels, and A is an integer greater than or equal to 2; the data processing device is configured to: perform a processing on all sub-pixels in the OSD sub-image data of n viewpoints. Rearrangement; different sub-pixels in the rearranged OSD sub-image data corresponding to n viewpoints are located in different rows.
  • the A sub-pixels include a first sub-pixel, the first sub-pixel is any one of the A sub-pixels, the first sub-pixel includes t sub-pixels, and t is an integer greater than or equal to 2.
  • the data processing device is configured to parse the OSD image data to obtain display parameters of the OSD image data and display data of the OSD image data; the display parameters of the OSD image data include the number of viewpoints and sub-display areas corresponding to the OSD image data. logo, and the capacity of display data of OSD image data. According to the display parameters of the OSD image data, the display data of the OSD image data is rearranged.
  • the data acquisition device is configured to: use n virtual cameras to shoot the first OSD model from different viewpoints to obtain OSD sub-image data of n viewpoints; or use three-dimensional modeling or an image processor to obtain the first OSD model. Two OSD models, and according to the second OSD model, OSD sub-image data of n viewpoints are obtained.
  • the position information of the OSD image data includes the starting position of the area to be displayed, and the length information and width information of the area to be displayed.
  • the data acquisition device is further configured to: obtain a first command.
  • the first command is used to switch the display mode of the OSD image from 3D display to 2D display, or to switch the display mode of the OSD image from 2D display. to 3D display.
  • two-dimensional OSD image data or three-dimensional OSD image data is acquired.
  • the data processing device is configured to control the display screen to display a two-dimensional OSD image according to an area to be displayed corresponding to the two-dimensional OSD image data; or to control the display screen to display a three-dimensional OSD image according to the area to be displayed corresponding to the three-dimensional OSD image data. .
  • the display device further includes a collection device configured to obtain the user's face image or pupil image to determine the user's main observation area; and/or obtain the user's gesture image to determine the user's main viewpoint area.
  • the data processing device is also configured to perform high-definition display of OSD image data and/or source image data in the main observation area and/or main viewpoint area; and to display OSD image data and/or source data outside the main observation area and/or main viewpoint area. Image data is displayed in low definition.
  • Figure 1 is a schematic diagram of a display area on a display screen according to some embodiments.
  • Figure 2 is a flow chart of a display method according to some embodiments.
  • Figure 3 is a schematic diagram of a method for obtaining multi-viewpoint OSD sub-image data according to some embodiments
  • Figure 4 is a 3D display scene diagram of a multi-viewpoint OSD image according to some embodiments.
  • Figure 5 is a flow chart of data parsing and rearrangement according to some embodiments.
  • Figure 6 is a structural diagram of OSD image data according to some embodiments.
  • Figure 7A is an arrangement diagram of OSD image data before and after rearrangement according to some embodiments.
  • Figure 7B is an arrangement diagram of source image data before and after rearrangement according to some embodiments.
  • Figure 7C is a schematic diagram of display data replacement in the area to be displayed of an OSD according to some embodiments.
  • Figure 7D is an arrangement diagram of data displayed on a display screen according to some embodiments.
  • Figure 7E is a schematic diagram of display data replacement in the area to be displayed of another OSD according to some embodiments.
  • Figure 7F is an arrangement diagram of data displayed on another display screen according to some embodiments.
  • Figure 8 is a flowchart of another display method according to some embodiments.
  • Figure 9 is a flow chart of yet another display method according to some embodiments.
  • Figure 10 is a flow chart of yet another display method according to some embodiments.
  • Figure 11 is a flow chart of yet another display method according to some embodiments.
  • Figure 12 is a flow chart of yet another display method according to some embodiments.
  • Figure 13 is a schematic diagram of a main observation area according to some embodiments.
  • Figure 14 is a flow chart of yet another display method according to some embodiments.
  • Figure 15 is an OSD interaction flow chart according to some embodiments.
  • Figure 16 is a structural diagram of a display device according to some embodiments.
  • first and second are used for descriptive purposes only and cannot be understood as indicating or implying relative importance or implicitly indicating the quantity of indicated technical features. Therefore, features defined as “first” and “second” may explicitly or implicitly include one or more of these features. In the description of the embodiments of the present disclosure, unless otherwise specified, "plurality" means two or more.
  • At least one of A, B and C has the same meaning as “at least one of A, B or C” and includes the following combinations of A, B and C: A only, B only, C only, A and B The combination of A and C, the combination of B and C, and the combination of A, B and C.
  • a and/or B includes the following three combinations: A only, B only, and a combination of A and B.
  • the term “if” is optionally interpreted to mean “when” or “in response to” or “in response to determining” or “in response to detecting,” depending on the context.
  • the phrase “if it is determined" or “if [stated condition or event] is detected” is optionally interpreted to mean “when it is determined" or “in response to the determination" or “on detection of [stated condition or event]” or “in response to detection of [stated condition or event]”.
  • Example embodiments are described herein with reference to cross-sectional illustrations and/or plan views that are idealized illustrations.
  • the thickness of layers and regions are exaggerated for clarity. Accordingly, variations from the shapes in the drawings due, for example, to manufacturing techniques and/or tolerances are contemplated.
  • example embodiments should not be construed as limited to the shapes of regions illustrated herein but are to include deviations in shapes that result from, for example, manufacturing. For example, an etched area shown as a rectangle will typically have curved features. Accordingly, the regions shown in the figures are schematic in nature and their shapes are not intended to illustrate the actual shapes of regions of the device and are not intended to limit the scope of the exemplary embodiments.
  • OSD is applied to a monitor and generates some special glyphs or graphics on the screen of the monitor to allow users to get some information. It is commonly seen on the display screen of home TVs or personal computers. When the user operates the TV to change channels or adjust the volume, picture quality, etc., the TV screen will display the current status to let the user know.
  • OSD image display solutions cannot achieve 3D display of OSD images, nor can they achieve switching between 2D display and 3D display of OSD images. In addition, they cannot achieve the effect of multi-viewpoint OSD images.
  • some embodiments of the present disclosure provide a display method, which is applied to a display device.
  • the display screen of the display device includes a plurality of sub-display areas.
  • the sub-display area may be one or more islands, and the present disclosure is not limited to how many pixel islands the sub-display area includes.
  • the display screen 10 includes 20 individually driveable sub-display areas 11 , each sub-display area 11 is an individually driveable area, and the OSD to-be-displayed area 20 includes 6 sub-display areas 11 .
  • the individually driveable area can be a parallel drive area, or the backlight hardware can be driven as an individually controllable area, or any of the areas can be partitioned for image quality processing when processing the picture content.
  • the display device may be a television, a personal computer, or other device with a display screen, and the disclosure does not limit the specific type of the display device.
  • Some embodiments of the present disclosure provide a display method, as shown in Figure 2, the method includes:
  • Step 201 Obtain image data to be displayed.
  • the image data to be displayed includes OSD image data.
  • the OSD image data includes OSD sub-image data from n viewpoints (views), where n is an integer greater than or equal to 2.
  • the method of obtaining OSD sub-image data of n viewpoints includes: using n cameras 31 to capture the first OSD model 30 from different viewpoints to obtain OSD sub-image data of n viewpoints.
  • the camera 31 may be a camera capable of acquiring photos, such as a virtual camera or a real camera. The present disclosure does not limit the type of the camera 31 .
  • the OSD sub-image data of the n viewpoints can be pre-stored in system-side circuits (such as FPGA (Field Programmable Gate Array) and SOC (System on Chip)).
  • system-side circuits such as FPGA (Field Programmable Gate Array) and SOC (System on Chip)
  • the OSD sub-image data of the n viewpoints are called from the memory of the system-side circuit.
  • OSD image data can be stored in the memory of the system-side circuit, and the present disclosure does not limit the number and type of OSD images corresponding to the OSD image data.
  • the method of obtaining OSD sub-image data of n viewpoints may also include: using three-dimensional modeling or an image processor to obtain a second OSD model, and obtaining OSD sub-images of n viewpoints based on the second OSD model. data.
  • the second OSD model can be generated in real time by a three-dimensional modeling or image processor. Based on the second OSD model, slicing or image generation and output are performed for different viewpoints to obtain OSD sub-image data of n viewpoints. .
  • the operation of the three-dimensional modeling or image processor can be completed by the internal system of the display device 10 or by an external system. This disclosure does not limit the specific method of using the three-dimensional modeling or image processor to obtain the second OSD model.
  • the image data to be displayed includes OSD image data, and the OSD image data includes OSD sub-image data from n viewpoints. That is to say, the image data to be displayed includes OSD sub-image data from n viewpoints.
  • the image data to be displayed includes OSD sub-image data from n viewpoints.
  • the display screen 10 can display OSD images from multiple viewpoints at the same time, thereby realizing 3D display of OSD.
  • the user can see different OSD images on the display screen 10 from different viewpoints.
  • the user stands at At the first viewpoint angle 401, it can be seen that the image on the to-be-displayed area 20 of the OSD is the OSD image 4011 of the first viewpoint.
  • the image on the to-be-displayed area 20 of the OSD is the OSD image 4021 of the second viewpoint, and the OSD image 4011 of the first viewpoint is the same as the OSD image 4021 of the second viewpoint.
  • the two-viewpoint OSD image 4021 is any two OSD images of different viewpoints among the n-viewpoint OSD images. The present disclosure does not limit the number of viewpoints of the OSD image that can be displayed in one frame.
  • Step 202 Analyze the OSD image data, and perform data rearrangement on the parsed OSD image data.
  • the OSD display area includes 6 sub-display areas 11 , therefore, the OSD image data may be distributed in at least one sub-display area 11 of the 6 sub-display areas 11 .
  • the OSD image data when the OSD image data is stored in the memory of the system-side circuit, there is no partition for storage. If the OSD image data is generated in real time, there is no partition for generation, and the OSD image data from the same viewpoint is not stored in the same location. in a storage address. Therefore, in order to ensure the correct display of OSD images, the OSD image data needs to be fragmented and reorganized on the system side (such as a graphics card), and the reorganized OSD image data needs to be input into the control system of the display device.
  • the control system of the display device parses and rearranges the data, and the control system of the display device outputs the rearranged OSD image data to the display screen 10 of the display device according to the display requirements, and then the 3D OSD image can be displayed on the display screen 10 .
  • the implementation method of step 202 may include:
  • Step 501 Analyze the OSD image data to obtain the display parameters of the OSD image data and the display data of the OSD image data.
  • the display parameters of the OSD image data include flag bits, compression rates, the number of viewpoints and sub-display area identification corresponding to the OSD image data, and the capacity of the display data of the OSD image data.
  • Flag bits are configured to handshake with front-end instructions.
  • the compression ratio is configured as the ratio of the size of the display data of the OSD image data after compression to the size before compression, so as to facilitate compression and decompression of the display data of the OSD image data.
  • the number of viewpoints corresponding to the OSD image data is configured to store which viewpoint among n viewpoints the viewpoint corresponding to the OSD image data is.
  • the sub-display area identification corresponding to the OSD image data is configured to transmit the position information of the sub-display area 11 corresponding to the OSD image data to the back end.
  • the position information of the sub-display area 11 can be obtained by numbering the sub-display area 11 of the display screen 10, for example, starting from the upper left corner, numbering in sequence from left to right, and from top to bottom, each one The number corresponds to a word display area.
  • the capacity of the display data of the OSD image data is configured to pass to the back end the size of the OSD image data that needs to be displayed.
  • the capacity of the display data of the OSD image data is larger than the size of the OSD image data that needs to be displayed of all viewpoints.
  • the display data capacity of the OSD image data should be small.
  • the backend can reasonably configure the data processing resources according to this parameter to avoid the waste of system resources.
  • the display parameters of the OSD image data include the pixel value of a certain sub-pixel.
  • Step 502 Perform data rearrangement on the display data of the OSD image data according to the display parameters of the OSD image data.
  • the OSD sub-image data of each viewpoint includes A sub-pixels, where A is an integer greater than or equal to 2.
  • the implementation method of step 502 may include: rearranging all sub-pixels in the OSD sub-image data of n viewpoints. Different sub-pixels in the rearranged OSD sub-image data corresponding to n viewpoints are located in different rows. Subpixels can include red, green, blue, black, white, and yellow.
  • the A sub-pixels include a first sub-pixel, the first sub-pixel is any one of the A sub-pixels, the first sub-pixel includes t sub-pixels, and t is an integer greater than or equal to 1.
  • the parsed OSD image data includes OSD sub-image data of n viewpoints.
  • the OSD sub-image data of each viewpoint includes 3 sub-pixels, and the 3 sub-pixels are respectively red (Red, R).
  • Sub-pixels, green (Green, G) sub-pixels and blue (Blue, B) sub-pixels are taken as examples.
  • each red sub-pixel includes t sub-pixels, such as R1-1, R1-2,..., R1-t
  • each green sub-pixel includes t sub-pixels, such as G1-1 , G1-2,...,G1-t
  • each blue sub-pixel includes t sub-pixels, such as B1-1, B1-2,..., B1-t.
  • the display data R1 -1 corresponds to the starting position of the area to be displayed 20 of the OSD
  • the display data Rt-n corresponds to the end position of the area to be displayed 20 . It can be understood that, as shown in Figure 1, when the area to be displayed 20 of the OSD is a rectangle, the starting point of the area 20 to be displayed of the OSD is the position D0 of the first pixel in the upper left corner of the rectangle.
  • the area 20 of the OSD to be displayed is The end position of is the position D1 of the last pixel in the lower right corner of the rectangle.
  • each sub-pixel includes 4 sub-pixels, and each sub-pixel has A pixel value
  • the area to be displayed 20 includes 12 rows and 5 columns of pixels, and the pixels correspond to the pixel values of the sub-pixels one-to-one as an example for a more intuitive explanation.
  • the OSD sub-image data of each viewpoint includes pixel values of 15 sub-pixels.
  • the OSD sub-image data 501 of the first viewpoint includes the pixel values of four sub-pixels of the red sub-pixel (such as R1-1, R1-2, R1-3 and R1-4), and the pixel values of the four sub-pixels of the green sub-pixel (such as R1-1, R1-2, R1-3 and R1-4).
  • R1-1, G1-2, G1-3 and G1-4 the pixel values of the four sub-pixels of the blue sub-pixel (such as B1-1, B1-2, B1-3, B1-4).
  • the OSD sub-image data 502 of the second viewpoint includes pixel values of four sub-pixels of the red sub-pixel (such as R2-1, R2-2, R2-3 and R2-4), and pixel values of the four sub-pixels of the green sub-pixel (such as R2-1, R2-2, R2-3 and R2-4). Such as G2-1, G2-2, G2-3 and G2-4) and the pixel values of the four sub-pixels of the blue sub-pixel (such as B2-1, B2-2, B2-3, B2-4).
  • the OSD sub-image data 505 of the fifth viewpoint includes the pixel values of four sub-pixels of the red sub-pixel (such as R5-1, R5-2, R5-3 and R5-4), and the four sub-pixels of the green sub-pixel.
  • the OSD image data is fragmented and reorganized on the system side, and the reorganized OSD image data is input to the control system of the display device, the pixel values of the sub-pixels are compressed and stored in the memory of the system-side circuit, not They are arranged in the order of display. Therefore, the OSD image data, that is, the sub-image data of the five viewpoints, needs to be analyzed and rearranged by the control system of the display device.
  • Table 1 Display data included in OSD sub-image data of 5 viewpoints
  • All sub-pixels in the OSD sub-image data of the five viewpoints are rearranged, that is, the red sub-pixels, green sub-pixels and blue sub-pixels in the OSD sub-image data of the five viewpoints are rearranged.
  • all the sub-pixels of the red sub-pixels in the OSD sub-image data of the 5 viewpoints are arranged in at least one row and parallel to each other, and all the sub-pixels of the blue sub-pixels in the OSD sub-image data of the 5 viewpoints are arranged in at least one row and Parallel to each other, all sub-pixels of the green sub-pixels in the OSD sub-image data of the five viewpoints are arranged in at least one row and parallel to each other.
  • any sub-pixel of the red sub-pixel in the OSD sub-image data of 5 viewpoints and any sub-pixel of the blue sub-pixel in the OSD sub-image data of 5 viewpoints or the green sub-pixel in the OSD sub-image data of 5 viewpoints Each sub-pixel of is arranged in a different row. That is, different sub-pixels in the rearranged OSD sub-image data are located in different rows.
  • the rearranged OSD sub-image data corresponding to the 5 viewpoints includes 12 rows of pixel values of sub-pixels, and the pixel values of the sub-pixels of the red sub-pixels in the OSD sub-image data of the 5 viewpoints R1-1, R1-2, R1-3, R1-4 and R1-5 are arranged in row 1, where the pixel value R1-1 of the sub-pixel corresponds to the starting position D0 of the area to be displayed 20.
  • the 5 viewpoints The pixel values G1-1, G1-2, G1-3, G1-4 and G1-5 of the sub-pixels of the green sub-pixel in the OSD sub-image data are arranged in the second row.
  • the pixel values B1-1, B1-2, B1-3, B1-4, and B1-5 of the sub-pixels of the blue sub-pixel in the OSD sub-image data of the five viewpoints are arranged in the third row.
  • the pixel values R4-1, R4-2, R4-3, R4-4 and R4-5 of the red sub-pixels in the OSD sub-image data of the five viewpoints are arranged in the 10th row.
  • the pixel values G4-1 and G4-2 and G4-3 and G4-4 and G4-5 of the green sub-pixel sub-pixels in the OSD sub-image data of the five viewpoints are arranged in the 11th row.
  • the pixel values B4-1, B4-2, B4-3, B4-4 and B4-5 of the sub-pixels of the blue sub-pixel in the OSD sub-image data of the 5 viewpoints are arranged in the 12th row, where the pixels of the sub-pixels
  • the value R4-5 corresponds to the end position D1 of the area to be displayed 20.
  • the pixel values G1-1, G1-2, G1-3, G1-4 and G1-5 of the green sub-pixels in the OSD sub-image data of the five viewpoints can also be arranged in the first row, Or arrange the pixel values B1-1, B1-2, B1-3, B1-4 and B1-5 of the sub-pixels of the blue sub-pixel in the OSD sub-image data of the five viewpoints in the first row.
  • the row of the sub-pixel of the red sub-pixel in the OSD sub-image data of n viewpoints is located and the row of the sub-pixel of the green sub-pixel in the OSD sub-pixel data of the n viewpoints, or the row of the sub-pixel of the green sub-pixel in the OSD sub-image data of n viewpoints.
  • the order of the rows in which the sub-pixels of the blue sub-pixel are located is not limited. That is to say, the rearranged order of multiple sub-pixels can include various forms, and as long as the rearranged different sub-pixels are located in different rows, they are all within the protection scope of the embodiments of the present disclosure.
  • Step 203 Determine the to-be-displayed area 20 corresponding to the OSD image data on the display screen 10 based on the position information of the OSD image data, and display the rearranged OSD image data in the to-be-displayed area 20.
  • the area to be displayed 20 includes at least one sub-display area 11 , and the rearranged OSD image data includes the rearranged OSD sub-image data of some viewpoints or all viewpoints.
  • the position information of the OSD image data includes the starting position D0 of the to-be-displayed area 20 of the OSD, and the length information and width information of the to-be-displayed area 20 of the OSD.
  • the end position D1 of the OSD to-be-displayed area 20 can be calculated from the starting point position D0 of 20 and the length information and width information of the OSD to-be-displayed area 20 .
  • the system-side circuit can configure or generate the coordinates of the starting position D0 of the OSD's to-be-displayed area 20 according to the display screen 10, as well as the length information and width information of the OSD's to-be-displayed area 20, the position of the sub-display area 11 to which the OSD image data belongs, etc. parameters, when the data stream of the display device reaches the starting position D0 of the OSD to-be-displayed area 20, the rearranged OSD sub-image data of the five viewpoints will be output to the OSD to-be-displayed area 20 according to the timing sequence, and the five viewpoints will be displayed.
  • OSD images when the data stream of the display device reaches the starting position D0 of the OSD to-be-displayed area 20, the rearranged OSD sub-image data of the five viewpoints will be output to the OSD to-be-displayed area 20 according to the timing sequence, and the five viewpoints will be displayed.
  • OSD images when the data stream of the display device reaches
  • the rearranged OSD image data may also include OSD sub-image data of some viewpoints.
  • OSD sub-image data of some viewpoints For example, during rearrangement, only the sub-image data of 3 viewpoints among the OSD sub-image data of 5 viewpoints are re-arranged. arranged, then when the data stream of the display device reaches the starting position D0 of the area to be displayed 20 of the OSD, the rearranged OSD sub-image data of the three viewpoints will be output to the area to be displayed 20 of the OSD according to the time sequence, and the three viewpoints will be displayed. OSD image.
  • OSD sub-image data of 5 viewpoints are all re-arranged, only the sub-image data of 3 viewpoints can be displayed during display.
  • the image data to be displayed also includes source image data
  • the display method further includes: parsing the source image data, and performing data rearrangement on the parsed image data.
  • the source image data may also include source image data from multiple viewpoints.
  • the embodiment of the present disclosure does not limit the method of obtaining the source image data.
  • the source image data includes source image data of n viewpoints, and the sub-image data of each viewpoint includes A sub-pixels.
  • the source image data of each viewpoint includes The type and number of sub-pixels are the same as the type and number of pixels included in the OSD sub-image data of each viewpoint.
  • Each sub-pixel of the source image data of each viewpoint includes m sub-pixels, and m is an integer greater than or equal to 1. It can be understood that m is the same as the above-mentioned each sub-pixel of the OSD sub-image data including t sub-pixels. The values of t can be the same or different.
  • the display range of the source image may be the entire display screen 10 or a partial display area of the display screen 10 .
  • Some embodiments of the present disclosure take the display range of the source image as the entire display screen 10 as an example for description.
  • the method of parsing and rearranging source image data is similar to the method of parsing and rearranging OSD image data, and will not be described again here.
  • the parsed source image data includes source image data of n viewpoints.
  • the source image data of each viewpoint includes 3 sub-pixels.
  • the 3 sub-pixels are red sub-pixels and green sub-pixels respectively. Take the blue sub-pixel as an example.
  • each red sub-pixel includes m sub-pixels, such as OR1-1, OR2-1,..., ORm-1
  • each green sub-pixel includes m sub-pixels, such as OG1-1, OG2-1,...,OGm-1
  • each blue sub-pixel includes m sub-pixels, such as OB1-1, OB2-1,...,OBm-1.
  • the display data OR1-1 corresponds to the starting position of the display screen 10
  • the display data ORm-n corresponds to the end position of the display screen 10.
  • the position of the first pixel in the upper left corner of the rectangle is the starting position of the display screen 10
  • the position of the last pixel in the lower right corner of the rectangle is That is the end position of the display screen 10 .
  • the rearranged source image data is input to the display screen 10 in a time-series manner, it can be displayed according to the rearranged source image data.
  • each sub-image includes 3 sub-pixels
  • each sub-pixel includes 4 sub-pixels
  • the display screen 10 includes 12 rows*5 columns of pixels.
  • the pixels are different from the source image.
  • the one-to-one correspondence of pixel values of data sub-pixels is used as an example to provide a more intuitive explanation of the rearranged source image data.
  • the rearranged source image data is shown in Table 3.
  • the pixel value OR1-1 of the sub-pixel corresponds to the first pixel in the upper left corner of the display screen 10
  • the pixel value OR1-1 of the sub-pixel corresponds to the last pixel in the lower right corner of the display screen 10.
  • the display area of the display screen 10 is larger than the to-be-displayed area 20 of the OSD.
  • the corresponding to-be-displayed area 20 on the display screen 10 displays the rearranged OSD image data in the to-be-displayed area 20 , including: using the rearranged OSD in the corresponding to-be-displayed area 20 on the display screen 10 according to the OSD image data.
  • the image data replaces the rearranged source image data corresponding to the area to be displayed 20 , or the rearranged OSD image data is superimposed on the rearranged source image data corresponding to the area to be displayed 20 .
  • the rearranged source image data (such as OR1-x and OR1-y) whose position information is located within the to-be-displayed area 20 of the OSD, x and y are both less than or an integer equal to m) is directly replaced with the rearranged OSD image data (such as R1-1 and R1-2).
  • the display data of the starting position D0 is replaced by OR1-x with R1-1, that is to say, in the OSD
  • the to-be-displayed area 20 does not display the source image, but only displays the OSD image.
  • the display data of the to-be-displayed area 20 of the OSD is the rearranged source image data as shown in FIG. 7B .
  • the rearranged OSD image data as shown in FIG. 7A is used as the display data of the OSD area 20 to be displayed. That is, in the area 20 of the OSD to be displayed, the rearranged OSD image data is used to replace the corresponding rearranged source image data. Therefore, as shown in FIG. 7D , the rearranged OSD image is displayed in the area 20 of the OSD to be displayed on the display screen 10 , and the rearranged OSD image is displayed on the display screen 10 outside the area 20 of the OSD to be displayed.
  • Source image data is displayed in the area 20 of the OSD to be displayed on the display screen 10 .
  • the rearranged source image data corresponding to the to-be-displayed area 20 of the OSD is superimposed on the rearranged source image data, and the source image data is displayed in the to-be-displayed area 20 of the OSD.
  • the overlay method may include setting transparency to the rearranged OSD image data, and superimposing the rearranged OSD image data with the transparency set on the rearranged source image data corresponding to the area to be displayed 20 .
  • the display data of the starting position D0 of the area to be displayed 20 of the OSD is replaced by the display data OR1-x of the source image at this position with R1-1', and R1-1' is the counterweight.
  • the rearranged OSD image data R1-1 is obtained by setting the transparency and then superimposing it on the source image data OR1-x.
  • R1-2' is obtained by setting the transparency on the rearranged OSD image data R2-1' and then superimposing it on the source image data OR2- x is obtained, and so on.
  • the display data of the to-be-displayed area 20 of the OSD is the rearranged source image data as shown in 7B, and the data flow reaches the to-be-displayed area of the OSD.
  • the starting point of the area 20 is D0, as shown in Figure 7E, the rearranged OSD image data is superimposed on the rearranged source image data corresponding to the area 20 of the OSD to be displayed, and the superimposed display data is obtained.
  • the superimposed display data is used as the display data of the area to be displayed 20 of the OSD. Therefore, as shown in FIG.
  • the OSD to-be-displayed area 20 of the display screen 10 displays the superimposed OSD image and the source image
  • the display screen 10 outside the OSD to-be-displayed area 20 displays the rearranged image.
  • the subsequent source image data is the display screen 10 outside the OSD to-be-displayed area 20 .
  • FIG. 8 and FIG. 9 are taken as examples to illustrate the overall process when the source image and the OSD image need to be displayed simultaneously.
  • the front-end transmits source image data through a data interface.
  • the front-end can be a certain part of the system or an external device. This disclosure does not limit the number and type of front-ends.
  • the data interface can include multiple, for example, the data interface can include the system interface (System Interface), or the internal interface (Inner Interface), or the display port (Display Port), or the high-definition multimedia interface (HDMI), or the built-in or embedded display interface (External/Embedded Display Port), or any one or more of Mobile Industry Processor Interface (Mobile Industry Processor Interface).
  • This disclosure does not limit the data source of the source image and the type of data interface. .
  • the system-side data receiving and conversion module After the system-side data receiving and conversion module receives the source image data, it outputs the converted source image data to the data analysis and rearrangement module. After data analysis and rearrangement, the rearranged source image data can be stored in the system-side circuit. DDR (Double Data Rate Synchronous Dynamic Random Access Memory, Double Rate Synchronous Dynamic Random Access Memory can also be abbreviated as DDR SDRAM,).
  • DDR Double Data Rate Synchronous Dynamic Random Access Memory, Double Rate Synchronous Dynamic Random Access Memory can also be abbreviated as DDR SDRAM,).
  • DDR SDRAM Double Rate Synchronous Dynamic Random Access Memory
  • the rearranged source image data can be called from the DDR, and the rearranged source image data can be processed for display data.
  • the data processing can include adjustment of specific pixel values.
  • the processed display data of the source image can be directly output to the data conversion and sending module and output to the back end, and the source image can be displayed under display timing control.
  • the back end may be the display screen 10 of the display device or the driving circuit of the display device. The present disclosure does not limit the type and quantity of the back end.
  • the system reads the switching command between 2D display and 3D display, and controls the OSD image data acquisition module through the OSD function control module to obtain the OSD image data corresponding to the 2D display or 3D display. .
  • the processed display data of the source image and the processed display data of the OSD image are output to the data fusion module for data fusion.
  • the fused display data of the source image and the display data of the OSD image are output to the backend through the data conversion and sending module.
  • the fused display data of the source image and the display data of the OSD image can be stored in the storage medium of the display device (such as EEPROM (Electrically Erasable Programmable Read-Only Memory), Flash (Flash EEPROM Memory) ) and ROM (Read-Only Memory, read-only memory).
  • the fused data is output from the storage medium, and the source image and OSD image are displayed under the control of the display timing.
  • the display data of the source image and the display data of the processed OSD image are fused in a data fusion manner, which may be replacement or superposition.
  • FIG. 9 is taken as an example to illustrate how the source image data and OSD image data are converted into display data on the display screen 10 and displayed on the display screen 10 .
  • the image data to be displayed is obtained.
  • the image data to be displayed includes source image data and OSD image data.
  • the OSD image data can be obtained from the memory of the system-side circuit.
  • the source image data can be obtained from the front end.
  • data analysis is performed on the source image data and OSD image data.
  • the analysis includes decompression. Through decompression, the display parameters of the source image data and OSD image data and the display data relative to the display parameters can be obtained respectively. Then, under the specifications of the clock, the display data of the image data is matched one-to-one with the display parameters.
  • the required display data may be display data of a certain area.
  • the data processing method can be the adjustment of pixel values, or the increase or decrease of rendering operations to obtain the processed display data.
  • the processed display data can be compared with the circuit on the system side.
  • the display data of the OSD image stored in is different. There is a one-to-one correspondence between the processed display data and the display parameters to which it belongs.
  • information about the display area in the display parameters corresponding to the processed display data is first obtained, that is, the display area corresponding to the processed display data is obtained.
  • the data flow of the processed display data reaches the starting point D0 of the OSD to-be-displayed area 20
  • the superimposed data obtained by the processed display data of the OSD image and the processed display data of the source image can be displayed, or only the processed data can be displayed.
  • the display data of the processed OSD image can be directly displayed in the display area other than the OSD to-be-displayed area 20 on the display screen 10 .
  • the fused data can be stored in the memory of the system-side circuit or the memory of the display device to facilitate next call or direct use.
  • this disclosure takes the OSD image data stored in the memory of the system-side circuit as an example to illustrate, it does not limit the actual use that the OSD image can only be displayed based on the OSD image data stored in the system-side circuit, and does not limit the actual use. Whether the source image data and OSD image data must be parsed, rearranged, processed and fused.
  • the rearranged source image data and the rearranged OSD image data are stored in the memory of the system-side circuit or the memory of the display device, after processing When the source image data and the processed OSD image data are fused, the required data can be directly called or used according to actual needs.
  • the display method further includes:
  • Step 1001 Obtain a first command.
  • the first command is used to switch the display mode of the OSD image from 3D display to 2D display, or to switch the display mode of the OSD image from 2D display to 3D display.
  • the first command may be an enable signal EN for switching between 3D display and 2D display.
  • the enable signal can be used to switch from 3D display to 2D display, or from 2D display to 2D display. 3D display.
  • the enable signal can be given by a switching button on the operating handle or remote control held by the user. Each time the user presses the switching button, the 3D display and the 2D display are switched. For example, when the button is not pressed, the enable signal is low, the system obtains the first command, and switches the display mode of the OSD image to 2D display, and the display mode of the OSD image is 2D display.
  • the enable signal When the user presses the switch button, the enable signal is set high, the system obtains the first command, and switches the display mode of the OSD image to 3D display, then the display mode of the OSD image is 3D display.
  • the enable signal EN can also be generated by the system side. For example, after the system side determines that a certain parameter or parameters reach a certain preset condition, the system side performs automatic conversion between 2D display and 3D display.
  • Step 1002 In response to the first command, obtain two-dimensional OSD image data or three-dimensional OSD image data.
  • the two-dimensional OSD image data can be obtained by de-3Dizing the OSD sub-image data of n viewpoints.
  • the OSD sub-image data of n viewpoints in the memory of the system side circuit can be called. image data, and perform de-3D data processing to obtain two-dimensional OSD image data.
  • the two-dimensional OSD image data can also be stored in the memory of the system-side circuit.
  • the system-side obtains the first command
  • the two-dimensional OSD image data in the memory of the system-side circuit can be called.
  • the two-dimensional OSD image data can also be the OSD sub-image data of a certain viewpoint among the OSD sub-image data of n viewpoints.
  • the system side When the system side obtains the first command, it can directly call the memory of the system side circuit.
  • the three-dimensional OSD image data includes the OSD sub-image data of the n viewpoints mentioned above, which can be obtained by taking pictures of multiple viewpoints of the first OSD model, or can be obtained in real time through the second OSD model.
  • Step 1003. Display the two-dimensional OSD image on the display screen according to the area to be displayed corresponding to the two-dimensional OSD image data; or display the three-dimensional OSD image on the display screen according to the area to be displayed corresponding to the three-dimensional OSD image data.
  • the to-be-displayed area 20 corresponding to the two-dimensional OSD image data can be configured or generated by the system, and the pixel value of each sub-pixel in the two-dimensional OSD image data is output to the area in accordance with the display timing.
  • the area to be displayed 20 can display a two-dimensional OSD image in the area to be displayed 20 .
  • the display method further includes:
  • Step 1101 Obtain the user's face image or pupil image to determine the user's main observation area; and/or obtain the user's gesture image to determine the user's main viewpoint area.
  • the user's face image, pupil image, or gesture image can be collected through a camera, which can be a visible light camera and/or an infrared camera.
  • a camera which can be a visible light camera and/or an infrared camera. This disclosure does not limit the acquisition method of the user's face image, pupil image, or gesture image.
  • the collected user's face image is subjected to face detection data processing, and then pupil detection is performed.
  • the position of the user's eyes and their position relative to the screen are analyzed.
  • Observation angle to determine the position of the user's eyes such as E1 and E2 or observation angle (such as ⁇ and ⁇ ) or gaze point.
  • the position of the user's eyes can be determined. For example, by calculating the angle between the eyebrow coordinates and the lateral boundary of each sub-display area 11 in the OSD display area (such as ⁇ and ⁇ ) based on the eyebrow coordinates, the user's observation angle or gaze point can be determined. Through the position E1 of the user's eyes or the observation angle ⁇ or the gaze point, the user's main observation area D11 can be determined. Through the position E1 of the user's eyes or the observation angle ⁇ or the gaze point, the user's main observation area D12 can be determined. On the display screen 10 The area D2 is the display area other than the main observation area D11 and the main observation area D12.
  • gesture recognition processing can also be performed on the collected user's gesture pictures, and the user's gesture actions or gesture action trends can be analyzed to determine gesture instructions and projection display positions. Therefore, it is possible to determine which viewpoint areas the user mainly pays attention to when interacting with the OSD image on the display screen 10 , that is, the user's main viewpoint area can be determined.
  • Step 1102 Display the OSD image data and/or source image data in the main observation area and/or the main viewpoint area in high definition. Perform low-definition display of OSD image data and/or source image data outside the main observation area and/or main viewpoint area.
  • the OSD image data and/or source image data displayed in the main observation area D11 and the main observation area D12 are displayed in high definition.
  • the OSD image data and/or source image data on the area D2 on the display screen 10 are displayed in low definition.
  • the sub-image data partitions of n viewpoints are processed.
  • the sub-image data partitions of n viewpoints are processed.
  • the OSD image data and/or source image data of the main observation area and/or the main viewpoint area can be displayed in high definition by focusing on processing the OSD image data and/or source images of the main observation area and/or the main viewpoint area.
  • the data processing may include emphatically adjusting the image quality and color temperature of the OSD image and/or the source image. For example, you can increase the OSD image data and/or the number of rendering layers of the source image in the main observation area and/or the main viewpoint area, or increase the complexity of the second OSD model generated in real time in the main observation area and/or the main viewpoint area.
  • the emphasis processing may also include emphatically adjusting the display brightness of the main observation area and/or the main viewpoint area.
  • the present disclosure explains how to process the OSD image data and/or source image data of the main observation area and/or the main viewpoint area by focusing on processing the OSD image data and/or source image data of the main observation area and/or the main viewpoint area.
  • the specific method of high-definition display is not limited.
  • low-definition display of OSD image data and/or source image data outside the main observation area and/or main viewpoint area can be performed by simply processing OSD image data and/or outside the main observation area and/or main viewpoint area.
  • simple processing may include simple adjustments to the OSD image and/or image quality and color temperature. For example, you can reduce the number of OSD image data and/or source image data rendering layers outside the main observation area and/or the main viewpoint area, or reduce the number of second OSD models generated in real time in the main observation area and/or the main viewpoint area.
  • the lookup table can be derived algorithmically and stored in the memory of the system-side circuit.
  • multiple lookup tables can be established in each sub-display area 11 to achieve different levels of image quality adjustment in the sub-display area 11 . For example, if a certain sub-pixel is in the first sub-display area, the image quality can be adjusted through the first lookup table to achieve simple image quality adjustment. If the certain sub-pixel is in the first sub-display area, the image quality can be adjusted through the second lookup table. Perform image quality adjustments to achieve complex image quality adjustments.
  • the first sub-display area may be any sub-display area 11 on the display screen 10 .
  • a brightness adjustment global coefficient can be generated at the system side according to the OSD's to-be-displayed area 20.
  • the first backlight value can be generated based on the global coefficient.
  • multiple adjustment coefficients corresponding to multiple areas can be generated.
  • the adjustment coefficients can be adjusted by multiplying the first fixed parameter, or adding the second fixed parameter, and can be based on The adjustment coefficient of the main viewing area and/or the main viewpoint area generates the second backlight value.
  • the pixel value of each sub-pixel in the main observation area and/or main viewpoint area can be multiplied by the first fixed parameter corresponding to the main observation area and/or main viewpoint area, or the main observation area and/or The pixel value of each sub-pixel on the main viewpoint area is added to the second fixed parameter corresponding to the main observation area and/or the main viewpoint area to change the pixel value of each sub-pixel on the main observation area and/or the main viewpoint area. , and then adjust the display brightness of the OSD image and/or the source image in the main observation area and/or the main viewpoint area.
  • the third backlight value can be generated based on the adjustment coefficient of the main observation area and/or the display area other than the main viewpoint area, and the display brightness adjustment method thereof is similar to the display brightness adjustment method of the main observation area and/or the main viewpoint area, which is not the case here.
  • the display brightness of the display screen 10 can be adjusted through the backlight value on the area to be displayed 20 of the OSD. It can be understood that through the global coefficient or the adjustment coefficient, the display brightness of any area on the display screen 10 can also be adjusted to grayscale display.
  • a local dimming (Local Dimming) algorithm can also be used to achieve high-definition display or low-definition display in different areas of the display screen 10 .
  • a simple dimming algorithm such as a single algorithm, can be used to display low-definition OSD image data and/or source image data outside the main observation area and/or main viewpoint area, and multiple dimming algorithms can be used.
  • the weighting of this algorithm performs high-definition display on the OSD image data and/or source image data in the main observation area and/or the main viewpoint area.
  • the display brightness of the main observation area and/or the main viewpoint area is strong, while the display brightness of the main observation area and/or the display area other than the main viewpoint area is weak, without any processing, the visual quality may be reduced.
  • the effect makes the user's experience worse, so the display data of all OSD to-be-displayed areas 20 can be screen compensated.
  • the main observation area and/or the main viewpoint area includes multiple sub-display areas 11, cross-area
  • the display data is processed by a transition algorithm.
  • the cross-region display data includes the display data of the main observation area and/or the main viewpoint area.
  • the processed display data of the to-be-displayed area 20 of the OSD is integrated and output, and is displayed in the main observation area and/or the main viewpoint area. Or the boundary position of the main viewpoint area and the connecting position of the sub-display area 11 can have better visual effects.
  • the display area where the OSD interactive content is located is forced to be determined as the main observation area and/or the main viewpoint area.
  • the main observation area and/or the main viewpoint area on the display screen 10 Or the display brightness of the main viewpoint area is increased, and the display brightness of the display areas other than the main observation area and/or the main viewpoint area on the display screen 10 is reduced, which can guide the viewer to move the line of sight to the OSD interactive content on the display screen 10 .
  • the OSD image data and/or source image data outside the main observation area and/or the main viewpoint area can be directly filtered out, and only the OSD image data in the main observation area and/or the main viewpoint area can be displayed in high definition using the dimming algorithm. , to reduce the amount of calculation.
  • the user can interact using the OSD image.
  • the system sends a display instruction of the OSD image and the display screen 10 receives the display instruction of the OSD image, it is displayed in the area 20 of the OSD to be displayed.
  • the OSD image is maintained for T1 seconds, and T1 is a positive number. If the user performs an interactive operation within T1 seconds, the display screen 10 can transmit the interactive operation back to the system side (such as a graphics card) in the form of instructions, and the system side determines Command type, and then perform 10 lines of data transmission with the display screen.
  • the system side such as a graphics card
  • the instruction may be a first instruction, and the first instruction is to switch to another OSD image. Then the system controls the display screen 10 of the display device to display another OSD image and maintains T2. seconds, T2 is a positive number, and it is judged again within T2 seconds whether there is an instruction issued.
  • the instruction may also be a second instruction. The second instruction is to enter the first sub-content of the OSD image. Then the system side controls the display screen 10 of the display device to display the first sub-content and maintain it for T3 seconds. T3 is a positive number. Determine again whether there is an instruction issued within T3 seconds.
  • the command can also be a third command. The third command is to enter the second sub-content of the OSD image.
  • T4 is a positive number. Determine again whether there is an instruction issued within T4 seconds.
  • the command may also be a fourth command, which is an exit command.
  • the system controls the display screen 10 of the display device to display the source image, and the OSD image disappears. If no instruction is issued within T1 seconds, T2 seconds, T3 seconds, or T4 seconds, the system side controls the display screen 10 of the display device to display the source image, and the OSD image disappears.
  • T1, T2, T3 and T4 may be the same or different, and the present disclosure is not limited to the specific numerical values of T1, T2, T3 and T4.
  • the second sub-content can be the sub-content of the OSD image or the sub-content of the first sub-content.
  • the sub-content of the OSD image can be any kind of sub-content.
  • the present disclosure provides specific types of the first sub-content and the second sub-content. And priorities and levels are not limited.
  • Some embodiments of the present disclosure provide a computer-readable storage medium (e.g., a non-transitory computer-readable storage medium) having computer program instructions stored therein, and the computer program instructions are displayed on a computer (e.g., display screen) When running on the device), the computer is caused to execute the display method as described in any of the above embodiments.
  • a computer-readable storage medium e.g., a non-transitory computer-readable storage medium
  • the computer program instructions are displayed on a computer (e.g., display screen)
  • the computer is caused to execute the display method as described in any of the above embodiments.
  • the above-mentioned computer-readable storage media may include, but are not limited to: magnetic storage devices (such as hard disks, floppy disks or tapes, etc.), optical disks (such as CD (Compact Disk, compressed disk), DVD (Digital Versatile Disk, etc.) Digital versatile disk), etc.), smart cards and flash memory devices (e.g., EPROM (Erasable Programmable Read-Only Memory, Erasable Programmable Read-Only Memory), cards, sticks or key drives, etc.).
  • the various computer-readable storage media described in this disclosure may represent one or more devices and/or other machine-readable storage media for storing information.
  • the term "machine-readable storage medium" may include, but is not limited to, wireless channels and various other media capable of storing, containing and/or carrying instructions and/or data.
  • Some embodiments of the present disclosure also provide a computer program product.
  • the computer program product includes computer program instructions.
  • the computer program instructions When the computer program instructions are executed on a computer (eg, a display device), the computer program instructions cause the computer to perform the display method as described in the above embodiments.
  • Some embodiments of the present disclosure also provide a computer program.
  • the computer program When the computer program is executed on a computer (for example, a display device), the computer program causes the computer to perform the display method as described in the above embodiment.
  • the display device 160 includes a display screen 10 , a data acquisition device 161 and a data processing device 162 .
  • the display screen 10 includes a plurality of sub-display areas 11 .
  • the data acquisition device 161 is configured to acquire image data to be displayed.
  • the image data to be displayed includes OSD image data.
  • the OSD image data includes OSD sub-image data from n viewpoints, where n is greater than or equal to 2. integer.
  • the data processing device 162 is configured to parse the OSD image data and perform data rearrangement on the parsed OSD image data.
  • the corresponding to-be-displayed area 20 of the OSD image data on the display screen 10 is determined, and the to-be-displayed area 20 on the display screen 10 is controlled to display the rearranged OSD image data;
  • the to-be-displayed area 20 includes at least A sub-display area 11, the rearranged OSD image data includes the rearranged OSD sub-image data of partial viewpoints or all viewpoints.
  • the data processing device may include a data receiving and converting module, a data parsing and rearranging module, a display data processing module, an OSD image display data processing module, a data fusion module, and a data converting and sending module. .
  • the image data to be displayed includes OSD image data
  • the OSD image data includes OSD sub-image data from n viewpoints
  • the image data to be displayed includes OSD sub-image data from n viewpoints
  • the n viewpoints The OSD sub-image data can be stored in the memory of the system-side circuit.
  • the image data needs to be scattered and reorganized on the system side (such as a graphics card), and the reorganized image data needs to be input into the control system of the display device.
  • the control system of the display device analyzes and rearranges the data.
  • the control system of the display device outputs the rearranged image data to the display screen 10 of the display device according to the display requirements. Then, in one frame of the display screen 10, Display OSD images from multiple viewpoints at the same time, thereby realizing 3D display of OSD.
  • the image data to be displayed also includes source image data
  • the data processing device 162 is further configured to: parse the source image data, and perform data rearrangement on the parsed source image data.
  • the display range of the source image may be the entire display screen 10 or a partial display area of the display screen 10,
  • the data processing device 162 is configured to replace the rearrangement of the area to be displayed 20 with the rearranged OSD image data according to the area 20 corresponding to the OSD image data. or superimpose the rearranged OSD image data on the rearranged source image data corresponding to the area to be displayed 20 .
  • the data processing device is configured to set transparency on the rearranged OSD image data, and superimpose and set transparency on the rearranged source image data corresponding to the area 20 to be displayed.
  • the rearranged OSD image data is configured to set transparency on the rearranged OSD image data, and superimpose and set transparency on the rearranged source image data corresponding to the area 20 to be displayed.
  • the OSD sub-image data of each viewpoint includes A sub-pixels, and A is an integer greater than or equal to 2; the data processing device is configured to: process the OSD sub-image data of n viewpoints All sub-pixels in are rearranged; different sub-pixels in the rearranged OSD sub-image data corresponding to n viewpoints are located in different rows.
  • Subpixels can include red, green, blue, black, white, and yellow.
  • the A sub-pixels include a first sub-pixel, the first sub-pixel is any one of the A sub-pixels, the first sub-pixel includes t sub-pixels, t is greater than Or an integer equal to 2.
  • Each sub-pixel of the source image data of each viewpoint includes m sub-pixels, where m is an integer greater than or equal to 1. It is understandable that the values of m and t may be the same or different.
  • the data processing device 162 is configured to parse the OSD image data to obtain the display parameters of the OSD image data and the display data of the OSD image data; the display parameters of the OSD image data include the viewpoint and sub-display area corresponding to the OSD image data. logo, and the capacity of display data of OSD image data. According to the display parameters of the OSD image data, the display data of the OSD image data is rearranged.
  • the sub-display area identification corresponding to the OSD image data can be obtained by labeling the sub-display area 11 of the display screen 10 .
  • the data acquisition device 161 is configured to: use n virtual cameras to capture the first OSD model from different viewpoints to obtain OSD sub-image data of n viewpoints; or, use three-dimensional modeling or an image processor to obtain second OSD model, and obtain OSD sub-image data of n viewpoints according to the second OSD model.
  • the second OSD model can be generated in real time by a three-dimensional modeling or image processor. Based on the second OSD model, slicing or image generation and output are performed for different viewpoints to obtain OSD sub-image data of n viewpoints.
  • the operation of the three-dimensional modeling or image processor may be completed by an internal system of the display device 10 or by an external system.
  • the position information of the OSD image data includes the starting position of the area to be displayed 20 , and the length information and width information of the area to be displayed 20 .
  • the range of the display area 20 can be determined based on the starting point position information of the area to be displayed 20 and the length information and width information of the area to be displayed 20 .
  • the data acquisition device is further configured to: obtain a first command.
  • the first command is used to switch the display mode of the OSD image from 3D display to 2D display, or to switch the display mode of the OSD image from 2D display. to 3D display.
  • two-dimensional OSD image data or three-dimensional OSD image data is acquired.
  • the data processing device is configured to control the display screen 10 to display the two-dimensional OSD image according to the to-be-displayed area 20 corresponding to the two-dimensional OSD image data; or to control the display screen 10 to display according to the to-be-displayed area 20 corresponding to the three-dimensional OSD image data.
  • Three-dimensional OSD image Therefore, the present disclosure can realize switching between 2D display and 3D display.
  • the display device further includes a collection device configured to obtain the user's face image or pupil image to determine the user's main observation area; and/or obtain the user's gesture image to determine the user's main viewpoint area.
  • the data processing device is further configured to perform high-definition display of OSD image data in the main observation area and/or main viewpoint area, and low-definition display of OSD image data outside the main observation area and/or main viewpoint area.
  • 3D display of OSD images can be realized, and users can see different OSD images from different viewing angles.
  • high-definition display or low-definition display can be performed in different areas. Clear display.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Controls And Circuits For Display Device (AREA)

Abstract

一种显示方法,应用于显示装置,所述显示装置的显示屏包括多个子显示区域,所述显示方法包括:首先,获取待显示图像数据,所述待显示图像数据包括OSD图像数据,所述OSD图像数据包括来自n个视点的OSD子图像数据,所述n为大于或等于2的整数;然后,解析所述OSD图像数据,并对解析后的所述OSD图像数据进行数据重排;最后,根据所述OSD图像数据的位置信息,确定所述OSD图像数据在所述显示屏上对应的待显示区域,并在所述待显示区域显示重排后的OSD图像数据;所述待显示区域包括至少一个所述子显示区域,所述重排后的OSD图像数据包括重排后的部分视点或全部视点的OSD子图像数据。

Description

显示方法及装置 技术领域
本公开涉及信号识别领域,尤其涉及一种显示方法及装置。
背景技术
随着显示技术的发展,显示装置被广泛地应用到各个领域,在使用显示装置的过程中,通常需要调用屏幕菜单式调节方式(On Screen Display,OSD)图像,以在显示装置上配置显示相关的参数。
发明内容
一方面,提供一种显示方法,应用于显示装置,显示装置的显示屏包括多个子显示区域,显示方法包括:首先,获取待显示图像数据,待显示图像数据包括OSD图像数据,OSD图像数据包括来自n个视点的OSD子图像数据,n为大于或等于2的整数。然后,解析OSD图像数据,并对解析后的OSD图像数据进行数据重排。最后,根据OSD图像数据的位置信息,确定OSD图像数据在显示屏上对应的待显示区域,并在待显示区域显示重排后的OSD图像数据;待显示区域包括至少一个子显示区域,重排后的OSD图像数据包括重排后的部分视点或全部视点的OSD子图像数据。
在一些实施例中,待显示图像数据还包括源图像数据,方法还包括:解析源图像数据,并对解析后的源图像数据进行数据重排。
在一些实施例中,根据OSD图像数据在显示屏上对应的待显示区域,在待显示区域显示重排后的OSD图像数据,包括:根据OSD图像数据在显示屏上对应的待显示区域,采用重排后的OSD图像数据替换待显示区域对应的重排后的源图像数据,或者,在待显示区域对应的重排后的源图像数据上叠加重排后的OSD图像数据。
在一些实施例中,在待显示区域对应的重排后的源图像数据上叠加重排后的OSD图像数据,包括:对重排后的OSD图像数据设置透明度,在待显示区域对应的重排后的源图像数据上叠加设置透明度后的重排后的OSD图像数据。
在一些实施例中,每个视点的OSD子图像数据包括A个亚像素,A为大于等于2的整数;对解析后的OSD图像数据进行数据重排,包括:对n个视点的OSD子图像数据中的所有亚像素进行重排;n个视点对应的重排后的OSD子图像数据中不同的亚像素位于不同的行。
在一些实施例中,A个亚像素包括第一亚像素,第一亚像素为A个亚像素中的任意一个亚像素,第一亚像素包括t个子像素,t为大于或等于2的整数。
在一些实施例中,解析OSD图像数据,并对解析后的OSD图像数据进行数据重排,包括:首先,解析OSD图像数据,得到OSD图像数据的显示参数和OSD图像数据的显示数据。OSD图像数据的显示参数包括OSD图像数据对应的视点数与子显示区域标识,以及OSD图像数据的显示数据的容量。然后,根据OSD图像数据的显示参数,对OSD图像数据的显示数据进行数据重排。
在一些实施例中,获取n个视点的OSD子图像数据的方法,包括:采用n个相机,从不同视点拍摄第一OSD模型,得到n个视点的OSD子图像数据。或者采用三维建模或图像处理器获取第二OSD模型,并根据第二OSD模型,得到n个视点的OSD子图像数据。
在一些实施例中,OSD图像数据的位置信息包括待显示区域的起点位置,以及待显示区域的长度信息及宽度信息。
在一些实施例中,显示方法还包括:首先,获取第一命令,第一命令用于将OSD图像的显示方式由3D显示切换至2D显示,或者,将OSD图像的显示方式由2D显示切换至3D显示。然后,响应于第一命令,获取二维的OSD图像数据。最后,根据二维的OSD图像数据对应的待显示区域,在显示屏上显示二维的OSD图像。
在一些实施例中,显示方法还包括:首先,获取用户的人脸图像或瞳孔图像,确定用户的主观测区;和/或获取用户的手势图像,确定用户的主视点区。然后对主观测区和/或主视点区的OSD图像数据和/或源图像数据进行高清显示;对主观测区和/或主视点区以外的OSD图像数据和/或源图像数据进行低清显示。
再一方面,提供一种计算机可读存储介质。计算机可读存储介质存储有计算机程序指令,计算机程序指令在计算机(例如,显示装置)上运行时,使得计算机执行如上述任一实施例所述的显示方法。
又一方面,提供一种计算机程序产品。计算机程序产品包括计算机程序指令,在计算机(例如,显示装置)上执行计算机程序指令时,计算机程序指令使计算机执行如上述任一实施例所述的显示方法。
又一方面,提供一种计算机程序。当计算机程序在计算机(例如,显示装置)上执行时,计算机程序使计算机执行如上述任一实施例所述的显示方 法。
又一方面,提供一种显示装置,包括显示屏、数据获取装置和数据处理装置。其中,显示屏,包括多个子显示区域。数据获取装置,被配置为获取待显示图像数据,待显示图像数据包括OSD图像数据,OSD图像数据包括来自n个视点的OSD子图像数据,n为大于或等于2的整数。数据处理装置,被配置为解析OSD图像数据,并对解析后的OSD图像数据进行数据重排。根据OSD图像数据的位置信息,确定OSD图像数据在显示屏上对应的待显示区域,并控制显示屏上的待显示区域显示重排后的OSD图像数据;待显示区域包括至少一个子显示区域,重排后的OSD图像数据包括重排后的部分视点或全部视点的OSD子图像数据。
在一些实施例中,待显示图像数据还包括源图像数据,数据处理装置还被配置为:解析源图像数据,并对解析后的源图像数据进行数据重排。
在一些实施例中,数据处理装置被配置为根据OSD图像数据对应的待显示区域,采用重排后的OSD图像数据替换待显示区域对应的重排后的源图像数据,或者,在待显示区域对应的重排后的源图像数据上叠加重排后的OSD图像数据。
在一些实施例中,数据处理装置被配置为对重排后的OSD图像数据设置透明度,在待显示区域对应的重排后的源图像数据上叠加设置透明度后的重排后的OSD图像数据。
在一些实施例中,每个视点的OSD子图像数据包括A个亚像素,A为大于等于2的整数;数据处理装置被配置为:对n个视点的OSD子图像数据中的所有亚像素进行重排;n个视点对应的重排后的OSD子图像数据中不同的亚像素位于不同的行。
在一些实施例中,A个亚像素包括第一亚像素,第一亚像素为A个亚像素中的任意一个亚像素,第一亚像素包括t个子像素,t为大于或等于2的整数。
在一些实施例中,数据处理装置被配置为解析OSD图像数据,得到OSD图像数据的显示参数和OSD图像数据的显示数据;OSD图像数据的显示参数包括OSD图像数据对应的视点数与子显示区域标识,以及OSD图像数据的显示数据的容量。根据OSD图像数据的显示参数,对OSD图像数据的显示数据进行数据重排。
在一些实施例中,数据获取装置被配置为:采用n个虚拟相机,从不同视点拍摄第一OSD模型,得到n个视点的OSD子图像数据;或者,采用三 维建模或图像处理器获取第二OSD模型,并根据第二OSD模型,得到n个视点的OSD子图像数据。
在一些实施例中,OSD图像数据的位置信息包括待显示区域的起点位置,以及待显示区域的长度信息及宽度信息。
在一些实施例中,数据获取装置还被配置为:获取第一命令,第一命令用于将OSD图像的显示方式由3D显示切换至2D显示,或者,将OSD图像的显示方式由2D显示切换至3D显示。响应于第一命令,获取二维的OSD图像数据或三维OSD图像数据。数据处理装置被配置为根据二维的OSD图像数据对应的待显示区域,控制显示屏显示二维的OSD图像;或根据三维的OSD图像数据对应的待显示区域,控制显示屏显示三维的OSD图像。
在一些实施例中,显示装置还包括采集装置,被配置为获取用户的人脸图像或瞳孔图像,确定用户的主观测区;和/或获取用户的手势图像,确定用户的主视点区。数据处理装置还被配置为对主观测区和/或主视点区的OSD图像数据和/或源图像数据进行高清显示;对主观测区和/或主视点区以外的OSD图像数据和/或源图像数据进行低清显示。
附图说明
为了更清楚地说明本公开中的技术方案,下面将对本公开的一些实施例中所需要使用的附图作简单地介绍,显而易见地,下面描述中的附图仅仅是本公开的一些实施例的附图,对于本领域普通技术人员来讲,还可以根据这些附图获得其他的附图。此外,以下描述中的附图可以视作示意图,并非对本公开实施例所涉及的产品的实际尺寸、方法的实际流程、信号的实际时序等的限制。
图1为根据一些实施例的一种显示屏上显示区域的示意图;
图2为根据一些实施例的一种显示方法的流程图;
图3为根据一些实施例的一种多个视点OSD子图像数据的获取方法示意图;
图4为根据一些实施例的一种多个视点OSD图像的3D显示场景图;
图5为根据一些实施例的一种数据解析与重排的流程图;
图6为根据一些实施例的一种OSD图像数据的结构图;
图7A为根据一些实施例的一种重排前后OSD图像数据的排列图;
图7B为根据一些实施例的一种重排前后源图像数据的排列图;
图7C为根据一些实施例的一种OSD的待显示区域的显示数据替换 示意图;
图7D为根据一些实施例的一种显示屏上显示数据的排列图;
图7E为根据一些实施例的另一种OSD的待显示区域的显示数据替换示意图;
图7F为根据一些实施例的另一种显示屏上显示数据的排列图;
图8为根据一些实施例的另一种显示方法的流程图;
图9为根据一些实施例的又一种显示方法的流程图;
图10为根据一些实施例的又一种显示方法的流程图;
图11为根据一些实施例的又一种显示方法的流程图;
图12为根据一些实施例的又一种显示方法的流程图;
图13为根据一些实施例的一种主观测区的示意图;
图14为根据一些实施例的又一种显示方法的流程图;
图15为根据一些实施例的一种OSD交互流程图;
图16为根据一些实施例的一种显示装置的结构图。
具体实施方式
下面将结合附图,对本公开的一些实施例中的技术方案进行清楚、完整地描述,显然,所描述的实施例仅仅是本公开一部分实施例,而不是全部的实施例。基于本公开所提供的实施例,本领域普通技术人员所获得的所有其他实施例,都属于本公开保护的范围。
除非上下文另有要求,否则,在整个说明书和权利要求书中,术语“包括(comprise)”及其其他形式例如第三人称单数形式“包括(comprises)”和现在分词形式“包括(comprising)”被解释为开放、包含的意思,即为“包含,但不限于”。在说明书的描述中,术语“一个实施例(one embodiment)”、“一些实施例(some embodiments)”、“示例性实施例(exemplary embodiments)”、“示例(example)”、“特定示例(specific example)”或“一些示例(some examples)”等旨在表明与该实施例或示例相关的特定特征、结构、材料或特性包括在本公开的至少一个实施例或示例中。上述术语的示意性表示不一定是指同一实施例或示例。此外,的特定特征、结构、材料或特点可以以任何适当方式包括在任何一个或多个实施例或示例中。
以下,术语“第一”、“第二”仅用于描述目的,而不能理解为指示或暗示相对重要性或者隐含指明所指示的技术特征的数量。由此,限定有“第一”、“第二”的特征可以明示或者隐含地包括一个或者更多个该特征。在本公开实施例的描述中,除非另有说明,“多个”的含义是两个或两个以上。
在描述一些实施例时,可能使用了“耦接”和“连接”及其衍伸的表达。例如,描述一些实施例时可能使用了术语“连接”以表明两个或两个以上部件彼此间有直接物理接触或电接触。又如,描述一些实施例时可能使用了术语“耦接”以表明两个或两个以上部件有直接物理接触或电接触。然而,术语“耦接”或“通信耦合(communicatively coupled)”也可能指两个或两个以上部件彼此间并无直接接触,但仍彼此协作或相互作用。这里所公开的实施例并不必然限制于本文内容。
“A、B和C中的至少一个”与“A、B或C中的至少一个”具有相同含义,均包括以下A、B和C的组合:仅A,仅B,仅C,A和B的组合,A和C的组合,B和C的组合,及A、B和C的组合。
“A和/或B”,包括以下三种组合:仅A,仅B,及A和B的组合。
如本文中所使用,根据上下文,术语“如果”任选地被解释为意思是“当……时”或“在……时”或“响应于确定”或“响应于检测到”。类似地,根据上下文,短语“如果确定……”或“如果检测到[所陈述的条件或事件]”任选地被解释为是指“在确定……时”或“响应于确定……”或“在检测到[所陈述的条件或事件]时”或“响应于检测到[所陈述的条件或事件]”。
本文中“适用于”或“被配置为”的使用意味着开放和包容性的语言,其不排除适用于或被配置为执行额外任务或步骤的设备。
另外,“基于”的使用意味着开放和包容性,因为“基于”一个或多个条件或值的过程、步骤、计算或其他动作在实践中可以基于额外条件或超出的值。
如本文所使用的那样,“约”或“近似”包括所阐述的值以及处于特定值的可接受偏差范围内的平均值,其中可接受偏差范围如由本领域普通技术人员考虑到正在讨论的测量以及与特定量的测量相关的误差(即,测量系统的局限性)所确定。
本文参照作为理想化示例性附图的剖视图和/或平面图描述了示例性实施方式。在附图中,为了清楚,放大了层和区域的厚度。因此,可设想到由于例如制造技术和/或公差引起的相对于附图的形状的变动。因此,示例性实施方式不应解释为局限于本文示出的区域的形状,而是包括因例如制造而引起的形状偏差。例如,示为矩形的蚀刻区域通常将具有弯曲的特征。因此,附图中所示的区域本质上是示意性的,且它们的形状并非旨在示出设备的区域的实际形状,并且并非旨在限制示例性实施方式的范围。
OSD是应用在显示器上,在显示器的屏幕中产生的一些特殊字形或图形,让使用者得到一些讯息。常见于家用电视机或个人电脑的显示屏幕上,当用 户操作电视机换台或调整音量、画质等,电视屏幕就会显示目前状态让使用者知道。
目前,OSD图像的显示方案无法实现OSD图像的3D显示,也无法实现OSD图像的2D显示与3D显示的切换,此外,无法实现多视点OSD图像的效果。
为解决上述问题,本公开的一些实施例提供一种显示方法,应用于显示装置,显示装置的显示屏包括多个子显示区域。子显示区域可以为一个或多个岛,本公开对于子显示区域包括多少个像素岛并不限定。如图1所示,以显示屏10包括20个可单独驱动的子显示区域11为例,每一个子显示区域11为一个可单独驱动的区域,OSD的待显示区域20包括6个子显示区域11。示例性地,可单独驱动的区域可以为并行的驱动区,或背光硬件驱动为可单独控制的区域,或处理画面内容时可以分区进行画质处理的区域中的任何一种区域。显示装置可以为电视机、个人电脑或其他具有显示屏的装置,本公开对显示装置的具体类型并不限定。
本公开的一些实施例提供一种显示方法,如图2所示,该方法包括:
步骤201,获取待显示图像数据,待显示图像数据包括OSD图像数据,OSD图像数据包括来自n个视点(view)的OSD子图像数据,n为大于或等于2的整数。
在一些实施例中,如图3所示,获取n个视点的OSD子图像数据的方法包括:采用n个相机31,从不同视点拍摄第一OSD模型30,得到n个视点的OSD子图像数据。示例性地,相机31可以为能够获取照片的一种相机,比如虚拟相机或真实相机,本公开对相机31的类型并不限定。
在一些实施例中,上述n个视点的OSD子图像数据可以预先存入系统端电路(比如FPGA(Field Programmable Gate Array,现场可编程逻辑门阵列)和SOC(System on Chip,系统级芯片))的存储器中,需要使用时,从系统端电路的存储器中调用该n个视点的OSD子图像数据。可以理解地,系统端电路的存储器中,可以存入多种OSD图像数据,本公开对于OSD图像数据对应的OSD图像的个数与种类并不限定。
在一些实施例中,获取n个视点的OSD子图像数据的方法也可以包括:采用三维建模或图像处理器获取第二OSD模型,并根据第二OSD模型,得到n个视点的OSD子图像数据。示例性地,第二OSD模型可以由三维建模或图像处理器实时生成,基于该第二OSD模型,针对不同的视点进行切片或图像生成与输出,即可得到n个视点的OSD子图像数据。三维建模或图像处 理器的操作可以由显示装置10的内部系统完成或外部系统完成,本公开对采用三维建模或图像处理器获取第二OSD模型的具体方法并不限定。
待显示图像数据包括OSD图像数据,OSD图像数据包括来自n个视点的OSD子图像数据,也就是说,待显示图像数据包括n个视点的OSD子图像数据,在显示屏10的1帧画面中,可以同时显示多个视点的OSD图像,进而实现OSD的3D显示。如此一来,如图4(a)与图4(b)所示,用户站在不同的视点角度,可以看到显示屏10上不同的OSD图像,如图4(a)所示,用户站在第一视点角度401,可以看到OSD的待显示区域20上的图像为第一视点的OSD图像4011。比如图4(b)所示,如果用户站在在第二视点角度402,可以看到OSD的待显示区域20上的图像为第二视点的OSD图像4021,第一视点的OSD图像4011与第二视点的OSD图像4021为n个视点的OSD图像中的任意两个不同视点的OSD图像。本公开对1帧画面中可以显示的OSD图像的视点个数并不限定。
步骤202,解析OSD图像数据,并对解析后的OSD图像数据进行数据重排。
在一些实施例中,如图1所示,OSD显示区域包括6个子显示区域11,因此,OSD图像数据可以分布在该6个子显示区域11中的至少一个子显示区域11中。但是当将OSD图像数据存储在系统端电路的存储器中时,并没有分区进行存储,若是实时生成的OSD图像数据,也并没有分区进行生成,并且同一个视点的OSD图像数据并未存储在同一个存储地址中。因此为了确保OSD图像的正确显示,需要将OSD图像数据在系统端(比如显卡)进行打散与重组,并将该重组后的OSD图像数据输入显示装置的控制系统。显示装置的控制系统对数据进行解析与重排,显示装置的控制系统根据显示需求将重排后的OSD图像数据输出至显示装置的显示屏10,则可在该显示屏10显示3D的OSD图像。
在一些实施例中,以OSD图像数据,即n个视点的OSD子图像数据存储在系统端电路的存储器中为例,如图5所示,步骤202的实现方法可以包括:
步骤501,解析OSD图像数据,得到OSD图像数据的显示参数和OSD图像数据的显示数据。
示例性地,如图6所示,OSD图像数据的显示参数包括标志位,压缩率,OSD图像数据对应的视点数与子显示区域标识,以及OSD图像数据的显示数据的容量。标志位被配置为与前端指令握手。压缩率被配置为将OSD图像数 据的显示数据进行压缩后的大小与压缩前的大小之比,便于对OSD图像数据的显示数据进行压缩与解压缩。OSD图像数据对应的视点数被配置为存储OSD图像数据所对应的视点为n个视点中的哪一个视点。OSD图像数据对应的子显示区域标识被配置为向后端传递OSD图像数据对应的子显示区域11的位置信息。示例性地,子显示区域11的位置信息可以通过对显示屏10的子显示区域11进行编号来获取,比如从左上角开始,按照从左到右,从上到下的顺序依次编号,每一个编号对应于一个字显示区域。OSD图像数据的显示数据的容量被配置为向后端传递需要显示OSD图像数据的大小,比如当仅需要显示部分视点的OED子图像数据时,OSD图像数据的显示数据的容量比需要显示全部视点的OED子图像数据时,OSD图像数据的显示数据的容量要小,后端可根据该参数合理配置数据处理的资源,以避免系统资源的浪费。示例性地,OSD图像数据的显示参数包括某一个子像素的像素值。
步骤502,根据OSD图像数据的显示参数,对OSD图像数据的显示数据进行数据重排。
在一些实施例中,每个视点的OSD子图像数据包括A个亚像素,A为大于等于2的整数。步骤502的实现方法可以包括:对n个视点的OSD子图像数据中的所有亚像素进行重排。n个视点对应的重排后的OSD子图像数据中不同的亚像素位于不同的行。亚像素可以包括红、绿、蓝、黑、白、黄。
在一些实施例中,A个亚像素包括第一亚像素,第一亚像素为A个亚像素中的任意一个亚像素,第一亚像素包括t个子像素,t为大于或等于1的整数。
如图7A所示,以解析后的OSD图像数据包括n个视点的OSD子图像数据,每个视点的OSD子图像数据包括3个亚像素,该3个亚像素分别为红色(Red,R)亚像素,绿色(Green,G)亚像素,蓝色(Blue,B)亚像素为例。在每个视点的OSD子图像数据中,每个红色亚像素包括t个子像素,比如R1-1,R1-2,…,R1-t,每个绿色亚像素包括t个子像素,比如G1-1,G1-2,…,G1-t,每个蓝色亚像素包括t个子像素,比如B1-1,B1-2,…,B1-t。重排时,根据OSD图像数据的每一个显示数据,比如子像素的像素值,对应的显示参数,按照该显示数据的位置信息进行重排,比如重排后的OSD图像数据中,显示数据R1-1对应于OSD的待显示区域20的起点位置,显示数据Rt-n对应于待显示区域20的终点位置。可以理解地,如图1所示,OSD的待显示区域20为长方形时,OSD的待显示区域20的起点位置即为该长方形左上角的第一个像素所在位置D0,OSD的待显示区域20的终点位置即为该长方形右下 角的最后一个像素所在位置D1。这样一来,重排后的OSD图像数据在时序控制性输入到显示屏10时,即可按照重排后的OSD图像数据进行显示。
示例性地,以n的取值为5,t的取值为4为例,也就是以OSD图像数据包括5个视点的OSD子图像数据,每个亚像素包括4个子像素,每个子像素具有一个像素值,且待显示区域20包括12行*5列像素,该像素与子像素的像素值一一对应为例,进行更直观的说明。
如表1所示,每个视点的OSD子图像数据包括15个子像素的像素值。第一视点的OSD子图像数据501包括红色亚像素的4个子像素的像素值(比如R1-1、R1-2、R1-3和R1-4),绿色亚像素的4个子像素的像素值(比如G1-1、G1-2、G1-3和G1-4)和蓝色亚像素的4个子像素的像素值(比如B1-1、B1-2、B1-3、B1-4)。第二视点的OSD子图像数据502包括红色亚像素的4个子像素的像素值(比如R2-1、R2-2、R2-3和R2-4),绿色亚像素的4个子像素的像素值(比如G2-1、G2-2、G2-3和G2-4)和蓝色亚像素的4个子像素的像素值(比如B2-1、B2-2、B2-3、B2-4)。以此类推,第五视点的OSD子图像数据505包括红色亚像素的4个子像素的像素值(比如R5-1、R5-2、R5-3和R5-4),绿色亚像素的4个子像素的像素值(比如G5-1、G5-2、G5-3和G5-4)和蓝色亚像素的4个子像素的像素值(比如B5-1、B5-2、B5-3、B5-4)。当将OSD图像数据在系统端进行打散与重组,并将该重组后的OSD图像数据输入显示装置的控制系统时,子像素的像素值是压缩后存入系统端电路的存储器中,并不是按照显示时的顺序进行排列的,因此,需要在显示装置的控制系统对OSD图像数据,即5个视点的子图像数据进行解析后重排。
表1 5个视点的OSD子图像数据所包括的显示数据
Figure PCTCN2022096698-appb-000001
Figure PCTCN2022096698-appb-000002
对5个视点的OSD子图像数据中的所有亚像素进行重排,即对5个视点的OSD子图像数据中的红色亚像素、绿色亚像素与蓝色亚像素进行重排。重排后,5个视点的OSD子图像数据中红色亚像素的所有子像素排列在至少一行且互相平行,5个视点的OSD子图像数据中蓝色亚像素的所有子像素排列在至少一行且互相平行,5个视点的OSD子图像数据中绿色亚像素的所有子像素排列在至少一行且互相平行。且5个视点的OSD子图像数据中红色亚像素的任一子像素与5个视点的OSD子图像数据中蓝色亚像素的任一子像素或5个视点的OSD子图像数据中绿色亚像素的任一子像素均排列在不同的行。即,重排后的OSD子图像数据中不同亚像素分别位于不同的行。
示例性地,如表2所示,5个视点对应的重排后的OSD子图像数据包括12行子像素的像素值,5个视点的OSD子图像数据中红色亚像素的子像素的像素值R1-1、R1-2、R1-3、R1-4和R1-5排列在第1行,其中,子像素的像素值R1-1对应于待显示区域20的起点位置D0。5个视点的OSD子图像数据中绿色亚像素的子像素的像素值G1-1、G1-2、G1-3、G1-4和G1-5排列在第2行。5个视点的OSD子图像数据中蓝色亚像素的子像素的像素值B1-1、B1-2、B1-3、B1-4和B1-5排列在第3行。以此类推,5个视点的OSD子图像数据中红色亚像素的子像素的像素值R4-1、R4-2、R4-3、R4-4和R4-5排列在第10行。5个视点的OSD子图像数据中绿色亚像素的子像素的像素值G4-1和G4-2和G4-3和G4-4和G4-5排列在第11行。5个视点的OSD子图像数据中蓝色亚像素的子像素的像素值B4-1、B4-2、B4-3、B4-4和B4-5排列在第12行,其中,子像素的像素值R4-5对应于待显示区域20的终点位置D1。
表2 重排后的OSD子图像数据
Figure PCTCN2022096698-appb-000003
可以理解地,也可以将5个视点的OSD子图像数据中绿色亚像素的子像素的像素值G1-1、G1-2、G1-3、G1-4和G1-5排列在第1行,或者将5个视点的OSD子图像数据中蓝色亚像素的子像素的像素值B1-1、B1-2、B1-3、B1-4和B1-5排列在第1行。本公开对于n个视点的OSD子图像数据中红色亚像素的子像素所在的行与n个视点的OSD子图像数据中绿色亚像素的子像素所在的行或n个视点的OSD子图像数据中蓝色亚像素的子像素所在的行的先后顺序并不限定。也就是说,多个亚像素重排后的顺序可以包括多种形式,只要重排后的不同亚像素分别位于不同的行就都在本公开实施例的保护范围内。通过将n个视点的OSD子图像数据进行重排,确保重排后的不同亚像素分别位于不同的行,从而能够保证较好的3D的显示效果,呈现在用户眼前的3D场景较为真实。
步骤203,根据OSD图像数据的位置信息,确定OSD图像数据在显示屏10上对应的待显示区域20,并在待显示区域20显示重排后的OSD图像数据。待显示区域20包括至少一个子显示区域11,重排后的OSD图像数据包括重排后的部分视点或全部视点的OSD子图像数据。
在一些实施例中,如图1所示,OSD图像数据的位置信息包括OSD的待显示区域20的起点位置D0,以及OSD的待显示区域20的长度信息及宽度信息,通过OSD的待显示区域20的起点位置D0以及OSD的待显示区域20的长度信息及宽度信息可以计算得到OSD待显示区域20的终点位置D1。系统端电路可根据显示屏10配置或者生成OSD的待显示区域20的起点位置D0的坐标,以及OSD的待显示区域20的长度信息及宽度信息,OSD图像数据所属的子显示区域11的位置等参数,当显示装置的数据流到达OSD的待显示区域20的起点位置D0,按照时序将重排后的5个视点的OSD子图像数据输出至OSD的待显示区域20,并显示5个视点的OSD图像。
示例性地,重排后的OSD图像数据也可以包括部分视点的OSD子图像数据,比如,重排时,仅对5个视点的OSD子图像数据中的3个视点的子图像数据进行了重排,则当显示装置的数据流到达OSD的待显示区域20的起点位置D0,按照时序将重排后的3个视点的OSD子图像数据输出至OSD的待显示区域20,并显示3个视点的OSD图像。再比如,重排时,虽然对5个视点的OSD子图像数据都进行重排,但是显示时可以仅显示3个视点的子图像数据。
在一些实施例中,待显示图像数据还包括源图像数据,显示方法还包括:解析源图像数据,并对解析后的图像数据进行数据重排。
示例性的,源图像数据也可以包括来自多个视点的源图像数据,本公开实施例对于源图像数据的获取方法并不进行限定。如图7B所示,源图像数据包括n个视点的源图像数据,每个视点的子图像数据包括A个亚像素,可以理解地,为了保证视觉效果,每个视点的源图像数据所包括的亚像素的种类与个数,与每个视点的OSD子图像数据所包括的像素的种类与个数相同。每个视点的源图像数据的每个亚像素包括m个子像素,m为大于等于1的整数,可以理解地,该m与上述的每个OSD子图像数据的每个亚像素包括t个子像素中的t,取值可以相同,也可以不同。
此外,源图像的显示范围可以为整块显示屏10也可以为显示屏10的部分显示区域,本公开的一些实施例以源图像的显示范围为整块显示屏10为例进行说明。对源图像数据的解析与重排的方法,与对OSD图像数据的解析与重排的方法相类似,此处不再赘述。
如图7B所示,以解析后的源图像数据包括n个视点的源图像数据,每个视点的源图像数据包括3个亚像素,该3个亚像素分别为红色亚像素,绿色亚像素,蓝色亚像素为例。在每个视点的源图像数据中,每个红色亚像素包括m个子像素,比如OR1-1,OR2-1,…,ORm-1,每个绿色亚像素包括m个子像素,比如OG1-1,OG2-1,…,OGm-1,每个蓝色亚像素包括m个子像素,比如OB1-1,OB2-1,…,OBm-1。
重排时,如图7B所示,根据源图像数据的每一个显示数据,比如子像素的像素值,对应的显示参数,按照该显示数据的位置信息进行重排,比如重排后的源图像数据中,显示数据OR1-1对应于显示屏10的起点位置,显示数据ORm-n对应于显示屏10的终点位置。如图1所示,可以理解地,当显示屏10为一个长方形,该长方形的左上角第一个像素所在的位置即为显示屏10的起点位置,该长方形的右下角最后一个像素所在的位置即为显示屏10的终点位置。这样一来,重排后的源图像数据在时序控制性输入到显示屏10时,即可按照重排后的源图像数据进行显示。
以解析后的源图像数据包括5个视点为例,每个子图像包括3个亚像素,每个亚像素包括4个子像素,显示屏10包括12行*5列像素为例,该像素与源图像数据子像素的像素值一一对应为例,对重排后的源图像数据进行更直观的说明。重排后的源图像数据如表3所示。其中,子像素的像素值OR1-1对应于显示屏10的左上角第一个像素,子像素的像素值OR1-1对应于显示屏10的右下角的最后一个像素。
表3 重排后的源图像数据
Figure PCTCN2022096698-appb-000004
在一些实施例中,如图1所示,显示屏10的显示区域大于OSD的待显示区域20,此时,当源图像与OSD图像均需要在显示屏10上进行显示时,根据OSD图像数据在显示屏10上对应的待显示区域20,在待显示区域20显示重排后的OSD图像数据,包括:根据OSD图像数据在显示屏10上对应的待显示区域20,采用重排后的OSD图像数据替换待显示区域20对应的重排后的源图像数据,或者,在待显示区域20对应的重排后的源图像数据上叠加重排后的OSD图像数据。
示例性地,如图7C与图7D所示,将位置信息位于OSD的待显示区域20范围内的重排后的源图像数据(比如OR1-x与OR1-y,x与y均为小于或等于m的整数)直接替换为重排后的OSD图像数据(比如R1-1与R1-2),比如将起点位置D0的显示数据由OR1-x替换为R1-1,也就是说,在OSD的待显示区域20不进行源图像的显示,仅显示OSD图像。
示例性地,在数据流到达OSD的待显示区域20的起点位置D0之前,OSD的待显示区域20的显示数据为如图7B所示的重排后的源图像数据,当数据流到达OSD的待显示区域20的起点位置D0时,从该起点位置D0起,将如图7A所示的重排后的OSD图像数据作为OSD的待显示区域20的显示数据。即,在OSD的待显示区域20中,采用重排后的OSD图像数据替换对应的重排后的源图像数据。因此,如图7D所示,在显示屏10的OSD的待显示区域20显示的是重排后的OSD图像,在OSD的待显示区域20以外的显示屏10上,显示的是重排后的源图像数据。
在一些实施例中,如图7E与图7F所示,在OSD的待显示区域20对应的重排后的源图像数据上叠加重排后的OSD图像数据,在OSD的待显示区域20显示源图像与OSD图像融合后的图像。
在一些实施例中,叠加的方式可以包括对重排后的OSD图像数据设置透明度,在待显示区域20对应的重排后的源图像数据上叠加设置透明度后的重排后的OSD图像数据。示例性地,如图7E所示,OSD的待显示区域20的起点位置D0的显示数据由该位置上源图像的显示数据OR1-x替换为为R1-1’,R1-1’是对重排后的OSD图像数据R1-1设置透明度后叠加至源图像数据OR1-x得到的,R1-2’是对重排后的OSD图像数据R2-1’设置透明度后叠加至源图像数据OR2-x得到的,以此类推。
示例性地,在数据流到达OSD的待显示区域20的起点位置D0之前,OSD的待显示区域20的显示数据为如7B所示的重排后的源图像数据,数据流到达OSD的待显示区域20的起点位置D0时,如图7E所示,将OSD的待显示区域20对应的重排后的源图像数据上叠加重排后的OSD图像数据,得到叠加后的显示数据,从OSD的待显示区域20的起点位置D0起,将该叠加后的显示数据作为OSD的待显示区域20的显示数据。因此如图7F所示,在显示屏10的OSD的待显示区域20显示的是OSD图像与源图像叠加后的图像,在OSD的待显示区域20以外的显示屏10上,显示的是重排后的源图像数据。
在一些实施例中,以图8和图9为例,说明需要对源图像与OSD图像同时显示时的整体流程。
如图8所示,首先,前端通过数据接口传输源图像数据,前端可以为系统端的某个部分,也可以为外部装置,本公开对前端的数量与类型并不限定。数据接口可以包括多个,比如数据接口可以包括系统接口(System Interface),或内部接口(Inner Interface),或显示端口(Display Port),或高清多媒体接口(High Definition Multimedia Interface,HDMI),或内置或嵌入式的显示接口(External/Embedded Display Port),或移动产业处理器接口(Mobile Industry Processor Interface)种的任何一个或多个,本公开对于源图像的数据来源以及数据接口的类型并不限定。系统端的数据接收与转换模块接收到源图像的数据后,将转换后的源图像数据输出至数据解析与重排模块,进行数据解析与重排重排后的源图像数据可存储于系统端电路的DDR(Double Data Rate Synchronous Dynamic Random Access Memory,双倍速率同步动态随机存储器也可以简写为DDR SDRAM,)中。当需要对源图像数据进行显示时,可从DDR中调用重排后的源图像数据,并对该重排后的源图像数据进行显示数据的处理,数据处理可以包括具体像素值的调节。若此时没有OSD显示指令,则可直接将该处理后的源图像的显示数据输出至数据转换与发送模块输 出至后端,并在显示时序控制下进行源图像的显示。后端可以为显示装置的显示屏10,也可以为显示装置的驱动电路,本公开对后端的类型与数量并不限定。
示例性地,当用户端发出OSD显示命令,系统端读取2D显示与3D显示的切换命令,并通过OSD功能控制模块控制OSD图像数据获取模块来获取2D显示或3D显示所对应的OSD图像数据。将获取到的OSD图像数据输出至OSD图像数据处理模块,由OSD图像数据处理模块对获取的OSD图像数据进行处理,该处理可以包括OSD图像数据的解析与重排,获取处理后的OSD图像数据。将处理后的源图像的显示数据与处理后的OSD图像的显示数据输出至数据融合模块进行数据融合,融合后的源图像的显示数据与OSD图像的显示数据经数据转换与发送模块输出至后端。融合后的源图像的显示数据与OSD图像的显示数据,可存储在显示装置的存储介质(如EEPROM(Electrically Erasable Programmable Read-Only Memory,电可擦编程只读存储器,Flash(Flash EEPROM Memory,闪存)和ROM(Read-Only Memory,只读存储器)中。当需要显示时,从该存储介质中输出融合后的数据,并在显示时序的控制下进行源图像与OSD图像的显示。对处理后的源图像的显示数据与处理后的OSD图像的显示数据进行数据融合的方式,可以为替换或者叠加的方式。
示例性地,以图9为例,说明源图像的数据和OSD图像数据如何转化为显示屏10上的显示数据,并在显示屏10上进行显示。
首先,获取待显示图像数据,该待显示图像数据包括源图像数据与OSD图像数据,OSD图像数据可由系统端电路的存储器中获取,如图8所示,源图像数据可由前端获取。然后,在数据解析时,对源图像数据与OSD图像数据进行数据解析,解析包括解压缩,通过解压缩可以分别得到源图像数据与OSD图像数据的显示参数和与显示参数相对于的显示数据,之后在时钟的规范下,将图像数据的显示数据与显示参数进行一一对应。
接着,进行数据重排,分别对源图像数据与OSD图像数据的一种或多种显示参数进行遍历,获取需要的显示数据,该需要的显示数据可以是某一个区域的显示数据。对该需要的显示数据进行数据处理,数据处理的方式可以是像素值的调整,可以是因为渲染操作的增加或减少,得到处理后的显示数据,该处理后的显示数据可以与在系统端电路中存储的OSD图像的显示数据不一样。将该处理后的显示数据与其所属的显示参数进行一一对应。
最后,在图像显示时,首先获取处理后的显示数据所对应的显示参数中 关于显示区域的信息,即获取处理后的显示数据对应的显示区域。当处理后的显示数据的数据流到达OSD的待显示区域20的起点位置D0,可以显示处理后的OSD图像的显示数据与处理后的源图像的显示数据得到的叠加数据,也可以只显示处理后的OSD图像的显示数据,对于显示屏10上OSD待显示区域20以外的显示区域,可以直接显示处理后的源图像的显示数据。
可以理解的,源图像数据与OSD图像数据、解析后的源图像数据与解析后的OSD图像数据、重排后的源图像数据与重排后的OSD图像数据,处理后的源图像数据与处理后的OSD图像数据。融合后的数据均可以存储于系统端电路的存储器中或者显示装置的存储器中,以便于下一次调用或直接使用。本公开虽然以OSD图像数据存储于系统端电路的存储器中为例进行说明,但是并不限定实际使用中只能基于存储在系统端电路的OSD图像数据进行OSD图像的显示,并且不限定实际使用中对于源图像数据和OSD图像数据是否必须进行解析、重排、处理与融合。也就是说,当系统端电路的存储器或显示装置的存储器中储存有解析后的源图像数据与解析后的OSD图像数据、重排后的源图像数据与重排后的OSD图像数据,处理后的源图像数据与处理后的OSD图像数据,融合后的数据时,可以根据实际需求直接调用或者使用所需要的数据。
在一些实施例中,如图10所示,显示方法还包括:
步骤1001、获取第一命令,第一命令用于将OSD图像的显示方式由3D显示切换至2D显示,或者,将OSD图像的显示方式由2D显示切换至3D显示。
示例性地,第一命令可以为切换3D显示与2D显示的使能信号EN,当需要显示OSD图像时,通过该使能信号,可以由3D显示切换至2D显示,也可以由2D显示切换为3D显示。本公开对第一命令的获取方式并不限定。示例性地,该使能信号可以由用户持有的操作手柄或遥控器上的切换按键给出,用户每按下该切换按键一次,3D显示与2D显示切换一次。比如,按键未按下时,使能信号为低,系统端获取到第一命令,并将OSD图像的显示方式切换为2D显示,OSD图像的显示方式为2D显示。当用户按下该切换按键,使能信号置高,系统端获取到第一命令,并将OSD图像的显示方式切换为3D显示,则OSD图像的显示方式为3D显示。示例性地,使能信号EN也可以由系统端生成,比如,系统端判定某个参数或多个参数达到某种预设的条件后,进行2D显示与3D显示的自动转换。
步骤1002、响应于第一命令,获取二维的OSD图像数据或三维OSD 图像数据。
示例性地,二维的OSD图像数据,可以通过对n个视点的OSD子图像数据去3D化得到,当系统端获取第一命令,可调用系统端电路的存储器中的n个视点的OSD子图像数据,并进行去3D化的数据处理,得到二维的OSD图像数据。示例性地,二维的OSD图像数据,也可以存入系统端电路的存储器中,当系统端获取第一命令,可调用系统端电路的存储器中的该二维的OSD图像数据。示例性地,二维的OSD图像数据,还可以为n个视点的OSD子图像数据中某一个视点的OSD子图像数据,当系统端获取第一命令,可直接调用系统端电路的存储器中的n个视点的OSD子图像数据中的该某一个视点的OSD子图像数据。本公开对二维的OSD图像数据的获取方式并不限定。三维OSD图像数据包括上述n个视点的OSD子图像数据,可以通过拍摄第一OSD模型的多个视点的图片得到,也可以通过第二OSD模型实时获取。
步骤1003、根据二维的OSD图像数据对应的待显示区域,在显示屏上显示二维的OSD图像;或,根据三维的OSD图像数据对应的待显示区域,在显示屏上显示三维的OSD图像。
示例性地,二维的OSD图像数据对应的待显示区域20可以有系统端配置或者生成,将二维的OSD图像数据中的每个亚像素的子像素的像素值,按照显示时序输出至该待显示区域20,则可以在该待显示区域20显示二维的OSD图像。
在一些实施例中,如图11所示,显示方法还包括:
步骤1101、获取用户的人脸图像或瞳孔图像,确定用户的主观测区;和/或获取用户的手势图像,确定用户的主视点区。
示例性地,如图12所示,可以通过摄像头采集用户的人脸图像,或瞳孔图像,或手势图像,该摄像头可以为可见光摄像头和/或红外摄像头。本公开对用户的人脸图像,或瞳孔图像,或手势图像的获取方式并不限定。
示例性地,以摄像头采集的方式举例说明,对采集得到的用户的人脸图像进行人脸检测的数据处理,然后进行瞳孔检测,如图13所示,分析用户眼睛的位置以及相对于屏幕的观测角度,以确定用户眼睛的位置(比如E1与E2)或观测角度(比如α与β)或注视点。也可以直接对采集得到的用户的瞳孔图像进行瞳孔检测的图像处理,分析用户眼睛的位置以及相对于屏幕的观测角度,以确定用户眼睛的位置(比如E1与E2)或观测角度(比如α与β)或注视点。比如,由人脸识别获取特征点,得到眉心坐标,再根据眉心与用户眼睛的距离推理得出用户眼睛眼坐标(比如E1与E2),则可以确定用户 眼睛的位置。比如,根据眉心坐标,计算眉心坐标与OSD的显示区域中每一个子显示区域11的横向边界的夹角(比如α与β),即可确定用户的观测角度或注视点。通过用户眼睛的位置E1或观测角度α或注视点,可确定用户的主观测区D11,通过用户眼睛的位置E1或观测角度β或注视点,可确定用户的主观测区D12,显示屏10上区域D2则为主观测区D11与主观测区D12以外的显示区域。
示例性地,如图12所示,还可以对采集得到的用户的手势图片进行手势识别的处理,分析用户手势的动作或者手势的动作趋势,以确定手势指令以及投射显示位置。因此可以确定用户在使用显示屏10上的OSD图像进行交互时主要关注的区域是哪些视点区,即可确定用户的主视点区。
步骤1102、对主观测区和/或主视点区的OSD图像数据和/或源图像数据进行高清显示。对主观测区和/或主视点区以外的OSD图像数据和/或源图像数据进行低清显示。
示例性地,如图13所示,对于在主观测区D11与主观测区D12显示的OSD图像数据和/或源图像数据进行高清显示。对显示屏10上区域D2上的OSD图像数据和/或源图像数据进行低清显示。
示例性地,如图12所示,解析n个视点的子图像数据,并获取数据的位置信息后,对n个视点的子图像数据分区进行处理。以实现对主观测区和/或主视点区的OSD图像数据和/或源图像数据进行高清显示,并对主观测区和/或主视点区以外的OSD图像数据和/或源图像数据进行低清显示。
示例性地,对主观测区和/或主视点区的OSD图像数据和/或源图像数据进行高清显示,可通过着重处理主观测区和/或主视点区的OSD图像数据和/或源图像数据,着重处理可以包括对OSD图像和/或源图像的画质、色温进行着重调节。比如可以增加主观测区和/或主视点区的OSD图像数据和/或源图像的渲染层数,或增加在主观测区和/或主视点区的实时生成的第二OSD模型的复杂程度,或者在系统端电路中调用在主观测区和/或主视点区所有视点的OSD子图像数据和/或源图像数据,或者传输至显示屏10时不压缩主观测区和/或主视点区的OSD图像数据和/或源图像数据。着重处理还可以包括对主观测区和/或主视点区的显示亮度进行着重调节。本公开对如何通过着重处理主观测区和/或主视点区的OSD图像数据和/或源图像数据,以实现对主观测区和/或主视点区的OSD图像数据和/或源图像数据进行高清显示的具体方式并不限定。
示例性地,对主观测区和/或主视点区以外的OSD图像数据和/或源图像 数据进行低清显示,可通过简单处理主观测区和/或主视点区以外的OSD图像数据和/或源图像数据,简单处理可以包括对OSD图像和/或的画质、色温进行简单调节。比如,可以减少主观测区和/或主视点区以外的OSD图像数据和/或源图像数据渲染层数,或减少在主观测区和/或主视点区的实时生成的第二OSD模型的模型复杂度,或对在系统端电路中调用的主观测区和/或主视点区以外的OSD图像数据和/或源图像数据进行压缩,或仅选择部分视点的OSD图像子数据作为主观测区和/或主视点区以外的OSD图像数据和/或源图像数据。可以理解地,该部分视点指的是小于n个视点,也就是说,也可以选择0个视点的OSD图像子数据和/或0个视点的源图像数据作为主观测区和/或主视点区以外的OSD图像数据和/或源图像数据,即对主观测区和/或主视点区以外的OSD图像数据不做显示。简单处理还可以包括对主观测区和/或主视点区的显示亮度进行简单调节。本公开对如何通过简单处理主观测区和/或主视点区以外的OSD图像数据,以实现对主观测区和/或主视点区以外的OSD图像数据和/或源图像数据进行低清显示的具体方式并不限定。
示例性地,当需要对OSD图像和或源图像的画质、色温进行调节,可通过建立不同的查找表,直接而快速地通过查找表进行画质调节。查找表可通过算法得出,并存储在系统端电路的存储器中。对应于每一个子像素,在每一个子显示区域11可建立多个查找表,以实现在子显示区域11进行不同程度的画质调节。比如,某一个子像素在第一子显示区域,可通过第一查找表进行画质调节,以实现简单的画质调节,该某一个子像素在第一子显示区域,可通过第二查找表进行画质调节,以实现复杂的画质调节。第一子显示区域可以为显示屏10上的任一个子显示区域11。
示例性地,如图14所示,当需要对OSD的待显示区域20的显示亮度进行调节,可根据OSD的待显示区域20,在系统端生成亮度调节全局系数,在OSD的待显示区域20,可基于该全局系数生成第一背光值。根据该全局系数以及对显示屏10的区域划分,可以生成对应于多个区域的多个调节系数,调节系数的调节方式可以是乘以第一固定参数,或者加上第二固定参数,可以基于主观测区和/或主视点区的调节系数生成第二背光值。比如,可以将该主观测区和/或主视点区上的每个子像素的像素值乘以该主观测区和/或主视点区对应的第一固定参数,或者将该主观测区和/或主视点区上的每个子像素的像素值加上该主观测区和/或主视点区对应的第二固定参数,以改变该主观测区和/或主视点区上的每个子像素的像素值,进而调节OSD图像和/或源图像在该主观测区和/或主视点区的显示亮度。可以基于主观测区和/或主视点区以 外的显示区域的调节系数生成第三背光值,其显示亮度调节的方式与主观测区和/或主视点区的显示亮度调节方式类似,此处不再赘述。对第一背光值、第二背光值和第三背光值分别进行处理后,整合并输出,即可在OSD的待显示区域20上通过背光值实现显示屏10显示亮度的调节。可以理解地,通过全局系数或调节系数,还可以将显示屏10上的任何一个区域的显示亮度调节为灰度显示。
示例性地,还可使用局域调光(Local Dimming)算法实现对显示屏10不同区域的高清显示或低清显示。例如,可以通过简单的调光(Dimming)算法,比如单一算法对主观测区和/或主视点区以外的OSD图像数据和/或源图像数据进行低清显示,且可以通过调光算法中多种算法的加权对主观测区和/或主视点区的OSD图像数据和/或源图像数据进行高清显示。
如图14所示,由于主观测区和/或主视点区的显示亮度较强,而主观测区和/或主视点区以外的显示区域的显示亮度较弱,不做任何处理可能会降低视觉效果,使得用户的体验感较差,因此可以对所有OSD的待显示区域20的显示数据进行画面补偿处理,若主观测区和/或主视点区包括多个子显示区域11,可对跨区域的显示数据进行过渡算法的处理,跨区域的显示数据包括主观测区和/或主视点区的显示数据,处理后的OSD的待显示区域20的显示数据经整合后输出,在主观测区和/或主视点区的边界位置以及子显示区域11相接的位置,则可以有更好的视觉效果。
在一些实施例中,当需要显示的OSD图像为OSD交互内容,该OSD交互内容所在的显示区域被强制判断为主观测区和/或主视点区,此时显示屏10上主观测区和/或主视点区的显示亮度提高,显示屏10上主观测区和/或主视点区的显示亮度以外的显示区域的显示亮度降低,可以引导观看者将视线移动至显示屏10上的OSD交互内容。此时,可以直接滤除主观测区和/或主视点区以外的OSD图像数据和/或源图像数据,仅使用调光算法对主观测区和/或主视点区的OSD图像数据进行高清显示,以减少计算量。
示例性地,如图15所示,用户可以使用OSD图像进行互动,当系统端发出OSD图像的显示指令,显示屏10接收到该OSD图像的显示指令后,即在OSD的待显示区域20显示OSD图像,维持T1秒,T1为正数,若在T1秒的时间内,用户进行了交互操作,显示屏10可以将该互动操作以指令的方式传回系统端(如显卡),系统端判断指令类型,再与显示屏进10行数据传输。
示例性地,如图15所示,指令可以为第一指令,该第一指令为切换至另 一种OSD图像,则系统端控制显示装置的显示屏10显示另一种OSD图像,并维持T2秒,T2为正数,在T2秒内再次判断是否有指令下发。指令也可以为第二指令,该第二指令为进入OSD图像的第一子内容,则系统端控制显示装置的显示屏10显示该第一子内容,并维持T3秒,T3为正数,在T3秒内再次判断是否有指令下发。指令也可以为第三指令,该第三指令为进入OSD图像的第二子内容,则系统端控制显示装置的显示屏10显示该第二子内容,并维持T4秒,T4为正数,在T4秒内再次判断是否有指令下发。指令也可以为第四指令,该第四指令为退出指令此时系统端控制显示装置的显示屏10显示源图像,OSD图像消失。若在T1秒或T2秒或T3秒或T4秒内,无任何指令发出,则系统端控制显示装置的显示屏10显示源图像,OSD图像消失。可以理解地,T1、T2、T3和T4的具体数值可以相同也可以不同,本公开对与T1、T2、T3和T4的具体数值并不限定。第二子内容可以为OSD图像的子内容,也可以为第一子内容的子内容,OSD图像的子内容可以为任意一种子内容,本公开对于第一子内容与第二子内容的具体种类以及优先级和层级并不限定。
本公开的一些实施例提供了一种计算机可读存储介质(例如,非暂态计算机可读存储介质),该计算机可读存储介质中存储有计算机程序指令,计算机程序指令在计算机(例如,显示装置)上运行时,使得计算机执行如上述实施例中任一实施例所述的显示方法。
示例性的,上述计算机可读存储介质可以包括,但不限于:磁存储器件(例如,硬盘、软盘或磁带等),光盘(例如,CD(Compact Disk,压缩盘)、DVD(Digital Versatile Disk,数字通用盘)等),智能卡和闪存器件(例如,EPROM(Erasable Programmable Read-Only Memory,可擦写可编程只读存储器)、卡、棒或钥匙驱动器等)。本公开描述的各种计算机可读存储介质可代表用于存储信息的一个或多个设备和/或其它机器可读存储介质。术语“机器可读存储介质”可包括但不限于,无线信道和能够存储、包含和/或承载指令和/或数据的各种其它介质。
本公开的一些实施例还提供了一种计算机程序产品。该计算机程序产品包括计算机程序指令,在计算机(例如,显示装置)上执行该计算机程序指令时,该计算机程序指令使计算机执行如上述实施例所述的显示方法。
本公开的一些实施例还提供了一种计算机程序。当该计算机程序在计算机(例如,显示装置)上执行时,该计算机程序使计算机执行如上述实施例所述的显示方法。
上述计算机可读存储介质、计算机程序产品及计算机程序的有益效果和上述一些实施例所述的显示方法的有益效果相同,此处不再赘述。
本公开的一些实施例还提供一种显示装置,如图16所示,显示装置160包括显示屏10、数据获取装置161和数据处理装置162。
如图1所示,显示屏10,包括多个子显示区域11。
如图16所示,数据获取装置161,被配置为获取待显示图像数据,待显示图像数据包括OSD图像数据,OSD图像数据包括来自n个视点的OSD子图像数据,n为大于或等于2的整数。
如图16所示,数据处理装置162,被配置为解析OSD图像数据,并对解析后的OSD图像数据进行数据重排。根据OSD图像数据的位置信息,确定OSD图像数据在显示屏10上对应的待显示区域20,并控制显示屏10上的待显示区域20显示重排后的OSD图像数据;待显示区域20包括至少一个子显示区域11,重排后的OSD图像数据包括重排后的部分视点或全部视点的OSD子图像数据。示例性地,如图8所示,数据处理装置可以包括数据接收与转换模块、数据解析与重排模块,显示数据处理模块、OSD图像的显示数据处理模块、数据融合模块以及数据转换与发送模块。
基于上述方案,由于待显示图像数据包括OSD图像数据,OSD图像数据包括来自n个视点的OSD子图像数据,也就是说,待显示图像数据包括n个视点的OSD子图像数据,该n个视点的OSD子图像数据可存储在系统端电路的存储器中。为了确保图像的正确显示,需要将图像数据在系统端(比如显卡)进行打散与重组,并将该重组后的图像数据输入显示装置的控制系统。显示装置的控制系统对数据进行解析与重排,显示装置的控制系统根据显示需求将重排后的图像数据输出至显示装置的显示屏10,则可在显示屏10的1帧画面中,可以同时显示多个视点的OSD图像,进而实现OSD的3D显示。
在一些实施例中,如图9所示,待显示图像数据还包括源图像数据,数据处理装置162还被配置为:解析源图像数据,并对解析后的源图像数据进行数据重排。源图像的显示范围可以为整块显示屏10也可以为显示屏10的部分显示区域,
在一些实施例中,如图7A至图7F所示,数据处理装置162被配置为根据OSD图像数据对应的待显示区域20,采用重排后的OSD图像数据替换待显示区域20对应的重排后的源图像数据,或者,在待显示区域20对应的重排后的源图像数据上叠加重排后的OSD图像数据。
在一些实施例中,如图7E和图7F所示,数据处理装置被配置为对重排 后的OSD图像数据设置透明度,在待显示区域20对应的重排后的源图像数据上叠加设置透明度后的重排后的OSD图像数据。
在一些实施例中,如图7A所示,每个视点的OSD子图像数据包括A个亚像素,A为大于等于2的整数;数据处理装置被配置为:对n个视点的OSD子图像数据中的所有亚像素进行重排;n个视点对应的重排后的OSD子图像数据中不同的亚像素位于不同的行。亚像素可以包括红、绿、蓝、黑、白、黄。
在一些实施例中,如图7A所示,A个亚像素包括第一亚像素,第一亚像素为A个亚像素中的任意一个亚像素,第一亚像素包括t个子像素,t为大于或等于2的整数。每个视点的源图像数据的每个亚像素包括m个子像素,m为大于等于1的整数,可以理解地,该m与t的取值可以相同,也可以不同。
在一些实施例中,数据处理装置162被配置为解析OSD图像数据,得到OSD图像数据的显示参数和OSD图像数据的显示数据;OSD图像数据的显示参数包括OSD图像数据对应的视点与子显示区域标识,以及OSD图像数据的显示数据的容量。根据OSD图像数据的显示参数,对OSD图像数据的显示数据进行数据重排。OSD图像数据对应的子显示区域标识可以通过对显示屏10的子显示区域11进行标号来获取。
在一些实施例中,数据获取装置161被配置为:采用n个虚拟相机,从不同视点拍摄第一OSD模型,得到n个视点的OSD子图像数据;或者,采用三维建模或图像处理器获取第二OSD模型,并根据第二OSD模型,得到n个视点的OSD子图像数据。第二OSD模型可以由三维建模或图像处理器实时生成,基于该第二OSD模型,针对不同的视点进行切片或图像生成与输出,即可得到n个视点的OSD子图像数据。三维建模或图像处理器的操作可以由显示装置10的内部系统完成或外部系统完成。
在一些实施例中,OSD图像数据的位置信息包括待显示区域20的起点位置,以及待显示区域20的长度信息及宽度信息。根据待显示区域20的起点位置信息以及待显示区域20的长度信息及宽度信息,就可以确定显示区域20的范围。
在一些实施例中,数据获取装置还被配置为:获取第一命令,第一命令用于将OSD图像的显示方式由3D显示切换至2D显示,或者,将OSD图像的显示方式由2D显示切换至3D显示。响应于第一命令,获取二维的OSD图像数据或三维OSD图像数据。数据处理装置被配置为根据二维的OSD图像数据对应的待显示区域20,控制显示屏10显示二维的OSD 图像;或根据三维的OSD图像数据对应的待显示区域20,控制显示屏10显示三维的OSD图像。因而,本公开可以实现2D显示与3D显示的切换。
在一些实施例中,显示装置还包括采集装置,被配置为获取用户的人脸图像或瞳孔图像,确定用户的主观测区;和/或获取用户的手势图像,确定用户的主视点区。数据处理装置还被配置为对主观测区和/或主视点区的OSD图像数据进行高清显示;对主观测区和/或主视点区以外的OSD图像数据进行低清显示。通过合理配置数据处理装置162,对不同的区域进行不同的数据处理,可以保证较好的视觉效果的同时,避免计算资源的浪费。
通过本公开的一些实施例提供的显示装置,可以实现OSD图像的3D显示,并且用户在不同的视角,可以看到不同的OSD图像,此外,对于不同的区域,可以分区域进行高清显示或低清显示。
以上,仅为本公开的具体实施方式,但本公开的保护范围并不局限于此,任何熟悉本技术领域的技术人员在本公开揭露的技术范围内,想到变化或替换,都应涵盖在本公开的保护范围之内。因此,本公开的保护范围应以权利要求的保护范围为准。

Claims (24)

  1. 一种显示方法,应用于显示装置,所述显示装置的显示屏包括多个子显示区域,所述显示方法包括:
    获取待显示图像数据,所述待显示图像数据包括OSD图像数据,所述OSD图像数据包括来自n个视点的OSD子图像数据,所述n为大于或等于2的整数;
    解析所述OSD图像数据,并对解析后的所述OSD图像数据进行数据重排;
    根据所述OSD图像数据的位置信息,确定所述OSD图像数据在所述显示屏上对应的待显示区域,并在所述待显示区域显示重排后的OSD图像数据;所述待显示区域包括至少一个所述子显示区域,所述重排后的OSD图像数据包括重排后的部分视点或全部视点的OSD子图像数据。
  2. 根据权利要求1所述的显示方法,其中,所述待显示图像数据还包括源图像数据,所述方法还包括:
    解析所述源图像数据,并对解析后的所述源图像数据进行数据重排。
  3. 根据权利要求1或2所述的显示方法,其中,所述根据所述OSD图像数据在所述显示屏上对应的待显示区域,在所述待显示区域显示所述重排后的OSD图像数据,包括:
    根据所述OSD图像数据在所述显示屏上对应的待显示区域,采用所述重排后的OSD图像数据替换所述待显示区域对应的重排后的所述源图像数据,或者,在所述待显示区域对应的重排后的所述源图像数据上叠加所述重排后的OSD图像数据。
  4. 根据权利要求3所述的显示方法,其中,所述在所述待显示区域对应的重排后的所述源图像数据上叠加所述重排后的OSD图像数据,包括:
    对所述重排后的OSD图像数据设置透明度,在所述待显示区域对应的重排后的所述源图像数据上叠加设置透明度后的所述重排后的OSD图像数据。
  5. 根据权利要求1~4中任一项所述的显示方法,其中,每个视点的OSD子图像数据包括A个亚像素,所述A为大于等于2的整数;对解析后的所述OSD图像数据进行数据重排,包括:
    对所述n个视点的OSD子图像数据中的所有亚像素进行重排;所述n个视点对应的重排后的OSD子图像数据中不同的亚像素位于不同的行。
  6. 根据权利要求5所述的显示方法,其中,所述A个亚像素包括第一亚像素,所述第一亚像素为所述A个亚像素中的任意一个亚像素,所述第一亚像素包括t个子像素,所述t为大于或等于2的整数。
  7. 根据权利要求1~6中任一项所述的显示方法,其中,所述解析所述OSD图像数据,并对解析后的所述OSD图像数据进行数据重排,包括:
    解析所述OSD图像数据,得到所述OSD图像数据的显示参数和所述OSD图像数据的显示数据;所述OSD图像数据的显示参数包括所述OSD图像数据对应的视点数与子显示区域标识,以及所述OSD图像数据的显示数据的容量;
    根据所述OSD图像数据的显示参数,对所述OSD图像数据的显示数据进行数据重排。
  8. 根据权利要求1~7中任一项所述的显示方法,其中,所述获取n个视点的OSD子图像数据的方法,包括:
    采用n个相机,从不同视点拍摄第一OSD模型,得到所述n个视点的OSD子图像数据;或者,
    采用三维建模或图像处理器获取第二OSD模型,并根据所述第二OSD模型,得到所述n个视点的OSD子图像数据。
  9. 根据权利要求1~7中任一项所述的显示方法,其中,所述OSD图像数据的位置信息包括所述待显示区域的起点位置,以及所述待显示区域的长度信息及宽度信息。
  10. 根据权利要求1~7中任一项所述的显示方法,还包括:
    获取第一命令,所述第一命令用于将OSD图像的显示方式由3D显示切换至2D显示,或者,将OSD图像的显示方式由2D显示切换至3D显示;
    响应于所述第一命令,获取二维的OSD图像数据或三维OSD图像数据;
    根据所述二维的OSD图像数据对应的待显示区域,在所述显示屏上显示二维的OSD图像;或根据所述三维的OSD图像数据对应的待显示区域,在所述显示屏上显示三维的OSD图像。
  11. 根据权利要求1~10中任一项所述的显示方法,还包括:
    获取用户的人脸图像或瞳孔图像,确定用户的主观测区;和/或获取用户的手势图像,确定用户的主视点区;
    对主观测区和/或主视点区的OSD图像数据和/或源图像数据进行高清显示;
    对主观测区和/或主视点区以外的OSD图像数据和/或源图像数据进行低清显示。
  12. 一种可读存储介质,所述计算机可读存储介质存储有计算机程序指令,所述计算机程序指令在计算机上运行时,使得所述计算机执行如权利要求1~11中任一项所述的显示方法。
  13. 一种计算机程序产品,所述计算机程序产品包括计算机程序指令,在计算机上执行所述计算机程序指令时,该所述计算机程序指令使所述计算机执行如权利要求1~11中任一项所述的显示方法。
  14. 一种显示装置,包括:
    显示屏,包括多个子显示区域;
    数据获取装置,被配置为获取待显示图像数据,所述待显示图像数据包括OSD图像数据,所述OSD图像数据包括来自n个视点的OSD子图像数据,所述n为大于或等于2的整数;
    数据处理装置,被配置为解析所述OSD图像数据,并对解析后的所述OSD图像数据进行数据重排;
    根据所述OSD图像数据的位置信息,确定所述OSD图像数据在所述显示屏上对应的待显示区域,并控制所述显示屏上的所述待显示区域显示重排后的OSD图像数据;所述待显示区域包括至少一个所述子显示区域,所述重排后的OSD图像数据包括重排后的部分视点或全部视点的OSD子图像数据。
  15. 根据权利要求14所述的显示装置,其中,所述待显示图像数据还包括源图像数据,所述数据处理装置还被配置为:解析所述源图像数据,并对解析后的所述源图像数据进行数据重排。
  16. 根据权利要求14或15所述的显示装置,其中,所述数据处理装置被配置为根据所述OSD图像数据对应的待显示区域,采用所述重排后的OSD图像数据替换所述待显示区域对应的重排后的所述源图像数据,或者,在所述待显示区域对应的重排后的所述源图像数据上叠加所述重排后的OSD图像数据。
  17. 根据权利要求14~16中任一项所述的显示装置,其中,所述数据处理装置被配置为对所述重排后的OSD图像数据设置透明度,在所述待显示区域对应的重排后的所述源图像数据上叠加设置透明度后的所述重排后的OSD图像数据。
  18. 根据权利要求14~17中任一项所述的显示装置,其中,每个视点的OSD子图像数据包括A个亚像素,所述A为大于等于2的整数;所述数据处理装置被配置为:
    对所述n个视点的OSD子图像数据中的所有亚像素进行重排;所述n个视点对应的重排后的OSD子图像数据中不同的亚像素位于不同的行。
  19. 根据权利要求14~18中任一项所述的显示装置,其中,A个所述亚像素包括第一亚像素,所述第一亚像素为所述A个亚像素中的任意一个亚像素,所述第一亚像素包括t个子像素,所述t为大于或等于2的整数。
  20. 根据权利要求14~18中任一项所述的显示装置,其中,所述数据处理装置被配置为解析所述OSD图像数据,得到所述OSD图像数据的显示参数和所述OSD图像数据的显示数据;所述OSD图像数据的显示参数包括所述OSD图像数据对应的视点数与子显示区域标识,以及所述OSD图像数据的显示数据的容量;
    根据所述OSD图像数据的显示参数,对所述OSD图像数据的显示数据进行数据重排。
  21. 根据权利要求14~20中任一项所述的显示装置,其中,所述数据获取装置被配置为:
    采用n个虚拟相机,从不同视点拍摄第一OSD模型,得到所述n个视点的OSD子图像数据;或者,
    采用三维建模或图像处理器获取第二OSD模型,并根据所述第二OSD模型,得到所述n个视点的OSD子图像数据。
  22. 根据权利要求14~20中任一项所述的显示装置,其中,所述OSD图像数据的位置信息包括所述待显示区域的起点位置,以及所述待显示区域的长度信息及宽度信息。
  23. 根据权利要求14~20中任一项所述的显示装置,所述数据获取装置还被配置为:获取第一命令,所述第一命令用于将OSD图像的显示方式由3D显示切换至2D显示,或者,将OSD图像的显示方式由2D显示切换至3D显示;响应于所述第一命令,获取二维的OSD图像数据或三维OSD图像数据;
    所述数据处理装置被配置为根据所述二维的OSD图像数据对应的待显示区域,控制所述显示屏显示二维的OSD图像;或,根据所述三维的OSD图像数据对应的待显示区域,控制所述显示屏显示三维的OSD图像。
  24. 根据权利要求14~23中任一项所述的显示装置,还包括采集装置,被配置为获取用户的人脸图像或瞳孔图像,确定用户的主观测区;和/或获取用户的手势图像,确定用户的主视点区;
    所述数据处理装置还被配置为对所述主观测区和/或所述主视点区的OSD图像数据和/或源图像数据进行高清显示;对所述主观测区和/或所述主视点区以外的OSD图像数据和/或源图像数据进行低清显示。
PCT/CN2022/096698 2022-06-01 2022-06-01 显示方法及装置 WO2023230965A1 (zh)

Priority Applications (2)

Application Number Priority Date Filing Date Title
PCT/CN2022/096698 WO2023230965A1 (zh) 2022-06-01 2022-06-01 显示方法及装置
CN202280001600.2A CN117529911A (zh) 2022-06-01 2022-06-01 显示方法及装置

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/CN2022/096698 WO2023230965A1 (zh) 2022-06-01 2022-06-01 显示方法及装置

Publications (1)

Publication Number Publication Date
WO2023230965A1 true WO2023230965A1 (zh) 2023-12-07

Family

ID=89026806

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2022/096698 WO2023230965A1 (zh) 2022-06-01 2022-06-01 显示方法及装置

Country Status (2)

Country Link
CN (1) CN117529911A (zh)
WO (1) WO2023230965A1 (zh)

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101656835A (zh) * 2008-08-21 2010-02-24 索尼株式会社 图像拾取设备、显示器和图像处理设备
CN103369335A (zh) * 2012-04-03 2013-10-23 Lg电子株式会社 图像显示设备及其操作方法
CN103826114A (zh) * 2013-11-15 2014-05-28 青岛海信电器股份有限公司 一种立体显示方法及自由立体显示装置

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101656835A (zh) * 2008-08-21 2010-02-24 索尼株式会社 图像拾取设备、显示器和图像处理设备
CN103369335A (zh) * 2012-04-03 2013-10-23 Lg电子株式会社 图像显示设备及其操作方法
CN103826114A (zh) * 2013-11-15 2014-05-28 青岛海信电器股份有限公司 一种立体显示方法及自由立体显示装置

Also Published As

Publication number Publication date
CN117529911A (zh) 2024-02-06

Similar Documents

Publication Publication Date Title
US11436787B2 (en) Rendering method, computer product and display apparatus
US8564647B2 (en) Color management of autostereoscopic 3D displays
US20130127989A1 (en) Conversion of 2-Dimensional Image Data into 3-Dimensional Image Data
CN103003870A (zh) 图像显示系统、图像处理设备及其控制方法
US20200029057A1 (en) Systems and methods for correcting color separation in field-sequential displays
WO2023202216A1 (zh) 图像处理方法、装置及存储介质
US11783445B2 (en) Image processing method, device and apparatus, image fitting method and device, display method and apparatus, and computer readable medium
WO2016033925A1 (zh) 显示方法及显示装置
WO2023230965A1 (zh) 显示方法及装置
US20130050183A1 (en) System and Method of Rendering Stereoscopic Images
KR20140004393A (ko) 디스플레이 장치 및 그 제어 방법
CN109194944B (zh) 图像处理方法、装置及系统、显示装置
CN103297791A (zh) 图像显示装置及图像显示方法
CN106454554A (zh) 一种氛围光电视时间同步方法和氛围光电视
US11917167B2 (en) Image compression method and apparatus, image display method and apparatus, and medium
US8289373B2 (en) Image processing method for multi-depth-of-field 3D-display
CN113949862B (zh) 3d画面显示测试方法和装置以及显示控制设备和系统
CN113132556B (zh) 视频处理方法、装置及系统和视频处理设备
US20130021324A1 (en) Method for improving three-dimensional display quality
KR102141556B1 (ko) 입체 디스플레이 장치의 구동 방법
US11800079B2 (en) Image processing method, image display method and image display system
CN113301411B (zh) 视频处理方法、装置及系统和视频处理设备
WO2024093893A1 (zh) 空间现实显示方法、空间现实显示系统以及非易失性计算机可读存储介质
CN106998461A (zh) 一种头戴显示系统
JP2014041455A (ja) 画像処理装置、画像処理方法、及びプログラム

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 22944296

Country of ref document: EP

Kind code of ref document: A1