CN106851249A - Image processing method and display device - Google Patents

Image processing method and display device Download PDF

Info

Publication number
CN106851249A
CN106851249A CN201710202490.7A CN201710202490A CN106851249A CN 106851249 A CN106851249 A CN 106851249A CN 201710202490 A CN201710202490 A CN 201710202490A CN 106851249 A CN106851249 A CN 106851249A
Authority
CN
China
Prior art keywords
viewing
image data
location parameter
beholder
location
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201710202490.7A
Other languages
Chinese (zh)
Inventor
雷小刚
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Lenovo Beijing Ltd
Original Assignee
Lenovo Beijing Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Lenovo Beijing Ltd filed Critical Lenovo Beijing Ltd
Priority to CN201710202490.7A priority Critical patent/CN106851249A/en
Publication of CN106851249A publication Critical patent/CN106851249A/en
Pending legal-status Critical Current

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/30Image reproducers
    • H04N13/332Displays for viewing with the aid of special glasses or head-mounted displays [HMD]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/10Processing, recording or transmission of stereoscopic or multi-view image signals
    • H04N13/106Processing image signals
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/30Image reproducers
    • H04N13/366Image reproducers using viewer tracking
    • H04N13/383Image reproducers using viewer tracking for tracking with gaze detection, i.e. detecting the lines of sight of the viewer's eyes

Abstract

The invention discloses a kind of image processing method and display device, methods described includes:Obtain source image data;Location parameter is obtained, the location parameter shows viewing location of the beholder in viewing area;The source image data is processed according to the location parameter, is obtained and display target image;Wherein, the target image includes viewing areas corresponding with the location parameter and the sensing region different from the viewing areas;Wherein, the source image data includes Part I corresponding with the viewing areas and Part II corresponding with the sensing region, the priority of the priority higher than the Part II of the Part I.The embodiment of the present invention can realize the Part II of the Part I according to the corresponding source image data of the corresponding viewing areas of viewing location and the source image data corresponding to the sensing region different from viewing areas, make the priority of Part I higher than the priority of Part II, so as to improve image processing efficiency.

Description

Image processing method and display device
Technical field
The present invention relates to technical field of image processing, more particularly to a kind of image processing method and display device.
Background technology
During image procossing, due to for precision, decrement, the aspect such as making up and have requirement higher, it is therefore desirable to by again The many and operand of complexity.
And, in scan picture, not only the quantity and complexity to computing are required, and for arithmetic speed There is requirement higher.For example, will be good by image procossing in real time while being imaged using mobile terminal, and show.Therefore, it is real When image processing techniques need stronger processor performance.
On the other hand, the essence of image is matrix, then picture size is bigger, and the operand of matrix is then bigger, at this moment Need stronger processor performance.
But, the need for the performance of current processor cannot meet raising image processing efficiency.
The content of the invention
In view of this, set it is an object of the invention to provide a kind of image processing method for improving image processing efficiency and display It is standby.
To achieve these goals, the invention provides a kind of image processing method, including:
Obtain source image data;
Location parameter is obtained, the location parameter shows viewing location of the beholder in viewing area;
The source image data is processed according to the location parameter, is obtained and display target image;Wherein, the target figure As including viewing areas corresponding with the location parameter and the sensing region different from the viewing areas;
Wherein, the source image data include Part I corresponding with the viewing areas and with the sensing region pair The Part II answered, the priority of the priority higher than the Part II of the Part I.
The embodiment of the present invention also provides a kind of display device, including:
Processing unit, is configured to obtain source image data;
Display device, is configured to constitute the viewing area of the display source image data;
Harvester, is configured to obtain location parameter, and the location parameter shows viewing of the beholder in viewing area Position;
The processing unit, is additionally configured to process the source image data according to the location parameter, obtains and show mesh Logo image;Wherein, the target image includes viewing areas corresponding with the location parameter and different from the viewing areas Sensing region;
Wherein, the source image data include Part I corresponding with the viewing areas and with the sensing region pair The Part II answered, the priority of the priority higher than the Part II of the Part I.
Can be seen that the embodiment of the present invention by the above embodiment of the present invention can realize according to beholder in viewing area Viewing location, Part I according to the corresponding source image data of the corresponding viewing areas of viewing location and different from viewing area The Part II of the source image data corresponding to the sensing region in domain, make during image procossing the priority of Part I to be higher than The priority of Part II, so as to improve image processing efficiency.
Brief description of the drawings
Fig. 1 is a kind of flow chart of embodiment of image processing method of the invention;
Fig. 2 is a schematic diagram of a scenario of image processing method of the invention;
Fig. 3 is another schematic diagram of a scenario of image processing method of the invention;
Fig. 4 is a kind of schematic diagram of embodiment of display device of the invention.
Specific embodiment
With reference to the accompanying drawings and examples, specific embodiment of the invention is described in further detail.Hereinafter implement Example is not limited to the scope of the present invention for illustrating the present invention.
Embodiment one
Viewing location of the present embodiment according to beholder in viewing area processes source image data, generates target figure Picture, wherein, target image includes viewing areas corresponding with viewing location and the sensing region different from the viewing areas, source View data includes Part I corresponding with viewing areas and Part II corresponding with the sensing region, and described the Priority of the priority of a part higher than the Part II.Specifically:
Fig. 1 is a kind of flow chart of embodiment of image processing method of the invention, as shown in figure 1, the figure of the present embodiment As processing method, specifically may include steps of:
Obtain source image data;
Location parameter is obtained, the location parameter shows viewing location of the beholder in viewing area;
The source image data is processed according to the location parameter, is obtained and display target image;Wherein, the target figure As including viewing areas corresponding with the location parameter and the sensing region different from the viewing areas;
Wherein, the source image data include Part I corresponding with the viewing areas and with the sensing region pair The Part II answered, the priority of the priority higher than the Part II of the Part I.
Specifically, the present embodiment first obtains source image data, and shows on the display apparatus after image is rendered.By In beholder when screen is watched, it can be seen that region within the vision, wherein, the region that can be paid close attention to is viewing area Domain, the region without emphasis relation is sensing region.Therefore, as shown in Fig. 2 the present embodiment in the specific implementation, is shown The location parameter of viewing location of the beholder in viewing area, and the eye focus for obtaining source images are calculated according to location parameter Part I, then Part I is processed source image data in the periphery of Part I, obtains target image, its In, the viewing areas of Part I correspondence target image, the sensing region of Part II correspondence target image.
Can be seen that the present invention can realize the viewing according to beholder in viewing area by the above embodiment of the present invention Position, the Part I according to the corresponding source image data of the corresponding viewing areas of viewing location and the sense different from viewing areas Know the Part II of the source image data corresponding to region, the priority for make during image procossing Part I is higher than second The priority divided, on the basis of the experience that ensure that beholder, improves image processing efficiency.
The present embodiment in the specific implementation, can apply to display device.For needing on large-sized display device When the image of display carries out real-time processing, number of calculations very big, complexity is very high so that image processing efficiency is relatively low.
Large scale is relative concept, for display device, for the viewing distance of respective usage scenario (such as project/ , more than 1 meter, within 1 meter, head-mounted display apparatus are within 10 centimetres for display for TV), viewing area is more than beholder's Area of visual field, can be referred to as large scale.Such as large scale projection, large size TV refer to more than 50 cun it is viewing area, Large scale computer display refers to more than 27 cun of viewing area, large scale head-mounted display apparatus and refers to more than 5 cun and shows Show region.
But beholder is actual watch image when, due to the visual characteristic of beholder, the fixation range reality of beholder The upper field range less than beholder, that is to say, that the viewing areas and viewing for focusing on and watching attentively are there is also in the visual field of beholder Sensing region beyond region.Therefore, if according to this visual characteristic of beholder, to the viewing areas in its visual field and sense Know processing mode of the area using differentiation, it is inevitable compared to relatively all being processed using identical processing mode whole viewing area The performance requirement to processor can be reduced, so as to improve image processing efficiency.
Embodiment two
Based on the method described in embodiment one, the source images number is processed this gives according to the location parameter According to several concrete modes, it is specific as follows:
Mode one:After the Part I and Part II for obtaining the location parameter of source images and determining source images, to source The Part I of view data is processed by the first processing mode, obtains the image of the viewing areas;Then to described The Part II of source image data is processed by second processing mode, obtains the image of the sensing region;It is wherein described Complexity of the complexity of the first processing mode higher than the second processing mode.
Further, can be processed using the first more complicated processing mode for Part I, using comparing Complicated algorithm is processed.
In an application scenarios, can be to the Part I of source image data using at more complicated rendering intent Reason so that the image display effect of the viewing areas obtained after Part I treatment is truer.And then adopted for Part II Processed with better simply second processing mode, if do not rendered, but directly invoked the Perception Area of previous frame image Area image shows as the sensing region of current frame image.Again for example, higher to the image procossing precision of Part I, and it is right The image procossing precision of Part II is relatively low, because the operand of the low needs of processing accuracy is little, therefore to the requirement of processor It is not high.
In another application scenarios, in order to embody spatial impression, render process will calculate each light source pair in scene The influence of object;Subsequently, in addition it is also necessary to calculate light source projects out according to whether the object of transparent material has been used in scene Shade come determine use depth pinup picture shade or use ray tracing shade.Subsequently, will also according to the material of object come Calculate color, the texture different equivalent of material of body surface really.This example is in the specific implementation, it is necessary to Part I and Two parts use different processing modes so that the display effect in viewing areas in target image is truer.For example, right Rendered by complex method in the Part I of image, such as strengthened Processing Algorithm and feature recognition algorithms, to the Two parts are only rendered by simple rendering intent, and Part II is processed only with the method for adjustment light such as Deng;Again for example, being processed using identical rendering intent in Part I and Part II, but to the image of Part I Processing accuracy is higher, can be reached when such as to image floating-point operation 7 after decimal point, and to the image procossing precision of Part II It is relatively low, only reached when such as to image floating-point operation 2 after decimal point.Due to processing accuracy it is relatively low when, required operand is not Greatly, therefore to the less demanding of processor.Again for example, image refreshing faster in the case of, rendered for viewing areas, Sensing region can not be rendered.
When image procossing is carried out, image processing program can by the screen that is rendered into of view projection images, in order to improve The effect for rendering, rendering frequency, to need to reach 90 frames per second or higher, so has higher to processor when to image procossing Requirement.And be only a part for whole viewing area due to Part I in the present embodiment, only to Part I at Reason, will greatly shorten the time of disposed of in its entirety image, so as to improve the image processing efficiency of image.
Mode two:After the Part I and Part II for obtaining the location parameter of source images and determining source images, wherein, The resolution ratio of the Part I of the source image data is identical with the resolution ratio of the Part II, but to Part I and Two parts carry out different treatment.For example, processing source image data, image is pressed in final processing stage then During contracting treatment, the high resolution of the viewing areas of the target image can be made in the resolution ratio of the sensing region.Such mesh The viewing areas of logo image seem more fine and smooth, and sensing region is more coarse.
Mode three:After the Part I and Part II for obtaining the location parameter of source images and determining source images, the One time started the Part I treatment to the source image data, obtained the image of the viewing areas;Opened in the second time The Part II treatment begun to the source image data, obtains the image of the sensing region;The very first time is earlier than described Second time.Specifically, the viewing areas of viewing area are focused on due to eye, it is inevitable in viewing display area first to see sight See region, therefore can first render the image of Part I, while Part I is rendered, Part II is still kept in display One two field picture part corresponding with Part II, the viewing effect of beholder is not interfered with so yet.Can be with for Part II Render after a while, if the image watched has been refreshed, the render process to Part II can also be omitted.Specifically render Method can be in the method cited by employing mode one.This method does not influence the viewing effect of beholder, also improves image Treatment effeciency.
The three of the above mode of the present embodiment both can individually be implemented, it is also possible to which combination is implemented.
It is worth noting that, in 3D rendering processing method involved in the application scenarios of aforesaid way one.Due to 3D rendering Display Technique range of application is wider, corresponding treatment can be done according to different application scenarios, for example, by wearing During formula display device viewing 3D rendering, because head-mounted display apparatus have two display screens corresponding with right and left eyes difference, two Individual display screen shows two images of correspondence right and left eyes, because the parallax that two images are present makes to show three-dimensional showing in human brain Show effect.Therefore, when processing this kind of 3D rendering, two images can also respectively be processed.That is, to first The processing procedure of individual image is that location parameter can be obtained by camera, and according to location parameter determine Part I and Part II, then to Part I and Part II, using three of the above mode, any one is processed so that target figure Then the priority of the viewing areas as in be presented higher than the priority of sensing region according to viewing areas in first image Picture material determine second viewing location of image, and second Part I of image is determined according to viewing location And Part II, and processed with first image identical mode, to cause that the priority of viewing areas is higher than Perception Area Domain.
Be can be seen that by the above embodiment of the present invention is carried out not by the Part I and Part II to source image data Same treatment, can improve image processing efficiency in the case of the viewing effect for not influenceing beholder.
Embodiment three
Based on the method described in embodiment one, this gives several concrete modes for obtaining location parameter, specifically It is as follows:
Acquisition includes the image information of the eyes of beholder;
Viewing location of the beholder in the viewing area is determined according to described image information;
Wherein, the viewing location is located in the viewing areas of the target image;
Wherein, obtaining includes the image information of the eyes of beholder, including:The sight is obtained in being spaced in Preset Time The eyes image of the person of seeing.
Specifically, the image information of the eyes of beholder can be obtained, according to the analysis to image information forever without pause of breathing, can To determine viewing location of the beholder in viewing area.Although image refreshing frequency in display is higher here, eyes Change the time interval of the time interval much larger than image update of viewing location, that is to say, that eyes change position comparatively fast, therefore Without consideration treatment, in the problem changed to Part I and Part II treatment retracement line.
In the specific implementation, because the eyes of beholder may at any time change viewing location, therefore, to need very short Time obtains an eyes image information for the eyes of beholder, for example, to make beholder to perceive viewing areas and sense Know the difference in region, the eyes image of beholder can be just obtained in each refreshed image.If it find that beholder Viewing location is changed, it is necessary to which the Part I to source image data is processed again, to the Part II of source images It is also carried out processing again.
In an application scenarios, obtaining location parameter can use eyeball tracking technology of the prior art.When people's Eyes see that during to different directions, eye can occur trickle change, and these changes can produce the feature that can be extracted, harvester These features can be extracted by picture catching or scanning, so that the change of real-time tracing eyes.
Further, in the present embodiment, the source image data includes:First source images subdata and the second source images Subdata, the first source images subdata and the second source images subdata can allow the beholder to see three-dimensional effect diagram Picture
Specifically, to make beholder obtain more preferable sensory experience, source image data can be 3D rendering data.Specific source View data includes the first source images subdata and the second source images subdata, the first source images subdata and the second source images There is certain difference in the corresponding content of data display, corresponding content coordinate in the picture, such beholder can see To the image that there is parallax, three-dimensional image effect is produced in the brain with beholder.
Further, location parameter includes:First sub- location parameter and the second sub- location parameter;Viewing area includes:The One sub-viewing areas and the second sub-viewing areas;Wherein, the described first sub- location parameter is that the first eye of the beholder exists The first viewing location in first sub-viewing areas, the second sub- location parameter be second of the beholder it is fine The second viewing location in second sub-viewing areas;
Processing the source image data acquisition target image includes:According to the first sub- location parameter treatment first source figure As subdata obtains first object image;The second source images subdata is processed according to the second sub- location parameter obtain the second mesh Logo image;
Display target image includes:The first object image is shown in first sub-viewing areas, described second Sub-viewing areas show second target image.
Specifically, the first sub- location parameter characterizes viewing location of the first eye eye in the first sub-viewing areas, second Sub- location parameter characterizes second fine viewing location in the second sub-viewing areas, and viewing areas are determined according to viewing location, And then source images are carried out with treatment formation first object image and the second target image.
Further, the described first sub- location parameter determines the first focal position in the first source images subdata, by institute The Part I that the first focal position determines the first source images subdata is stated,
The second sub- location parameter determines the second focal position in the second source images subdata, by second focus Position determines the Part I of the second source images subdata.
According to first focal position and preset range parameter, the Part I is determined;
According to second focal position and preset range parameter, the Part II is determined.
Can determine focal position in a different manner when it is determined that watching focal position.For example, in an applied field Jing Zhong, as shown in figure 3, the content that can be shown according to viewing areas determines the first focal position of the first source images subdata With the second focal position of the second source image data.It is identical with the method that eye image determines for left-eye image, do not do herein Distinguish.Display content of the beholder in viewing areas, such as lightning are determined first, and the first source images is determined further according to display content First position parameter in data, according to first position parameter, determines the display content pair in the first source images subdata The coordinate answered, then determines position coordinates of the display content in the second source images subdata, further according to default model again Enclose parameter and determine the corresponding Part I of the first source images subdata and Part II respectively, further according to the side that embodiment two is introduced Method is respectively processed to Part I and Part II.
In another application scenarios, can come true according to first eye to the viewing location of the first source images subdata Viewing location of fixed second eyes to the second source images subdata.For example, it is first determined beholder is in the first source images subdata Viewing location, further according to preset range parameter, determine the Part I and Part II of the first source images subdata, Ran Houzai According to the first source image data and the coordinate difference of the second source image data, it is determined that giving the Part I of the first source images subdata The Part I and Part II of the second source images subdata corresponding with Part II, then again to Part I and second The method introduced according to embodiment two is divided to be respectively processed.
Can be seen that the present invention can realize the viewing according to beholder in viewing area by the above embodiment of the present invention Position, the Part I according to the corresponding source image data of the corresponding viewing areas of viewing location and the sense different from viewing areas Know the Part II of the source image data corresponding to region, the priority for make during image procossing Part I is higher than second The priority divided, so as to improve rendering efficiency.
Example IV
A kind of display device is present embodiments provided, the display device is provided with display device, harvester and treatment Device.Display device configurations are acquisition source image data;Display device, is configured to constitute the display of the display source image data Region;Harvester, is configured to obtain location parameter, and the location parameter shows viewing position of the beholder in viewing area Put;The processing unit, is additionally configured to process the source image data according to the location parameter, obtains and display target figure Picture;Wherein, the target image includes viewing areas corresponding with the location parameter and the sense different from the viewing areas Know region;Wherein, the source image data include Part I corresponding with the viewing areas and with the sensing region pair The Part II answered, the priority of the priority higher than the Part II of the Part I.
Wherein in one embodiment, display device can be wearable device, and such as wear-type virtual reality (VR) shows Show equipment or wear-type augmented reality (AR) display device.Display device is arranged at the visual line of sight of beholder's eyes after wearing Interior, if wearable device is wear-type VR display devices, the display device is preferably lighttight device, and wear-type VR shows The camera of equipment is set towards eyes, in order to gather viewing location of the beholder in viewing area;If wearable Equipment is wear-type AR display devices, then display device can be the eyeglass module of printing opacity, the environment of wear-type AR display devices Camera is consistent with visual line of sight, in order to gather viewing location of the beholder described in image in viewing area.Wherein, source figure As data can be obtained by main frame with download or online mode, or, for that can be come using wear-type AR display devices From the image of environment camera collection.
Viewing area can be divided into sight by the display device of the present embodiment by the viewing location of the eyes to beholder Viewing areas and the non-sensing region paid close attention to that the person's of seeing eyes are paid close attention to, are processed by source images so that seen See that the priority in region is preferential higher than sensing region.Further, as shown in figure 4, the display device include processing unit, Display device and harvester;
Processing unit, is configured to obtain source image data;
Display device, is configured to constitute the viewing area of the display source image data;
Harvester, is configured to obtain location parameter, and the location parameter shows viewing of the beholder in viewing area Position;
The processing unit, is additionally configured to process the source image data according to the location parameter, obtains and show mesh Logo image;Wherein, the target image includes viewing areas corresponding with the location parameter and different from the viewing areas Sensing region;
Wherein, the source image data include Part I corresponding with the viewing areas and with the sensing region pair The Part II answered, the priority of the priority higher than the Part II of the Part I.
In another specific embodiment, the harvester includes the first harvester and the second harvester.Described One harvester is arranged at the marginal position of first display device, and second harvester is similarly provided at described second The edge of display device, and the direction that two harvesters show both facing to image.
Wherein, first harvester is configured to the image information of the left eye for gathering beholder, to determine the viewing The viewing location of the left eye of person;
Second harvester is configured to gather the image information of beholder's right eye, to determine beholder's right eye Viewing location;
Wherein, the display device includes the first display device and the second display device,
First display device configurations are to constitute the viewing area of beholder's left eye viewing;
First display device configurations are to constitute the viewing area of beholder's right eye viewing.
In another specific embodiment, wherein, the resolution ratio of the Part I of the source image data and described second Partial resolution ratio is identical, and the viewing areas of the target image high resolution in the resolution ratio of the sensing region.
In another specific embodiment, wherein,
Start the Part I treatment to the source image data in the very first time, obtain the image of the viewing areas;
Start the Part II treatment to the source image data in the second time, obtain the image of the sensing region;
The very first time is earlier than second time.
In another specific embodiment, the harvester concrete configuration is:
Acquisition includes the image information of the eyes of beholder;
Viewing location of the beholder in the viewing area is determined according to described image information;
Wherein, the viewing location is located in the viewing areas of the target image.
In another specific embodiment, wherein,
The source image data includes:First source images subdata and the second source images subdata, first source images Subdata and the second source images subdata are the images that the beholder can be allowed to see stereoeffect;
Location parameter includes:First sub- location parameter and the second sub- location parameter;Viewing area includes:First sub- viewing area Domain and the second sub-viewing areas;Wherein, the described first sub- location parameter is the first eye of the beholder in the described first son The first viewing location in viewing area, the second sub- location parameter is second eyes of the beholder in the described second son The second viewing location in viewing area;
Processing the source image data acquisition target image includes:According to the first sub- location parameter treatment first source figure As subdata obtains first object image;The second source images subdata is processed according to the second sub- location parameter obtain the second mesh Logo image;
Display target image includes:The first object image is shown in first sub-viewing areas, described second Sub-viewing areas show second target image.
In another specific embodiment, wherein,
First focal position is determined in the first source images subdata according to the described first sub- location parameter;
Second focal position is determined in the second source images subdata according to the described second sub- location parameter.
In another specific embodiment, wherein,
According to first focal position, first content is determined, by the first content and preset range parameter determination institute State Part I;
According to second focal position, the second content is determined, by second content and preset range parameter determination institute State Part II.
It need to be noted that be:Above display device implements the description of item, is similar, tool with above method description There is same embodiment of the method identical beneficial effect, therefore do not repeat.For what is do not disclosed in display device embodiment of the present invention Ins and outs, those skilled in the art refer to the description of the inventive method embodiment and understand, to save length, here not Repeat again.
In several embodiments provided herein, it should be understood that disclosed apparatus and method, can be by it Its mode is realized.Apparatus embodiments described above are only schematical, for example, the division of the unit, is only A kind of division of logic function, can have other dividing mode, such as when actually realizing:Multiple units or component can be combined, or Another system is desirably integrated into, or some features can be ignored, or do not perform.In addition, shown or discussed each composition portion Coupling point each other or direct-coupling or communication connection can be the INDIRECT COUPLINGs of equipment or unit by some interfaces Or communication connection, can be electrical, machinery or other forms.
The above-mentioned unit that is illustrated as separating component can be or may not be it is physically separate, it is aobvious as unit The part for showing can be or may not be physical location, you can with positioned at a place, it is also possible to be distributed to multiple network lists In unit;Part or all of unit therein can be according to the actual needs selected to realize the purpose of this embodiment scheme.
In addition, each functional unit in various embodiments of the present invention can be fully integrated into a processing unit, also may be used Being each unit individually as a unit, it is also possible to which two or more units are integrated in a unit;It is above-mentioned Integrated unit can both be realized in the form of hardware, it would however also be possible to employ hardware adds the form of SFU software functional unit to realize.
One of ordinary skill in the art will appreciate that:Realizing all or part of step of above method embodiment can pass through Programmed instruction related hardware is completed, and foregoing program can be stored in a computer read/write memory medium, the program Upon execution, the step of including above method embodiment is performed;And foregoing storage medium includes:It is movable storage device, read-only Memory (ROM, Read Only Memory), random access memory (RAM, Random Access Memory), magnetic disc or Person's CD etc. is various can be with the medium of store program codes.
Or, if the above-mentioned integrated unit of the present invention is to realize in the form of software function module and as independent product When selling or using, it is also possible to which storage is in a computer read/write memory medium.Based on such understanding, the present invention is implemented The part that the technical scheme of example substantially contributes to prior art in other words can be embodied in the form of software product, The computer software product is stored in a storage medium, including some instructions are used to so that computer equipment (can be with It is personal computer, server or network equipment etc.) perform all or part of each embodiment methods described of the invention. And foregoing storage medium includes:Movable storage device, read-only storage (ROM, Read Only Memory), arbitrary access are deposited Reservoir (RAM, Random Access Memory), magnetic disc or CD etc. are various can be with the medium of store program codes.
The above, specific embodiment only of the invention, but protection scope of the present invention is not limited thereto, and it is any Those familiar with the art the invention discloses technical scope in, change or replacement can be readily occurred in, should all contain Cover within protection scope of the present invention.Therefore, protection scope of the present invention should be based on the protection scope of the described claims.

Claims (10)

1. a kind of image processing method, including:
Obtain source image data;
Location parameter is obtained, the location parameter shows viewing location of the beholder in viewing area;
The source image data is processed according to the location parameter, is obtained and display target image;Wherein, the target image bag Include viewing areas corresponding with the location parameter and the sensing region different from the viewing areas;
Wherein, the source image data includes Part I corresponding with the viewing areas and corresponding with the sensing region Part II, the priority of the priority higher than the Part II of the Part I.
2. method according to claim 1, wherein, the source image data is processed according to the location parameter, including:
Part I to the source image data is processed by the first processing mode, obtains the figure of the viewing areas Picture;
Part II to the source image data is processed by second processing mode, obtains the figure of the sensing region Picture;
Complexity of the complexity of first processing mode higher than the second processing mode.
3. method according to claim 1, wherein, the resolution ratio of the Part I of the source image data and described second Partial resolution ratio is identical, and the viewing areas of the target image high resolution in the resolution ratio of the sensing region.
4. method according to claim 1, wherein,
Start the Part I treatment to the source image data in the very first time, obtain the image of the viewing areas;
Start the Part II treatment to the source image data in the second time, obtain the image of the sensing region;
The very first time is earlier than second time.
5. method according to claim 1, obtains location parameter, including:
Acquisition includes the image information of the eyes of beholder;
Viewing location of the beholder in the viewing area is determined according to described image information;
Wherein, the viewing location is located in the viewing areas of the target image.
6. method according to claim 5, wherein,
The source image data includes:First source images subdata and the second source images subdata, the first source images subnumber It is image that the beholder can be allowed to see stereoeffect according to the second source images subdata;
Location parameter includes:First sub- location parameter and the second sub- location parameter;Viewing area includes:First sub-viewing areas and Second sub-viewing areas;Wherein, the described first sub- location parameter is the first eye of the beholder in the described first son display The first viewing location in region, the second sub- location parameter is second eyes of the beholder in the described second son display The second viewing location in region;
Processing the source image data acquisition target image includes:The first source images is processed according to the first sub- location parameter Data obtain first object image;The second source images subdata is processed according to the second sub- location parameter obtain the second target figure Picture;
Display target image includes:The first object image is shown in first sub-viewing areas, it is aobvious in the described second son Show that region shows second target image.
7. method according to claim 6, wherein,
First focal position is determined in the first source images subdata according to the described first sub- location parameter;
Second focal position is determined in the second source images subdata according to the described second sub- location parameter.
8. method according to claim 7, wherein,
According to first focal position, first content is determined, as described in the first content and preset range parameter determination A part;
According to second focal position, the second content is determined, as described in second content and preset range parameter determination Two parts.
9. a kind of display device, including:
Processing unit, is configured to obtain source image data;
Display device, is configured to constitute the viewing area of the display source image data;
Harvester, is configured to obtain location parameter, and the location parameter shows viewing location of the beholder in viewing area;
The processing unit, is additionally configured to process the source image data according to the location parameter, obtains and display target figure Picture;Wherein, the target image includes viewing areas corresponding with the location parameter and the sense different from the viewing areas Know region;
Wherein, the source image data includes Part I corresponding with the viewing areas and corresponding with the sensing region Part II, the priority of the priority higher than the Part II of the Part I.
10. display device according to claim 9, wherein, the harvester is adopted including the first harvester and second Acquisition means,
First harvester is configured to the image information of the left eye for gathering beholder, to determine the left eye of the beholder Viewing location;
Second harvester is configured to gather the image information of beholder's right eye, to determine the viewing of beholder's right eye Position;
Wherein, the display device includes the first display device and the second display device,
First display device configurations are to constitute the viewing area of beholder's left eye viewing;
First display device configurations are to constitute the viewing area of beholder's right eye viewing.
CN201710202490.7A 2017-03-30 2017-03-30 Image processing method and display device Pending CN106851249A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201710202490.7A CN106851249A (en) 2017-03-30 2017-03-30 Image processing method and display device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201710202490.7A CN106851249A (en) 2017-03-30 2017-03-30 Image processing method and display device

Publications (1)

Publication Number Publication Date
CN106851249A true CN106851249A (en) 2017-06-13

Family

ID=59141581

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201710202490.7A Pending CN106851249A (en) 2017-03-30 2017-03-30 Image processing method and display device

Country Status (1)

Country Link
CN (1) CN106851249A (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109522866A (en) * 2018-11-29 2019-03-26 宁波视睿迪光电有限公司 Naked eye 3D rendering processing method, device and equipment
WO2020034669A1 (en) * 2018-08-16 2020-02-20 北京七鑫易维信息技术有限公司 Image obtaining device control method and apparatus
CN110856035A (en) * 2018-07-24 2020-02-28 顶级公司 Processing image data to perform object detection
CN111988525A (en) * 2020-08-25 2020-11-24 Oppo广东移动通信有限公司 Image processing method and related device

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101655984A (en) * 2008-05-12 2010-02-24 美国西门子医疗解决公司 System for adaptively processing medical image data
CN102347016A (en) * 2010-07-28 2012-02-08 佳能株式会社 Display control apparatus for displaying image, display control method
CN103930817A (en) * 2011-06-20 2014-07-16 谷歌公司 Systems and methods for adaptive transmission of data
CN104767992A (en) * 2015-04-13 2015-07-08 北京集创北方科技有限公司 Head-wearing type display system and image low-bandwidth transmission method

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101655984A (en) * 2008-05-12 2010-02-24 美国西门子医疗解决公司 System for adaptively processing medical image data
CN102347016A (en) * 2010-07-28 2012-02-08 佳能株式会社 Display control apparatus for displaying image, display control method
CN103930817A (en) * 2011-06-20 2014-07-16 谷歌公司 Systems and methods for adaptive transmission of data
CN104767992A (en) * 2015-04-13 2015-07-08 北京集创北方科技有限公司 Head-wearing type display system and image low-bandwidth transmission method

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110856035A (en) * 2018-07-24 2020-02-28 顶级公司 Processing image data to perform object detection
CN110856035B (en) * 2018-07-24 2024-01-26 Arm有限公司 Processing image data to perform object detection
WO2020034669A1 (en) * 2018-08-16 2020-02-20 北京七鑫易维信息技术有限公司 Image obtaining device control method and apparatus
US11853471B2 (en) 2018-08-16 2023-12-26 Beijing 7Invensun Technology Co., Ltd. Method and apparatus for controlling image acquisition device
CN109522866A (en) * 2018-11-29 2019-03-26 宁波视睿迪光电有限公司 Naked eye 3D rendering processing method, device and equipment
CN111988525A (en) * 2020-08-25 2020-11-24 Oppo广东移动通信有限公司 Image processing method and related device

Similar Documents

Publication Publication Date Title
US20210075963A1 (en) Method and apparatus for obtaining binocular panoramic image, and storage medium
CN106484116B (en) The treating method and apparatus of media file
CN110378914A (en) Rendering method and device, system, display equipment based on blinkpunkt information
CN106851249A (en) Image processing method and display device
Blum et al. The effect of out-of-focus blur on visual discomfort when using stereo displays
CN109522866A (en) Naked eye 3D rendering processing method, device and equipment
CA2709417A1 (en) Method of simulating blur in digitally processed images
US10885651B2 (en) Information processing method, wearable electronic device, and processing apparatus and system
CN109901290B (en) Method and device for determining gazing area and wearable device
CN111880654A (en) Image display method and device, wearable device and storage medium
US11956415B2 (en) Head mounted display apparatus
CN111275801A (en) Three-dimensional picture rendering method and device
US20230239457A1 (en) System and method for corrected video-see-through for head mounted displays
US11212502B2 (en) Method of modifying an image on a computational device
JP2023515205A (en) Display method, device, terminal device and computer program
Tong et al. Optical distortions in VR bias the perceived slant of moving surfaces
CN115202475A (en) Display method, display device, electronic equipment and computer-readable storage medium
JP2016532147A (en) How to identify the visual effects of ophthalmic lenses
CN108881892B (en) Anti-dizziness method and system for desktop virtual reality system
US20230244307A1 (en) Visual assistance
CN111587397A (en) Image generation device, spectacle lens selection system, image generation method, and program
CN115327782B (en) Display control method and device, head-mounted display equipment and readable storage medium
EP3961572A1 (en) Image rendering system and method
US11270475B2 (en) Variable rendering system and method
Gurrieri Improvements in the visualization of stereoscopic 3D imagery

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20170613

RJ01 Rejection of invention patent application after publication