CN106851249A - Image processing method and display device - Google Patents
Image processing method and display device Download PDFInfo
- Publication number
- CN106851249A CN106851249A CN201710202490.7A CN201710202490A CN106851249A CN 106851249 A CN106851249 A CN 106851249A CN 201710202490 A CN201710202490 A CN 201710202490A CN 106851249 A CN106851249 A CN 106851249A
- Authority
- CN
- China
- Prior art keywords
- viewing
- image data
- location parameter
- beholder
- location
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/30—Image reproducers
- H04N13/332—Displays for viewing with the aid of special glasses or head-mounted displays [HMD]
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/106—Processing image signals
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/30—Image reproducers
- H04N13/366—Image reproducers using viewer tracking
- H04N13/383—Image reproducers using viewer tracking for tracking with gaze detection, i.e. detecting the lines of sight of the viewer's eyes
Abstract
The invention discloses a kind of image processing method and display device, methods described includes:Obtain source image data;Location parameter is obtained, the location parameter shows viewing location of the beholder in viewing area;The source image data is processed according to the location parameter, is obtained and display target image;Wherein, the target image includes viewing areas corresponding with the location parameter and the sensing region different from the viewing areas;Wherein, the source image data includes Part I corresponding with the viewing areas and Part II corresponding with the sensing region, the priority of the priority higher than the Part II of the Part I.The embodiment of the present invention can realize the Part II of the Part I according to the corresponding source image data of the corresponding viewing areas of viewing location and the source image data corresponding to the sensing region different from viewing areas, make the priority of Part I higher than the priority of Part II, so as to improve image processing efficiency.
Description
Technical field
The present invention relates to technical field of image processing, more particularly to a kind of image processing method and display device.
Background technology
During image procossing, due to for precision, decrement, the aspect such as making up and have requirement higher, it is therefore desirable to by again
The many and operand of complexity.
And, in scan picture, not only the quantity and complexity to computing are required, and for arithmetic speed
There is requirement higher.For example, will be good by image procossing in real time while being imaged using mobile terminal, and show.Therefore, it is real
When image processing techniques need stronger processor performance.
On the other hand, the essence of image is matrix, then picture size is bigger, and the operand of matrix is then bigger, at this moment
Need stronger processor performance.
But, the need for the performance of current processor cannot meet raising image processing efficiency.
The content of the invention
In view of this, set it is an object of the invention to provide a kind of image processing method for improving image processing efficiency and display
It is standby.
To achieve these goals, the invention provides a kind of image processing method, including:
Obtain source image data;
Location parameter is obtained, the location parameter shows viewing location of the beholder in viewing area;
The source image data is processed according to the location parameter, is obtained and display target image;Wherein, the target figure
As including viewing areas corresponding with the location parameter and the sensing region different from the viewing areas;
Wherein, the source image data include Part I corresponding with the viewing areas and with the sensing region pair
The Part II answered, the priority of the priority higher than the Part II of the Part I.
The embodiment of the present invention also provides a kind of display device, including:
Processing unit, is configured to obtain source image data;
Display device, is configured to constitute the viewing area of the display source image data;
Harvester, is configured to obtain location parameter, and the location parameter shows viewing of the beholder in viewing area
Position;
The processing unit, is additionally configured to process the source image data according to the location parameter, obtains and show mesh
Logo image;Wherein, the target image includes viewing areas corresponding with the location parameter and different from the viewing areas
Sensing region;
Wherein, the source image data include Part I corresponding with the viewing areas and with the sensing region pair
The Part II answered, the priority of the priority higher than the Part II of the Part I.
Can be seen that the embodiment of the present invention by the above embodiment of the present invention can realize according to beholder in viewing area
Viewing location, Part I according to the corresponding source image data of the corresponding viewing areas of viewing location and different from viewing area
The Part II of the source image data corresponding to the sensing region in domain, make during image procossing the priority of Part I to be higher than
The priority of Part II, so as to improve image processing efficiency.
Brief description of the drawings
Fig. 1 is a kind of flow chart of embodiment of image processing method of the invention;
Fig. 2 is a schematic diagram of a scenario of image processing method of the invention;
Fig. 3 is another schematic diagram of a scenario of image processing method of the invention;
Fig. 4 is a kind of schematic diagram of embodiment of display device of the invention.
Specific embodiment
With reference to the accompanying drawings and examples, specific embodiment of the invention is described in further detail.Hereinafter implement
Example is not limited to the scope of the present invention for illustrating the present invention.
Embodiment one
Viewing location of the present embodiment according to beholder in viewing area processes source image data, generates target figure
Picture, wherein, target image includes viewing areas corresponding with viewing location and the sensing region different from the viewing areas, source
View data includes Part I corresponding with viewing areas and Part II corresponding with the sensing region, and described the
Priority of the priority of a part higher than the Part II.Specifically:
Fig. 1 is a kind of flow chart of embodiment of image processing method of the invention, as shown in figure 1, the figure of the present embodiment
As processing method, specifically may include steps of:
Obtain source image data;
Location parameter is obtained, the location parameter shows viewing location of the beholder in viewing area;
The source image data is processed according to the location parameter, is obtained and display target image;Wherein, the target figure
As including viewing areas corresponding with the location parameter and the sensing region different from the viewing areas;
Wherein, the source image data include Part I corresponding with the viewing areas and with the sensing region pair
The Part II answered, the priority of the priority higher than the Part II of the Part I.
Specifically, the present embodiment first obtains source image data, and shows on the display apparatus after image is rendered.By
In beholder when screen is watched, it can be seen that region within the vision, wherein, the region that can be paid close attention to is viewing area
Domain, the region without emphasis relation is sensing region.Therefore, as shown in Fig. 2 the present embodiment in the specific implementation, is shown
The location parameter of viewing location of the beholder in viewing area, and the eye focus for obtaining source images are calculated according to location parameter
Part I, then Part I is processed source image data in the periphery of Part I, obtains target image, its
In, the viewing areas of Part I correspondence target image, the sensing region of Part II correspondence target image.
Can be seen that the present invention can realize the viewing according to beholder in viewing area by the above embodiment of the present invention
Position, the Part I according to the corresponding source image data of the corresponding viewing areas of viewing location and the sense different from viewing areas
Know the Part II of the source image data corresponding to region, the priority for make during image procossing Part I is higher than second
The priority divided, on the basis of the experience that ensure that beholder, improves image processing efficiency.
The present embodiment in the specific implementation, can apply to display device.For needing on large-sized display device
When the image of display carries out real-time processing, number of calculations very big, complexity is very high so that image processing efficiency is relatively low.
Large scale is relative concept, for display device, for the viewing distance of respective usage scenario (such as project/
, more than 1 meter, within 1 meter, head-mounted display apparatus are within 10 centimetres for display for TV), viewing area is more than beholder's
Area of visual field, can be referred to as large scale.Such as large scale projection, large size TV refer to more than 50 cun it is viewing area,
Large scale computer display refers to more than 27 cun of viewing area, large scale head-mounted display apparatus and refers to more than 5 cun and shows
Show region.
But beholder is actual watch image when, due to the visual characteristic of beholder, the fixation range reality of beholder
The upper field range less than beholder, that is to say, that the viewing areas and viewing for focusing on and watching attentively are there is also in the visual field of beholder
Sensing region beyond region.Therefore, if according to this visual characteristic of beholder, to the viewing areas in its visual field and sense
Know processing mode of the area using differentiation, it is inevitable compared to relatively all being processed using identical processing mode whole viewing area
The performance requirement to processor can be reduced, so as to improve image processing efficiency.
Embodiment two
Based on the method described in embodiment one, the source images number is processed this gives according to the location parameter
According to several concrete modes, it is specific as follows:
Mode one:After the Part I and Part II for obtaining the location parameter of source images and determining source images, to source
The Part I of view data is processed by the first processing mode, obtains the image of the viewing areas;Then to described
The Part II of source image data is processed by second processing mode, obtains the image of the sensing region;It is wherein described
Complexity of the complexity of the first processing mode higher than the second processing mode.
Further, can be processed using the first more complicated processing mode for Part I, using comparing
Complicated algorithm is processed.
In an application scenarios, can be to the Part I of source image data using at more complicated rendering intent
Reason so that the image display effect of the viewing areas obtained after Part I treatment is truer.And then adopted for Part II
Processed with better simply second processing mode, if do not rendered, but directly invoked the Perception Area of previous frame image
Area image shows as the sensing region of current frame image.Again for example, higher to the image procossing precision of Part I, and it is right
The image procossing precision of Part II is relatively low, because the operand of the low needs of processing accuracy is little, therefore to the requirement of processor
It is not high.
In another application scenarios, in order to embody spatial impression, render process will calculate each light source pair in scene
The influence of object;Subsequently, in addition it is also necessary to calculate light source projects out according to whether the object of transparent material has been used in scene
Shade come determine use depth pinup picture shade or use ray tracing shade.Subsequently, will also according to the material of object come
Calculate color, the texture different equivalent of material of body surface really.This example is in the specific implementation, it is necessary to Part I and
Two parts use different processing modes so that the display effect in viewing areas in target image is truer.For example, right
Rendered by complex method in the Part I of image, such as strengthened Processing Algorithm and feature recognition algorithms, to the
Two parts are only rendered by simple rendering intent, and Part II is processed only with the method for adjustment light such as
Deng;Again for example, being processed using identical rendering intent in Part I and Part II, but to the image of Part I
Processing accuracy is higher, can be reached when such as to image floating-point operation 7 after decimal point, and to the image procossing precision of Part II
It is relatively low, only reached when such as to image floating-point operation 2 after decimal point.Due to processing accuracy it is relatively low when, required operand is not
Greatly, therefore to the less demanding of processor.Again for example, image refreshing faster in the case of, rendered for viewing areas,
Sensing region can not be rendered.
When image procossing is carried out, image processing program can by the screen that is rendered into of view projection images, in order to improve
The effect for rendering, rendering frequency, to need to reach 90 frames per second or higher, so has higher to processor when to image procossing
Requirement.And be only a part for whole viewing area due to Part I in the present embodiment, only to Part I at
Reason, will greatly shorten the time of disposed of in its entirety image, so as to improve the image processing efficiency of image.
Mode two:After the Part I and Part II for obtaining the location parameter of source images and determining source images, wherein,
The resolution ratio of the Part I of the source image data is identical with the resolution ratio of the Part II, but to Part I and
Two parts carry out different treatment.For example, processing source image data, image is pressed in final processing stage then
During contracting treatment, the high resolution of the viewing areas of the target image can be made in the resolution ratio of the sensing region.Such mesh
The viewing areas of logo image seem more fine and smooth, and sensing region is more coarse.
Mode three:After the Part I and Part II for obtaining the location parameter of source images and determining source images, the
One time started the Part I treatment to the source image data, obtained the image of the viewing areas;Opened in the second time
The Part II treatment begun to the source image data, obtains the image of the sensing region;The very first time is earlier than described
Second time.Specifically, the viewing areas of viewing area are focused on due to eye, it is inevitable in viewing display area first to see sight
See region, therefore can first render the image of Part I, while Part I is rendered, Part II is still kept in display
One two field picture part corresponding with Part II, the viewing effect of beholder is not interfered with so yet.Can be with for Part II
Render after a while, if the image watched has been refreshed, the render process to Part II can also be omitted.Specifically render
Method can be in the method cited by employing mode one.This method does not influence the viewing effect of beholder, also improves image
Treatment effeciency.
The three of the above mode of the present embodiment both can individually be implemented, it is also possible to which combination is implemented.
It is worth noting that, in 3D rendering processing method involved in the application scenarios of aforesaid way one.Due to
3D rendering Display Technique range of application is wider, corresponding treatment can be done according to different application scenarios, for example, by wearing
During formula display device viewing 3D rendering, because head-mounted display apparatus have two display screens corresponding with right and left eyes difference, two
Individual display screen shows two images of correspondence right and left eyes, because the parallax that two images are present makes to show three-dimensional showing in human brain
Show effect.Therefore, when processing this kind of 3D rendering, two images can also respectively be processed.That is, to first
The processing procedure of individual image is that location parameter can be obtained by camera, and according to location parameter determine Part I and
Part II, then to Part I and Part II, using three of the above mode, any one is processed so that target figure
Then the priority of the viewing areas as in be presented higher than the priority of sensing region according to viewing areas in first image
Picture material determine second viewing location of image, and second Part I of image is determined according to viewing location
And Part II, and processed with first image identical mode, to cause that the priority of viewing areas is higher than Perception Area
Domain.
Be can be seen that by the above embodiment of the present invention is carried out not by the Part I and Part II to source image data
Same treatment, can improve image processing efficiency in the case of the viewing effect for not influenceing beholder.
Embodiment three
Based on the method described in embodiment one, this gives several concrete modes for obtaining location parameter, specifically
It is as follows:
Acquisition includes the image information of the eyes of beholder;
Viewing location of the beholder in the viewing area is determined according to described image information;
Wherein, the viewing location is located in the viewing areas of the target image;
Wherein, obtaining includes the image information of the eyes of beholder, including:The sight is obtained in being spaced in Preset Time
The eyes image of the person of seeing.
Specifically, the image information of the eyes of beholder can be obtained, according to the analysis to image information forever without pause of breathing, can
To determine viewing location of the beholder in viewing area.Although image refreshing frequency in display is higher here, eyes
Change the time interval of the time interval much larger than image update of viewing location, that is to say, that eyes change position comparatively fast, therefore
Without consideration treatment, in the problem changed to Part I and Part II treatment retracement line.
In the specific implementation, because the eyes of beholder may at any time change viewing location, therefore, to need very short
Time obtains an eyes image information for the eyes of beholder, for example, to make beholder to perceive viewing areas and sense
Know the difference in region, the eyes image of beholder can be just obtained in each refreshed image.If it find that beholder
Viewing location is changed, it is necessary to which the Part I to source image data is processed again, to the Part II of source images
It is also carried out processing again.
In an application scenarios, obtaining location parameter can use eyeball tracking technology of the prior art.When people's
Eyes see that during to different directions, eye can occur trickle change, and these changes can produce the feature that can be extracted, harvester
These features can be extracted by picture catching or scanning, so that the change of real-time tracing eyes.
Further, in the present embodiment, the source image data includes:First source images subdata and the second source images
Subdata, the first source images subdata and the second source images subdata can allow the beholder to see three-dimensional effect diagram
Picture
Specifically, to make beholder obtain more preferable sensory experience, source image data can be 3D rendering data.Specific source
View data includes the first source images subdata and the second source images subdata, the first source images subdata and the second source images
There is certain difference in the corresponding content of data display, corresponding content coordinate in the picture, such beholder can see
To the image that there is parallax, three-dimensional image effect is produced in the brain with beholder.
Further, location parameter includes:First sub- location parameter and the second sub- location parameter;Viewing area includes:The
One sub-viewing areas and the second sub-viewing areas;Wherein, the described first sub- location parameter is that the first eye of the beholder exists
The first viewing location in first sub-viewing areas, the second sub- location parameter be second of the beholder it is fine
The second viewing location in second sub-viewing areas;
Processing the source image data acquisition target image includes:According to the first sub- location parameter treatment first source figure
As subdata obtains first object image;The second source images subdata is processed according to the second sub- location parameter obtain the second mesh
Logo image;
Display target image includes:The first object image is shown in first sub-viewing areas, described second
Sub-viewing areas show second target image.
Specifically, the first sub- location parameter characterizes viewing location of the first eye eye in the first sub-viewing areas, second
Sub- location parameter characterizes second fine viewing location in the second sub-viewing areas, and viewing areas are determined according to viewing location,
And then source images are carried out with treatment formation first object image and the second target image.
Further, the described first sub- location parameter determines the first focal position in the first source images subdata, by institute
The Part I that the first focal position determines the first source images subdata is stated,
The second sub- location parameter determines the second focal position in the second source images subdata, by second focus
Position determines the Part I of the second source images subdata.
According to first focal position and preset range parameter, the Part I is determined;
According to second focal position and preset range parameter, the Part II is determined.
Can determine focal position in a different manner when it is determined that watching focal position.For example, in an applied field
Jing Zhong, as shown in figure 3, the content that can be shown according to viewing areas determines the first focal position of the first source images subdata
With the second focal position of the second source image data.It is identical with the method that eye image determines for left-eye image, do not do herein
Distinguish.Display content of the beholder in viewing areas, such as lightning are determined first, and the first source images is determined further according to display content
First position parameter in data, according to first position parameter, determines the display content pair in the first source images subdata
The coordinate answered, then determines position coordinates of the display content in the second source images subdata, further according to default model again
Enclose parameter and determine the corresponding Part I of the first source images subdata and Part II respectively, further according to the side that embodiment two is introduced
Method is respectively processed to Part I and Part II.
In another application scenarios, can come true according to first eye to the viewing location of the first source images subdata
Viewing location of fixed second eyes to the second source images subdata.For example, it is first determined beholder is in the first source images subdata
Viewing location, further according to preset range parameter, determine the Part I and Part II of the first source images subdata, Ran Houzai
According to the first source image data and the coordinate difference of the second source image data, it is determined that giving the Part I of the first source images subdata
The Part I and Part II of the second source images subdata corresponding with Part II, then again to Part I and second
The method introduced according to embodiment two is divided to be respectively processed.
Can be seen that the present invention can realize the viewing according to beholder in viewing area by the above embodiment of the present invention
Position, the Part I according to the corresponding source image data of the corresponding viewing areas of viewing location and the sense different from viewing areas
Know the Part II of the source image data corresponding to region, the priority for make during image procossing Part I is higher than second
The priority divided, so as to improve rendering efficiency.
Example IV
A kind of display device is present embodiments provided, the display device is provided with display device, harvester and treatment
Device.Display device configurations are acquisition source image data;Display device, is configured to constitute the display of the display source image data
Region;Harvester, is configured to obtain location parameter, and the location parameter shows viewing position of the beholder in viewing area
Put;The processing unit, is additionally configured to process the source image data according to the location parameter, obtains and display target figure
Picture;Wherein, the target image includes viewing areas corresponding with the location parameter and the sense different from the viewing areas
Know region;Wherein, the source image data include Part I corresponding with the viewing areas and with the sensing region pair
The Part II answered, the priority of the priority higher than the Part II of the Part I.
Wherein in one embodiment, display device can be wearable device, and such as wear-type virtual reality (VR) shows
Show equipment or wear-type augmented reality (AR) display device.Display device is arranged at the visual line of sight of beholder's eyes after wearing
Interior, if wearable device is wear-type VR display devices, the display device is preferably lighttight device, and wear-type VR shows
The camera of equipment is set towards eyes, in order to gather viewing location of the beholder in viewing area;If wearable
Equipment is wear-type AR display devices, then display device can be the eyeglass module of printing opacity, the environment of wear-type AR display devices
Camera is consistent with visual line of sight, in order to gather viewing location of the beholder described in image in viewing area.Wherein, source figure
As data can be obtained by main frame with download or online mode, or, for that can be come using wear-type AR display devices
From the image of environment camera collection.
Viewing area can be divided into sight by the display device of the present embodiment by the viewing location of the eyes to beholder
Viewing areas and the non-sensing region paid close attention to that the person's of seeing eyes are paid close attention to, are processed by source images so that seen
See that the priority in region is preferential higher than sensing region.Further, as shown in figure 4, the display device include processing unit,
Display device and harvester;
Processing unit, is configured to obtain source image data;
Display device, is configured to constitute the viewing area of the display source image data;
Harvester, is configured to obtain location parameter, and the location parameter shows viewing of the beholder in viewing area
Position;
The processing unit, is additionally configured to process the source image data according to the location parameter, obtains and show mesh
Logo image;Wherein, the target image includes viewing areas corresponding with the location parameter and different from the viewing areas
Sensing region;
Wherein, the source image data include Part I corresponding with the viewing areas and with the sensing region pair
The Part II answered, the priority of the priority higher than the Part II of the Part I.
In another specific embodiment, the harvester includes the first harvester and the second harvester.Described
One harvester is arranged at the marginal position of first display device, and second harvester is similarly provided at described second
The edge of display device, and the direction that two harvesters show both facing to image.
Wherein, first harvester is configured to the image information of the left eye for gathering beholder, to determine the viewing
The viewing location of the left eye of person;
Second harvester is configured to gather the image information of beholder's right eye, to determine beholder's right eye
Viewing location;
Wherein, the display device includes the first display device and the second display device,
First display device configurations are to constitute the viewing area of beholder's left eye viewing;
First display device configurations are to constitute the viewing area of beholder's right eye viewing.
In another specific embodiment, wherein, the resolution ratio of the Part I of the source image data and described second
Partial resolution ratio is identical, and the viewing areas of the target image high resolution in the resolution ratio of the sensing region.
In another specific embodiment, wherein,
Start the Part I treatment to the source image data in the very first time, obtain the image of the viewing areas;
Start the Part II treatment to the source image data in the second time, obtain the image of the sensing region;
The very first time is earlier than second time.
In another specific embodiment, the harvester concrete configuration is:
Acquisition includes the image information of the eyes of beholder;
Viewing location of the beholder in the viewing area is determined according to described image information;
Wherein, the viewing location is located in the viewing areas of the target image.
In another specific embodiment, wherein,
The source image data includes:First source images subdata and the second source images subdata, first source images
Subdata and the second source images subdata are the images that the beholder can be allowed to see stereoeffect;
Location parameter includes:First sub- location parameter and the second sub- location parameter;Viewing area includes:First sub- viewing area
Domain and the second sub-viewing areas;Wherein, the described first sub- location parameter is the first eye of the beholder in the described first son
The first viewing location in viewing area, the second sub- location parameter is second eyes of the beholder in the described second son
The second viewing location in viewing area;
Processing the source image data acquisition target image includes:According to the first sub- location parameter treatment first source figure
As subdata obtains first object image;The second source images subdata is processed according to the second sub- location parameter obtain the second mesh
Logo image;
Display target image includes:The first object image is shown in first sub-viewing areas, described second
Sub-viewing areas show second target image.
In another specific embodiment, wherein,
First focal position is determined in the first source images subdata according to the described first sub- location parameter;
Second focal position is determined in the second source images subdata according to the described second sub- location parameter.
In another specific embodiment, wherein,
According to first focal position, first content is determined, by the first content and preset range parameter determination institute
State Part I;
According to second focal position, the second content is determined, by second content and preset range parameter determination institute
State Part II.
It need to be noted that be:Above display device implements the description of item, is similar, tool with above method description
There is same embodiment of the method identical beneficial effect, therefore do not repeat.For what is do not disclosed in display device embodiment of the present invention
Ins and outs, those skilled in the art refer to the description of the inventive method embodiment and understand, to save length, here not
Repeat again.
In several embodiments provided herein, it should be understood that disclosed apparatus and method, can be by it
Its mode is realized.Apparatus embodiments described above are only schematical, for example, the division of the unit, is only
A kind of division of logic function, can have other dividing mode, such as when actually realizing:Multiple units or component can be combined, or
Another system is desirably integrated into, or some features can be ignored, or do not perform.In addition, shown or discussed each composition portion
Coupling point each other or direct-coupling or communication connection can be the INDIRECT COUPLINGs of equipment or unit by some interfaces
Or communication connection, can be electrical, machinery or other forms.
The above-mentioned unit that is illustrated as separating component can be or may not be it is physically separate, it is aobvious as unit
The part for showing can be or may not be physical location, you can with positioned at a place, it is also possible to be distributed to multiple network lists
In unit;Part or all of unit therein can be according to the actual needs selected to realize the purpose of this embodiment scheme.
In addition, each functional unit in various embodiments of the present invention can be fully integrated into a processing unit, also may be used
Being each unit individually as a unit, it is also possible to which two or more units are integrated in a unit;It is above-mentioned
Integrated unit can both be realized in the form of hardware, it would however also be possible to employ hardware adds the form of SFU software functional unit to realize.
One of ordinary skill in the art will appreciate that:Realizing all or part of step of above method embodiment can pass through
Programmed instruction related hardware is completed, and foregoing program can be stored in a computer read/write memory medium, the program
Upon execution, the step of including above method embodiment is performed;And foregoing storage medium includes:It is movable storage device, read-only
Memory (ROM, Read Only Memory), random access memory (RAM, Random Access Memory), magnetic disc or
Person's CD etc. is various can be with the medium of store program codes.
Or, if the above-mentioned integrated unit of the present invention is to realize in the form of software function module and as independent product
When selling or using, it is also possible to which storage is in a computer read/write memory medium.Based on such understanding, the present invention is implemented
The part that the technical scheme of example substantially contributes to prior art in other words can be embodied in the form of software product,
The computer software product is stored in a storage medium, including some instructions are used to so that computer equipment (can be with
It is personal computer, server or network equipment etc.) perform all or part of each embodiment methods described of the invention.
And foregoing storage medium includes:Movable storage device, read-only storage (ROM, Read Only Memory), arbitrary access are deposited
Reservoir (RAM, Random Access Memory), magnetic disc or CD etc. are various can be with the medium of store program codes.
The above, specific embodiment only of the invention, but protection scope of the present invention is not limited thereto, and it is any
Those familiar with the art the invention discloses technical scope in, change or replacement can be readily occurred in, should all contain
Cover within protection scope of the present invention.Therefore, protection scope of the present invention should be based on the protection scope of the described claims.
Claims (10)
1. a kind of image processing method, including:
Obtain source image data;
Location parameter is obtained, the location parameter shows viewing location of the beholder in viewing area;
The source image data is processed according to the location parameter, is obtained and display target image;Wherein, the target image bag
Include viewing areas corresponding with the location parameter and the sensing region different from the viewing areas;
Wherein, the source image data includes Part I corresponding with the viewing areas and corresponding with the sensing region
Part II, the priority of the priority higher than the Part II of the Part I.
2. method according to claim 1, wherein, the source image data is processed according to the location parameter, including:
Part I to the source image data is processed by the first processing mode, obtains the figure of the viewing areas
Picture;
Part II to the source image data is processed by second processing mode, obtains the figure of the sensing region
Picture;
Complexity of the complexity of first processing mode higher than the second processing mode.
3. method according to claim 1, wherein, the resolution ratio of the Part I of the source image data and described second
Partial resolution ratio is identical, and the viewing areas of the target image high resolution in the resolution ratio of the sensing region.
4. method according to claim 1, wherein,
Start the Part I treatment to the source image data in the very first time, obtain the image of the viewing areas;
Start the Part II treatment to the source image data in the second time, obtain the image of the sensing region;
The very first time is earlier than second time.
5. method according to claim 1, obtains location parameter, including:
Acquisition includes the image information of the eyes of beholder;
Viewing location of the beholder in the viewing area is determined according to described image information;
Wherein, the viewing location is located in the viewing areas of the target image.
6. method according to claim 5, wherein,
The source image data includes:First source images subdata and the second source images subdata, the first source images subnumber
It is image that the beholder can be allowed to see stereoeffect according to the second source images subdata;
Location parameter includes:First sub- location parameter and the second sub- location parameter;Viewing area includes:First sub-viewing areas and
Second sub-viewing areas;Wherein, the described first sub- location parameter is the first eye of the beholder in the described first son display
The first viewing location in region, the second sub- location parameter is second eyes of the beholder in the described second son display
The second viewing location in region;
Processing the source image data acquisition target image includes:The first source images is processed according to the first sub- location parameter
Data obtain first object image;The second source images subdata is processed according to the second sub- location parameter obtain the second target figure
Picture;
Display target image includes:The first object image is shown in first sub-viewing areas, it is aobvious in the described second son
Show that region shows second target image.
7. method according to claim 6, wherein,
First focal position is determined in the first source images subdata according to the described first sub- location parameter;
Second focal position is determined in the second source images subdata according to the described second sub- location parameter.
8. method according to claim 7, wherein,
According to first focal position, first content is determined, as described in the first content and preset range parameter determination
A part;
According to second focal position, the second content is determined, as described in second content and preset range parameter determination
Two parts.
9. a kind of display device, including:
Processing unit, is configured to obtain source image data;
Display device, is configured to constitute the viewing area of the display source image data;
Harvester, is configured to obtain location parameter, and the location parameter shows viewing location of the beholder in viewing area;
The processing unit, is additionally configured to process the source image data according to the location parameter, obtains and display target figure
Picture;Wherein, the target image includes viewing areas corresponding with the location parameter and the sense different from the viewing areas
Know region;
Wherein, the source image data includes Part I corresponding with the viewing areas and corresponding with the sensing region
Part II, the priority of the priority higher than the Part II of the Part I.
10. display device according to claim 9, wherein, the harvester is adopted including the first harvester and second
Acquisition means,
First harvester is configured to the image information of the left eye for gathering beholder, to determine the left eye of the beholder
Viewing location;
Second harvester is configured to gather the image information of beholder's right eye, to determine the viewing of beholder's right eye
Position;
Wherein, the display device includes the first display device and the second display device,
First display device configurations are to constitute the viewing area of beholder's left eye viewing;
First display device configurations are to constitute the viewing area of beholder's right eye viewing.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201710202490.7A CN106851249A (en) | 2017-03-30 | 2017-03-30 | Image processing method and display device |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201710202490.7A CN106851249A (en) | 2017-03-30 | 2017-03-30 | Image processing method and display device |
Publications (1)
Publication Number | Publication Date |
---|---|
CN106851249A true CN106851249A (en) | 2017-06-13 |
Family
ID=59141581
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201710202490.7A Pending CN106851249A (en) | 2017-03-30 | 2017-03-30 | Image processing method and display device |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN106851249A (en) |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109522866A (en) * | 2018-11-29 | 2019-03-26 | 宁波视睿迪光电有限公司 | Naked eye 3D rendering processing method, device and equipment |
WO2020034669A1 (en) * | 2018-08-16 | 2020-02-20 | 北京七鑫易维信息技术有限公司 | Image obtaining device control method and apparatus |
CN110856035A (en) * | 2018-07-24 | 2020-02-28 | 顶级公司 | Processing image data to perform object detection |
CN111988525A (en) * | 2020-08-25 | 2020-11-24 | Oppo广东移动通信有限公司 | Image processing method and related device |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101655984A (en) * | 2008-05-12 | 2010-02-24 | 美国西门子医疗解决公司 | System for adaptively processing medical image data |
CN102347016A (en) * | 2010-07-28 | 2012-02-08 | 佳能株式会社 | Display control apparatus for displaying image, display control method |
CN103930817A (en) * | 2011-06-20 | 2014-07-16 | 谷歌公司 | Systems and methods for adaptive transmission of data |
CN104767992A (en) * | 2015-04-13 | 2015-07-08 | 北京集创北方科技有限公司 | Head-wearing type display system and image low-bandwidth transmission method |
-
2017
- 2017-03-30 CN CN201710202490.7A patent/CN106851249A/en active Pending
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101655984A (en) * | 2008-05-12 | 2010-02-24 | 美国西门子医疗解决公司 | System for adaptively processing medical image data |
CN102347016A (en) * | 2010-07-28 | 2012-02-08 | 佳能株式会社 | Display control apparatus for displaying image, display control method |
CN103930817A (en) * | 2011-06-20 | 2014-07-16 | 谷歌公司 | Systems and methods for adaptive transmission of data |
CN104767992A (en) * | 2015-04-13 | 2015-07-08 | 北京集创北方科技有限公司 | Head-wearing type display system and image low-bandwidth transmission method |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110856035A (en) * | 2018-07-24 | 2020-02-28 | 顶级公司 | Processing image data to perform object detection |
CN110856035B (en) * | 2018-07-24 | 2024-01-26 | Arm有限公司 | Processing image data to perform object detection |
WO2020034669A1 (en) * | 2018-08-16 | 2020-02-20 | 北京七鑫易维信息技术有限公司 | Image obtaining device control method and apparatus |
US11853471B2 (en) | 2018-08-16 | 2023-12-26 | Beijing 7Invensun Technology Co., Ltd. | Method and apparatus for controlling image acquisition device |
CN109522866A (en) * | 2018-11-29 | 2019-03-26 | 宁波视睿迪光电有限公司 | Naked eye 3D rendering processing method, device and equipment |
CN111988525A (en) * | 2020-08-25 | 2020-11-24 | Oppo广东移动通信有限公司 | Image processing method and related device |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US20210075963A1 (en) | Method and apparatus for obtaining binocular panoramic image, and storage medium | |
CN106484116B (en) | The treating method and apparatus of media file | |
CN110378914A (en) | Rendering method and device, system, display equipment based on blinkpunkt information | |
CN106851249A (en) | Image processing method and display device | |
Blum et al. | The effect of out-of-focus blur on visual discomfort when using stereo displays | |
CN109522866A (en) | Naked eye 3D rendering processing method, device and equipment | |
CA2709417A1 (en) | Method of simulating blur in digitally processed images | |
US10885651B2 (en) | Information processing method, wearable electronic device, and processing apparatus and system | |
CN109901290B (en) | Method and device for determining gazing area and wearable device | |
CN111880654A (en) | Image display method and device, wearable device and storage medium | |
US11956415B2 (en) | Head mounted display apparatus | |
CN111275801A (en) | Three-dimensional picture rendering method and device | |
US20230239457A1 (en) | System and method for corrected video-see-through for head mounted displays | |
US11212502B2 (en) | Method of modifying an image on a computational device | |
JP2023515205A (en) | Display method, device, terminal device and computer program | |
Tong et al. | Optical distortions in VR bias the perceived slant of moving surfaces | |
CN115202475A (en) | Display method, display device, electronic equipment and computer-readable storage medium | |
JP2016532147A (en) | How to identify the visual effects of ophthalmic lenses | |
CN108881892B (en) | Anti-dizziness method and system for desktop virtual reality system | |
US20230244307A1 (en) | Visual assistance | |
CN111587397A (en) | Image generation device, spectacle lens selection system, image generation method, and program | |
CN115327782B (en) | Display control method and device, head-mounted display equipment and readable storage medium | |
EP3961572A1 (en) | Image rendering system and method | |
US11270475B2 (en) | Variable rendering system and method | |
Gurrieri | Improvements in the visualization of stereoscopic 3D imagery |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20170613 |
|
RJ01 | Rejection of invention patent application after publication |