CN105867615A - Information processing method and electronic device - Google Patents
Information processing method and electronic device Download PDFInfo
- Publication number
- CN105867615A CN105867615A CN201610173581.8A CN201610173581A CN105867615A CN 105867615 A CN105867615 A CN 105867615A CN 201610173581 A CN201610173581 A CN 201610173581A CN 105867615 A CN105867615 A CN 105867615A
- Authority
- CN
- China
- Prior art keywords
- image
- virtual scene
- virtual
- display parameters
- scene
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/017—Gesture based interaction, e.g. based on a set of recognized hand gestures
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T19/00—Manipulating 3D models or images for computer graphics
- G06T19/006—Mixed reality
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2203/00—Indexing scheme relating to G06F3/00 - G06F3/048
- G06F2203/01—Indexing scheme relating to G06F3/01
- G06F2203/012—Walk-in-place systems for allowing a user to walk in a virtual environment while constraining him to a given position in the physical environment
Landscapes
- Engineering & Computer Science (AREA)
- General Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Computer Graphics (AREA)
- Computer Hardware Design (AREA)
- Software Systems (AREA)
- Human Computer Interaction (AREA)
- Processing Or Creating Images (AREA)
Abstract
The present invention discloses an information processing method and an electronic device. The method comprises: based on a current environment in which a first object is located, acquiring a first image, wherein the first image comprises first object information corresponding to the first object; acquiring a second image corresponding to a first virtual scene, wherein the second image comprises at least one virtual object; and based on the first image and the second image, generating a third image corresponding to the first virtual scene, wherein the third image comprises the first object information.
Description
Technical field
The present invention relates to electronic technology field, particularly to a kind of information processing method and electronic equipment.
Background technology
Along with the development of science and technology, in prior art, various electronic equipments, especially wear-type show
Equipment, such as: augmented reality glasses, virtual implementing helmet etc. are all being constantly updated, and bring different to user
Experience effect.
At present, the virtual scene built by virtual reality or augmented reality, it is possible to allow user incorporate
Virtual scene obtains colourful experience.But, when the object in reality scene is taken pictures by user,
Can only obtain the picture that this object is in reality scene, such as user uses mobile phone to autodyne, and can only obtain
To the user oneself picture with true environment as background, and cannot autodyning user and abundant virtual scene
Combining, Consumer's Experience is more single.
Summary of the invention
The embodiment of the present invention provides a kind of information processing and electronic equipment, by taking pictures with virtual in reality scene
Scene combines, and promotes Consumer's Experience.
On the one hand, the embodiment of the present application provides a kind of information processing method, comprises the following steps:
Obtaining the first image of current environment residing for the first object, described first image includes first right with described
As the first corresponding object information;
Obtaining second image corresponding with the first virtual scene, described second image includes that at least one is virtual
Object;
Based on described first image and described second image, generate corresponding with described first virtual scene
Three images, wherein, described 3rd image includes described first object information.
Optionally, the second image that described acquisition is corresponding with the first virtual scene, including:
Obtain second virtual scene different from described first virtual scene;
Adjust at least one virtual objects in described first virtual scene and described second void in the same scale
Intend at least one virtual objects in scene, it is thus achieved that the first adjusted virtual scene and the adjusted second void
Intend scene;
Splice described the first adjusted virtual scene and described the second adjusted virtual scene, it is thus achieved that the 3rd
Virtual scene;
Obtain second image corresponding with described 3rd virtual scene.
Optionally, described based on described first image with described second image, generate and the first virtual scene phase
The 3rd corresponding image, including:
Based on described first object information, extract from described first image and comprise the of described first object
One subgraph;
Synthesize described first subgraph and described second image, generate corresponding with described first virtual scene
3rd image.
Optionally, after described first subgraph of described synthesis and described second image, described method is also wrapped
Include:
Obtain the first display parameters and second display parameters of described second image of described first subgraph;
If described first display parameters do not mate with described second display parameters, then adjust described first display ginseng
Several and/or described second display parameters so that the second display ginseng after the first display parameters after adjustment and adjustment
Number matches.
Optionally, described first display parameters and described second display parameters include brightness of image, saturation degree or
At least one in transparency.
On the other hand, the embodiment of the present application also provides for a kind of electronic equipment, including:
Housing;
Processing means, is arranged in described housing, for obtaining the first figure of current environment residing for the first object
Picture, described first image includes first object information corresponding with described first object;Obtain virtual with first
The second image that scene is corresponding, described second image includes at least one virtual objects;Based on described first
Image and described second image, generate threeth image corresponding with described first virtual scene, wherein, institute
State the 3rd image and include described first object information.
Optionally, described processing means is used for:
Obtain second virtual scene different from described first virtual scene;
Adjust at least one virtual objects in described first virtual scene and described second void in the same scale
Intend at least one virtual objects in scene, it is thus achieved that the first adjusted virtual scene and the adjusted second void
Intend scene;
Splice described the first adjusted virtual scene and described the second adjusted virtual scene, it is thus achieved that the 3rd
Virtual scene;
Obtain second image corresponding with described 3rd virtual scene.
Optionally, described processing means is used for:
Based on described first object information, extract from described first image and comprise the of described first object
One subgraph;
Synthesize described first subgraph and described second image, generate corresponding with described first virtual scene
3rd image.
Optionally, after described first subgraph of described synthesis and described second image, described processing means
It is additionally operable to:
Obtain the first display parameters and second display parameters of described second image of described first subgraph;
If described first display parameters do not mate with described second display parameters, then adjust described first display ginseng
Several and/or described second display parameters so that the second display ginseng after the first display parameters after adjustment and adjustment
Number matches.
Optionally, described first display parameters and described second display parameters include brightness of image, saturation degree or
At least one in transparency.
On the other hand, the embodiment of the present application also provides for a kind of electronic equipment, including:
First acquiring unit, for obtaining the first image of current environment, described first figure residing for the first object
As including first object information corresponding with described first object;
Second acquisition unit, for obtaining second image corresponding with the first virtual scene, described second figure
As including at least one virtual objects;
First signal generating unit, for based on described first image and described second image, generates and described first
The 3rd image that virtual scene is corresponding, wherein, described 3rd image includes described first object information.
Said one in the embodiment of the present application or multiple technical scheme, at least have one or more skills following
Art effect:
One, the technical scheme provided due to the embodiment of the present application, obtains the of current environment residing for the first object
One image, described first image includes first object information corresponding with described first object;Obtain and first
The second image that virtual scene is corresponding, described second image includes at least one virtual objects;Based on described
First image and described second image, generate threeth image corresponding with described first virtual scene, wherein,
Described 3rd image includes described first object information.I.e. will not be as, in prior art, user is to reality scene
In object when taking pictures, the picture that this object is in reality scene, such as user can only be obtained and use
Mobile phone is autodyned, and can only obtain the user oneself picture with true environment as background, and cannot be by user's
Autodyning and combine with abundant virtual scene, Consumer's Experience is more single.And in the technical program, it is possible to
According to user's image in reality scene and generate corresponding with virtual scene with the image in virtual scene
3rd image, obtains the background image more horn of plenty of image, to reach to improve the technique effect of user experience.
Two, due to the technical scheme in the embodiment of the present application, different from described first virtual scene is obtained
Two virtual scenes;Adjust at least one virtual objects in described first virtual scene and institute in the same scale
State at least one virtual objects in the second virtual scene, it is thus achieved that the first adjusted virtual scene and adjusted
The second virtual scene;Splice described the first adjusted virtual scene and described the second adjusted virtual field
Scape, it is thus achieved that the 3rd virtual scene;Obtain second image corresponding with described 3rd virtual scene.I.e. in this skill
In art scheme, current virtual scene and previously saved virtual scene can be spliced, more to obtain
The image of the virtual scene of form, and then reach to improve the technique effect of user experience.
Three, due to the technical scheme in the embodiment of the present application, the first display ginseng of described first subgraph is obtained
Several and the second display parameters of described second image;If described first display parameters and described second display parameters
Do not mate, then adjust described first display parameters and/or described second display parameters so that first after adjustment
The second display parameters after display parameters and adjustment match.I.e. in the technical program, at synthesis the first son
After image and the second image, also the display parameters of the first subgraph and/or the second image are adjusted,
So that first subgraph the second image after He Cheng more meets the visual effect of user, and then reach to improve user
The technique effect of Experience Degree.
Accompanying drawing explanation
A kind of information processing method that Fig. 1 provides for the embodiment of the present application one implement flow chart;
In a kind of information processing method that Fig. 2 provides for the embodiment of the present application one, step S102 implements stream
Cheng Tu;
The structural representation of a kind of electronic equipment that Fig. 3 provides for the embodiment of the present application two;
The structural representation of a kind of electronic equipment that Fig. 4 provides for the embodiment of the present application three.
Detailed description of the invention
The technical scheme that the embodiment of the present application provides, is used for solving to exist in prior art to pass through image at needs
When collecting unit obtains user images, the technical problem of user's image in reality scene can only be obtained, with
Realize the technique effect that actual situation is autodyned.
Technical scheme in the embodiment of the present application is for solving above-mentioned technical problem, and general thought is as follows:
Obtaining the first image of current environment residing for the first object, described first image includes first right with described
As the first corresponding object information;
Obtaining second image corresponding with the first virtual scene, described second image includes that at least one is virtual
Object;
Based on described first image and described second image, generate corresponding with described first virtual scene
Three images, wherein, described 3rd image includes described first object information.
In technique scheme, obtain the first image of current environment residing for the first object, described first figure
As including first object information corresponding with described first object;Obtain corresponding with the first virtual scene
Two images, described second image includes at least one virtual objects;Based on described first image and described second
Image, generates threeth image corresponding with described first virtual scene, and wherein, described 3rd image includes
Described first object information.I.e. will not be as, in prior art, the object in reality scene is taken pictures by user
Time, the picture that this object is in reality scene, such as user can only be obtained and use mobile phone to autodyne, only
The user oneself picture with true environment as background can be obtained, and cannot be by virtual with enrich for autodyning of user
Scene combines, and Consumer's Experience is more single.And in the technical program, it is possible to according to user in reality field
Image in scape and threeth image corresponding with virtual scene with the image generation in virtual scene, obtain figure
The background image of picture more horn of plenty, to reach to improve the technique effect of user experience.
In order to be better understood from technique scheme, below by accompanying drawing and specific embodiment to skill of the present invention
Art scheme is described in detail, it should be understood that the specific features in the embodiment of the present application and embodiment is to this
The detailed description of inventive technique scheme rather than the restriction to technical solution of the present invention, in the feelings do not conflicted
Under condition, the technical characteristic in the embodiment of the present application and embodiment can be mutually combined.
Embodiment one
Refer to Fig. 1, a kind of information processing method that the embodiment of the present application provides, including:
S101: obtain the first image of current environment residing for the first object, described first image includes with described
The first object information that first object is corresponding;
S102: obtain second image corresponding with the first virtual scene, described second image includes at least one
Individual virtual objects;
S103: based on described first image and described second image, generates relative with described first virtual scene
The 3rd image answered, wherein, described 3rd image includes described first object information.
A kind of information processing method that the embodiment of the present application one provides, it is possible to be applied to an electronic equipment, this electricity
Subset can be head-mounted display apparatus, such as: augmented reality glasses, virtual implementing helmet or be not
Headset equipment, here, schematically illustrate the most one by one.
In the embodiment of the present application, step S101 is first carried out: obtain of current environment residing for the first object
One image, described first image includes first object information corresponding with described first object.
In the embodiment of the present application, it is thus achieved that the first image be specifically as follows the image by the second electronic equipment
Collecting unit obtains, and sends to this electronic equipment, and is stored in the image in this electronic equipment, concrete real
During Xian, this image can obtain at current time, it is also possible to when being before current time
Obtain.In the embodiment of the present application, the second electronic equipment be specifically as follows with front-facing camera and/or after
Put the smart mobile phone of camera, panel computer or notebook computer etc., or other electronic equipment.
In the embodiment of the present application, the first image is specifically as follows two dimensional image or 3-D view, and specifically
During realization, 3-D view can be built based on the two dimensional image obtained, to obtain each object in this image
Depth information, further image is processed.
In the embodiment of the present application, the first object is specifically as follows who object, such as: the user of electronic equipment,
The household of user or friend;Or be animal target;Or it is plant object.Accordingly, implementing
Cheng Zhong, if the first object is personage, the first object information is then the face feature information of personage, such as: nose,
The characteristic informations such as face, chin, forehead;If the first object is plant, the first object information then this plant
The characteristic informations such as flower, rhizome, leaf.
In the embodiment of the present application, if the first image is two dimensional image, then can be based on the first image and first
The depth image of image, determines the first object information;If the second image is 3-D view, then can be based on
The depth information of the first image, determines the first object information, for above two embodiment, this area
Those of ordinary skill can be selected according to actual conditions, is not especially limited in the embodiment of the present application.
After execution of step S101, then perform step S102: obtain corresponding with the first virtual scene
Second image, described second image includes at least one virtual objects.
In the embodiment of the present application, before obtaining the second image, it is also possible to according to the actual demand pair of user
At least one virtual objects in first virtual scene is adjusted.During implementing, electronic equipment
User can pass through the gesture display parameters at least one virtual objects in virtual scene, such as: display position
Put, display size or display direction etc. are adjusted.Concrete, virtual objects, as a example by strange beast, works as identification
When gesture operation is specially the operation from top to bottom carried out in virtual scene, it is i.e. by virtual objects strange beast,
Move to the second place being positioned at immediately below current location from the primary importance being currently located;When identifying gesture behaviour
Make to be specially the position different from virtual scene of two fingers of hand close to each other time, be i.e. by virtual right
As strange beast reduces according to preset ratio, such as: be contracted to original 3/4ths or 1/2nd or other ratio
Example.
In the embodiment of the present application, when by gesture operation, at least one virtual objects in virtual scene being entered
During Row sum-equal matrix, the image of the hand of user can be gathered by the image acquisition units of electronic equipment, then pass through
Analysis to collection image, determines the gesture of user, and enters the virtual objects in virtual scene according to gesture
Row sum-equal matrix;Further, experience more really to provide the user with, it is also possible to according to gathering image, build
The three dimensional virtual models of one hand, shows three dimensional virtual models in virtual scene, when collecting user's hand
During gesture, the hand in virtual scene will make corresponding actions, and the hand of oneself is being adjusted to make user feel,
And then reach to improve the technique effect of user experience.
Further, in the embodiment of the present application, not only can be by the image of current virtual scene and first
Image synthesizes, it is also possible to provides the user more rich selection, i.e. obtains and include that current virtual scene exists
In multiple different virtual scenes, will multiple virtual scenes synthesize after the image that generates, the most again with first
Image synthesizes so that the material of picture is abundanter, to improve the Experience Degree of user.
Then corresponding step S102 implement process, refer to Fig. 2, specifically include following steps:
S201: obtain second virtual scene different from described first virtual scene;
S202: adjust at least one virtual objects in described first virtual scene and described in the same scale
At least one virtual objects in second virtual scene, it is thus achieved that the first adjusted virtual scene and adjusted
Second virtual scene;
S203: splice described adjusted first virtual scene and described adjusted second virtual scene, it is thus achieved that the
Three virtual scenes;
S204: obtain second image corresponding with described 3rd virtual scene.
During implementing, if the virtual scene that the first virtual scene is a game, then second is virtual
Scene can be the virtual scene on seashore;If the first virtual scene is the virtual scene in river, then second is virtual
Scene can be the virtual scene on grassland, or is other different virtual scene, here, the most one by one
Schematically illustrate.
After obtaining the first virtual scene and the second virtual scene, in order to make the size of virtual scene after merging
Equivalently-sized with the first virtual scene or the second virtual scene, then adjust the first virtual scene with same ratio
In at least one virtual objects and at least one virtual objects in the second virtual scene, concrete, first
At least one virtual objects in virtual scene and at least one object in the second virtual scene are all contracted to former
Two point one come, thus obtain the first adjusted virtual scene and the second adjusted virtual scene.
After obtaining the first adjusted virtual scene and the second adjusted virtual scene, adjusted
First virtual scene and the second adjusted virtual scene are stitched together, to obtain spliced virtual scene.
In concrete splicing, the first adjusted virtual scene and the second adjusted virtual scene can be according to
Connecting method, left and right connecting method or other connecting method up and down, those of ordinary skill in the art can root
Select according to being actually needed, be not especially limited in the embodiment of the present application.
After obtaining spliced virtual scene, obtain second image corresponding with spliced virtual scene.
Equally, the second image obtained based on spliced virtual scene is specifically as follows two dimensional image or 3-D view.
By the technical scheme in the embodiment of the present application, synthesized by the splicing of multiple virtual scenes, can give
User provides the material of more horn of plenty so that user can select according to the custom of oneself or needs, enters
And reach to improve the technique effect of user experience.
After execution of step S102, then perform step S103: based on described first image and described second
Image, generates threeth image corresponding with described first virtual scene, and wherein, described 3rd image includes
Described first object information.
In the embodiment of the present application, during step S103 is implemented, specifically include following steps:
Based on the first object information, from described first image, extract the first son comprising described first object
Image;
Synthesize described first subgraph and described second image, generate corresponding with described first virtual scene
3rd image.
In the embodiment of the present application, after obtaining the first image and the second image, then will be based on the first image
With the second image, generate threeth image corresponding with the first virtual scene.During implementing, the
One virtual scene is specially adjusted rear first virtual scene or for through spliced virtual scene.
In the embodiment of the present application, owing to the first image comprising the first object information.Therefore, concrete real
During Xian, the first son comprising the first object can be extracted from the first image according to the first object information
Image.During implementing, if the first object is personage, then by the first object information, including but
It is not limited to the characteristic information points such as the glasses of personage, eyebrow, maxilla, chin, ear, carries from the first image
Take out these pixels as the first object subgraph.
In the embodiment of the present application, after obtaining the first subgraph, then merge the first subgraph and the second image,
The most after merging, then also needing to analyze further in merging process, the first subgraph is at the second image
In position, during implementing, can based on according to user selection operate place, such as:
Detect that user first area on the second image carries out for determining that the first subgraph is at the second image
On the operation of position time, then the first subgraph is placed on the area;Or according to preset rules, such as:
First subgraph is placed on the second image blank region;Or determine that the first subgraph exists otherwise
Position on second image, those of ordinary skill in the art can select according to actual needs, in this Shen
Embodiment please be not especially limited.
Further, after the first subgraph and the second image are synthesized, also to further determine that, the first subgraph
Whether the size of picture mates with the current size of the second image, if coupling, then without adjusting;If not mating,
As: the size of the first image is more than the size of the second image or the oversize of the second image, the most then need
Adjust the first subgraph and/or the size of the second image, concrete: according to preset ratio amplify, reduce and/or
Cutting etc..After adjustment terminates, then generate threeth image corresponding with the first virtual scene.
In the embodiment of the present application, the 3rd image is specifically as follows two dimensional image, 3-D view or other dimension
Image, be not especially limited in the embodiment of the present application.
In the embodiment of the present application, after execution of step S103, described method also comprises the steps:
Obtain the first display parameters and second display parameters of described second image of described first subgraph;
If described first display parameters do not mate with described second display parameters, then adjust described first display ginseng
Several and/or described second display parameters so that the second display ginseng after the first display parameters after adjustment and adjustment
Number matches.
In the embodiment of the present application, described first display parameters and described second display parameters include brightness of image,
At least one in saturation degree or transparency.
During implementing, the first display parameters and the second display parameters include but do not limit above-mentioned image
Brightness, saturation degree or transparency, it is also possible to include colour temperature, tone, resolution ratio etc., or other display ginseng
Number, here, schematically illustrate the most one by one.
In the embodiment of the present application, the first display parameters and the second display parameters are specifically as a example by brightness, to this
Application embodiment illustrates, during implementing, specifically includes following steps:
Obtain the first brightness and second brightness of described second image of described first subgraph;
If described first brightness and described second brightness are not mated, then adjust described first brightness and/or described
Two brightness so that the first brightness after adjustment and the second brightness after adjustment match.
During implementing, owing to the first subgraph and the second image are not to obtain under Same Scene
Image, therefore, the brightness of the first subgraph and the second image there may exist difference, therefore, by
After one subgraph and the synthesis of the second image, then to obtain the first brightness of the first subgraph, such as: 20cd/m2(bank
De La every square metre), the brightness of the second image, such as: 30cd/m2。
After obtaining the brightness of the first subgraph and the second image, first determine whether the first subgraph and the second image
Brightness whether mate, i.e. judge that the luminance difference of the first subgraph and the second subgraph is whether more than presetting model
Enclose, such as: 3cd/m2、5cd/m2Or 6cd/m2, those of ordinary skill in the art can be according to actual needs
It is set, is not especially limited in the embodiment of the present application.
During implementing, if the first brightness of the first subgraph is 20cd/m2(candela every square
Rice), the brightness of the second image is 30cd/m2, and predetermined luminance difference value is 5cd/m2For, then luminance difference
10cd/m2A value poor more than predetermined luminance.So in order to match between the first brightness and the second brightness,
The first brightness can be adjusted, as being adjusted to 30cd/m during implementing2So that it is equal with the second brightness
Weighing apparatus;Or adjust the second brightness, be reduced to 20cd/m2So that it is balance each other with the first brightness;Or by first
Brightness rises to 25cd/m2, it is 25cd/m by the second luminance-reduction simultaneously2So that the first brightness after adjustment
Balance each other with the second brightness after adjusting.
The technical scheme provided by the embodiment of the present application, can not only make the image that synthesis needs, additionally it is possible to
Further the image after synthesis is processed, to better meet the visual effect of user, and then reach to carry
The technique effect of high user experience.
Embodiment two
Refer to Fig. 3, based on same inventive concept, the embodiment of the present application two provides a kind of electronic equipment, bag
Include:
Housing 30;
Processing means 31, is arranged in described housing, obtains the first image of current environment residing for the first object,
Described first image includes first object information corresponding with described first object;Obtain and the first virtual scene
The second corresponding image, described second image includes at least one virtual objects;And based on described first figure
Picture and described second image, generate threeth image corresponding with described first virtual scene, wherein, described
3rd image includes described first object information.
In the embodiment of the present application, processing means 31 is specifically as follows a processor or two processors, its
In, when have have two processors time, such as: CPU1 and CPU2, CPU1 be exclusively used in process based on first
Current environment residing for object, it is thus achieved that the first image, described first image includes corresponding with described first object
The first object information;Obtaining second image corresponding with the first virtual scene, described second image includes
At least one virtual objects;CPU2 is exclusively used in process based on described first image and described second image, generation
Threeth image corresponding with described first virtual scene, wherein, described 3rd image includes described first right
Image information.Or being allocated otherwise, those of ordinary skill in the art can be according to actual needs
It is set, is not especially limited in the embodiment of the present application.
Optionally, described processing means 31 is used for:
Obtain second virtual scene different from described first virtual scene;
Adjust at least one virtual objects in described first virtual scene and described second void in the same scale
Intend at least one virtual objects in scene, it is thus achieved that the first adjusted virtual scene and the adjusted second void
Intend scene;
Splice described the first adjusted virtual scene and described the second adjusted virtual scene, it is thus achieved that the 3rd
Virtual scene;
Obtain second image corresponding with described 3rd virtual scene.
Optionally, described processing means 31 is used for:
Based on described first object information, extract from described first image and comprise the of described first object
One subgraph;
Synthesize described first subgraph and described second image, generate corresponding with described first virtual scene
3rd image.
Optionally, after described first subgraph of described synthesis and described second image, described processing means
31 are additionally operable to:
Obtain the first display parameters and second display parameters of described second image of described first subgraph;
If described first display parameters do not mate with described second display parameters, then adjust described first display ginseng
Several and/or described second display parameters so that the second display ginseng after the first display parameters after adjustment and adjustment
Number matches.
Optionally, described first display parameters and described second display parameters include brightness of image, saturation degree or
At least one in transparency.
Embodiment three
Refer to Fig. 4, based on same inventive concept, the embodiment of the present application provides a kind of electronic equipment, including:
First acquiring unit 40, for obtaining acquisition first image of current environment residing for the first object, described
First image includes first object information corresponding with described first object;
Second acquisition unit 41, for obtaining second image corresponding with the first virtual scene, described second
Image includes at least one virtual objects;
First signal generating unit 42, for based on described first image and described second image, generates and described the
The 3rd image that one virtual scene is corresponding, wherein, described 3rd image includes described first object information.
Optionally, described second acquisition unit 41, including:
First acquisition module, for obtaining second virtual scene different from described first virtual scene;
First adjusting module, empty for adjusting at least one in described first virtual scene in the same scale
Intend at least one virtual objects in object and described second virtual scene, it is thus achieved that the first adjusted virtual field
Scape and the second adjusted virtual scene;
First concatenation module, for splicing described the first adjusted virtual scene and described adjusted second
Virtual scene, it is thus achieved that the 3rd virtual scene;
Obtain second image corresponding with described 3rd virtual scene.
Optionally, described first signal generating unit 42, including:
First extraction module, for based on described first object information, extracts bag from described first image
The first subgraph containing described first object;
First synthesis module, is used for synthesizing described first subgraph and described second image, generates and described the
The 3rd image that one virtual scene is corresponding.
Optionally, after described first subgraph of described synthesis and described second image, described electronic equipment
Also include:
3rd acquiring unit, for obtaining the first display parameters of described first subgraph and described second image
The second display parameters;
First adjustment unit, if not mating, then with described second display parameters for described first display parameters
Adjust described first display parameters and/or described second display parameters so that the first display parameters after adjustment and
The second display parameters after adjustment match.
By the one or more technical schemes in the embodiment of the present application, it is possible to achieve following one or more skills
Art effect:
One, the technical scheme provided due to the embodiment of the present application, obtains the of current environment residing for the first object
One image, described first image includes first object information corresponding with described first object;Obtain and first
The second image that virtual scene is corresponding, described second image includes at least one virtual objects;Based on described
First image and described second image, generate threeth image corresponding with described first virtual scene, wherein,
Described 3rd image includes described first object information.I.e. will not be as, in prior art, user is to reality scene
In object when taking pictures, the picture that this object is in reality scene, such as user can only be obtained and use
Mobile phone is autodyned, and can only obtain the user oneself picture with true environment as background, and cannot be by user's
Autodyning and combine with abundant virtual scene, Consumer's Experience is more single.And in the technical program, it is possible to
According to user's image in reality scene and generate corresponding with virtual scene with the image in virtual scene
3rd image, obtains the background image more horn of plenty of image, to reach to improve the technique effect of user experience.
Two, due to the technical scheme in the embodiment of the present application, different from described first virtual scene is obtained
Two virtual scenes;Adjust at least one virtual objects in described first virtual scene and institute in the same scale
State at least one virtual objects in the second virtual scene, it is thus achieved that the first adjusted virtual scene and adjusted
The second virtual scene;Splice described the first adjusted virtual scene and described the second adjusted virtual field
Scape, it is thus achieved that the 3rd virtual scene;Obtain second image corresponding with described 3rd virtual scene.I.e. in this skill
In art scheme, current virtual scene and previously saved virtual scene can be spliced, more to obtain
The image of the virtual scene of form, and then reach to improve the technique effect of user experience.
Three, due to the technical scheme in the embodiment of the present application, the first display ginseng of described first subgraph is obtained
Several and the second display parameters of described second image;If described first display parameters and described second display parameters
Do not mate, then adjust described first display parameters and/or described second display parameters so that first after adjustment
The second display parameters after display parameters and adjustment match.I.e. in the technical program, at synthesis the first son
After image and the second image, also the display parameters of the first subgraph and/or the second image are adjusted,
So that first subgraph the second image after He Cheng more meets the visual effect of user, and then reach to improve user
The technique effect of Experience Degree.
Those skilled in the art are it should be appreciated that embodiments of the invention can be provided as method, system or meter
Calculation machine program product.Therefore, the present invention can use complete hardware embodiment, complete software implementation or knot
The form of the embodiment in terms of conjunction software and hardware.And, the present invention can use and wherein wrap one or more
Computer-usable storage medium containing computer usable program code (include but not limited to magnetic disc store,
CD-ROM, optical memory etc.) form of the upper computer program implemented.
The present invention is with reference to method, equipment (system) and computer program according to embodiments of the present invention
Flow chart and/or block diagram describe.It should be understood that can by computer program instructions flowchart and/
Or flow process in each flow process in block diagram and/or square frame and flow chart and/or block diagram and/or
The combination of square frame.These computer program instructions can be provided to all-purpose computer, special-purpose computer, embedded
The processor of processor or other programmable data processing device is to produce a machine so that pass through computer
Or the instruction that the processor of other programmable data processing device performs produces for realizing at flow chart one stream
The device of the function specified in journey or multiple flow process and/or one square frame of block diagram or multiple square frame.
These computer program instructions may be alternatively stored in and can guide computer or other programmable data processing device
In the computer-readable memory worked in a specific way so that be stored in this computer-readable memory
Instruction produces the manufacture including command device, and this command device realizes at one flow process of flow chart or multiple stream
The function specified in journey and/or one square frame of block diagram or multiple square frame.
These computer program instructions also can be loaded in computer or other programmable data processing device, makes
Sequence of operations step must be performed to produce computer implemented place on computer or other programmable devices
Reason, thus the instruction performed on computer or other programmable devices provides for realizing flow chart one
The step of the function specified in flow process or multiple flow process and/or one square frame of block diagram or multiple square frame.
Specifically, the computer program instructions that the information processing method in the embodiment of the present application is corresponding can be by
It is stored in CD, hard disk, on the storage medium such as USB flash disk, corresponding with information processing method when in storage medium
Computer program instructions read or when being performed by an electronic equipment, comprise the steps:
Obtaining the first image of current environment residing for the first object, described first image includes first right with described
As the first corresponding object information;
Obtaining second image corresponding with the first virtual scene, described second image includes that at least one is virtual
Object;
Based on described first image and described second image, generate corresponding with described first virtual scene
Three images, wherein, described 3rd image includes described first object information.
Optionally, in described storage medium storage with step: described acquisition is corresponding with the first virtual scene
The second image, corresponding computer instruction, during being specifically performed, specifically includes following steps:
Obtain second virtual scene different from described first virtual scene;
Adjust at least one virtual objects in described first virtual scene and described second void in the same scale
Intend at least one virtual objects in scene, it is thus achieved that the first adjusted virtual scene and the adjusted second void
Intend scene;
Splice described the first adjusted virtual scene and described the second adjusted virtual scene, it is thus achieved that the 3rd
Virtual scene;
Obtain second image corresponding with described 3rd virtual scene.
Optionally, in described storage medium storage and step: described based on described first image with described the
Two images, generate threeth image corresponding with the first virtual scene, and corresponding computer instruction is at specifically quilt
During execution, specifically include following steps:
Based on described first object information, extract from described first image and comprise the of described first object
One subgraph;
Synthesize described first subgraph and described second image, generate corresponding with described first virtual scene
3rd image.
Optionally, in described storage medium, also storage has other computer instruction, and this other calculates
Machine instructs with step: the computer corresponding at described first subgraph of described synthesis and described second image refers to
Order is performed and is performed afterwards, and this other computer instruction, during being specifically performed, specifically includes
Following steps:
Obtain the first display parameters and second display parameters of described second image of described first subgraph;
If described first display parameters do not mate with described second display parameters, then adjust described first display ginseng
Several and/or described second display parameters so that the second display ginseng after the first display parameters after adjustment and adjustment
Number matches.
Although preferred embodiments of the present invention have been described, but those skilled in the art once know base
This creativeness concept, then can make other change and amendment to these embodiments.So, appended right is wanted
Ask and be intended to be construed to include preferred embodiment and fall into all changes and the amendment of the scope of the invention.
Obviously, those skilled in the art can carry out various change and modification without deviating from this to the present invention
Bright spirit and scope.So, if the present invention these amendment and modification belong to the claims in the present invention and
Within the scope of its equivalent technologies, then the present invention is also intended to comprise these change and modification.
Claims (11)
1. an information processing method, including:
Obtaining the first image of current environment residing for the first object, described first image includes first right with described
As the first corresponding object information;
Obtaining second image corresponding with the first virtual scene, described second image includes that at least one is virtual
Object;
Based on described first image and described second image, generate corresponding with described first virtual scene
Three images, wherein, described 3rd image includes described first object information.
2. the method for claim 1, it is characterised in that described acquisition is relative with the first virtual scene
The second image answered, including:
Obtain second virtual scene different from described first virtual scene;
Adjust at least one virtual objects in described first virtual scene and described second void in the same scale
Intend at least one virtual objects in scene, it is thus achieved that the first adjusted virtual scene and the adjusted second void
Intend scene;
Splice described the first adjusted virtual scene and described the second adjusted virtual scene, it is thus achieved that the 3rd
Virtual scene;
Obtain second image corresponding with described 3rd virtual scene.
3. method as claimed in claim 1 or 2, it is characterised in that described based on described first image and
Described second image, generates threeth image corresponding with the first virtual scene, including:
Based on described first object information, extract from described first image and comprise the of described first object
One subgraph;
Synthesize described first subgraph and described second image, generate corresponding with described first virtual scene
3rd image.
4. method as claimed in claim 3, it is characterised in that at described first subgraph of described synthesis and
After described second image, described method also includes:
Obtain the first display parameters and second display parameters of described second image of described first subgraph;
If described first display parameters do not mate with described second display parameters, then adjust described first display ginseng
Several and/or described second display parameters so that the second display ginseng after the first display parameters after adjustment and adjustment
Number matches.
5. method as claimed in claim 4, it is characterised in that described first display parameters and described the
Two display parameters include at least one in brightness of image, saturation degree or transparency.
6. an electronic equipment, including:
Housing;
Processing means, is arranged in described housing, for obtaining the first figure of current environment residing for the first object
Picture, described first image includes first object information corresponding with described first object;Obtain virtual with first
The second image that scene is corresponding, described second image includes at least one virtual objects;Based on described first
Image and described second image, generate threeth image corresponding with described first virtual scene, wherein, institute
State the 3rd image and include described first object information.
7. electronic equipment as claimed in claim 6, it is characterised in that described processing means is used for:
Obtain second virtual scene different from described first virtual scene;
Adjust at least one virtual objects in described first virtual scene and described second void in the same scale
Intend at least one virtual objects in scene, it is thus achieved that the first adjusted virtual scene and the adjusted second void
Intend scene;
Splice described the first adjusted virtual scene and described the second adjusted virtual scene, it is thus achieved that the 3rd
Virtual scene;
Obtain second image corresponding with described 3rd virtual scene.
Electronic equipment the most as claimed in claims 6 or 7, it is characterised in that described processing means is used for:
Based on described first object information, extract from described first image and comprise the of described first object
One subgraph;
Synthesize described first subgraph and described second image, generate corresponding with described first virtual scene
3rd image.
9. electronic equipment as claimed in claim 8, it is characterised in that at described first subgraph of described synthesis
After picture and described second image, described processing means is additionally operable to:
Obtain the first display parameters and second display parameters of described second image of described first subgraph;
If described first display parameters do not mate with described second display parameters, then adjust described first display ginseng
Several and/or described second display parameters so that the second display ginseng after the first display parameters after adjustment and adjustment
Number matches.
10. electronic equipment as claimed in claim 9, it is characterised in that described first display parameters and institute
State the second display parameters and include at least one in brightness of image, saturation degree or transparency.
11. 1 kinds of electronic equipments, including:
First acquiring unit, for obtaining the first image of current environment, described first figure residing for the first object
As including first object information corresponding with described first object;
Second acquisition unit, for obtaining second image corresponding with the first virtual scene, described second figure
As including at least one virtual objects;
First signal generating unit, for based on described first image and described second image, generates and described first
The 3rd image that virtual scene is corresponding, wherein, described 3rd image includes described first object information.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201610173581.8A CN105867615A (en) | 2016-03-24 | 2016-03-24 | Information processing method and electronic device |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201610173581.8A CN105867615A (en) | 2016-03-24 | 2016-03-24 | Information processing method and electronic device |
Publications (1)
Publication Number | Publication Date |
---|---|
CN105867615A true CN105867615A (en) | 2016-08-17 |
Family
ID=56625666
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201610173581.8A Pending CN105867615A (en) | 2016-03-24 | 2016-03-24 | Information processing method and electronic device |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN105867615A (en) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108958460A (en) * | 2017-05-19 | 2018-12-07 | 深圳市掌网科技股份有限公司 | Building sand table methods of exhibiting and system based on virtual reality |
CN109829964A (en) * | 2019-02-11 | 2019-05-31 | 北京邮电大学 | The rendering method and device of Web augmented reality |
CN110933393A (en) * | 2019-12-09 | 2020-03-27 | 中国人民解放军陆军装甲兵学院 | Parallax image sequence synthesis method and system for holographic stereogram printing |
Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101668126A (en) * | 2008-09-02 | 2010-03-10 | 新奥特(北京)视频技术有限公司 | Method for realizing unlimited blue-box function used in virtual studio systems |
CN102098452A (en) * | 2009-12-09 | 2011-06-15 | 北京瑞诚时代投资顾问有限公司 | Method for manufacturing remote controller-controlled interactive visual photographic system |
CN102110298A (en) * | 2009-12-25 | 2011-06-29 | 新奥特(北京)视频技术有限公司 | Method and device for projecting three-dimensional model in virtual studio system |
CN102118574A (en) * | 2009-12-30 | 2011-07-06 | 新奥特(北京)视频技术有限公司 | Method for sports event live broadcast |
CN102118568A (en) * | 2009-12-30 | 2011-07-06 | 新奥特(北京)视频技术有限公司 | Graphics generation system for sports competitions |
CN103701772A (en) * | 2013-11-26 | 2014-04-02 | 北京航空航天大学 | Method for constructing digital media content resource cloud system |
CN104537716A (en) * | 2015-01-20 | 2015-04-22 | 湖南化身科技有限公司 | System for synthesizing three-dimensional digital human image and virtual scene |
-
2016
- 2016-03-24 CN CN201610173581.8A patent/CN105867615A/en active Pending
Patent Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101668126A (en) * | 2008-09-02 | 2010-03-10 | 新奥特(北京)视频技术有限公司 | Method for realizing unlimited blue-box function used in virtual studio systems |
CN102098452A (en) * | 2009-12-09 | 2011-06-15 | 北京瑞诚时代投资顾问有限公司 | Method for manufacturing remote controller-controlled interactive visual photographic system |
CN102110298A (en) * | 2009-12-25 | 2011-06-29 | 新奥特(北京)视频技术有限公司 | Method and device for projecting three-dimensional model in virtual studio system |
CN102118574A (en) * | 2009-12-30 | 2011-07-06 | 新奥特(北京)视频技术有限公司 | Method for sports event live broadcast |
CN102118568A (en) * | 2009-12-30 | 2011-07-06 | 新奥特(北京)视频技术有限公司 | Graphics generation system for sports competitions |
CN103701772A (en) * | 2013-11-26 | 2014-04-02 | 北京航空航天大学 | Method for constructing digital media content resource cloud system |
CN104537716A (en) * | 2015-01-20 | 2015-04-22 | 湖南化身科技有限公司 | System for synthesizing three-dimensional digital human image and virtual scene |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108958460A (en) * | 2017-05-19 | 2018-12-07 | 深圳市掌网科技股份有限公司 | Building sand table methods of exhibiting and system based on virtual reality |
CN109829964A (en) * | 2019-02-11 | 2019-05-31 | 北京邮电大学 | The rendering method and device of Web augmented reality |
CN109829964B (en) * | 2019-02-11 | 2021-08-13 | 北京邮电大学 | Web augmented reality rendering method and device |
CN110933393A (en) * | 2019-12-09 | 2020-03-27 | 中国人民解放军陆军装甲兵学院 | Parallax image sequence synthesis method and system for holographic stereogram printing |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN102812497B (en) | The image experiencing image subsequently can be provided to provide device, image to provide method | |
CN102834849B (en) | Carry out the image displaying device of the description of three-dimensional view picture, image drawing method, image depiction program | |
CN106598229B (en) | Virtual reality scene generation method and device and virtual reality system | |
US20180332222A1 (en) | Method and apparatus for obtaining binocular panoramic image, and storage medium | |
CN106355153A (en) | Virtual object display method, device and system based on augmented reality | |
KR20140108128A (en) | Method and apparatus for providing augmented reality | |
CN108668050B (en) | Video shooting method and device based on virtual reality | |
JP6126271B1 (en) | Method, program, and recording medium for providing virtual space | |
CN109598796A (en) | Real scene is subjected to the method and apparatus that 3D merges display with dummy object | |
CN106296789B (en) | It is a kind of to be virtually implanted the method and terminal that object shuttles in outdoor scene | |
CN106648098A (en) | User-defined scene AR projection method and system | |
CN106683193A (en) | Three-dimensional model design method and design device | |
CN108043027A (en) | Storage medium, electronic device, the display methods of game picture and device | |
CN105867615A (en) | Information processing method and electronic device | |
KR100720718B1 (en) | Apparatus and method for augmented reality | |
CN113253842A (en) | Scene editing method and related device and equipment | |
US11579746B2 (en) | Dynamic image capturing apparatus and method using arbitrary viewpoint image generation technology | |
CN112230765A (en) | AR display method, AR display device, and computer-readable storage medium | |
EP3346375B1 (en) | Program, recording medium, content provision device, and control method | |
JP4892405B2 (en) | Image processing apparatus and method | |
CN112511815B (en) | Image or video generation method and device | |
CN106200973A (en) | A kind of method and device playing virtual reality file based on external image | |
US11961190B2 (en) | Content distribution system, content distribution method, and content distribution program | |
US20230244354A1 (en) | 3d models for displayed 2d elements | |
JP2018198025A (en) | Image processing device, image processing device control method, and program |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20160817 |
|
RJ01 | Rejection of invention patent application after publication |