CN108905208A - A kind of electronic gaming method and device based on augmented reality - Google Patents
A kind of electronic gaming method and device based on augmented reality Download PDFInfo
- Publication number
- CN108905208A CN108905208A CN201810640651.5A CN201810640651A CN108905208A CN 108905208 A CN108905208 A CN 108905208A CN 201810640651 A CN201810640651 A CN 201810640651A CN 108905208 A CN108905208 A CN 108905208A
- Authority
- CN
- China
- Prior art keywords
- augmented reality
- reality model
- picture
- current scene
- depth
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Classifications
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F13/00—Video games, i.e. games using an electronically generated display having two or more dimensions
- A63F13/60—Generating or modifying game content before or while executing the game program, e.g. authoring tools specially adapted for game development or game-integrated level editor
- A63F13/65—Generating or modifying game content before or while executing the game program, e.g. authoring tools specially adapted for game development or game-integrated level editor automatically by game devices or servers from real world data, e.g. measurement in live racing competition
- A63F13/655—Generating or modifying game content before or while executing the game program, e.g. authoring tools specially adapted for game development or game-integrated level editor automatically by game devices or servers from real world data, e.g. measurement in live racing competition by importing photos, e.g. of the player
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F13/00—Video games, i.e. games using an electronically generated display having two or more dimensions
- A63F13/50—Controlling the output signals based on the game progress
- A63F13/52—Controlling the output signals based on the game progress involving aspects of the displayed game scene
Abstract
The present invention proposes that a kind of electronic gaming method based on augmented reality includes the following steps for the equipment with image pick-up device and display:The image data of current scene is obtained, and calculates the depth of field data of object in current environment according to image data;Augmented reality model is imported, and based on image data positioning augmented reality model data at current scene position, wherein the position data includes at least augmented reality model in the coordinate position of current scene and towards angle;According to the position data of the depth of field data of current scene and augmented reality model, the transition matrix of augmented reality model is calculated in real time;Augmented reality model is superimposed to current display picture by transition matrix.The present invention also proposes a kind of electronic game computer based on augmented reality.Beneficial effects of the present invention are:By calculating the depth of field data of current environment and the transition matrix of augmented reality model in real time, the object in electronic game is accurately added to and is shown in equipment.
Description
Technical field
The present invention relates to Video Composition technical field more particularly to a kind of electronic gaming methods and dress based on augmented reality
It sets.
Background technique
Electronic video games are as current popular entertainment way, by the way that illusory magnificent game generation is presented over the display
Boundary and game player is immersed wherein, to obtain good game experiencing.However, even if gaming world be again it is true to nature,
It makes it difficult for game player to obtain the complete sense of reality, it is made to be immersed in gaming world wherein.
Augmented reality (Augmented Reality, abbreviation AR) is a kind of by real world information and virtual world information
" seamless " integrated new technology.It is by the way that virtual threedimensional model to be added to the mode in display environment, originally in reality
Can not be experienced in the world it is true bring game player with the illusory feeling of immersion combined, to reach the body of exceeding reality
It tests.True environment is added to the emerging experience mode in the same space by game player in real time to this with illusory object
The enthusiasm of performance processing height.According to E3 sponsor of the U.S., it is expected that in the future being expected, the year market of reality game will reach
To 10,000,000,000 dollars.
However at present augmented reality due to the three-dimensional animation technology cycle of synthesis it is long, to avoid the augmented reality mould after synthesis
Type seems lofty, so that the electronic game of augmented reality is mostly simple trivial games at present, content is relatively dull, and there is no sufficiently
Embody the effect of augmented reality.
Summary of the invention
The purpose of the application is to solve the deficiencies in the prior art, provide a kind of electronic gaming method based on augmented reality and
Device, to solve the above technical problems, game player will be enabled through augmented reality equipment with can be convenient in gaming, it is seen that from
There is the object in electronic game in true environment (in such as bedroom or parlor) belonging to oneself, thus generates the strong sense of reality and pressure
Compel sense.
To achieve the goals above, the following technical solution is employed by the present invention:
Firstly, the application proposes a kind of electronic gaming method based on augmented reality, for having image pick-up device and display
Equipment on, include the following steps:The image data of current scene is obtained, and object in current environment is calculated according to image data
Depth of field data;Augmented reality model is imported, and is counted at current scene position based on image data positioning augmented reality model
According to wherein above-mentioned position data includes at least augmented reality model in the coordinate position of current scene and towards angle;According to working as
The depth of field data of preceding scene and the position data of augmented reality model calculate the transition matrix of augmented reality model in real time;Pass through
Augmented reality model is superimposed to current display picture by transition matrix.
Further, above-mentioned image data includes at least the shooting picture of multiple current scenes shot from different perspectives.
Still further, the calculating of the depth of field data of each object is also wrapped in current environment in the above method of the application
Include following sub-step:Corresponding object in multiple shooting pictures is marked, and it is shared in each shooting picture to calculate each object
Some number of pixels;The number of pixels having shared by picture is shot at each according to each object, extracts main object in current scene
Body;According to primary objects in each shooting picture position difference, calculate current scene in primary objects depth of field data.
Further, above-mentioned image data includes at least the shooting picture and corresponding flight time picture of current scene.
Still further, the calculating of the depth of field data of each object is also wrapped in current environment in the above method of the application
Include following sub-step:Corresponding object in multiple shooting pictures is marked, and it is shared in each shooting picture to calculate each object
Some number of pixels;The number of pixels having shared by picture is shot at each according to each object, extracts main object in current scene
Body;According to primary objects in each shooting picture position difference, calculate current scene in primary objects depth of field data.
Further, augmented reality model being shown to, being superimposed to current display picture further includes following sub-step:According to working as
Data, calculating augmented reality model are hidden at current scene position for the depth of field data of object and augmented reality model in preceding environment
Stopper point;It is blocked partial correction augmented reality model according to augmented reality model, and the augmented reality model after correction is led to
It crosses transition matrix and is superimposed to current display picture.
Secondly, the application also proposes a kind of electronic game computer based on augmented reality, for having image pick-up device and display
In the equipment of device, comprise the following modules:Module is obtained, for obtaining the image data of current scene, and according to image data meter
Calculate the depth of field data of each object in current environment;Import modul is positioned for importing augmented reality model, and based on image data
Augmented reality model data at current scene position, wherein the position data, which includes at least augmented reality model, is working as front court
The coordinate position of scape and towards angle;Conversion module, for according to the depth of field data of current scene and the position of augmented reality model
Data are set, calculate the transition matrix of augmented reality model in real time;Display module, for passing through transition matrix for augmented reality model
Display is superimposed to current display picture.
Further, above-mentioned image data includes at least the shooting picture of multiple current scenes shot from different perspectives.
Still further, obtaining module further includes following submodule in the above-mentioned apparatus of the application:First label mould
Block for marking corresponding object in multiple shooting pictures, and calculates each object and shoots the picture occupied in picture at each
Plain number;First extraction module, for shooting the number of pixels having shared by picture at each according to each object, front court is worked as in extraction
Primary objects in scape;First depth of field module, for front court to be worked as in position difference, calculating in each shooting picture according to primary objects
The depth of field data of primary objects in scape.
Further, above-mentioned image data includes at least the shooting picture and corresponding flight time picture of current scene.
Still further, obtaining module further includes following submodule in the above-mentioned apparatus of the application:Second label mould
Block for marking corresponding object in multiple shooting pictures, and calculates each object and shoots the picture occupied in picture at each
Plain number;Second extraction module, for shooting the number of pixels having shared by picture at each according to each object, front court is worked as in extraction
Primary objects in scape;Second depth of field module, for front court to be worked as in position difference, calculating in each shooting picture according to primary objects
The depth of field data of primary objects in scape.
Further, display module further includes following submodule:Occlusion test module, for according to object in current environment
Depth of field data and augmented reality model data at current scene position, calculate augmented reality model and be blocked part;It blocks
Correction module, for being blocked partial correction augmented reality model according to augmented reality model, and by the augmented reality after correction
Model is superimposed to current display picture by transition matrix.
Finally, it is stored thereon with computer instruction the invention also discloses a kind of computer readable storage medium, the instruction
It realizes when being executed by processor such as the step of aforementioned described in any item methods.
Beneficial effects of the present invention are:By the conversion for calculating the depth of field data and augmented reality model of current environment in real time
Matrix allows the object in electronic game to be accurately added to and shows in equipment, widen augmented reality in electronics
The scope of application of field of play.
Detailed description of the invention
Fig. 1 show the overview flow chart of the electronic gaming method disclosed in the present application based on augmented reality;
Fig. 2 show the method flow diagram for calculating the first embodiment of the depth of field data of each object in current environment;
Fig. 3 show the method flow diagram for calculating the second embodiment of the depth of field data of each object in current environment;
Fig. 4 show the sub-step method flow diagram of superposition augmented reality model to current display picture;
Fig. 5 show the function structure chart of the electronic game computer disclosed in the present application based on augmented reality.
Specific embodiment
It is only to be not intended to be limiting the disclosure merely for for the purpose of describing particular embodiments in the term that the disclosure uses.
The "an" of the singular used in disclosure and the accompanying claims book, " described " and "the" are also intended to including majority
Form, unless the context clearly indicates other meaning.It is also understood that term "and/or" used herein refers to and wraps
It may be combined containing one or more associated any or all of project listed.
It will be appreciated that though various modules, but this may be described using term first, second, third, etc. in the disclosure
A little modules should not necessarily be limited by these terms.These terms are only used to for same type of module being distinguished from each other out.For example, not departing from
In the case where disclosure range, the first module can also be referred to as the second module, and similarly, the second module can also be referred to as
One module.
Unless otherwise expressly stated, signified image pick-up device can be general shooting camera (such as intelligence hand herein
Video camera on machine) and/or time-of-flight camera (Time of Flight Camera).Correspondingly, shooting picture refers to use
General shooting camera shoots resulting photo.It is consistent with the current scene that game player's naked eyes are watched, and also works as with equipment
The picture of preceding display is consistent.Correspondingly, flight time picture refers to by using time-of-flight camera shooting, with current
The picture of the depth of field data of scene.Wherein the pixel value on flight time picture indicates the corresponding points in current scene to image pick-up device
Distance.
Fig. 1 show the overview flow chart of the electronic gaming method based on augmented reality.This method embodiment includes following
Step:The image data of current scene is obtained, and calculates the depth of field data of object in current environment according to image data;It imports and increases
Strong real model, and augmented reality model data at current scene position are positioned based on image data, wherein the positional number
According to including at least augmented reality model in the coordinate position of current scene and towards angle;According to the depth of field data of current scene and
The position data of augmented reality model calculates the transition matrix of augmented reality model in real time;By transition matrix by augmented reality
Model is superimposed to current display picture.The wherein augmented reality model that the coordinate system based on model itself and position data determine
Transition matrix can according to the conventional algorithm of this field determine, the present invention to this not limit.
In one or more embodiments of the application, image data can be works as including what multiple shot from different perspectives
The shooting picture of preceding scene.For example, the shooting picture for the current scene that the image pick-up device of more different locations is shot simultaneously.Wherein,
Relative position between the image pick-up device of more different locations can be known before shooting.So as to according to corresponding object in difference
The difference of shooting figure on piece determines the depth of field data of object in current scene.This skill can be used in specific depth of field data calculating
The conventional algorithm in art field realizes that the present invention not limits this.
It further, is the operand for reducing system, some unessential objects in scene can be omitted, not as folded
Add augmented reality model to be added to and currently shows the foundation of picture.Specifically, referring to submethod flow chart shown in Fig. 2, at this
In one or more embodiments of application, the calculating of the depth of field data of each object further includes following sub-step in current environment:Mark
Remember corresponding object in multiple shooting pictures, and calculates each object and shoot the number of pixels occupied in picture at each;Root
The number of pixels having shared by picture is shot at each according to each object, extracts primary objects in current scene;According to primary objects
The position difference in each shooting picture calculates the depth of field data of primary objects in current scene.For example, when object is clapped at each
When taking the photograph the number of pixels occupied in picture and being both less than some preset threshold value, which is considered not being the master in current scene
Object is wanted, to not calculate its depth of field data, to reduce operand.It is same for the label for corresponding to object in multiple shooting pictures
It can be realized using the conventional algorithm of the art, the present invention not limits this.
Similarly, in one or more embodiments of the application, image data can also be the bat including current scene
Take the photograph picture and corresponding flight time picture.For example, shooting camera and time-of-flight camera that relative position has determined.Shooting
The depth of field data of object can be determined according to flight time photograph in current scene in picture.
It further, is the operand for reducing system, some unessential objects in scene can be omitted, not as folded
Add augmented reality model to be added to and currently shows the foundation of picture.Specifically, referring to submethod flow chart shown in Fig. 3, at this
In one or more embodiments of application, the calculating of the depth of field data of each object further includes following sub-step in current environment:Mark
Correspondence object in note shooting picture and flight time picture, and calculate the pixel that each object is occupied in shooting picture
Number;The number of pixels having shared by picture is being shot according to each object, is extracting primary objects in current scene;According to primary objects
Position in flight time picture calculates the depth of field data of primary objects.For example, occupied in shooting picture when object
When number of pixels is both less than some preset threshold value, which is considered not being the primary objects in current scene, to disregard
Its depth of field data is calculated, to reduce operand.
Further to promote the sense of reality that augmented reality model is superimposed to current picture, such as the object on some plant doorway
Body can be corrected augmented reality model, the part of augmented reality model is stashed more to meet current scene.Tool
Body, augmented reality model is shown in one or more embodiments of the application referring to submethod flow chart shown in Fig. 4
Show that being superimposed to current display picture further includes following sub-step:According to the depth of field data of object in current environment and augmented reality mould
Type data at current scene position calculate extension real model and are blocked part;It is blocked part according to augmented reality model
Augmented reality model is corrected, and the augmented reality model after correction is superimposed to current display picture by transition matrix.
Fig. 5 show the function structure chart of the electronic game computer based on augmented reality.The device comprises the following modules:It obtains
Modulus block, for obtaining the image data of current scene, and according to the depth of field number of each object in image data calculating current environment
According to;Import modul, for importing augmented reality model, and based on image data positioning augmented reality model in the position of current scene
Data are set, wherein the position data includes at least augmented reality model in the coordinate position of current scene and towards angle;Turn
Block is changed the mold, for calculating extend real mould in real time according to the depth of field data of current scene and the position data of augmented reality model
The transition matrix of type;Display module is superimposed to current display picture for showing augmented reality model by transition matrix.Its
In can be according to this field based on the transition matrix of augmented reality model that determines of coordinate system and position data of model itself
Conventional algorithm determine, the present invention to this not limit.
In one or more embodiments of the application, image data can be works as including what multiple shot from different perspectives
The shooting picture of preceding scene.For example, the shooting picture for the current scene that the image pick-up device of more different locations is shot simultaneously.Wherein,
Relative position between the image pick-up device of more different locations can be known before shooting.So as to according to corresponding object in difference
The difference of shooting figure on piece determines the depth of field data of object in current scene.This skill can be used in specific depth of field data calculating
The conventional algorithm in art field realizes that the present invention not limits this.
It further, is the operand for reducing system, some unessential objects in scene can be omitted, not as folded
Add augmented reality model to be added to and currently shows the foundation of picture.Specifically, it in one or more embodiments of the application, obtains
Modulus block further includes following submodule:First mark module for marking corresponding object in multiple shooting pictures, and calculates each
A object shoots the number of pixels occupied in picture at each;First extraction module, for being clapped according to each object at each
The number of pixels that picture is occupied is taken the photograph, primary objects in current scene are extracted;First depth of field module, for being existed according to primary objects
Position difference in each shooting picture calculates the depth of field data of primary objects in current scene.For example, when object is shot at each
When the number of pixels occupied in picture is both less than some preset threshold value, which is considered not being main in current scene
Object, to not calculate its depth of field data, to reduce operand.Label for corresponding to object in multiple shooting pictures equally may be used
To be realized using the conventional algorithm of the art, the present invention not limits this.
Similarly, in one or more embodiments of the application, image data can also be the bat including current scene
Take the photograph picture and corresponding flight time picture.For example, shooting camera and time-of-flight camera that relative position has determined.Shooting
The depth of field data of object can be determined according to flight time photograph in current scene in picture.
It further, is the operand for reducing system, some unessential objects in scene can be omitted, not as folded
Add augmented reality model to be added to and currently shows the foundation of picture.Specifically, it in one or more embodiments of the application, obtains
Modulus block may include following submodule:Second mark module, for marking the counterpart in shooting picture and flight time picture
Body, and calculate the number of pixels that each object is occupied in shooting picture;Second extraction module, for being existed according to each object
The number of pixels having shared by shooting picture, extracts primary objects in current scene;Second depth of field module, for according to primary objects
Position in flight time picture calculates the depth of field data of primary objects.For example, occupied in shooting picture when object
When number of pixels is both less than some preset threshold value, which is considered not being the primary objects in current scene, to disregard
Its depth of field data is calculated, to reduce operand.
Further to promote the sense of reality that augmented reality model is superimposed to current picture, such as the object on some plant doorway
Body can be corrected augmented reality model, the part of augmented reality model is stashed more to meet current scene.Tool
Body, in one or more embodiments of the application, display module further includes following submodule:Technology modules are blocked, are used for
According to the depth of field data of object in current environment and augmented reality model, data, calculating extend real mould at current scene position
Type is blocked part;Correction module is blocked, for being blocked partial correction augmented reality model according to augmented reality model, and will
Augmented reality model after correction is superimposed to current display picture by transition matrix.
It should be appreciated that the embodiment of the present invention can be by computer hardware, the combination of hardware and software or by depositing
The computer instruction in non-transitory computer-readable memory is stored up to be effected or carried out.Standard volume can be used in the above method
Journey technology-includes that the non-transitory computer-readable storage media configured with computer program is realized in computer program,
In configured in this way storage medium computer is operated in a manner of specific and is predefined --- according in a particular embodiment
The method and attached drawing of description.Each program can with the programming language of level process or object-oriented come realize with department of computer science
System communication.However, if desired, the program can be realized with compilation or machine language.Under any circumstance, which can be volume
The language translated or explained.In addition, the program can be run on the specific integrated circuit of programming for this purpose.
In addition, the operation of process described herein can be performed in any suitable order, unless herein in addition instruction or
Otherwise significantly with contradicted by context.Process described herein (or modification and/or combination thereof) can be held being configured with
It executes, and is can be used as jointly on the one or more processors under the control of one or more computer systems of row instruction
The code (for example, executable instruction, one or more computer program or one or more application) of execution, by hardware or its group
It closes to realize.Above-mentioned computer program includes the multiple instruction that can be performed by one or more processors.
Further, the above method can be realized in being operably coupled to suitable any kind of computing platform, wrap
Include but be not limited to PC, mini-computer, main frame, work station, network or distributed computing environment, individual or integrated
Computer platform or communicated with charged particle tool or other imaging devices etc..Each aspect of the present invention can be to deposit
The machine readable code on non-transitory storage medium or equipment is stored up to realize no matter be moveable or be integrated to calculating
Platform, such as hard disk, optical reading and/or write-in storage medium, RAM, ROM, so that it can be read by programmable calculator, when
Storage medium or equipment can be used for configuration and operation computer to execute process described herein when being read by computer.This
Outside, machine readable code, or part thereof can be transmitted by wired or wireless network.When such media include combining microprocessor
Or other data processors are when realizing the instruction or program of above-mentioned steps above, hereinbefore invention include these and other not
The non-transitory computer-readable storage media of same type.When methods and techniques above-mentioned according to the present invention program, the present invention
It further include computer itself.
Computer program can be applied to input data to execute hereinbefore function, to convert input data with life
At storing to the output data of nonvolatile memory.Output information can also be applied to one or more output equipments as shown
Device.In the preferred embodiment of the invention, the data of conversion indicate physics and tangible object, including the object generated on display
Reason and the particular visual of physical objects are described.
The above, only presently preferred embodiments of the present invention, the invention is not limited to above embodiment, as long as
It reaches technical effect of the invention with identical means, all within the spirits and principles of the present invention, any modification for being made,
Equivalent replacement, improvement etc., should be included within the scope of the present invention.Its technical solution within the scope of the present invention
And/or embodiment can have a variety of different modifications and variations.
Claims (8)
1. a kind of electronic gaming method based on augmented reality, for the equipment with image pick-up device and display, feature exists
In including the following steps:
The image data of current scene is obtained, and calculates the depth of field data of object in current environment according to image data;
Augmented reality model is imported, and augmented reality model data at current scene position are positioned based on image data, wherein
The position data includes at least augmented reality model in the coordinate position of current scene and towards angle;
According to the position data of the depth of field data of current scene and augmented reality model, the conversion of augmented reality model is calculated in real time
Matrix;
Augmented reality model is superimposed to current display picture by transition matrix.
2. being clapped from different perspectives the method according to claim 1, wherein the image data includes at least multiple
The shooting picture for the current scene taken the photograph.
3. according to the method described in claim 2, it is characterized in that, in current environment the calculating of the depth of field data of each object also wrap
Include following sub-step:
Corresponding object in multiple shooting pictures is marked, and calculates each object and shoots the pixel occupied in picture at each
Number;
The number of pixels having shared by picture is shot at each according to each object, extracts primary objects in current scene;
According to primary objects in each shooting picture position difference, calculate current scene in primary objects depth of field data.
4. the method according to claim 1, wherein the image data includes at least the shooting figure of current scene
Piece and corresponding flight time picture.
5. according to the method described in claim 4, it is characterized in that, in current environment the calculating of the depth of field data of each object also wrap
Include following sub-step:
Correspondence object in label shooting picture and flight time picture, and calculate what each object was occupied in shooting picture
Number of pixels;
The number of pixels having shared by picture is being shot according to each object, is extracting primary objects in current scene;
According to position of the primary objects in flight time picture, the depth of field data of primary objects is calculated.
6. being superimposed to current display picture the method according to claim 1, wherein augmented reality model is shown
It further include following sub-step:
According to the depth of field data of object in current environment and augmented reality model, data, calculating enhancing are existing at current scene position
Real mould is blocked part;
It is blocked partial correction augmented reality model according to augmented reality model, and by the augmented reality model after correction by turning
It changes matrix and is superimposed to current display picture.
7. a kind of electronic game computer based on augmented reality, for the equipment with image pick-up device and display, feature exists
In comprising the following modules:
Module is obtained, calculates each object in current environment for obtaining the image data of current scene, and according to image data
Depth of field data;
Import modul, for importing augmented reality model, and based on image data positioning augmented reality model in current scene
Position data, wherein the position data includes at least augmented reality model in the coordinate position of current scene and towards angle;
Conversion module, for calculating enhancing in real time according to the depth of field data of current scene and the position data of augmented reality model
The transition matrix of real model;
Display module is superimposed to current display picture for showing augmented reality model by transition matrix.
8. a kind of computer readable storage medium, is stored thereon with computer instruction, it is characterised in that the instruction is held by processor
It realizes when row such as the step of any one of claims 1 to 6 the method.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201810640651.5A CN108905208A (en) | 2018-06-21 | 2018-06-21 | A kind of electronic gaming method and device based on augmented reality |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201810640651.5A CN108905208A (en) | 2018-06-21 | 2018-06-21 | A kind of electronic gaming method and device based on augmented reality |
Publications (1)
Publication Number | Publication Date |
---|---|
CN108905208A true CN108905208A (en) | 2018-11-30 |
Family
ID=64420307
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201810640651.5A Pending CN108905208A (en) | 2018-06-21 | 2018-06-21 | A kind of electronic gaming method and device based on augmented reality |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN108905208A (en) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111429543A (en) * | 2020-02-28 | 2020-07-17 | 苏州叠纸网络科技股份有限公司 | Material generation method and device, electronic equipment and medium |
CN113838201A (en) * | 2021-09-23 | 2021-12-24 | 北京百度网讯科技有限公司 | Model adaptation method and device, electronic equipment and readable storage medium |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2950228B2 (en) * | 1996-02-15 | 1999-09-20 | 株式会社セガ・エンタープライゼス | Game image display method and game device |
CN101084672A (en) * | 2004-12-21 | 2007-12-05 | 索尼株式会社 | Image processing device, learning device, and coefficient generating device and method |
CN106548519A (en) * | 2016-11-04 | 2017-03-29 | 上海玄彩美科网络科技有限公司 | Augmented reality method based on ORB SLAM and the sense of reality of depth camera |
CN106683195A (en) * | 2016-12-30 | 2017-05-17 | 上海网罗电子科技有限公司 | AR scene rendering method based on indoor location |
CN107016704A (en) * | 2017-03-09 | 2017-08-04 | 杭州电子科技大学 | A kind of virtual reality implementation method based on augmented reality |
-
2018
- 2018-06-21 CN CN201810640651.5A patent/CN108905208A/en active Pending
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2950228B2 (en) * | 1996-02-15 | 1999-09-20 | 株式会社セガ・エンタープライゼス | Game image display method and game device |
CN101084672A (en) * | 2004-12-21 | 2007-12-05 | 索尼株式会社 | Image processing device, learning device, and coefficient generating device and method |
CN106548519A (en) * | 2016-11-04 | 2017-03-29 | 上海玄彩美科网络科技有限公司 | Augmented reality method based on ORB SLAM and the sense of reality of depth camera |
CN106683195A (en) * | 2016-12-30 | 2017-05-17 | 上海网罗电子科技有限公司 | AR scene rendering method based on indoor location |
CN107016704A (en) * | 2017-03-09 | 2017-08-04 | 杭州电子科技大学 | A kind of virtual reality implementation method based on augmented reality |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111429543A (en) * | 2020-02-28 | 2020-07-17 | 苏州叠纸网络科技股份有限公司 | Material generation method and device, electronic equipment and medium |
CN113838201A (en) * | 2021-09-23 | 2021-12-24 | 北京百度网讯科技有限公司 | Model adaptation method and device, electronic equipment and readable storage medium |
CN113838201B (en) * | 2021-09-23 | 2022-06-07 | 北京百度网讯科技有限公司 | Model adaptation method and device, electronic equipment and readable storage medium |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN112348969B (en) | Display method and device in augmented reality scene, electronic equipment and storage medium | |
WO2018119889A1 (en) | Three-dimensional scene positioning method and device | |
CN107308645B (en) | A kind of play has an X-rayed the method and game client of plug-in detection | |
CN102735100B (en) | Individual light weapon shooting training method and system by using augmented reality technology | |
CN104781849B (en) | Monocular vision positions the fast initialization with building figure (SLAM) simultaneously | |
CN107315470B (en) | Graphic processing method, processor and virtual reality system | |
KR101251701B1 (en) | Stereo video for gaming | |
KR101615719B1 (en) | Apparatus and method for extracting user's third dimension facial expression | |
CN111694430A (en) | AR scene picture presentation method and device, electronic equipment and storage medium | |
CN103136744B (en) | For the apparatus and method for the three-dimensional position for calculating characteristic point | |
CN105074776B (en) | Planar texture target is formed in situ | |
KR100490726B1 (en) | Apparatus and method for video based shooting game | |
JP5795250B2 (en) | Subject posture estimation device and video drawing device | |
CN109598796A (en) | Real scene is subjected to the method and apparatus that 3D merges display with dummy object | |
CN108230384A (en) | Picture depth computational methods, device, storage medium and electronic equipment | |
CN109035415B (en) | Virtual model processing method, device, equipment and computer readable storage medium | |
CN108043027B (en) | Storage medium, electronic device, game screen display method and device | |
CN106373182A (en) | Augmented reality-based human face interaction entertainment method | |
CN108325208A (en) | Augmented reality implementation method applied to field of play | |
CN112950751A (en) | Gesture action display method and device, storage medium and system | |
CN107862733A (en) | Large scale scene real-time three-dimensional method for reconstructing and system based on sight more new algorithm | |
CN108905208A (en) | A kind of electronic gaming method and device based on augmented reality | |
CN108986232A (en) | A method of it is shown in VR and AR environment picture is presented in equipment | |
CN110460794A (en) | A kind of method and device of video record | |
CN109847360B (en) | 3D effect processing method and device of game props, electronic equipment and medium |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20181130 |