CN103268589A - Somatosensory interaction method based on front-facing camera of mobile terminal - Google Patents
Somatosensory interaction method based on front-facing camera of mobile terminal Download PDFInfo
- Publication number
- CN103268589A CN103268589A CN2013101738210A CN201310173821A CN103268589A CN 103268589 A CN103268589 A CN 103268589A CN 2013101738210 A CN2013101738210 A CN 2013101738210A CN 201310173821 A CN201310173821 A CN 201310173821A CN 103268589 A CN103268589 A CN 103268589A
- Authority
- CN
- China
- Prior art keywords
- amplitude
- translation
- variation
- rotation
- optical flow
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Images
Landscapes
- Image Analysis (AREA)
Abstract
The invention discloses a somatosensory interaction method based on a front-facing camera of a mobile terminal. The method includes the following steps: (1) acquiring a current scene image, namely a video sequence, through the front-facing camera, (2) judging whether the acquired current scene image is a first-frame image, and if yes, returning the process to carry out the step (1), (3) carrying out optical flow field analysis on image data of a current frame and image data of a previous frame to acquire optical flow variation between the two frames and coordinates of corresponding points, (4) selecting a coordinate set of the points with obvious optical flow variation between the two frames, and calculating the variation tendency and the variation amplitude of scenes between the frames as the tendency and the amplitude of translation, rotation and scaling correspondingly, and (5) carrying out the step (1) to the step (4) circularly, calculating tendency data and amplitude data of the translation, the rotation and the scaling, and providing the data to an application program. The somatosensory interaction method based on the front-facing camera of the mobile terminal converts the variation of the scenes into the motions of translation, rotation, scaling and the like, and is more universal.
Description
Technical field
The present invention relates to intelligent movable equipment application, particularly a kind of body sense exchange method based on the preposition camera of mobile phone.
Background technology
On mobile terminal devices such as mobile phone, with mobile phone alternately mainly by touch-screen and physical keyboard.More and more higher along with handset capability, application program is more and more abundanter, and the diversity of interactive mode is also had higher requirement.
The preposition camera of portable terminals such as mobile phone can obtain abundant scene information, but existing application to preposition camera is main still in purposes such as video calling, auto heterodyne, recognitions of face.
Utilize methods such as the detection of people's face, gestures detection to follow the trail of the variation of scene, though have higher accuracy, real-time and applicability are very poor, lose efficacy easily under complex scene.
Under general common scene, the video of preposition camera is carried out the detection and tracking of specific objective, have very high algorithm complex, verification and measurement ratio is also lower, can't satisfy the demand of real-time, interactive.
In most cases, do not need point-device detection and tracking precision, but real-time and applicability there is very high requirement, patent of the present invention has provided a kind of method and has addressed this problem, the variation of scene is converted to actions such as translation, rotation and convergent-divergent, realizes more general body sense interactive mode.
Summary of the invention
In view of problems of the prior art, the object of the invention is to provide a kind of body sense exchange method based on the preposition camera of mobile phone, may further comprise the steps:
(1) by preposition camera, obtains the current scene image, i.e. video sequence;
(2) judge whether the current scene image that obtains is first two field picture, if return execution in step (1);
(3) to the view data of present frame and former frame, carry out the optical flow field analysis, obtain light stream variation between two frames and the coordinate of corresponding point;
(4) choose the point coordinate set that two interframe light streams are changed significantly, calculate variation tendency and the amplitude of interframe scene, correspond to trend and the amplitude of translation, rotation and convergent-divergent;
(5) circulation execution in step (1)~(4), and simultaneously trend and the amplitude data of the translation, rotation and the convergent-divergent that calculate offered application program.
Preferably, between step (2) and (3), also comprise step: the scene image that gets access to is carried out down-sampling, obtain the view data of small in resolution.
Preferably, the Downsapling method of employing is the nearest neighbor pixels method, makes the gray-scale value of pixel after the conversion equal gray-scale value apart from its nearest input pixel.
Preferably, use the KLT algorithm to calculate optical flow field, obtain the corresponding point between consecutive frame.
Description of drawings
Fig. 1 illustration a kind of body sense exchange method process flow diagram based on the preposition camera of mobile phone of the embodiment of the invention;
Fig. 2 illustration corresponding point set graph of a relation between embodiment of the invention image.
Embodiment
For above-mentioned purpose of the present invention, feature and advantage are become apparent more, the present invention is further detailed explanation below in conjunction with the drawings and specific embodiments.
Fig. 1 illustration a kind of body sense exchange method process flow diagram based on the preposition camera of mobile phone of providing of the embodiment of the invention.
As shown in Figure 1, said method may further comprise the steps:
(1) by preposition camera, obtains the current scene image, i.e. video sequence;
(2) judge whether the current scene image that obtains is first two field picture, if return execution in step (1);
(3) scene image that gets access to is carried out down-sampling, obtain the view data of small in resolution;
(4) to the view data of present frame and former frame, carry out the optical flow field analysis, obtain light stream variation between two frames and the coordinate of corresponding point;
(5) choose the point coordinate set that two interframe light streams are changed significantly, calculate variation tendency and the amplitude of interframe scene, correspond to trend and the amplitude of translation, rotation and convergent-divergent;
(6) circulation execution in step (1)~(5), and simultaneously trend and the amplitude data of the translation, rotation and the convergent-divergent that calculate offered application program.
Pass through said method, use the optical flow tracking algorithm, obtain mean change trend and the amplitude of moving region in the scene, and be converted to actions such as translation, rotation and convergent-divergent, thereby realize a kind of real-time, general body sense interactive mode, application program can be utilized trend and the corresponding various processing operations of amplitude data of the translation, rotation and the convergent-divergent that obtain.
For more detailed and the present invention will be described best and the protection, below the illustrated method of Fig. 1 is carried out the more explanation of details, but it will be appreciated by those skilled in the art that it is not construed as limiting the invention.
In step shown in Figure 1 (3), the video data that gets access to is carried out down-sampling, obtain the image of small in resolution, purpose is to reduce the computational data amount.In the method, do not need to obtain very high-precision gesture, so not very high requirement of the quality of data, on the image of little resolution, still can effectively calculate, but can reduce computation complexity significantly.Therefore, one skilled in the art can appreciate that step (3) is the preferred steps that makes that the present invention more optimizes, is not steps necessary, and under the situation of not carrying out the down-sampling processing, the present invention still can realize, only is that computation complexity is higher.
Among the present invention, the Downsapling method of employing can be the nearest neighbor pixels method for example, makes the gray-scale value of pixel after the conversion equal gray-scale value apart from its nearest input pixel.
In step shown in Figure 1 (4), use the KLT algorithm to calculate optical flow field, obtain the corresponding point between consecutive frame.
In the KLT algorithm, following prerequisite hypothesis is arranged: (a) brightness constancy, (b) time continuous, (c) space unanimity.On a window W of image, all points (x, y) all to a direction translation (dx dy), thereby obtains (x ', y '), i.e. t
0Constantly (x is y) at t
1Constantly be (x ', y '), be converted into following formula is sought minimum value so seek the problem of corresponding point:
Use the KLT algorithm, obtain the corresponding point set of adjacent two interframe respectively, for { Pi}, { Pi ' }, the distribution on image is similar to shown in Figure 2.
In step (2) and (3), the corresponding point set by step (1) obtains calculates its central point
,
, wherein: try to achieve each unique point to the mean distance of central point Di}, { Di ' }, wherein:
Try to achieve each unique point to the angle { α of central point
i, { α
i', wherein:
By above result calculated, can obtain the amplitude of translation, convergent-divergent, rotation respectively:
It more than is the detailed description that the preferred embodiments of the present invention are carried out, but those of ordinary skill in the art is to be appreciated that, within the scope of the present invention, and guided by the spirit, various improvement, interpolation and replacement all are possible, for example use that the different programming language (as C, C++, Java etc.) of algorithm, use that can realize functional purpose of the same race is realized etc.These are all in the protection domain that claim of the present invention limits.
Claims (4)
1. body sense exchange method based on the preposition camera of mobile phone may further comprise the steps:
(1) by preposition camera, obtains the current scene image, i.e. video sequence;
(2) judge whether the current scene image that obtains is first two field picture, if return execution in step (1);
(3) to the view data of present frame and former frame, carry out the optical flow field analysis, obtain light stream variation between two frames and the coordinate of corresponding point;
(4) choose the point coordinate set that two interframe light streams are changed significantly, calculate variation tendency and the amplitude of interframe scene, correspond to trend and the amplitude of translation, rotation and convergent-divergent;
(5) circulation execution in step (1)~(4), and simultaneously trend and the amplitude data of the translation, rotation and the convergent-divergent that calculate offered application program.
2. the method for claim 1 is characterized in that also comprising step between step (2) and (3): the scene image that gets access to is carried out down-sampling, obtain the view data of small in resolution.
3. method as claimed in claim 2 is characterized in that, the Downsapling method of employing is the nearest neighbor pixels method, makes the gray-scale value of pixel after the conversion equal gray-scale value apart from its nearest input pixel.
4. as arbitrary described method among the claim 1-3, it is characterized in that, use the KLT algorithm to calculate optical flow field, obtain the corresponding point between consecutive frame.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN2013101738210A CN103268589A (en) | 2013-05-10 | 2013-05-10 | Somatosensory interaction method based on front-facing camera of mobile terminal |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN2013101738210A CN103268589A (en) | 2013-05-10 | 2013-05-10 | Somatosensory interaction method based on front-facing camera of mobile terminal |
Publications (1)
Publication Number | Publication Date |
---|---|
CN103268589A true CN103268589A (en) | 2013-08-28 |
Family
ID=49012216
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN2013101738210A Pending CN103268589A (en) | 2013-05-10 | 2013-05-10 | Somatosensory interaction method based on front-facing camera of mobile terminal |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN103268589A (en) |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH11110566A (en) * | 1997-10-07 | 1999-04-23 | Ntt Data Corp | Method and device for image recognition |
CN101710993A (en) * | 2009-11-30 | 2010-05-19 | 北京大学 | Block-based self-adaptive super-resolution video processing method and system |
CN101881615A (en) * | 2010-05-28 | 2010-11-10 | 清华大学 | Method for detecting visual barrier for driving safety |
WO2010151215A1 (en) * | 2009-06-22 | 2010-12-29 | Imint Image Intelligence Ab | Real time video stabilization |
-
2013
- 2013-05-10 CN CN2013101738210A patent/CN103268589A/en active Pending
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH11110566A (en) * | 1997-10-07 | 1999-04-23 | Ntt Data Corp | Method and device for image recognition |
WO2010151215A1 (en) * | 2009-06-22 | 2010-12-29 | Imint Image Intelligence Ab | Real time video stabilization |
CN101710993A (en) * | 2009-11-30 | 2010-05-19 | 北京大学 | Block-based self-adaptive super-resolution video processing method and system |
CN101881615A (en) * | 2010-05-28 | 2010-11-10 | 清华大学 | Method for detecting visual barrier for driving safety |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US9965861B2 (en) | Method and system of feature matching for multiple images | |
US9041834B2 (en) | Systems and methods for reducing noise in video streams | |
Kong et al. | Fastflownet: A lightweight network for fast optical flow estimation | |
US9743016B2 (en) | Techniques for improved focusing of camera arrays | |
CN107079100A (en) | The method and system that Lens shift for camera array is corrected | |
US10554928B2 (en) | Telepresence device | |
US11656722B1 (en) | Method and apparatus for creating an adaptive bayer pattern | |
JP6109956B2 (en) | Utilize encoder hardware to pre-process video content | |
CN108491763B (en) | Unsupervised training method and device for three-dimensional scene recognition network and storage medium | |
US11089265B2 (en) | Telepresence devices operation methods | |
WO2015112652A1 (en) | Image demosaicing | |
US9807313B2 (en) | Method and system of increasing integer disparity accuracy for camera images with a diagonal layout | |
US9584806B2 (en) | Using depth information to assist motion compensation-based video coding | |
US10742868B2 (en) | Imaging device and focusing control method | |
Santamaria et al. | A comparison of block-matching motion estimation algorithms | |
CN112991381B (en) | Image processing method and device, electronic equipment and storage medium | |
CN108171222B (en) | Real-time video classification method and device based on multi-stream neural network | |
US11004266B2 (en) | Articulated model registration apparatus and method | |
CN104349197A (en) | Data processing method and device | |
CN116385260B (en) | Image processing method, device, chip, electronic equipment and medium | |
Wang et al. | Freeman: Towards benchmarking 3d human pose estimation in the wild | |
CN103268589A (en) | Somatosensory interaction method based on front-facing camera of mobile terminal | |
CN113361360B (en) | Multi-person tracking method and system based on deep learning | |
CN100371861C (en) | Interactive input control method based on computer image and pure color object | |
CN106331890A (en) | Processing method and device for video communication image |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20130828 |
|
RJ01 | Rejection of invention patent application after publication |