CN103268589A - Somatosensory interaction method based on front-facing camera of mobile terminal - Google Patents

Somatosensory interaction method based on front-facing camera of mobile terminal Download PDF

Info

Publication number
CN103268589A
CN103268589A CN2013101738210A CN201310173821A CN103268589A CN 103268589 A CN103268589 A CN 103268589A CN 2013101738210 A CN2013101738210 A CN 2013101738210A CN 201310173821 A CN201310173821 A CN 201310173821A CN 103268589 A CN103268589 A CN 103268589A
Authority
CN
China
Prior art keywords
amplitude
translation
variation
rotation
optical flow
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN2013101738210A
Other languages
Chinese (zh)
Inventor
朱巍巍
陈继
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
ThunderSoft Co Ltd
Original Assignee
ThunderSoft Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by ThunderSoft Co Ltd filed Critical ThunderSoft Co Ltd
Priority to CN2013101738210A priority Critical patent/CN103268589A/en
Publication of CN103268589A publication Critical patent/CN103268589A/en
Pending legal-status Critical Current

Links

Images

Landscapes

  • Image Analysis (AREA)

Abstract

The invention discloses a somatosensory interaction method based on a front-facing camera of a mobile terminal. The method includes the following steps: (1) acquiring a current scene image, namely a video sequence, through the front-facing camera, (2) judging whether the acquired current scene image is a first-frame image, and if yes, returning the process to carry out the step (1), (3) carrying out optical flow field analysis on image data of a current frame and image data of a previous frame to acquire optical flow variation between the two frames and coordinates of corresponding points, (4) selecting a coordinate set of the points with obvious optical flow variation between the two frames, and calculating the variation tendency and the variation amplitude of scenes between the frames as the tendency and the amplitude of translation, rotation and scaling correspondingly, and (5) carrying out the step (1) to the step (4) circularly, calculating tendency data and amplitude data of the translation, the rotation and the scaling, and providing the data to an application program. The somatosensory interaction method based on the front-facing camera of the mobile terminal converts the variation of the scenes into the motions of translation, rotation, scaling and the like, and is more universal.

Description

A kind of body sense exchange method based on the preposition camera of portable terminal
Technical field
The present invention relates to intelligent movable equipment application, particularly a kind of body sense exchange method based on the preposition camera of mobile phone.
Background technology
On mobile terminal devices such as mobile phone, with mobile phone alternately mainly by touch-screen and physical keyboard.More and more higher along with handset capability, application program is more and more abundanter, and the diversity of interactive mode is also had higher requirement.
The preposition camera of portable terminals such as mobile phone can obtain abundant scene information, but existing application to preposition camera is main still in purposes such as video calling, auto heterodyne, recognitions of face.
Utilize methods such as the detection of people's face, gestures detection to follow the trail of the variation of scene, though have higher accuracy, real-time and applicability are very poor, lose efficacy easily under complex scene.
Under general common scene, the video of preposition camera is carried out the detection and tracking of specific objective, have very high algorithm complex, verification and measurement ratio is also lower, can't satisfy the demand of real-time, interactive.
In most cases, do not need point-device detection and tracking precision, but real-time and applicability there is very high requirement, patent of the present invention has provided a kind of method and has addressed this problem, the variation of scene is converted to actions such as translation, rotation and convergent-divergent, realizes more general body sense interactive mode.
Summary of the invention
In view of problems of the prior art, the object of the invention is to provide a kind of body sense exchange method based on the preposition camera of mobile phone, may further comprise the steps:
(1) by preposition camera, obtains the current scene image, i.e. video sequence;
(2) judge whether the current scene image that obtains is first two field picture, if return execution in step (1);
(3) to the view data of present frame and former frame, carry out the optical flow field analysis, obtain light stream variation between two frames and the coordinate of corresponding point;
(4) choose the point coordinate set that two interframe light streams are changed significantly, calculate variation tendency and the amplitude of interframe scene, correspond to trend and the amplitude of translation, rotation and convergent-divergent;
(5) circulation execution in step (1)~(4), and simultaneously trend and the amplitude data of the translation, rotation and the convergent-divergent that calculate offered application program.
Preferably, between step (2) and (3), also comprise step: the scene image that gets access to is carried out down-sampling, obtain the view data of small in resolution.
Preferably, the Downsapling method of employing is the nearest neighbor pixels method, makes the gray-scale value of pixel after the conversion equal gray-scale value apart from its nearest input pixel.
Preferably, use the KLT algorithm to calculate optical flow field, obtain the corresponding point between consecutive frame.
Description of drawings
Fig. 1 illustration a kind of body sense exchange method process flow diagram based on the preposition camera of mobile phone of the embodiment of the invention;
Fig. 2 illustration corresponding point set graph of a relation between embodiment of the invention image.
Embodiment
For above-mentioned purpose of the present invention, feature and advantage are become apparent more, the present invention is further detailed explanation below in conjunction with the drawings and specific embodiments.
Fig. 1 illustration a kind of body sense exchange method process flow diagram based on the preposition camera of mobile phone of providing of the embodiment of the invention.
As shown in Figure 1, said method may further comprise the steps:
(1) by preposition camera, obtains the current scene image, i.e. video sequence;
(2) judge whether the current scene image that obtains is first two field picture, if return execution in step (1);
(3) scene image that gets access to is carried out down-sampling, obtain the view data of small in resolution;
(4) to the view data of present frame and former frame, carry out the optical flow field analysis, obtain light stream variation between two frames and the coordinate of corresponding point;
(5) choose the point coordinate set that two interframe light streams are changed significantly, calculate variation tendency and the amplitude of interframe scene, correspond to trend and the amplitude of translation, rotation and convergent-divergent;
(6) circulation execution in step (1)~(5), and simultaneously trend and the amplitude data of the translation, rotation and the convergent-divergent that calculate offered application program.
Pass through said method, use the optical flow tracking algorithm, obtain mean change trend and the amplitude of moving region in the scene, and be converted to actions such as translation, rotation and convergent-divergent, thereby realize a kind of real-time, general body sense interactive mode, application program can be utilized trend and the corresponding various processing operations of amplitude data of the translation, rotation and the convergent-divergent that obtain.
For more detailed and the present invention will be described best and the protection, below the illustrated method of Fig. 1 is carried out the more explanation of details, but it will be appreciated by those skilled in the art that it is not construed as limiting the invention.
In step shown in Figure 1 (3), the video data that gets access to is carried out down-sampling, obtain the image of small in resolution, purpose is to reduce the computational data amount.In the method, do not need to obtain very high-precision gesture, so not very high requirement of the quality of data, on the image of little resolution, still can effectively calculate, but can reduce computation complexity significantly.Therefore, one skilled in the art can appreciate that step (3) is the preferred steps that makes that the present invention more optimizes, is not steps necessary, and under the situation of not carrying out the down-sampling processing, the present invention still can realize, only is that computation complexity is higher.
Among the present invention, the Downsapling method of employing can be the nearest neighbor pixels method for example, makes the gray-scale value of pixel after the conversion equal gray-scale value apart from its nearest input pixel.
In step shown in Figure 1 (4), use the KLT algorithm to calculate optical flow field, obtain the corresponding point between consecutive frame.
In the KLT algorithm, following prerequisite hypothesis is arranged: (a) brightness constancy, (b) time continuous, (c) space unanimity.On a window W of image, all points (x, y) all to a direction translation (dx dy), thereby obtains (x ', y '), i.e. t 0Constantly (x is y) at t 1Constantly be (x ', y '), be converted into following formula is sought minimum value so seek the problem of corresponding point:
∈ ( d ) = ∈ ( d x , d y ) = Σ x = u x - w x u x + w x Σ y = y y - w y u y + w y ( I ( x , y ) - J ( x + d x , y + d y ) ) 2 - - - ( 1 )
Use the KLT algorithm, obtain the corresponding point set of adjacent two interframe respectively, for { Pi}, { Pi ' }, the distribution on image is similar to shown in Figure 2.
In step (2) and (3), the corresponding point set by step (1) obtains calculates its central point
Figure BDA00003175320800034
,
Figure BDA00003175320800035
, wherein: try to achieve each unique point to the mean distance of central point Di}, { Di ' }, wherein:
Di = Pi - C → 0 - - - ( 4 )
Di ′ = Pi ′ - C → 1 - - - ( 5 )
Try to achieve each unique point to the angle { α of central point i, { α i', wherein:
α i = arctan ( Pi - C → 0 ) - - - ( 6 )
α i ′ = arctan ( Pi ′ - C 1 → ) - - - ( 7 )
By above result calculated, can obtain the amplitude of translation, convergent-divergent, rotation respectively:
D ( t 0 , t 1 ) = C → 1 - C → 0 - - - ( 8 )
Z ( t 0 , t 1 ) = Σ D i Σ D i ′ - - - ( 9 )
R ( t 0 , t 1 ) = Σ ( α i ′ - α i ) / n - - - ( 10 )
It more than is the detailed description that the preferred embodiments of the present invention are carried out, but those of ordinary skill in the art is to be appreciated that, within the scope of the present invention, and guided by the spirit, various improvement, interpolation and replacement all are possible, for example use that the different programming language (as C, C++, Java etc.) of algorithm, use that can realize functional purpose of the same race is realized etc.These are all in the protection domain that claim of the present invention limits.

Claims (4)

1. body sense exchange method based on the preposition camera of mobile phone may further comprise the steps:
(1) by preposition camera, obtains the current scene image, i.e. video sequence;
(2) judge whether the current scene image that obtains is first two field picture, if return execution in step (1);
(3) to the view data of present frame and former frame, carry out the optical flow field analysis, obtain light stream variation between two frames and the coordinate of corresponding point;
(4) choose the point coordinate set that two interframe light streams are changed significantly, calculate variation tendency and the amplitude of interframe scene, correspond to trend and the amplitude of translation, rotation and convergent-divergent;
(5) circulation execution in step (1)~(4), and simultaneously trend and the amplitude data of the translation, rotation and the convergent-divergent that calculate offered application program.
2. the method for claim 1 is characterized in that also comprising step between step (2) and (3): the scene image that gets access to is carried out down-sampling, obtain the view data of small in resolution.
3. method as claimed in claim 2 is characterized in that, the Downsapling method of employing is the nearest neighbor pixels method, makes the gray-scale value of pixel after the conversion equal gray-scale value apart from its nearest input pixel.
4. as arbitrary described method among the claim 1-3, it is characterized in that, use the KLT algorithm to calculate optical flow field, obtain the corresponding point between consecutive frame.
CN2013101738210A 2013-05-10 2013-05-10 Somatosensory interaction method based on front-facing camera of mobile terminal Pending CN103268589A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN2013101738210A CN103268589A (en) 2013-05-10 2013-05-10 Somatosensory interaction method based on front-facing camera of mobile terminal

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN2013101738210A CN103268589A (en) 2013-05-10 2013-05-10 Somatosensory interaction method based on front-facing camera of mobile terminal

Publications (1)

Publication Number Publication Date
CN103268589A true CN103268589A (en) 2013-08-28

Family

ID=49012216

Family Applications (1)

Application Number Title Priority Date Filing Date
CN2013101738210A Pending CN103268589A (en) 2013-05-10 2013-05-10 Somatosensory interaction method based on front-facing camera of mobile terminal

Country Status (1)

Country Link
CN (1) CN103268589A (en)

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH11110566A (en) * 1997-10-07 1999-04-23 Ntt Data Corp Method and device for image recognition
CN101710993A (en) * 2009-11-30 2010-05-19 北京大学 Block-based self-adaptive super-resolution video processing method and system
CN101881615A (en) * 2010-05-28 2010-11-10 清华大学 Method for detecting visual barrier for driving safety
WO2010151215A1 (en) * 2009-06-22 2010-12-29 Imint Image Intelligence Ab Real time video stabilization

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH11110566A (en) * 1997-10-07 1999-04-23 Ntt Data Corp Method and device for image recognition
WO2010151215A1 (en) * 2009-06-22 2010-12-29 Imint Image Intelligence Ab Real time video stabilization
CN101710993A (en) * 2009-11-30 2010-05-19 北京大学 Block-based self-adaptive super-resolution video processing method and system
CN101881615A (en) * 2010-05-28 2010-11-10 清华大学 Method for detecting visual barrier for driving safety

Similar Documents

Publication Publication Date Title
US9965861B2 (en) Method and system of feature matching for multiple images
US9041834B2 (en) Systems and methods for reducing noise in video streams
Kong et al. Fastflownet: A lightweight network for fast optical flow estimation
US9743016B2 (en) Techniques for improved focusing of camera arrays
CN107079100A (en) The method and system that Lens shift for camera array is corrected
US10554928B2 (en) Telepresence device
US11656722B1 (en) Method and apparatus for creating an adaptive bayer pattern
JP6109956B2 (en) Utilize encoder hardware to pre-process video content
CN108491763B (en) Unsupervised training method and device for three-dimensional scene recognition network and storage medium
US11089265B2 (en) Telepresence devices operation methods
WO2015112652A1 (en) Image demosaicing
US9807313B2 (en) Method and system of increasing integer disparity accuracy for camera images with a diagonal layout
US9584806B2 (en) Using depth information to assist motion compensation-based video coding
US10742868B2 (en) Imaging device and focusing control method
Santamaria et al. A comparison of block-matching motion estimation algorithms
CN112991381B (en) Image processing method and device, electronic equipment and storage medium
CN108171222B (en) Real-time video classification method and device based on multi-stream neural network
US11004266B2 (en) Articulated model registration apparatus and method
CN104349197A (en) Data processing method and device
CN116385260B (en) Image processing method, device, chip, electronic equipment and medium
Wang et al. Freeman: Towards benchmarking 3d human pose estimation in the wild
CN103268589A (en) Somatosensory interaction method based on front-facing camera of mobile terminal
CN113361360B (en) Multi-person tracking method and system based on deep learning
CN100371861C (en) Interactive input control method based on computer image and pure color object
CN106331890A (en) Processing method and device for video communication image

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20130828

RJ01 Rejection of invention patent application after publication