US20140254870A1 - Method for recognizing motion gesture commands - Google Patents
Method for recognizing motion gesture commands Download PDFInfo
- Publication number
- US20140254870A1 US20140254870A1 US14/171,029 US201414171029A US2014254870A1 US 20140254870 A1 US20140254870 A1 US 20140254870A1 US 201414171029 A US201414171029 A US 201414171029A US 2014254870 A1 US2014254870 A1 US 2014254870A1
- Authority
- US
- United States
- Prior art keywords
- image
- gesture
- computer
- blurred
- frame
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/017—Gesture based interaction, e.g. based on a set of recognized hand gestures
-
- G06K9/00355—
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/20—Movements or behaviour, e.g. gesture recognition
- G06V40/28—Recognition of hand or arm movements, e.g. recognition of deaf sign language
Definitions
- the present invention relates to gesture commands in general, and particularly to a method for recognizing motion gesture commands for controlling computer operations.
- a desktop computer, a laptop computer (laptop PC), a tablet computer, or a smartphone is commonly equipped with a web camera capable of projecting real-time video on a monitor to enable a user to make a video phone call or to take a photograph or a video.
- One prior art discloses a method for recognizing a motion pattern of a person from the parameters of a motion blur taken with a camera.
- the extent of the motion blur can be utilized as an indicator to indicate the speed of an object.
- the trajectory of the object can be obtained from variations in the direction of the motion blur.
- the motion blur is associated with a gesture.
- Another prior art discloses a method for detecting a moving object with high accuracy even when a photographing environment changes in terms of the brightness of an object or when the pictures is taken during a camera shake.
- the moving object can be detected from a differential image corresponding to a density difference between a first image and a second image input immediately before the first image.
- a reference pattern is initially defined.
- a frame is subsequently received from an image sensor that has captured a blurred image of an object.
- the shape of the blurred image included in the received frame is identified.
- the reference pattern are then comparing to the identified shape of the blurred image to determine whether or not the identified shape represents a gesture command.
- a corresponding command for controlling an operation of a computer is generated.
- FIG. 1 is a block diagram of a laptop PC
- FIG. 2 is a block diagram of a camera system within the laptop PC from FIG. 1 ;
- FIG. 3 is a block diagram depicting the software configuration of the camera system from FIG. 2 ;
- FIGS. 4A-4B are diagrams depicting a relationship between frame rate and exposure time
- FIG. 5 is a block diagram depicting the configuration of a gesture recognition application
- FIG. 6 is a flowchart of a procedure through which the camera system from FIG. 2 recognizes a gesture
- FIG. 7 is a diagram abstractly depicting a state of a background image and a person present in the foreground, who swings an arm in a cyclic way;
- FIGS. 8A-8D are diagrams depicting the shapes of reference patterns registered in a reference image registering section
- FIG. 9 is a diagram depicting an example of generating a differential image group and a composite image group from a background image and a frame sequence
- FIG. 10 is a diagram depicting a relationship between frame transfer period Tf and gesture cycle Tj.
- FIG. 11 is a diagram depicting an example of generating a differential image data group and a composite image data group from a frame sequence.
- FIG. 1 is a block diagram of a camera system 10 implemented in a laptop PC.
- a CPU 11 incorporates a memory controller and a PCI Express controller.
- the CPU 11 is connected to a main memory 13 , a video card 15 , and a chipset 19 .
- An LCD 17 is connected to the video card 15 .
- the chipset 19 incorporates controllers, such as SATA, USB, PCI Express, and LPC, and also incorporates Real-Time Clock (RTC).
- RTC Real-Time Clock
- An HDD 21 is connected to the SATA controller.
- the USB controller is configured to include multiple hubs that make up a USB host controller, a root hub, and I/O ports.
- a camera module 100 is a USB device conforming to USB2.0 or USB3.0 standards.
- the camera module 100 is connected to a USB port of the USB controller through a pair or three pairs of USB buses 50 to transfer data using a differential signal.
- the USB port to which the camera module 100 is connected may share a hub with another USB device. However, it is preferred that the USB port be connected to a hub dedicated to the camera module 100 to control the power of the camera module 100 effectively using a selective suspend mechanism of a USB system.
- a request packet is sent to a USB bus driver.
- the USB bus driver makes USB buses connected to the USB hub transit to an idle state, and USB devices that have detected the state transition of the buses transit to the selective suspend state. Therefore, as the camera module 100 stays in the idle state for a longer period of time, the period for which the USB buses stay in the idle state becomes longer, and as a result, the period of the transition of the camera module 100 to the selective suspend state becomes longer.
- the camera module 100 may be of the type of being incorporated in the housing of the laptop PC or an external type module that is connected to a USB connector mounted in the housing of the laptop PC.
- the camera module 100 may also be connected to the laptop PC by wireless USB.
- the camera system 10 according to the present invention can transfer image data in a gesture mode introduced by the present invention in addition to a conventional video mode for taking a video.
- the gesture mode is a transfer mode used by the gesture recognition application 201 in FIG. 3 to acquire frames in order to recognize a gesture.
- the camera module 100 transfers frames at a frame rate lower than the video mode to increase the exposure time so that each frame can capture a motion blur of the gesture.
- the frame corresponds to one still image unit in image data composed of all horizontal lines.
- one frame is composed of two fields. The details of the gesture mode will be described later.
- the camera module 100 In addition to the video mode and the gesture mode, the camera module 100 also operates in a monitoring mode for checking an image upon taking a still image. In the monitoring mode, the number of pixels in a frame can be reduced compared to that in the video mode.
- the hardware of the laptop PC 10 such as the CPU 11 , the chipset 19 , and the main memory 13 , and software such as the gesture recognition application 201 shown in FIG. 3 , respective layers of device drivers 207 , 209 , 211 , and 213 , a still image transfer service 203 , and an operating system 205 are the main components of the camera system 10 .
- An EC 25 is a microcontroller for controlling temperature inside the housing of the computer 10 and controlling the power of devices.
- the EC 25 operates independently of the CPU 11 .
- An acceleration sensor 27 and a DC/DC converter 29 are connected to the EC 25 , and a keyboard, a mouse, a charger, an exhaust fan, and the like, which are not shown, are also connected to the EC 25 .
- a battery or an AC/DC adapter, not shown, supplies power to the DC/DC converter 29 .
- the DC/DC converter 29 supplies power to the devices that make up the computer 10 .
- FIG. 2 is a detailed block diagram of the camera module 100 .
- FIG. 3 is a block diagram depicting the software configuration of the camera system 10 .
- the camera module 100 can transfer frames having a screen resolution (number of pixels) such as VGA (640 ⁇ 480), QVGA (320 ⁇ 240), WVGA (800 ⁇ 480), or WQVGA (400 ⁇ 240).
- An optical mechanism 101 is made up of an optical lens, an optical filter, and the like to focus the image of a subject onto an image sensor 103 .
- the image sensor 103 is made up of a CMOS image sensor for converting, to an electric signal, an electric charge corresponding to the amount of light accumulated in a photodiode that forms part of a pixel, and outputting the electric signal, a CDS circuit for reducing noise, an AGC circuit for adjusting gain, an AD conversion circuit for converting an analog signal to a digital signal, and the like.
- the image sensor 103 outputs a digital signal corresponding to the brightness of the subject image.
- the image sensor 103 can transfer image data at a frame rate of 30 fps.
- the CMOS image sensor includes an electronic shutter called a rolling shutter.
- the rolling shutter uses one line or a few lines as one block unless otherwise specified by the system to control the exposure time so that the exposure time will become best for the photographing environment.
- the rolling shutter rests, in the process of photographing, a signal charge accumulated during one field period in the photodiode that forms part of each pixel to control the exposure time.
- the image sensor 103 can also use a CCD image sensor instead of the CMOS image sensor.
- An image signal processor (ISP) 105 is an image signal processing circuit for performing processing, such as correction processing for correcting a defective pixel and shading, white balance processing for correcting the spectral characteristics of the image sensor 103 according to the human visibility, interpolation processing for outputting common RGB data from signals of an RGB Bayer array, color correction processing for approximating the spectral characteristics of color filters in the image sensor 103 to ideal characteristics, and the like.
- the ISP 105 further performs processing, such as contour correction processing for increasing the sense of resolution of the subject, gamma processing for correcting the non-linear input/output characteristics of the LCD 17 , and the like.
- An encoder 107 compresses image data received from the ISP 105 .
- An end point buffer 109 temporarily stores data to be transferred with the system bi-directionally to form multiple pipes for USB transfer.
- a serial interface engine (SIE) 111 packetizes image data received from the end point buffer 109 to meet the USB standards and sends the packetized image data to a transceiver 113 , or analyzes a packet received from the transceiver 113 and sends a payload to an MPU 115 .
- SIE serial interface engine
- the SIE 111 interrupts the MPU 115 to make the USB bus 50 transit to the selective suspend state. Then, the SIE 111 operates the MPU 115 suspended when the USB bus 50 is resumed.
- the transceiver 113 includes a transmission transceiver and a reception transceiver for USB communication.
- the MPU 115 performs enumeration for USB transfer, and controls the operation of the camera module 100 for further taking an image and transferring image data.
- the camera module 100 conforms to the power management specified in the USB standards.
- the MPU 115 can disrupt an internal clock to make the camera module 100 including itself transit to the suspended state.
- the MPU 115 can return the camera module 100 to the power-on state or a photographing state when the USB bus 50 is resumed.
- the MPU 115 interprets a command received from the system to control the operation of each component in order to transfer image data in the video mode (including the monitoring mode) or the gesture mode.
- the MPU 115 sends a resume signal to the USB bus driver 211 to release the idle state.
- the gesture recognition application 201 can specify various parameters to set an exposure time.
- the gesture recognition application 201 specifies only the frame rate for transferring still images to the camera module 100 .
- the MPU 115 sets the maximum exposure time available for the current photographing conditions on the frame rate specified.
- the MPU 115 further sets, in the image sensor 103 and a predetermined register in the ISP 105 , other parameters to be optimized for the current photographing conditions on the set exposure time.
- the gesture recognition application 201 specifies several parameters that affect the taking of a blurred image, such as exposure time, ACG gain, gamma curve, contrast, and aperture value.
- the MPU 115 sets other parameters to be optimized for the current photographing conditions on the specified parameters.
- the MPU 115 can set a frame rate according to the exposure time.
- FIG. 4A shows a state of transferring image data at a frame rate F 1 of 30 fps in the video mode
- FIG. 4B shows a state of transferring image data at a frame rate F 2 of 5 fps as an example in the gesture mode.
- a description will be made by taking a non-interlaced system as an example, but the present invention may also employ a camera module for outputting image data in an interlaced system.
- first to N lines are scanned sequentially in each frame transfer period Tf 1 (1/F 1 ) to output data.
- the time required for each line is composed of an exposure time 151 and a reading time 153 .
- first to N lines are scanned sequentially in a frame transfer period Tf 2 (1/F 2 ) to output data.
- Each line is composed of an exposure time 155 and a reading time 153 .
- the exposure time can be prolonged up to about sixfold.
- the longer the exposure time the larger the area of a blurred image when a moving object is shot.
- the area of the blurred image corresponds to the amount of information used for pattern recognition.
- the blurred image is an image formed by the contour of the moving object shot in one frame.
- the gesture recognition application 201 identifies the shape of the blurred image from the contour of the blurred image.
- the moving object is an arm of a person who makes a gesture as an example, but the present invention is not limited thereto.
- the moving object can be any other object such as a moving machine or natural object.
- the camera module 100 transits to the selective suspend state when no frame is transferred, and this may result in loss of the set parameters.
- the MPU 115 can also store, in a flash ROM 119 , the parameters set once, and set the parameters by reading them from the flash ROM 119 in each frame transfer period.
- the flash ROM stores a program executed by the MPU 115 .
- the MPU 115 can interpret a command received from the system to operate the camera module 100 in the gesture mode or the video mode.
- the camera module 100 is a bus-powered device that operates by receiving the supply of power from the USB bus 50 .
- the camera module 100 may also be a self-powered device that operates with power supplied by itself.
- the MPU 115 controls the power supplied independently to follow the state of the USB bus 50 .
- the camera module 100 transits between the selective suspend state and the power-on state based on the USB standards.
- the camera module 100 may transit to two or more low-power states during the period of the selective suspend state.
- the camera module 100 can operate only the image sensor 103 during a period of sending no frame to transit to a low-power state with least power consumption.
- FIGS. 1 and 2 simply show the main hardware configuration and connection relations associated with the embodiment in a simplified manner to describe the embodiment.
- many other devices are used to make up the camera system 10 .
- the interface of the camera module 100 may be configured as a Mobile Industry Processor Interface (MIPI) and the system may be configured as a system-on-a-chip (SOC).
- MIPI Mobile Industry Processor Interface
- SOC system-on-a-chip
- a moving image application 204 is a known program for acquiring video data from the camera module 100 to display the video on the LCD 17 or recording it in the HDD 21 .
- the moving image application 204 can be a video-phone program or a moving image shooting program.
- the moving image application 204 can also be a gesture engine for performing sophisticated recognition of a gesture by a conventional method.
- the gesture recognition application 201 is a novel program for acquiring a frame sequence from the camera module 100 in the gesture mode, performing pattern recognition of a blurred image, and outputting a command corresponding to the recognized gesture.
- the frame sequence means a frame group of continuous still images transferred at a frame rate low enough to produce pictures unnatural as video data.
- the power state of the laptop PC 10 can be controlled or the moving image application 204 can be started by the command output from the gesture recognition application 201 .
- the gesture recognition application 201 can include a user interface for allowing a user to set parameters of the camera system 10 to be applied to the gesture mode, set the gesture mode to enabled, and the like.
- the gesture recognition application 201 may be executed only when the laptop PC is receiving the supply of power from the AC/DC adapter. Since the camera system 10 operates with low power consumption, this is particularly suitable for operating the camera system 10 while power is being supplied from a battery pack.
- a streaming service 206 is a service program provided by the OS 205 to send frames to the moving image application 204 in the video mode or the monitoring mode, and pass, to the USB camera driver 207 , a command issued by the moving image application 204 .
- a still image transfer service 203 is a program that runs in a user mode of the OS 205 to send the gesture recognition application 201 a frame sequence transferred in the gesture mode, and send the USB camera driver 207 a command and parameters received from the gesture recognition application 201 .
- the USB camera driver 207 is a device driver for controlling the operation of the camera module 100 and controlling data transfer.
- the USB class driver 209 is a device driver for performing common processing defined in the USB video class.
- the USB bus driver 211 controls the operation of the USB buses connected to the USB controller.
- the USB bus driver 211 When receiving an instruction from the USB camera driver 207 to make a transition of the camera module 100 to selective suspend, the USB bus driver 211 makes the USB bus 50 connected to the camera module 100 transit to the idle state, while when receiving an instruction to resume the USB bus 50 or an instruction for data transfer, the USB bus driver 211 makes the USB bus transit to an active state (resumed state).
- the USB host controller driver 213 controls data transfer to the camera module 100 and the operation of the USB host controller.
- FIG. 5 is a block diagram depicting the configuration of the gesture recognition application 201 .
- An interface section 251 sends a differential image generating section 257 frames received from the still image transfer service 203 sequentially in the gesture mode.
- the OS 205 gives a notification indicating that the background image to be taken with the camera module 100 has changed due to a change in the position of the laptop PC.
- the interface section 251 that received the notification stores, in a background image storing section 255 , multiple frames received from the still image transfer service 203 .
- the interface section 251 acquires multiple frames periodically in the gesture mode and stores the acquired frames in the background image storing section 255 .
- the interface section 251 sends the still image transfer service 203 parameters set by the user, such as enable/disable setting of the gesture mode, frame rate, exposure time, contrast, and AGC gain.
- a background image generating section 253 generates a background image from the frames received from the interface section 251 .
- the background image generating section 253 calculates and binarizes a difference between gradation values of corresponding pixels of received multiple frames possible to include the blurred image to identify the blurred image.
- the background image generating section 253 calculates a logical sum of multiple background images except the blurred image identified from the frames, generates a background image without including the blurred image, and stores the background image in the background image storing section 255 .
- the differential image generating section 257 calculates a gradation difference between corresponding pixels of the frame received from the interface section 251 and the frame of the background image received from the background image storing section 255 , generates a binarized differential image, and sends the differential image to a composite image generating section 259 .
- the composite image generating section 259 that received the differential image calculates a logical sum to combine differential images in order to generate a composite image, and sends the composite image to a pattern recognition section 261 .
- a reference image registering section 265 registers multiple reference patterns associated with predetermined commands.
- the pattern recognition section 261 extracts, using a known algorithm, the features of the composite image generated by the composite image generating section 259 , and compares the composite image with the multiple reference patterns registered in the reference image registering section 263 . When determining that the composite image matches either of the reference patterns, the pattern recognition section 261 outputs, to the OS 205 , a command associated with the reference pattern.
- Each component shown in FIG. 5 is a hardware component that cooperates with hardware such as the CPU 11 and the main memory 13 to provide the laptop PC with a specific function.
- FIG. 7 abstractly shows a state of a background image 291 and a person present in the foreground, who swings an arm 293 in a cyclic way as an example of the moving object that generates a blurred image.
- FIG. 8A-8D show the shapes of reference patterns registered in the reference image registering section 263 .
- FIG. 8A corresponds to a reference pattern of a blurred image in which a person facing a camera swings the arm 293 from side to side with the arm down using a shoulder as a pivot point.
- FIG. 8B corresponds to a reference pattern of a blurred image in which the person facing the camera swings the arm 293 from side to side toward the camera module 100 , where the arm extends forward with the elbow bent.
- FIGS. 8C and 8D correspond to reference patterns of blurred images in each of which the person with the face turned to the left side or right side of the camera module 100 swings the arm back and forth toward himself or herself, where the arm extends forward with the elbow bent.
- FIG. 9 is a diagram depicting an example of generating a differential image group 350 and a composite image group 400 from the background image 291 and a frame sequence 300 .
- the interface section 251 sends the camera module 100 a command for enabling the gesture mode and a parameter associated with the exposure time.
- the camera module 100 shifts into the gesture mode and sets the specified parameter and other appropriate parameters.
- the camera module 100 sets the frame rate, for example, to 5 fps to set the exposure time to the maximum value available in the range in order to start the transfer of the shot frame sequence 300 .
- the user makes a gesture toward the camera module 100 at proper timing. Since the timing of starting the gesture is not synchronized with the timing of transferring frames, taking images of the gesture is started at any time with respect to the frame transfer period.
- the procedure shifts to block 631 .
- the timing of updating the background image comes when the acceleration sensor 27 detects the movement of the laptop PC or in predetermined regular periods.
- the background image generating section 253 stores, in the background image storing section 255 , multiple frames newly generated.
- the differential image generating section 257 receives frames 301 to 309 constituting the frame sequence 300 in order.
- the frames 301 to 309 are images of a gesture taken when the user extends the arm forward in front of the camera and moves the arm once from right to left using the elbow as a pivot point.
- each frame contains each of blurred images 301 a to 309 a corresponding to the motion of the arm that makes the gesture.
- the differential image generating section 257 calculates gradation differences of each of the frames 301 to 309 from corresponding pixels of the latest background image stored in the background image storing section 255 at that point to generate binarized differential images 351 to 359 .
- the calculated gradation differences show large values for pixels the representation of which are changing like the blurred images 301 a to 309 a and small values for pixels to represent the background image 291 with no change. Therefore, before taking the images of the gesture, since the interface section 251 receives only the background image 291 , the gradation of differential images take meaninglessly small values on most of the pixels.
- the differential image generating section 257 sets a fixed threshold value for each pixel for which a gradation difference is calculated to binarize the gradation of each pixel. For example, the differential image generating section 257 sets logical value 1 for pixels that exceed the threshold value and logical value 0 for pixels less than the threshold value.
- the background image 291 is removed from the differential images 351 to 359 , and the blurred images 301 a to 309 a are converted to binarized differential images 351 a to 359 a.
- the differential image generating section 257 sends the differential images 351 to 359 to the composite image generating section 259 .
- the differential images 351 to 359 also contain blurred images (noise blur) as noise caused when the frames 301 to 309 are shot along with the motion of any other object unrelated to the gesture.
- the noise blur reduces the accuracy of pattern recognition.
- the reference patterns 281 to 287 have relatively simple shapes distinguishable from the noise blur, the influence thereof is eliminated in the process of pattern recognition.
- the differential image generating section 257 generates each of the differential images 351 to 359 and sends the generated differential image to the composite image generating section 259 each time the differential image generating section 257 receives each of the frames 301 to 309 in the return path from block 615 to block 605 until the gesture is recognized.
- the composite image generating section 259 calculates a logical sum of logical values set for each pixel of the received one of the differential images 351 to 359 and each pixel of the differential image last received to generate the composite image group 400 . Since an image with which the differential image 351 is calculated to obtain the logical sum does not contain any binarized blurred image, the differential image 351 is exactly the same as a composite image 401 . After generating the composite image 401 , the composite image generating section 259 calculates a logical sum of the differential image 353 next received and the composite image 401 last generated to generate a new composite image 403 . Then, each of composite images 405 to 409 is generated each time each of the differential images 355 to 357 is added to increase the amount of information on blurred images 301 a to 309 a, approaching the shape of a reference pattern 283 .
- the composite image 409 has a shape almost close to the shape of the reference pattern 283 .
- the pattern recognition section 261 compares, using a known algorithm, the features of a composite image with each of the reference patterns 281 to 287 each time each of the composite images 401 to 409 is received.
- the features can include, for example, the direction of being opened into a fan shape, and the center angle or the area. Since the reference patterns 281 to 287 are relatively simple patterns, the load on the CPU 11 that performs pattern recognition can be reduced.
- the pattern recognition section 261 can also complement portions missing in the composite image 405 or 407 to determine that the gesture matches the reference pattern 283 .
- the pattern recognition section 261 sends the OS 205 a command associated with the reference pattern in block 633 .
- the OS 205 activates the gesture engine that performs sophisticated processing for manipulating the laptop PC.
- the interface section 251 further receives a frame sequence.
- the pattern recognition section 261 can ignore a predetermined number of composite images generated after the command is output. After that, the pattern recognition section 261 starts the recognition of a new gesture again at the time of receiving a composite image without the influence of the binarized blurred images as the basis for outputting the previous command.
- the gesture mode since the exposure time is made long intentionally, a gesture can be recognized without being affected by photographing environments ranging from bright places to dark places.
- the recognition target in the present invention is the shape formed by the entire contour of a blurred image.
- the speed of the gesture and the exposure time do not affect the shape of the contour directly, when they are selected properly, it is advantageous in terms of the number of times of a gesture required to recognize the gesture, the recognition time, the power consumption, and photographing in darkness.
- FIG. 10 is a diagram for describing a relationship between frame transfer period Tf and gesture cycle Tj.
- the gesture is the motion of an arm swinging from position P 1 to position P 2 in the outward passage and swinging from position P 2 to position P 1 in the homeward passage.
- the frame transfer period Tf as the reciprocal of the frame rate F is almost equal to the exposure time
- the amount of information usable for comparison with the reference pattern of a blurred image included in one frame increases as the frame transfer period Tf is lengthened.
- the longer the frame transfer period Tf the longer the time from the start of the gesture until the gesture recognition application 201 acquires the first frame, resulting in a delay in the recognition of the gesture.
- a blurred image the inside of which is filled over the entire part of the outward passage (from position P 1 to position P 2 ) or the homeward passage (from position P 2 to position P 1 ) of the gesture is an adequate amount of information necessary for comparison with the reference pattern.
- the recognition can be performed accurately without necessarily providing such an amount of information.
- an upper limit can be set to the frame transfer period Tf because only a delay is caused.
- the upper limit of the frame transfer period Tf can be defined as the minimum time capable of obtaining an adequate amount of information usable for recognition. Note that the time t 0 for starting the gesture cannot be synchronized with the frame transfer timing. In the example of FIG. 10 , a frame for which a frame transfer period Tf 1 shorter than the gesture cycle Tj is set, and the shooting of which is started at time t 1 and finished at time t 2 lacks information on a blurred image from position P 1 to position P 3 of the gesture in the outward passage and the homeward passage.
- the information can obtain at time t 3 at which shooting of the gesture in the second outward passage is finished.
- a frame for which a frame transfer period Tf 2 equal to the gesture cycle Tj is set, and the shooting of which is started at time t 1 and finished at time t 4 contains information on the blurred image obtained by shooting a total extent of the gesture in the first homeward passage. Therefore, even if the frame transfer period Tf is set longer than the gesture cycle Tj, information more useful to add to a composite image than the information already obtained cannot be obtained, leading only to lengthening the time for recognition.
- the maximum value of the frame transfer period Tf (the minimum value of the frame rate) can be so determined that the gesture cycle Tj will match the exposure time.
- the present invention can also recognize the direction of the gesture to increase the number of commands.
- the gesture recognition application 201 recognizes which of a gesture starting from position P 1 or a gesture starting from position P 2 in FIG. 10 has been made.
- the maximum value of the frame transfer period Tf can be set to Tf ⁇ Tj/4 to recognize the gesture starting position.
- the minimum value of the frame transfer period Tf can be determined for each system in consideration of these points. Since the parameters set in the camera module 100 to recognize a gesture depends on the performance of the camera system 10 such as the camera module 100 and the pattern recognition section 261 , the photographing environment, the number of reference patterns, the allowable number of times of a gesture, and the like, the parameters can be determined by experiment for each camera system.
- the frame rate of the camera module 100 is set low in the gesture mode.
- the frame rate in the gesture mode is 5 fps, which is 1 ⁇ 6 of the frame rate of 30 fps in the video mode.
- the time for transferring data using the USB bus 50 becomes 1 ⁇ 6, the period of time during which the USB bus is in the idle state becomes long.
- the load on the CPU 11 to perform pattern recognition is reduced, and this allows the CPU 11 to make a transition to a deeper sleep state.
- the camera system 10 is suitable for applications requiring that a state of being able to shoot the motion of a moving object performed at unspecified timing is maintained over a long time with low power consumption.
- the gesture recognition method according to the present invention can be applied to various applications.
- One application is to control the power state of an electronic device.
- the power consumption of a device in the idle state is reduced in power-on state (S0 state) specified by ACPI by implementing runtime idle detection or by employing a technique called S0 idle proposed by Microsoft®.
- the CPU 11 transits between the active state and the sleep state periodically or in response to an interrupt while the LCD 17 and many other devices transit to a power saving state when the system is in the idle state.
- the camera system 10 operating in the gesture mode can recognize a gesture to wake up the system, or forcibly make the system transit to the idle state.
- the target to be recognized by the camera system 10 is not limited to a gesture as the motion of a part of the human body, and the present invention can be applied to all aspects of the motion of moving objects. For example, specific motion, such as the figure of an animal passing through a specific place, the motion of a machine, or a change in a natural environment, can be detected as a blurred image to issue a command in order to take a still image or moving image.
- FIG. 8 shows the multiple reference patterns 281 to 287 , but the number of reference patterns may also be one.
- the pattern recognition section 261 has only to be able to distinguish between noise blur and a blurred image of the gesture, the load on the CPU 11 to perform pattern recognition is further reduced.
- the gesture recognition application 201 may be stored in the flash ROM 119 of the camera module 100 so that the MPU 115 will perform recognition and output a command.
- the laptop PC 10 can also be woken up from the suspend state (S3 state) in which the power supply of the CPU 11 is turned off.
- FIG. 11 shows the frame sequence 300 received at the interface section 251 like in FIG. 9 , a differential image group 500 generated by the differential image generating section 257 , and a composite image group 550 generated by the composite image generating section 259 .
- the method in FIG. 11 does not require the background image generating section 253 and the background image storing section 255 .
- the differential image generating section 257 receives each of frames 301 to 309 in order to generate differential images 501 to 509 . At this time, the differential image generating section 257 generates a differential image from two consecutive frames. For example, the differential image 501 is generated from the frame of the background image and the frame 301 , and the differential image 503 is generated from the frame 301 and the frame 303 .
- the composite image generating section 259 calculates a logical sum of a differential image with the latest composite image repeatedly each time receiving each of the differential images 501 to 509 to generate composite images 551 to 559 .
- this method there is no need to generate and update the background image, but a pixel area in which a blurred image of a frame with a shot of the gesture in the outward passage overlaps a blurred image of a frame with a shot of the gesture in the homeward passage cannot be extracted as a binarized blurred image.
- the exposure time can be lengthened so that recognition can be done from a one-way gesture, or the gesture cycle in the outward passage and the homeward passage can be changed.
- the present disclosure provides a method for recognizing motion gesture commands for controlling computer operations.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- General Engineering & Computer Science (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Health & Medical Sciences (AREA)
- Computer Vision & Pattern Recognition (AREA)
- General Health & Medical Sciences (AREA)
- Psychiatry (AREA)
- Social Psychology (AREA)
- Multimedia (AREA)
- Image Analysis (AREA)
- User Interface Of Digital Computer (AREA)
- Studio Devices (AREA)
Abstract
A computer capable of recognizing gesture commands is disclosed. Suppose that a user makes a gesture of swinging a hand from side to side in front of a camera associated with a computer. A camera module receives frames with shots of the gesture in order. The camera module calculates a gradation difference between corresponding pixels of each frame and a background image to generate a set of binarized differential images. The camera module then combines differential images to generate composite images. In response to a determination that any of the composite images matches a reference pattern, the camera module outputs a computer command. The computer command can be used to control the power state of the computer or start a specific application within the computer.
Description
- The present application claims benefit of priority under 35 U.S.C. §§120, 365 to the previously filed Japanese Patent Application No. JP2013-047459 with a priority date of Mar. 11, 2013, which is incorporated by reference herein.
- 1. Technical Field
- The present invention relates to gesture commands in general, and particularly to a method for recognizing motion gesture commands for controlling computer operations.
- 2. Description of Related Art
- A desktop computer, a laptop computer (laptop PC), a tablet computer, or a smartphone is commonly equipped with a web camera capable of projecting real-time video on a monitor to enable a user to make a video phone call or to take a photograph or a video.
- One prior art discloses a method for recognizing a motion pattern of a person from the parameters of a motion blur taken with a camera. For example, the extent of the motion blur can be utilized as an indicator to indicate the speed of an object. In addition, the trajectory of the object can be obtained from variations in the direction of the motion blur. Furthermore, the motion blur is associated with a gesture.
- Another prior art discloses a method for detecting a moving object with high accuracy even when a photographing environment changes in terms of the brightness of an object or when the pictures is taken during a camera shake. The moving object can be detected from a differential image corresponding to a density difference between a first image and a second image input immediately before the first image.
- In accordance with a preferred embodiment of the present invention, a reference pattern is initially defined. A frame is subsequently received from an image sensor that has captured a blurred image of an object. The shape of the blurred image included in the received frame is identified. The reference pattern are then comparing to the identified shape of the blurred image to determine whether or not the identified shape represents a gesture command. In response to a determination that the identified shape represents a gesture command, a corresponding command for controlling an operation of a computer is generated.
- All features and advantages of the present disclosure will become apparent in the following detailed written description.
- The disclosure itself, as well as a preferred mode of use, further objects, and advantages thereof, will best be understood by reference to the following detailed description of an illustrative embodiment when read in conjunction with the accompanying drawings, wherein:
-
FIG. 1 is a block diagram of a laptop PC; -
FIG. 2 is a block diagram of a camera system within the laptop PC fromFIG. 1 ; -
FIG. 3 is a block diagram depicting the software configuration of the camera system fromFIG. 2 ; -
FIGS. 4A-4B are diagrams depicting a relationship between frame rate and exposure time; -
FIG. 5 is a block diagram depicting the configuration of a gesture recognition application; -
FIG. 6 is a flowchart of a procedure through which the camera system fromFIG. 2 recognizes a gesture; -
FIG. 7 is a diagram abstractly depicting a state of a background image and a person present in the foreground, who swings an arm in a cyclic way; -
FIGS. 8A-8D are diagrams depicting the shapes of reference patterns registered in a reference image registering section; -
FIG. 9 is a diagram depicting an example of generating a differential image group and a composite image group from a background image and a frame sequence; -
FIG. 10 is a diagram depicting a relationship between frame transfer period Tf and gesture cycle Tj; and -
FIG. 11 is a diagram depicting an example of generating a differential image data group and a composite image data group from a frame sequence. -
FIG. 1 is a block diagram of acamera system 10 implemented in a laptop PC. ACPU 11 incorporates a memory controller and a PCI Express controller. TheCPU 11 is connected to amain memory 13, avideo card 15, and achipset 19. AnLCD 17 is connected to thevideo card 15. Thechipset 19 incorporates controllers, such as SATA, USB, PCI Express, and LPC, and also incorporates Real-Time Clock (RTC). AnHDD 21 is connected to the SATA controller. The USB controller is configured to include multiple hubs that make up a USB host controller, a root hub, and I/O ports. - A
camera module 100 is a USB device conforming to USB2.0 or USB3.0 standards. Thecamera module 100 is connected to a USB port of the USB controller through a pair or three pairs ofUSB buses 50 to transfer data using a differential signal. The USB port to which thecamera module 100 is connected may share a hub with another USB device. However, it is preferred that the USB port be connected to a hub dedicated to thecamera module 100 to control the power of thecamera module 100 effectively using a selective suspend mechanism of a USB system. - Here, in the selective suspend of the USB system, when a USB client driver determines that a USB device is idle, a request packet is sent to a USB bus driver. When all USB devices connected to the USB hub become idle, the USB bus driver makes USB buses connected to the USB hub transit to an idle state, and USB devices that have detected the state transition of the buses transit to the selective suspend state. Therefore, as the
camera module 100 stays in the idle state for a longer period of time, the period for which the USB buses stay in the idle state becomes longer, and as a result, the period of the transition of thecamera module 100 to the selective suspend state becomes longer. - The
camera module 100 may be of the type of being incorporated in the housing of the laptop PC or an external type module that is connected to a USB connector mounted in the housing of the laptop PC. Thecamera module 100 may also be connected to the laptop PC by wireless USB. Thecamera system 10 according to the present invention can transfer image data in a gesture mode introduced by the present invention in addition to a conventional video mode for taking a video. The gesture mode is a transfer mode used by thegesture recognition application 201 inFIG. 3 to acquire frames in order to recognize a gesture. - In the gesture mode, the
camera module 100 transfers frames at a frame rate lower than the video mode to increase the exposure time so that each frame can capture a motion blur of the gesture. Here, the frame corresponds to one still image unit in image data composed of all horizontal lines. In an interlaced system, one frame is composed of two fields. The details of the gesture mode will be described later. - In addition to the video mode and the gesture mode, the
camera module 100 also operates in a monitoring mode for checking an image upon taking a still image. In the monitoring mode, the number of pixels in a frame can be reduced compared to that in the video mode. The hardware of the laptop PC 10 such as theCPU 11, thechipset 19, and themain memory 13, and software such as thegesture recognition application 201 shown inFIG. 3 , respective layers ofdevice drivers image transfer service 203, and anoperating system 205 are the main components of thecamera system 10. - An EC 25 is a microcontroller for controlling temperature inside the housing of the
computer 10 and controlling the power of devices. TheEC 25 operates independently of theCPU 11. Anacceleration sensor 27 and a DC/DC converter 29 are connected to theEC 25, and a keyboard, a mouse, a charger, an exhaust fan, and the like, which are not shown, are also connected to theEC 25. A battery or an AC/DC adapter, not shown, supplies power to the DC/DC converter 29. The DC/DC converter 29 supplies power to the devices that make up thecomputer 10. -
FIG. 2 is a detailed block diagram of thecamera module 100.FIG. 3 is a block diagram depicting the software configuration of thecamera system 10. In the gesture mode, thecamera module 100 can transfer frames having a screen resolution (number of pixels) such as VGA (640×480), QVGA (320×240), WVGA (800×480), or WQVGA (400×240). Anoptical mechanism 101 is made up of an optical lens, an optical filter, and the like to focus the image of a subject onto animage sensor 103. - The
image sensor 103 is made up of a CMOS image sensor for converting, to an electric signal, an electric charge corresponding to the amount of light accumulated in a photodiode that forms part of a pixel, and outputting the electric signal, a CDS circuit for reducing noise, an AGC circuit for adjusting gain, an AD conversion circuit for converting an analog signal to a digital signal, and the like. Theimage sensor 103 outputs a digital signal corresponding to the brightness of the subject image. When operating in the video mode, theimage sensor 103 can transfer image data at a frame rate of 30 fps. - The CMOS image sensor includes an electronic shutter called a rolling shutter. The rolling shutter uses one line or a few lines as one block unless otherwise specified by the system to control the exposure time so that the exposure time will become best for the photographing environment. During one frame period, or in the case of an interlaced system, the rolling shutter rests, in the process of photographing, a signal charge accumulated during one field period in the photodiode that forms part of each pixel to control the exposure time. Note that the
image sensor 103 can also use a CCD image sensor instead of the CMOS image sensor. - An image signal processor (ISP) 105 is an image signal processing circuit for performing processing, such as correction processing for correcting a defective pixel and shading, white balance processing for correcting the spectral characteristics of the
image sensor 103 according to the human visibility, interpolation processing for outputting common RGB data from signals of an RGB Bayer array, color correction processing for approximating the spectral characteristics of color filters in theimage sensor 103 to ideal characteristics, and the like. TheISP 105 further performs processing, such as contour correction processing for increasing the sense of resolution of the subject, gamma processing for correcting the non-linear input/output characteristics of theLCD 17, and the like. - An
encoder 107 compresses image data received from theISP 105. Anend point buffer 109 temporarily stores data to be transferred with the system bi-directionally to form multiple pipes for USB transfer. A serial interface engine (SIE) 111 packetizes image data received from theend point buffer 109 to meet the USB standards and sends the packetized image data to atransceiver 113, or analyzes a packet received from thetransceiver 113 and sends a payload to anMPU 115. - When a
USB bus 50 is in the idle state for more than a predetermined period of time, theSIE 111 interrupts theMPU 115 to make theUSB bus 50 transit to the selective suspend state. Then, theSIE 111 operates theMPU 115 suspended when theUSB bus 50 is resumed. Thetransceiver 113 includes a transmission transceiver and a reception transceiver for USB communication. TheMPU 115 performs enumeration for USB transfer, and controls the operation of thecamera module 100 for further taking an image and transferring image data. Thecamera module 100 conforms to the power management specified in the USB standards. When interrupted by theSIE 111, theMPU 115 can disrupt an internal clock to make thecamera module 100 including itself transit to the suspended state. - The
MPU 115 can return thecamera module 100 to the power-on state or a photographing state when theUSB bus 50 is resumed. TheMPU 115 interprets a command received from the system to control the operation of each component in order to transfer image data in the video mode (including the monitoring mode) or the gesture mode. When theUSB bus 50 is in the idle state upon transferring frames in the gesture mode, theMPU 115 sends a resume signal to the USB bus driver 211 to release the idle state. - When sending the MPU 115 a command for enabling the gesture mode, the
gesture recognition application 201 can specify various parameters to set an exposure time. In one method, thegesture recognition application 201 specifies only the frame rate for transferring still images to thecamera module 100. After performing calibration, such as the exposure time of the rolling shutter, the white balance, and the gain of the AGC circuit, theMPU 115 sets the maximum exposure time available for the current photographing conditions on the frame rate specified. - The
MPU 115 further sets, in theimage sensor 103 and a predetermined register in theISP 105, other parameters to be optimized for the current photographing conditions on the set exposure time. In another method, thegesture recognition application 201 specifies several parameters that affect the taking of a blurred image, such as exposure time, ACG gain, gamma curve, contrast, and aperture value. TheMPU 115 sets other parameters to be optimized for the current photographing conditions on the specified parameters. When an exposure time is specified, theMPU 115 can set a frame rate according to the exposure time. - The situation will be described with reference to
FIGS. 4A and 4B .FIG. 4A shows a state of transferring image data at a frame rate F1 of 30 fps in the video mode, andFIG. 4B shows a state of transferring image data at a frame rate F2 of 5 fps as an example in the gesture mode. Here, a description will be made by taking a non-interlaced system as an example, but the present invention may also employ a camera module for outputting image data in an interlaced system. - In
FIG. 4A , first to N lines are scanned sequentially in each frame transfer period Tf1 (1/F1) to output data. The time required for each line is composed of anexposure time 151 and areading time 153. InFIG. 4B , first to N lines are scanned sequentially in a frame transfer period Tf2 (1/F2) to output data. Each line is composed of anexposure time 155 and areading time 153. - When the frame rate F2 is ⅙ of F1 and the frame transfer period Tf2 is sixfold, the exposure time can be prolonged up to about sixfold. The longer the exposure time, the larger the area of a blurred image when a moving object is shot. As will be described later, the area of the blurred image corresponds to the amount of information used for pattern recognition. Here, the blurred image is an image formed by the contour of the moving object shot in one frame. The
gesture recognition application 201 identifies the shape of the blurred image from the contour of the blurred image. In the embodiment, the moving object is an arm of a person who makes a gesture as an example, but the present invention is not limited thereto. The moving object can be any other object such as a moving machine or natural object. - In the gesture mode, the
camera module 100 transits to the selective suspend state when no frame is transferred, and this may result in loss of the set parameters. TheMPU 115 can also store, in aflash ROM 119, the parameters set once, and set the parameters by reading them from theflash ROM 119 in each frame transfer period. The flash ROM stores a program executed by theMPU 115. TheMPU 115 can interpret a command received from the system to operate thecamera module 100 in the gesture mode or the video mode. - The
camera module 100 is a bus-powered device that operates by receiving the supply of power from theUSB bus 50. In this regard, however, thecamera module 100 may also be a self-powered device that operates with power supplied by itself. In the case of a self-powered device, theMPU 115 controls the power supplied independently to follow the state of theUSB bus 50. - The
camera module 100 transits between the selective suspend state and the power-on state based on the USB standards. Thecamera module 100 may transit to two or more low-power states during the period of the selective suspend state. When operating in the gesture mode, if the frame transfer period is long, thecamera module 100 can operate only theimage sensor 103 during a period of sending no frame to transit to a low-power state with least power consumption. - Note that
FIGS. 1 and 2 simply show the main hardware configuration and connection relations associated with the embodiment in a simplified manner to describe the embodiment. In addition to those mentioned thus far, many other devices are used to make up thecamera system 10. However, since these devices are known to those skilled in the art, the description thereof will not be given in detail here. The present invention contains a configuration in which multiple blocks shown in the figure are formed as one integrated circuit or device, and conversely, a configuration in which one block is divided into multiple integrated circuits or devices as long as an arbitrary selection can be made by those skilled in the art. For example, the interface of thecamera module 100 may be configured as a Mobile Industry Processor Interface (MIPI) and the system may be configured as a system-on-a-chip (SOC). - Software shown in
FIG. 3 is stored in theHDD 21, loaded into themain memory 13, and executed by theCPU 11. A movingimage application 204 is a known program for acquiring video data from thecamera module 100 to display the video on theLCD 17 or recording it in theHDD 21. The movingimage application 204 can be a video-phone program or a moving image shooting program. The movingimage application 204 can also be a gesture engine for performing sophisticated recognition of a gesture by a conventional method. - The
gesture recognition application 201 is a novel program for acquiring a frame sequence from thecamera module 100 in the gesture mode, performing pattern recognition of a blurred image, and outputting a command corresponding to the recognized gesture. Note that the frame sequence means a frame group of continuous still images transferred at a frame rate low enough to produce pictures unnatural as video data. As an example, the power state of thelaptop PC 10 can be controlled or the movingimage application 204 can be started by the command output from thegesture recognition application 201. - The
gesture recognition application 201 can include a user interface for allowing a user to set parameters of thecamera system 10 to be applied to the gesture mode, set the gesture mode to enabled, and the like. Thegesture recognition application 201 may be executed only when the laptop PC is receiving the supply of power from the AC/DC adapter. Since thecamera system 10 operates with low power consumption, this is particularly suitable for operating thecamera system 10 while power is being supplied from a battery pack. - A
streaming service 206 is a service program provided by theOS 205 to send frames to the movingimage application 204 in the video mode or the monitoring mode, and pass, to theUSB camera driver 207, a command issued by the movingimage application 204. A stillimage transfer service 203 is a program that runs in a user mode of theOS 205 to send the gesture recognition application 201 a frame sequence transferred in the gesture mode, and send the USB camera driver 207 a command and parameters received from thegesture recognition application 201. - The
USB camera driver 207 is a device driver for controlling the operation of thecamera module 100 and controlling data transfer. TheUSB class driver 209 is a device driver for performing common processing defined in the USB video class. The USB bus driver 211 controls the operation of the USB buses connected to the USB controller. - When receiving an instruction from the
USB camera driver 207 to make a transition of thecamera module 100 to selective suspend, the USB bus driver 211 makes theUSB bus 50 connected to thecamera module 100 transit to the idle state, while when receiving an instruction to resume theUSB bus 50 or an instruction for data transfer, the USB bus driver 211 makes the USB bus transit to an active state (resumed state). The USBhost controller driver 213 controls data transfer to thecamera module 100 and the operation of the USB host controller. -
FIG. 5 is a block diagram depicting the configuration of thegesture recognition application 201. Aninterface section 251 sends a differentialimage generating section 257 frames received from the stillimage transfer service 203 sequentially in the gesture mode. When the acceleration value detected by theacceleration sensor 27 is lowered to a value less than or equal to a predetermined value after exceeding the predetermined value, theOS 205 gives a notification indicating that the background image to be taken with thecamera module 100 has changed due to a change in the position of the laptop PC. - The
interface section 251 that received the notification stores, in a backgroundimage storing section 255, multiple frames received from the stillimage transfer service 203. In addition to this, theinterface section 251 acquires multiple frames periodically in the gesture mode and stores the acquired frames in the backgroundimage storing section 255. Theinterface section 251 sends the stillimage transfer service 203 parameters set by the user, such as enable/disable setting of the gesture mode, frame rate, exposure time, contrast, and AGC gain. - A background
image generating section 253 generates a background image from the frames received from theinterface section 251. When a moving object is shot, still images taken in the gesture mode contain a blurred image. The backgroundimage generating section 253 calculates and binarizes a difference between gradation values of corresponding pixels of received multiple frames possible to include the blurred image to identify the blurred image. The backgroundimage generating section 253 calculates a logical sum of multiple background images except the blurred image identified from the frames, generates a background image without including the blurred image, and stores the background image in the backgroundimage storing section 255. - The differential
image generating section 257 calculates a gradation difference between corresponding pixels of the frame received from theinterface section 251 and the frame of the background image received from the backgroundimage storing section 255, generates a binarized differential image, and sends the differential image to a compositeimage generating section 259. The compositeimage generating section 259 that received the differential image calculates a logical sum to combine differential images in order to generate a composite image, and sends the composite image to apattern recognition section 261. A reference image registering section 265 registers multiple reference patterns associated with predetermined commands. - The
pattern recognition section 261 extracts, using a known algorithm, the features of the composite image generated by the compositeimage generating section 259, and compares the composite image with the multiple reference patterns registered in the referenceimage registering section 263. When determining that the composite image matches either of the reference patterns, thepattern recognition section 261 outputs, to theOS 205, a command associated with the reference pattern. Each component shown inFIG. 5 is a hardware component that cooperates with hardware such as theCPU 11 and themain memory 13 to provide the laptop PC with a specific function. - Next, a procedure for recognizing a gesture by a simple method according to the embodiment to enable the
camera system 10 to activate a gesture engine for performing sophisticated gesture recognition requiring large power consumption will be described with reference to a flowchart ofFIG. 6 .FIG. 7 abstractly shows a state of abackground image 291 and a person present in the foreground, who swings anarm 293 in a cyclic way as an example of the moving object that generates a blurred image.FIG. 8A-8D show the shapes of reference patterns registered in the referenceimage registering section 263.FIG. 8A corresponds to a reference pattern of a blurred image in which a person facing a camera swings thearm 293 from side to side with the arm down using a shoulder as a pivot point. -
FIG. 8B corresponds to a reference pattern of a blurred image in which the person facing the camera swings thearm 293 from side to side toward thecamera module 100, where the arm extends forward with the elbow bent.FIGS. 8C and 8D correspond to reference patterns of blurred images in each of which the person with the face turned to the left side or right side of thecamera module 100 swings the arm back and forth toward himself or herself, where the arm extends forward with the elbow bent.FIG. 9 is a diagram depicting an example of generating adifferential image group 350 and acomposite image group 400 from thebackground image 291 and aframe sequence 300. - In
block 601, theinterface section 251 sends the camera module 100 a command for enabling the gesture mode and a parameter associated with the exposure time. Inblock 603, thecamera module 100 shifts into the gesture mode and sets the specified parameter and other appropriate parameters. Inblock 605, thecamera module 100 sets the frame rate, for example, to 5 fps to set the exposure time to the maximum value available in the range in order to start the transfer of theshot frame sequence 300. - The user makes a gesture toward the
camera module 100 at proper timing. Since the timing of starting the gesture is not synchronized with the timing of transferring frames, taking images of the gesture is started at any time with respect to the frame transfer period. Inblock 607, at the timing of updating the background image, the procedure shifts to block 631. The timing of updating the background image comes when theacceleration sensor 27 detects the movement of the laptop PC or in predetermined regular periods. Inblock 631, the backgroundimage generating section 253 stores, in the backgroundimage storing section 255, multiple frames newly generated. - In
block 609, the differentialimage generating section 257 receivesframes 301 to 309 constituting theframe sequence 300 in order. Theframes 301 to 309 are images of a gesture taken when the user extends the arm forward in front of the camera and moves the arm once from right to left using the elbow as a pivot point. When the exposure time is set properly for the cycle of the gesture, each frame contains each of blurredimages 301 a to 309 a corresponding to the motion of the arm that makes the gesture. The differentialimage generating section 257 calculates gradation differences of each of theframes 301 to 309 from corresponding pixels of the latest background image stored in the backgroundimage storing section 255 at that point to generate binarizeddifferential images 351 to 359. - The calculated gradation differences show large values for pixels the representation of which are changing like the
blurred images 301 a to 309 a and small values for pixels to represent thebackground image 291 with no change. Therefore, before taking the images of the gesture, since theinterface section 251 receives only thebackground image 291, the gradation of differential images take meaninglessly small values on most of the pixels. - The differential
image generating section 257 sets a fixed threshold value for each pixel for which a gradation difference is calculated to binarize the gradation of each pixel. For example, the differentialimage generating section 257 setslogical value 1 for pixels that exceed the threshold value and logical value 0 for pixels less than the threshold value. Thebackground image 291 is removed from thedifferential images 351 to 359, and theblurred images 301 a to 309 a are converted to binarizeddifferential images 351 a to 359 a. - The differential
image generating section 257 sends thedifferential images 351 to 359 to the compositeimage generating section 259. Thedifferential images 351 to 359 also contain blurred images (noise blur) as noise caused when theframes 301 to 309 are shot along with the motion of any other object unrelated to the gesture. The noise blur reduces the accuracy of pattern recognition. However, since thereference patterns 281 to 287 have relatively simple shapes distinguishable from the noise blur, the influence thereof is eliminated in the process of pattern recognition. The differentialimage generating section 257 generates each of thedifferential images 351 to 359 and sends the generated differential image to the compositeimage generating section 259 each time the differentialimage generating section 257 receives each of theframes 301 to 309 in the return path fromblock 615 to block 605 until the gesture is recognized. - In
block 613, the compositeimage generating section 259 calculates a logical sum of logical values set for each pixel of the received one of thedifferential images 351 to 359 and each pixel of the differential image last received to generate thecomposite image group 400. Since an image with which thedifferential image 351 is calculated to obtain the logical sum does not contain any binarized blurred image, thedifferential image 351 is exactly the same as acomposite image 401. After generating thecomposite image 401, the compositeimage generating section 259 calculates a logical sum of thedifferential image 353 next received and thecomposite image 401 last generated to generate a newcomposite image 403. Then, each ofcomposite images 405 to 409 is generated each time each of thedifferential images 355 to 357 is added to increase the amount of information on blurredimages 301 a to 309 a, approaching the shape of areference pattern 283. - The
composite image 409 has a shape almost close to the shape of thereference pattern 283. Inblock 615, thepattern recognition section 261 compares, using a known algorithm, the features of a composite image with each of thereference patterns 281 to 287 each time each of thecomposite images 401 to 409 is received. The features can include, for example, the direction of being opened into a fan shape, and the center angle or the area. Since thereference patterns 281 to 287 are relatively simple patterns, the load on theCPU 11 that performs pattern recognition can be reduced. - The
pattern recognition section 261 can also complement portions missing in thecomposite image reference pattern 283. When determining that any of thecomposite images 401 to 409 matches either of thereference patterns 281 to 287, thepattern recognition section 261 sends the OS 205 a command associated with the reference pattern inblock 633. Inblock 635, theOS 205 activates the gesture engine that performs sophisticated processing for manipulating the laptop PC. - Returning from
block 635 to block 605, theinterface section 251 further receives a frame sequence. Immediately after a command is output, since the composite images that have been generated until then remain in the compositeimage generating section 259, there is a need to exclude the composite images from recognition targets. Thepattern recognition section 261 can ignore a predetermined number of composite images generated after the command is output. After that, thepattern recognition section 261 starts the recognition of a new gesture again at the time of receiving a composite image without the influence of the binarized blurred images as the basis for outputting the previous command. In the gesture mode, since the exposure time is made long intentionally, a gesture can be recognized without being affected by photographing environments ranging from bright places to dark places. - The recognition target in the present invention is the shape formed by the entire contour of a blurred image. Although the speed of the gesture and the exposure time do not affect the shape of the contour directly, when they are selected properly, it is advantageous in terms of the number of times of a gesture required to recognize the gesture, the recognition time, the power consumption, and photographing in darkness.
FIG. 10 is a diagram for describing a relationship between frame transfer period Tf and gesture cycle Tj. - Suppose that the gesture is the motion of an arm swinging from position P1 to position P2 in the outward passage and swinging from position P2 to position P1 in the homeward passage. In this case, if it is assumed that the frame transfer period Tf as the reciprocal of the frame rate F is almost equal to the exposure time, the amount of information usable for comparison with the reference pattern of a blurred image included in one frame increases as the frame transfer period Tf is lengthened. However, the longer the frame transfer period Tf, the longer the time from the start of the gesture until the
gesture recognition application 201 acquires the first frame, resulting in a delay in the recognition of the gesture. - It can be said that a blurred image the inside of which is filled over the entire part of the outward passage (from position P1 to position P2) or the homeward passage (from position P2 to position P1) of the gesture is an adequate amount of information necessary for comparison with the reference pattern. However, when a complementary method is employed or the number of reference patterns is small, the recognition can be performed accurately without necessarily providing such an amount of information. When information other than the information already obtained in the outward passage or the homeward passage cannot be obtained even if the frame transfer period Tf is set longer, an upper limit can be set to the frame transfer period Tf because only a delay is caused.
- Here, the upper limit of the frame transfer period Tf can be defined as the minimum time capable of obtaining an adequate amount of information usable for recognition. Note that the time t0 for starting the gesture cannot be synchronized with the frame transfer timing. In the example of
FIG. 10 , a frame for which a frame transfer period Tf1 shorter than the gesture cycle Tj is set, and the shooting of which is started at time t1 and finished at time t2 lacks information on a blurred image from position P1 to position P3 of the gesture in the outward passage and the homeward passage. - The information can obtain at time t3 at which shooting of the gesture in the second outward passage is finished. On the other hand, a frame for which a frame transfer period Tf2 equal to the gesture cycle Tj is set, and the shooting of which is started at time t1 and finished at time t4 contains information on the blurred image obtained by shooting a total extent of the gesture in the first homeward passage. Therefore, even if the frame transfer period Tf is set longer than the gesture cycle Tj, information more useful to add to a composite image than the information already obtained cannot be obtained, leading only to lengthening the time for recognition. Although there is no need to limit the invention, the maximum value of the frame transfer period Tf (the minimum value of the frame rate) can be so determined that the gesture cycle Tj will match the exposure time.
- The present invention can also recognize the direction of the gesture to increase the number of commands. To this end, the
gesture recognition application 201 recognizes which of a gesture starting from position P1 or a gesture starting from position P2 inFIG. 10 has been made. In this case, the maximum value of the frame transfer period Tf can be set to Tf<Tj/4 to recognize the gesture starting position. - When the frame transfer period Tf is set shorter, it is difficult to recognize the gesture in a dark place, increasing the power consumption of the system. Acceptable power consumption of the camera system differs from system to system. Therefore, the minimum value of the frame transfer period Tf can be determined for each system in consideration of these points. Since the parameters set in the
camera module 100 to recognize a gesture depends on the performance of thecamera system 10 such as thecamera module 100 and thepattern recognition section 261, the photographing environment, the number of reference patterns, the allowable number of times of a gesture, and the like, the parameters can be determined by experiment for each camera system. - Next, the power saving function of the
camera system 10 will be described. The frame rate of thecamera module 100 is set low in the gesture mode. In the example ofFIG. 4 , the frame rate in the gesture mode is 5 fps, which is ⅙ of the frame rate of 30 fps in the video mode. As a result, the period of time during which thecamera module 100 does not transfer any frame is increased to lengthen the period of time during which thecamera module 100 is in the selective suspend state, reducing the power consumption. - Further, since the time for transferring data using the
USB bus 50 becomes ⅙, the period of time during which the USB bus is in the idle state becomes long. In addition to the low frame rate in the gesture mode, when the reference patterns are limited to relatively simple several patterns, the load on theCPU 11 to perform pattern recognition is reduced, and this allows theCPU 11 to make a transition to a deeper sleep state. Thus, since the power consumption of thecamera module 100, theUSB bus 50, and theCPU 11 is reduced, thecamera system 10 is suitable for applications requiring that a state of being able to shoot the motion of a moving object performed at unspecified timing is maintained over a long time with low power consumption. - Using the feature of being able to perform pattern recognition while standing by with low power consumption without being affected by such a photographing environment, the gesture recognition method according to the present invention can be applied to various applications. One application is to control the power state of an electronic device. In the case of a laptop PC, there is a case where the power consumption of a device in the idle state is reduced in power-on state (S0 state) specified by ACPI by implementing runtime idle detection or by employing a technique called S0 idle proposed by Microsoft®.
- In detecting a runtime idle state, the
CPU 11 transits between the active state and the sleep state periodically or in response to an interrupt while theLCD 17 and many other devices transit to a power saving state when the system is in the idle state. At this time, thecamera system 10 operating in the gesture mode can recognize a gesture to wake up the system, or forcibly make the system transit to the idle state. The target to be recognized by thecamera system 10 is not limited to a gesture as the motion of a part of the human body, and the present invention can be applied to all aspects of the motion of moving objects. For example, specific motion, such as the figure of an animal passing through a specific place, the motion of a machine, or a change in a natural environment, can be detected as a blurred image to issue a command in order to take a still image or moving image. -
FIG. 8 shows themultiple reference patterns 281 to 287, but the number of reference patterns may also be one. In such a case, since thepattern recognition section 261 has only to be able to distinguish between noise blur and a blurred image of the gesture, the load on theCPU 11 to perform pattern recognition is further reduced. While the example of realizing thecamera system 10 in the laptop PC has been described, thegesture recognition application 201 may be stored in theflash ROM 119 of thecamera module 100 so that theMPU 115 will perform recognition and output a command. In such a case, thelaptop PC 10 can also be woken up from the suspend state (S3 state) in which the power supply of theCPU 11 is turned off. - The example of generating the
differential image group 350 and thecomposite image group 400 from thebackground image 291 and theframe sequence 300 is described above with reference toFIG. 9 . The following will describe another method of generating a composite image with reference toFIG. 11 .FIG. 11 shows theframe sequence 300 received at theinterface section 251 like inFIG. 9 , adifferential image group 500 generated by the differentialimage generating section 257, and acomposite image group 550 generated by the compositeimage generating section 259. The method inFIG. 11 does not require the backgroundimage generating section 253 and the backgroundimage storing section 255. - The differential
image generating section 257 receives each offrames 301 to 309 in order to generatedifferential images 501 to 509. At this time, the differentialimage generating section 257 generates a differential image from two consecutive frames. For example, thedifferential image 501 is generated from the frame of the background image and theframe 301, and thedifferential image 503 is generated from theframe 301 and theframe 303. - The composite
image generating section 259 calculates a logical sum of a differential image with the latest composite image repeatedly each time receiving each of thedifferential images 501 to 509 to generatecomposite images 551 to 559. In this method, there is no need to generate and update the background image, but a pixel area in which a blurred image of a frame with a shot of the gesture in the outward passage overlaps a blurred image of a frame with a shot of the gesture in the homeward passage cannot be extracted as a binarized blurred image. In this case, the exposure time can be lengthened so that recognition can be done from a one-way gesture, or the gesture cycle in the outward passage and the homeward passage can be changed. - As has been described, the present disclosure provides a method for recognizing motion gesture commands for controlling computer operations.
- While the disclosure has been particularly shown and described with reference to a preferred embodiment, it will be understood by those skilled in the art that various changes in form and detail may be made therein without departing from the spirit and scope of the disclosure.
Claims (14)
1. A method comprising:
defining a reference pattern;
receiving a frame from an image sensor that has captured a blurred image of an object;
identifying a shape of said blurred image included in said received frame;
comparing said identified shape of said blurred image with said reference pattern to determine whether or not said identified shape represents a gesture command; and
in response to a determination that said identified shape represents a gesture command, generating a corresponding command for controlling an operation of a computer.
2. The method of claim 1 , wherein said receiving further includes receiving a plurality of consecutive frames of said object taken at a predetermined frame rate.
3. The method of claim 2 , wherein said identifying further includes
identifying shapes of blurred images in said plurality of consecutive frames;
combining said shapes of said blurred images respectively included in said plurality of consecutive frames to generate a composite image.
4. The method of claim 3 , wherein said identifying further includes comparing a background image with said images in said consecutive frames to determine a difference between gradation values of corresponding pixels in order to generate a binarized differential image.
5. The method of claim 3 , wherein said identifying further includes comparing said frames with each other to determine a difference between gradation values of corresponding pixels in order to generate a binarized differential image.
6. The method of claim 5 , wherein said comparing further includes determining a logical sum for said binarized differential image to generate said composite image.
7. A method comprising:
defining a reference pattern with which a computer command is associated;
setting an exposure time of an image sensor to capture a blurred image of a motion gesture;
receiving, from said image sensor, a frame with said motion gesture shot therein;
identifying a shape of said blurred image from said frame;
determining whether or not said shape of said blurred image matches said reference pattern; and
in response to a determination that said shape of said blurred image matches said reference pattern, sending a corresponding computer command to a computer.
8. The method of claim 7 , further comprising changing a power state of said computer in response to said computer command.
9. The method of claim 7 , further comprising causing said computer to start executing an application program in response to said computer command.
10. The method of claim 7 , wherein a shape of said reference pattern is a fan shape that forms a blurred image as a result of reciprocating motion of an arm using an elbow or a shoulder as a pivot point.
11. A computer comprising:
a camera system;
a reference image registering block for storing a reference pattern with which a command is associated;
a blurred image processing block for identifying a shape of a blurred image of a moving object from a frame with the object shot therein; and
a pattern recognition block for
comparing said identified shape of said blurred image with said reference pattern;
determining whether or not said shape of said blurred image matches said reference pattern; and
generating, in response to a determination that said shape of said blurred image matches said reference pattern, a corresponding computer command.
12. The computer of claim 11 , wherein said blurred image processing block includes
a differential image generating section for determining a gradation difference between corresponding pixels of frames to generate a binarized differential image; and
a composite image generating block for determining a logical sum for the differential image to generate a composite image.
13. The computer of claim 11 , further comprising:
a background image generating section for generating a background image of said object;
a differential image generating section for calculating a gradation difference between corresponding pixels of said background image and each frame to generate a binarized differential image; and
a composite image generating section for calculating a logical sum for said differential image to generate a composite image.
14. The computer of claim 13 , wherein when said computer detects an acceleration greater than or equal to a predetermined value, said background image generating block updates the background image.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US15/582,338 US9864436B2 (en) | 2013-03-11 | 2017-04-28 | Method for recognizing motion gesture commands |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2013-047459 | 2013-03-11 | ||
JP2013047459A JP5782061B2 (en) | 2013-03-11 | 2013-03-11 | Method for recognizing movement of moving object and portable computer |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/582,338 Division US9864436B2 (en) | 2013-03-11 | 2017-04-28 | Method for recognizing motion gesture commands |
Publications (1)
Publication Number | Publication Date |
---|---|
US20140254870A1 true US20140254870A1 (en) | 2014-09-11 |
Family
ID=51487882
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/171,029 Abandoned US20140254870A1 (en) | 2013-03-11 | 2014-02-03 | Method for recognizing motion gesture commands |
US15/582,338 Active US9864436B2 (en) | 2013-03-11 | 2017-04-28 | Method for recognizing motion gesture commands |
Family Applications After (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/582,338 Active US9864436B2 (en) | 2013-03-11 | 2017-04-28 | Method for recognizing motion gesture commands |
Country Status (2)
Country | Link |
---|---|
US (2) | US20140254870A1 (en) |
JP (1) | JP5782061B2 (en) |
Cited By (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
TWI614608B (en) * | 2014-12-16 | 2018-02-11 | 英特爾公司 | Apparatus, system and method for communication of touch sensor information |
US10146318B2 (en) | 2014-06-13 | 2018-12-04 | Thomas Malzbender | Techniques for using gesture recognition to effectuate character selection |
US20180348841A1 (en) * | 2015-11-11 | 2018-12-06 | Ams Ag | Method, optical sensor arrangement and computer program product for passive optical motion detection |
US20190020813A1 (en) * | 2017-07-14 | 2019-01-17 | Casio Computer Co., Ltd. | Image Recording Apparatus, Image Recording Method, and Computer-Readable Storage Medium |
CN109348124A (en) * | 2018-10-23 | 2019-02-15 | Oppo广东移动通信有限公司 | Image transfer method, device, electronic equipment and storage medium |
US10870436B2 (en) | 2018-07-24 | 2020-12-22 | Honda Motor Co., Ltd. | Operation assistance system and operation assistance method |
US10963064B2 (en) * | 2018-06-19 | 2021-03-30 | Honda Motor Co., Ltd. | Control device and control method |
US11320914B1 (en) * | 2020-11-30 | 2022-05-03 | EMC IP Holding Company LLC | Computer interaction method, device, and program product |
Families Citing this family (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP5980879B2 (en) * | 2014-10-28 | 2016-08-31 | レノボ・シンガポール・プライベート・リミテッド | Gesture recognition method, gesture input system, and electronic device |
CN107533359B (en) * | 2015-05-20 | 2019-04-23 | 三菱电机株式会社 | Information processing unit and interlocking control method |
JP6868478B2 (en) * | 2017-06-16 | 2021-05-12 | 株式会社Nttドコモ | Code video linkage device and code video linkage program |
JP7109736B2 (en) * | 2018-06-28 | 2022-08-01 | 株式会社東京精密 | Auxiliary device and method |
Citations (39)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4878843A (en) * | 1988-06-08 | 1989-11-07 | Kuch Nina J | Process and apparatus for conveying information through motion sequences |
US6128003A (en) * | 1996-12-20 | 2000-10-03 | Hitachi, Ltd. | Hand gesture recognition system and method |
US20020036617A1 (en) * | 1998-08-21 | 2002-03-28 | Timothy R. Pryor | Novel man machine interfaces and applications |
US20020118880A1 (en) * | 2000-11-02 | 2002-08-29 | Che-Bin Liu | System and method for gesture interface |
US6647131B1 (en) * | 1999-08-27 | 2003-11-11 | Intel Corporation | Motion detection using normal optical flow |
US6654483B1 (en) * | 1999-12-22 | 2003-11-25 | Intel Corporation | Motion detection using normal optical flow |
US6766036B1 (en) * | 1999-07-08 | 2004-07-20 | Timothy R. Pryor | Camera based man machine interfaces |
US20040193413A1 (en) * | 2003-03-25 | 2004-09-30 | Wilson Andrew D. | Architecture for controlling a computer using hand gestures |
US6819782B1 (en) * | 1999-06-08 | 2004-11-16 | Matsushita Electric Industrial Co., Ltd. | Device and method for recognizing hand shape and position, and recording medium having program for carrying out the method recorded thereon |
US20050271279A1 (en) * | 2004-05-14 | 2005-12-08 | Honda Motor Co., Ltd. | Sign based human-machine interaction |
US7036094B1 (en) * | 1998-08-10 | 2006-04-25 | Cybernet Systems Corporation | Behavior recognition system |
US20060210188A1 (en) * | 2004-11-29 | 2006-09-21 | Seiko Epson Corporation | Image correcting method, image correcting apparatus, and storage medium having program stored thereon |
US20070124702A1 (en) * | 2005-11-25 | 2007-05-31 | Victor Company Of Japan, Ltd. | Method and apparatus for entering desired operational information to devices with the use of human motions |
US20080024619A1 (en) * | 2006-07-27 | 2008-01-31 | Hiroaki Ono | Image Processing Apparatus, Image Processing Method and Program |
US20080085048A1 (en) * | 2006-10-05 | 2008-04-10 | Department Of The Navy | Robotic gesture recognition system |
US20080136775A1 (en) * | 2006-12-08 | 2008-06-12 | Conant Carson V | Virtual input device for computing |
US20090110292A1 (en) * | 2007-10-26 | 2009-04-30 | Honda Motor Co., Ltd. | Hand Sign Recognition Using Label Assignment |
US20110135148A1 (en) * | 2009-12-08 | 2011-06-09 | Micro-Star Int'l Co., Ltd. | Method for moving object detection and hand gesture control method based on the method for moving object detection |
US20110142349A1 (en) * | 2009-06-30 | 2011-06-16 | Sony Corporation | Information processing apparatus and information processing method |
US20110239118A1 (en) * | 2010-03-25 | 2011-09-29 | Sony Corporation | Gesture input device, gesture input method, and program |
US20110286676A1 (en) * | 2010-05-20 | 2011-11-24 | Edge3 Technologies Llc | Systems and related methods for three dimensional gesture recognition in vehicles |
US20120019728A1 (en) * | 2010-07-26 | 2012-01-26 | Darnell Janssen Moore | Dynamic Illumination Compensation For Background Subtraction |
US20120019684A1 (en) * | 2009-01-30 | 2012-01-26 | Thomson Licensing | Method for controlling and requesting information from displaying multimedia |
US20120129605A1 (en) * | 2010-11-19 | 2012-05-24 | Total Immersion | Method and device for detecting and tracking non-rigid objects in movement, in real time, in a video stream, enabling a user to interact with a computer system |
US20120200494A1 (en) * | 2009-10-13 | 2012-08-09 | Haim Perski | Computer vision gesture based control of a device |
US20120275686A1 (en) * | 2011-04-29 | 2012-11-01 | Microsoft Corporation | Inferring spatial object descriptions from spatial gestures |
US20130027300A1 (en) * | 2011-07-29 | 2013-01-31 | Toshiaki Nakasu | Recognition apparatus, method, and computer program product |
US20130050076A1 (en) * | 2011-08-22 | 2013-02-28 | Research & Business Foundation Sungkyunkwan University | Method of recognizing a control command based on finger motion and mobile device using the same |
US20130077831A1 (en) * | 2011-09-26 | 2013-03-28 | Sony Corporation | Motion recognition apparatus, motion recognition method, operation apparatus, electronic apparatus, and program |
US20130154919A1 (en) * | 2011-12-20 | 2013-06-20 | Microsoft Corporation | User control gesture detection |
US20130273968A1 (en) * | 2008-08-19 | 2013-10-17 | Digimarc Corporation | Methods and systems for content processing |
US20130293722A1 (en) * | 2012-05-07 | 2013-11-07 | Chia Ming Chen | Light control systems and methods |
US20130329948A1 (en) * | 2012-06-06 | 2013-12-12 | Fujitsu Limited | Subject tracking device and subject tracking method |
US20140118257A1 (en) * | 2012-10-29 | 2014-05-01 | Amazon Technologies, Inc. | Gesture detection systems |
US20140118244A1 (en) * | 2012-10-25 | 2014-05-01 | Pointgrab Ltd. | Control of a device by movement path of a hand |
US20140337807A1 (en) * | 2011-12-09 | 2014-11-13 | Sony Corporation | Information processing apparatus, information processing method, and recording medium |
US8891868B1 (en) * | 2011-08-04 | 2014-11-18 | Amazon Technologies, Inc. | Recognizing gestures captured by video |
US20150092040A1 (en) * | 2013-10-01 | 2015-04-02 | Broadcom Corporation | Gesture-Based Industrial Monitoring |
US9164589B2 (en) * | 2011-11-01 | 2015-10-20 | Intel Corporation | Dynamic gesture based short-range human-machine interaction |
Family Cites Families (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20100046796A1 (en) * | 2005-06-30 | 2010-02-25 | Koninklijke Philips Electronics, N.V. | method of recognizing a motion pattern of an object |
WO2008139399A2 (en) * | 2007-05-15 | 2008-11-20 | Philips Intellectual Property & Standards Gmbh | Method of determining motion-related features and method of performing motion classification |
US8565535B2 (en) * | 2007-08-20 | 2013-10-22 | Qualcomm Incorporated | Rejecting out-of-vocabulary words |
-
2013
- 2013-03-11 JP JP2013047459A patent/JP5782061B2/en active Active
-
2014
- 2014-02-03 US US14/171,029 patent/US20140254870A1/en not_active Abandoned
-
2017
- 2017-04-28 US US15/582,338 patent/US9864436B2/en active Active
Patent Citations (39)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4878843A (en) * | 1988-06-08 | 1989-11-07 | Kuch Nina J | Process and apparatus for conveying information through motion sequences |
US6128003A (en) * | 1996-12-20 | 2000-10-03 | Hitachi, Ltd. | Hand gesture recognition system and method |
US7036094B1 (en) * | 1998-08-10 | 2006-04-25 | Cybernet Systems Corporation | Behavior recognition system |
US20020036617A1 (en) * | 1998-08-21 | 2002-03-28 | Timothy R. Pryor | Novel man machine interfaces and applications |
US6819782B1 (en) * | 1999-06-08 | 2004-11-16 | Matsushita Electric Industrial Co., Ltd. | Device and method for recognizing hand shape and position, and recording medium having program for carrying out the method recorded thereon |
US6766036B1 (en) * | 1999-07-08 | 2004-07-20 | Timothy R. Pryor | Camera based man machine interfaces |
US6647131B1 (en) * | 1999-08-27 | 2003-11-11 | Intel Corporation | Motion detection using normal optical flow |
US6654483B1 (en) * | 1999-12-22 | 2003-11-25 | Intel Corporation | Motion detection using normal optical flow |
US20020118880A1 (en) * | 2000-11-02 | 2002-08-29 | Che-Bin Liu | System and method for gesture interface |
US20040193413A1 (en) * | 2003-03-25 | 2004-09-30 | Wilson Andrew D. | Architecture for controlling a computer using hand gestures |
US20050271279A1 (en) * | 2004-05-14 | 2005-12-08 | Honda Motor Co., Ltd. | Sign based human-machine interaction |
US20060210188A1 (en) * | 2004-11-29 | 2006-09-21 | Seiko Epson Corporation | Image correcting method, image correcting apparatus, and storage medium having program stored thereon |
US20070124702A1 (en) * | 2005-11-25 | 2007-05-31 | Victor Company Of Japan, Ltd. | Method and apparatus for entering desired operational information to devices with the use of human motions |
US20080024619A1 (en) * | 2006-07-27 | 2008-01-31 | Hiroaki Ono | Image Processing Apparatus, Image Processing Method and Program |
US20080085048A1 (en) * | 2006-10-05 | 2008-04-10 | Department Of The Navy | Robotic gesture recognition system |
US20080136775A1 (en) * | 2006-12-08 | 2008-06-12 | Conant Carson V | Virtual input device for computing |
US20090110292A1 (en) * | 2007-10-26 | 2009-04-30 | Honda Motor Co., Ltd. | Hand Sign Recognition Using Label Assignment |
US20130273968A1 (en) * | 2008-08-19 | 2013-10-17 | Digimarc Corporation | Methods and systems for content processing |
US20120019684A1 (en) * | 2009-01-30 | 2012-01-26 | Thomson Licensing | Method for controlling and requesting information from displaying multimedia |
US20110142349A1 (en) * | 2009-06-30 | 2011-06-16 | Sony Corporation | Information processing apparatus and information processing method |
US20120200494A1 (en) * | 2009-10-13 | 2012-08-09 | Haim Perski | Computer vision gesture based control of a device |
US20110135148A1 (en) * | 2009-12-08 | 2011-06-09 | Micro-Star Int'l Co., Ltd. | Method for moving object detection and hand gesture control method based on the method for moving object detection |
US20110239118A1 (en) * | 2010-03-25 | 2011-09-29 | Sony Corporation | Gesture input device, gesture input method, and program |
US20110286676A1 (en) * | 2010-05-20 | 2011-11-24 | Edge3 Technologies Llc | Systems and related methods for three dimensional gesture recognition in vehicles |
US20120019728A1 (en) * | 2010-07-26 | 2012-01-26 | Darnell Janssen Moore | Dynamic Illumination Compensation For Background Subtraction |
US20120129605A1 (en) * | 2010-11-19 | 2012-05-24 | Total Immersion | Method and device for detecting and tracking non-rigid objects in movement, in real time, in a video stream, enabling a user to interact with a computer system |
US20120275686A1 (en) * | 2011-04-29 | 2012-11-01 | Microsoft Corporation | Inferring spatial object descriptions from spatial gestures |
US20130027300A1 (en) * | 2011-07-29 | 2013-01-31 | Toshiaki Nakasu | Recognition apparatus, method, and computer program product |
US8891868B1 (en) * | 2011-08-04 | 2014-11-18 | Amazon Technologies, Inc. | Recognizing gestures captured by video |
US20130050076A1 (en) * | 2011-08-22 | 2013-02-28 | Research & Business Foundation Sungkyunkwan University | Method of recognizing a control command based on finger motion and mobile device using the same |
US20130077831A1 (en) * | 2011-09-26 | 2013-03-28 | Sony Corporation | Motion recognition apparatus, motion recognition method, operation apparatus, electronic apparatus, and program |
US9164589B2 (en) * | 2011-11-01 | 2015-10-20 | Intel Corporation | Dynamic gesture based short-range human-machine interaction |
US20140337807A1 (en) * | 2011-12-09 | 2014-11-13 | Sony Corporation | Information processing apparatus, information processing method, and recording medium |
US20130154919A1 (en) * | 2011-12-20 | 2013-06-20 | Microsoft Corporation | User control gesture detection |
US20130293722A1 (en) * | 2012-05-07 | 2013-11-07 | Chia Ming Chen | Light control systems and methods |
US20130329948A1 (en) * | 2012-06-06 | 2013-12-12 | Fujitsu Limited | Subject tracking device and subject tracking method |
US20140118244A1 (en) * | 2012-10-25 | 2014-05-01 | Pointgrab Ltd. | Control of a device by movement path of a hand |
US20140118257A1 (en) * | 2012-10-29 | 2014-05-01 | Amazon Technologies, Inc. | Gesture detection systems |
US20150092040A1 (en) * | 2013-10-01 | 2015-04-02 | Broadcom Corporation | Gesture-Based Industrial Monitoring |
Non-Patent Citations (1)
Title |
---|
"The Recognition of Human Movement Using Temporal Template," Aaron F. Bobick et al., IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, VOL. 23, NO. 3, MARCH 2001, pages 257-267. * |
Cited By (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10146318B2 (en) | 2014-06-13 | 2018-12-04 | Thomas Malzbender | Techniques for using gesture recognition to effectuate character selection |
TWI614608B (en) * | 2014-12-16 | 2018-02-11 | 英特爾公司 | Apparatus, system and method for communication of touch sensor information |
US20180348841A1 (en) * | 2015-11-11 | 2018-12-06 | Ams Ag | Method, optical sensor arrangement and computer program product for passive optical motion detection |
US10635153B2 (en) * | 2015-11-11 | 2020-04-28 | Ams Ag | Method, optical sensor arrangement and computer program product for passive optical motion detection |
US20190020813A1 (en) * | 2017-07-14 | 2019-01-17 | Casio Computer Co., Ltd. | Image Recording Apparatus, Image Recording Method, and Computer-Readable Storage Medium |
US10616479B2 (en) * | 2017-07-14 | 2020-04-07 | Casio Computer Co., Ltd. | Image recording apparatus, image recording method, and computer-readable storage medium |
US10963064B2 (en) * | 2018-06-19 | 2021-03-30 | Honda Motor Co., Ltd. | Control device and control method |
US10870436B2 (en) | 2018-07-24 | 2020-12-22 | Honda Motor Co., Ltd. | Operation assistance system and operation assistance method |
CN109348124A (en) * | 2018-10-23 | 2019-02-15 | Oppo广东移动通信有限公司 | Image transfer method, device, electronic equipment and storage medium |
US11320914B1 (en) * | 2020-11-30 | 2022-05-03 | EMC IP Holding Company LLC | Computer interaction method, device, and program product |
Also Published As
Publication number | Publication date |
---|---|
US20170308174A1 (en) | 2017-10-26 |
JP5782061B2 (en) | 2015-09-24 |
US9864436B2 (en) | 2018-01-09 |
JP2014174776A (en) | 2014-09-22 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US9864436B2 (en) | Method for recognizing motion gesture commands | |
CN108337433B (en) | Photographing method, mobile terminal and computer readable storage medium | |
US11770619B2 (en) | Generating static images with an event camera | |
US11611715B2 (en) | System and method for event camera data processing | |
US10356320B2 (en) | Information processing device and image input device | |
WO2020038087A1 (en) | Method and apparatus for photographic control in super night scene mode and electronic device | |
CN102999298B (en) | The method of a kind of camera module transmission view data and computer | |
US20200084387A1 (en) | Low power mode for one or more cameras of a multiple camera system | |
WO2017173585A1 (en) | Photographing method and terminal | |
CN108513069B (en) | Image processing method, image processing device, storage medium and electronic equipment | |
US20210176405A1 (en) | Electronic device, controller device, and control method | |
US20190045099A1 (en) | Electronic device for playing video based on movement information and operating method thereof | |
US20200077019A1 (en) | Electronic device for obtaining images by controlling frame rate for external moving object through point of interest, and operating method thereof | |
WO2024001506A1 (en) | Display method and electronic device | |
EP4376433A1 (en) | Camera switching method and electronic device | |
US11200653B2 (en) | Local histogram matching with global regularization and motion exclusion for multi-exposure image fusion | |
US10958833B2 (en) | Electronic device for controlling frame rate of image sensor and method thereof | |
CN211557362U (en) | Front-end image acquisition device capable of adapting to image scene | |
CN111212221A (en) | Front-end image processing method capable of adapting to image scene and storage medium | |
US9946956B2 (en) | Differential image processing |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: LENOVO (SINGAPORE) PTE. LTD., SINGAPORE Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:TSUKAMOTO, YASUSHI;SUGIYAMA, JUN;REEL/FRAME:032119/0301 Effective date: 20140106 |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- AFTER EXAMINER'S ANSWER OR BOARD OF APPEALS DECISION |