WO2011096571A1 - Input device - Google Patents

Input device Download PDF

Info

Publication number
WO2011096571A1
WO2011096571A1 PCT/JP2011/052591 JP2011052591W WO2011096571A1 WO 2011096571 A1 WO2011096571 A1 WO 2011096571A1 JP 2011052591 W JP2011052591 W JP 2011052591W WO 2011096571 A1 WO2011096571 A1 WO 2011096571A1
Authority
WO
WIPO (PCT)
Prior art keywords
image
processing
area
activity
input device
Prior art date
Application number
PCT/JP2011/052591
Other languages
French (fr)
Japanese (ja)
Inventor
堪亮 坂本
慶 木下
Original Assignee
株式会社ネクステッジテクノロジー
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 株式会社ネクステッジテクノロジー filed Critical 株式会社ネクステッジテクノロジー
Publication of WO2011096571A1 publication Critical patent/WO2011096571A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/017Gesture based interaction, e.g. based on a set of recognized hand gestures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/0304Detection arrangements using opto-electronic means

Definitions

  • the present invention is used by being connected to an information device such as an information terminal device or a personal computer, takes an operation image of an operator (user) by a camera, and controls cursor operation of the information device, selection and execution of an application program, and the like.
  • An input device in particular, video input that simplifies the algorithm and reduces the amount of processing data as much as possible to reduce the amount of computation and memory usage, as well as controlling the cursor of the personal computer in real time. Relates to the device.
  • FIG. 32 is a block diagram of an operation input device for explaining such a conventional image input device (see Patent Document 1).
  • the operation input device 101 shown in this figure uses the visible light camera 106 (see FIG. 33) and the like to image the user and output a color image, and analyzes the color image output from the imaging unit 102.
  • Hand region detecting means 103 for detecting the shape of the user's hand
  • hand operation determining means for comparing the hand shape registered in advance with the hand shape output from the hand region detecting means 103 to determine the operation instruction content 104
  • a selection menu expression unit 105 that informs the user of the selection menu by voice or a project image based on the determination content of the manual operation determination unit 104.
  • the operation input device 101 extracts a hand region from a color image photographed by the visible light camera 106 and determines what shape the hand is in (for example, tilting a hand or bending a finger). The judgment is made, and a manual operation instruction corresponding to the judgment content is notified to the user by voice or a project image.
  • both the background image processed by the hand region detection means 103 and the current image have high resolution, so the number of pixels and the amount of data are enormous. Therefore, the capacity of the hard disk, the memory capacity, etc. have to be increased by that amount, and there is a problem that the entire operation input device 101 becomes expensive.
  • the hand region detection unit 103 As shown in FIG. 33 as the hand region detection unit 103, the difference region extraction unit 110, the skin color region extraction unit 111, the binarization correction unit 112, the distance calculation unit 113, the center weight correction Means 114, hand region candidate detection means 115, contour length / area calculation means 116, hand region determination means 117, etc., which not only have to perform complicated calculations, but also use a fairly fast CPU or dedicated circuit. Otherwise, there was a problem that the shape of the hand could not be detected in real time.
  • the conventional operation input device 101 has a problem that when a person other than the user moves within the shooting range of the visible light camera 106 and moves his / her hand, this is detected and malfunctions.
  • the present invention enables the use of a low-resolution and inexpensive camera, and greatly reduces the cost of the entire apparatus by greatly reducing the amount of calculation and memory, and the movement of the user. It is an object of the present invention to provide an input device that can remotely detect a cursor of a personal computer and can operate a personal computer cursor and the like, and does not malfunction even when a person other than the user moves within the shooting range of the camera.
  • the purpose is to provide an input device that can remotely control the cursor of the computer, the scroll of the operation target screen, etc., and that does not malfunction even if a person other than the user is in the shooting range of the camera and moves his / her hand. To do.
  • an input device that can remotely control operations such as enlargement / reduction of the operation target screen and rotation of the operation target screen, and that does not malfunction even if a person other than the user moves within the shooting range of the camera. For the purpose.
  • the cost of the entire apparatus is greatly reduced, and the user's hand moves.
  • An object of the present invention is to provide an input device that does not malfunction even when a person other than the user is present and moves his / her hand.
  • the binocular parallax method can be used while significantly reducing the cost of the entire device by significantly reducing the amount of computation, memory, or circuit scale while allowing the use of low-resolution and inexpensive cameras.
  • An object of the present invention is to provide an input device with improved accuracy by correcting the measurement distance from the camera to the subject.
  • the binocular parallax method can be used while significantly reducing the cost of the entire device by significantly reducing the amount of computation, memory, or circuit scale while allowing the use of low-resolution and inexpensive cameras.
  • An object of the present invention is to provide an input device that speeds up processing when correcting a measurement distance from a camera to a subject.
  • an input device of the present invention is an input device that processes an operator's image obtained by a video camera and generates an operation instruction according to the operation content of the operator.
  • the color image output from the right-eye color camera is subjected to graying processing, image division / binarization processing, interframe difference processing, histogram processing, and active rectangular area extraction processing, and the right eye side of the operator
  • a right eye side image processing program for extracting an active rectangular area and a color image output from the left eye color camera, graying processing, image division / binarization processing, inter-frame difference processing, hysteresis Gram processing, activity rectangle area extraction processing, the left eye side image processing program for extracting the left eye side activity rectangle area of the operator, the right eye side activity rectangle area obtained by the right eye side image processing program,
  • the input device of the present invention is an input device that processes an image of an operator obtained by a video camera, generates an operation instruction according to the operation content of the operator, and controls the operation of the remote operation target device.
  • An input device case formed in a box shape, a color camera body for a right eye that is attached to the front left side of the input device case and takes an image of an operator, and attached to the front right side of the input device case
  • a color camera body for the left eye that captures an image of the operator, and a graying processing circuit, an image division / binarization processing circuit, an inter-frame difference processing circuit, and a histogram processing circuit, which are disposed in the input device casing.
  • a right-eye side image processing board for processing a color image output from the right-eye color camera main body by the activity rectangular area extraction processing circuit and extracting the right-eye side activity rectangular area of the operator; Arranged in the apparatus casing and output from the color camera body for the left eye by a graying processing circuit, an image division / binarization processing circuit, an inter-frame difference processing circuit, a histogram processing circuit, and an active rectangular area extraction processing circuit
  • a left-eye side image processing board for processing a color image and extracting the left-eye side active rectangular area of the operator, and an active rectangular area selection processing circuit, a virtual cursor control process / screen disposed in the input device casing Activity using the binocular parallax method on the right-eye side active rectangular area obtained on the right-eye side image processing board and the left-eye side active rectangular area obtained on the left-eye side image processing board by the control processing circuit Performs rectangular area selection processing, virtual cursor control processing / screen control processing, detects the movement of the operator's hand or fingertip, generate
  • the virtual cursor control process / screen control process or the virtual cursor control process / screen control processing circuit has one active rectangular area group on the virtual cursor active area image, Based on the shape and presence / absence of movement, a Carlsol control instruction or a screen scroll instruction is generated.
  • the virtual cursor control process / screen control process or the virtual cursor control process / screen control processing circuit has two active rectangular area groups on the virtual cursor active area image, Based on the moving direction, one of a screen rotation instruction, a screen enlargement instruction, and a screen reduction instruction is generated.
  • the activity rectangular area extraction process uses a histogram statistical processing result to generate a virtual cursor activity area image, a virtual button click activity area from the histogram. It is characterized by creating images.
  • the activity rectangle area extraction process, or the activity rectangle area extraction processing circuit is a multi-stage rectangle object extraction process for the virtual cursor activity area image or the virtual button click activity area image. And removing noise components.
  • the input device of the present invention adds an enlargement / reduction rectangle mask creation process or an enlargement / reduction rectangle mask creation processing circuit, and performs the enlargement / reduction rectangle mask creation process or the enlargement / reduction rectangle mask creation processing circuit.
  • the image corresponding to the change area rectangle on the virtual cursor activity area image or the change area rectangle of the virtual button click activity area image is extracted from the color images obtained by the color camera and the color camera body. The other image is cut to remove noise components.
  • the active rectangular area extraction process or the active rectangular area extraction processing circuit is the latest difference image of the multi-stage difference image data generated by the histogram processing or the histogram processing circuit.
  • An invalid activity rectangle area is determined based on a comparison between the activity rectangle area extracted from the data and the virtual button click activity area image.
  • the input apparatus may be configured such that the viewing angle of the color camera is a constant and the right eye activity rectangular area is input between the center point of the left eye activity rectangular area.
  • the distance from the color camera to the subject is corrected according to the center coordinate distance in the horizontal direction.
  • the input device is characterized in that the distance to the subject corrected in accordance with the center coordinate distance is stored in advance as table data.
  • an inexpensive camera with low resolution and to detect a user's movement while greatly reducing the cost of the entire apparatus by greatly reducing the amount of calculation and memory.
  • An input device that can remotely control a cursor of a personal computer or the like and that does not malfunction even when a person other than the user moves within the shooting range of the camera and moves his / her hand can be realized.
  • FIG. 1 is a block diagram showing a first embodiment of an input device according to the present invention.
  • 3 is a flowchart illustrating a detailed operation example of the input device illustrated in FIG. 1.
  • 3 is a flowchart showing a detailed operation example of grayed / binarized image processing shown in FIG. 2.
  • 3 is a flowchart showing a detailed operation example of inter-frame difference / histogram creation processing shown in FIG. 2. It is a flowchart which shows the detailed operation example of the activity rectangular area extraction process shown in FIG. It is a flowchart which shows the detailed operation example of the active rectangle area
  • region selection process shown in FIG. 3 is a flowchart showing detailed operation examples of a virtual cursor control process and a screen control process shown in FIG. 2.
  • FIG. 3 is a schematic diagram showing an example of an active rectangular area group selected by an active rectangular area selection process in the input device shown in FIG. 1.
  • FIG. 3 is a schematic diagram showing an example of an active rectangular area group selected by an active rectangular area selection process in the input device shown in FIG. 1.
  • FIG. 1 It is a schematic diagram which shows an example of the multistage object extraction process used with the input device shown in FIG. It is a schematic diagram which shows an example of the multistage object extraction process used with the input device shown in FIG. It is a schematic diagram which shows the outline
  • FIG. 2 it is a schematic diagram which shows the relationship between a right active rectangular area and a left active rectangular area after correction by the binocular parallax method in the input device shown in FIG. 1.
  • FIG. 2 is a schematic diagram showing a relationship between a right active rectangular area and a left active rectangular area after correction by the binocular parallax method in the input device shown in FIG. 1.
  • FIG. 1 it is a schematic diagram which shows the actual example of the right activity rectangular area after correcting by the binocular parallax method, and each left activity rectangular area. It is a schematic diagram which shows the example of a relationship between the motion of the user's hand image
  • FIG. 31 is a schematic diagram illustrating an operation example of the flowchart illustrated in FIG. 30. It is a block diagram which shows an example of the operation input apparatus known conventionally. It is a block diagram which shows the detailed circuit structural example of the hand area
  • region detection means shown in FIG. FIG. 6 shows a step 41 ′ to be processed instead of the step 41 in the flowchart shown in FIG. 5.
  • Step 43 ′ to be processed is shown instead of step 43 in the flowchart shown in FIG. 5.
  • FIG. 7 shows a step 57 ′ to be processed instead of the step 57 in the flowchart shown in FIG. 6.
  • FIG. 2 is a schematic diagram for explaining an operation for removing noise generated by an operation unintended by a user in an extraction activity region in the input device shown in FIG. 1. It is a schematic diagram explaining the relationship between the camera viewing angle, the subject width, and the distance to the subject in one camera. It is a schematic diagram explaining the measurement of the exact distance from a camera to a to-be-photographed object by the binocular parallax method by two cameras of the input device shown in FIG.
  • FIG. 1 is a block diagram showing a first embodiment of an input device according to the present invention.
  • the input device 1a shown in this figure includes a built-in web camera (a left-eye color camera of claim 1) 4 provided in the display unit 3 of the personal computer 2, a video capture 5 provided in the personal computer 2, and a personal computer 2
  • the CPU 9 is provided inside and the memory 10 is provided in the personal computer 2.
  • the input device 1a analyzes the color images obtained by the web cameras 4 and 6, and has a predetermined distance range from the installation position of the web cameras 4 and 6, for example, “0.3 m” to “0.8 m”.
  • a virtual cursor 25 [FIG. 19] displayed on the display unit 3 of the personal computer 2 is detected by detecting only movements of the user's hand, fingertips, etc. while distinguishing between a user within the range and another person at a distance other than that. (See (b))], the operation target screen (OS screen, application screen) and the like are controlled, and the currently running application is controlled.
  • the web camera 4 is a color camera having a resolution of about 320 pixels ⁇ 240 pixels, and supplies the video capture 5 with a color video signal obtained by photographing the user when a shooting instruction is issued from the video capture 5. To do.
  • the video capture 5 controls the web camera 4 to shoot a user, captures a color video signal obtained by the shooting operation, and receives an RGB signal. It is converted into a color image of the format and supplied to the CPU 9.
  • the web camera 6 is a color camera having a resolution of about 320 pixels ⁇ 240 pixels attached to the upper edge of the display unit 3 at a predetermined distance in the horizontal direction from the web camera 4.
  • the YUV signal obtained by photographing the user is supplied to the USB interface 7.
  • the USB interface 7 controls the web camera 6 to take a user's image and captures the YUV signal obtained by the shooting operation. Is converted into a color image in the RGB signal format.
  • the hard disk 8 includes an OS storage area 13 in which an OS (Operating System) and constant data are stored, an application storage area 14 in which application programs such as an Internet Explorer program and a browser program are stored, and image processing used in the present invention.
  • An image processing program storage area 15 for storing a program (a right eye side image processing program, a left eye side image processing program, an image processing program) according to claim 1 and an HSV (hue / saturation / lightness) method are set in advance.
  • a color mask, a binarized image, a histogram, a virtual cursor active area image 27 (see FIG. 9), a virtual button click active area image, and the like necessary for extracting a specific color image (for example, skin color).
  • the CPU 9 When a read instruction is output from the CPU 9, the CPU 9 captures this via the system bus 12 and stores the OS, constant data, application program, image processing program, binary image, histogram, The virtual cursor activity area image 27, the virtual button click activity area image, and the like are read out and supplied to the CPU 9 via the system bus 12. Further, when a write instruction and data are output from the CPU 9, these are taken in via the system bus 12 and stored in an area designated by the write instruction, for example, the image storage area 16.
  • the CPU 9 generates display data specified by the OS, constant data, application program and the like stored in the hard disk 8 and supplies the display data to the display interface 11 connected to the system bus 12, and displays the operation target screen on the display unit 3. Display.
  • the image processing described in the right eye side image processing program, the left eye side image processing program, the image processing program, etc. is performed, the control of the size and position of the virtual cursor displayed on the operation target screen, click control, Scroll control, screen rotation control, screen enlargement control, screen reduction control, etc.
  • the memory 10 has a capacity of about several hundred megabytes to several gigabytes, and temporary data when the CPU 9 performs processing specified by an application program, a right eye side image processing program, a left eye side image processing program, an image processing program, or the like. Used as storage area.
  • the USB interface 7 is controlled by the CPU 9 so that the YUV signal obtained by the photographing operation of the web camera 6 is taken in and converted into a color image in the RGB signal format and temporarily stored in the memory 10 or the like. Stored (step S2).
  • a user is selected from the personal computer 2 side.
  • a color image obtained by a web camera corresponding to the right eye, for example, an external web camera 6 is read (step S3).
  • step S4 the CPU 9 starts graying / binarized image processing. That is, as shown in the flow chart of FIG. 3, the color image obtained by the external web camera 6 is masked by the color mask stored in the image storage rear 16 of the hard disk 8, and preset from the color images. A color image (skin color image) of a specific color (for example, skin color) is extracted (step S21), and the color image obtained by the external web camera 6 is gray-processed and set in advance. The image is converted into a monochrome image of a certain gradation, and the image capacity for one frame is reduced (step S22).
  • the CPU 9 checks whether or not a screen division instruction is set. If there is a screen division instruction, the monochrome image is divided into a plurality of areas (each area is composed of several to several tens of pixels). If there is no screen division instruction, the division process is skipped, and then the monochrome image is binarized by the maximum likelihood threshold method to create a binarized image (step S23).
  • step S24 the logical sum of the binarized image and the skin color image is taken by the CPU 9 and the skin color portion in the binarized image is extracted (step S24), and this is the binarized image for one frame (on the right eye side). (Binary image) is stored in the image storage area 16 of the hard disk 8 (step S25).
  • the CPU 9 starts graying / binarized image processing (step S6). That is, as shown in the flowchart of FIG. 3, the color image obtained by the built-in web camera 4 is masked by the color mask stored in the image storage area 16 of the hard disk 8 and preset from the color image. A color image (skin color image) of a specific color (for example, skin color) is extracted (step S21), and the color image obtained by the built-in web camera 4 is subjected to gray processing, and is set in advance. The image is converted into a tone monochrome image, and the image capacity for one frame is reduced (step S22).
  • the CPU 9 checks whether or not a screen division instruction is set. If there is a screen division instruction, the monochrome image is divided into a plurality of areas (each area is composed of several to several tens of pixels). If there is no screen division instruction, the division process is skipped, and then the monochrome image is binarized by the maximum likelihood threshold method to create a binarized image (step S23).
  • step S24 the logical sum of the binarized image and the skin color image is obtained by the CPU 9 and the skin color portion in the binarized image is extracted (step S24), and this is converted into a binarized image for one frame (on the left eye side). (Binary image) is stored in the image storage area 16 of the hard disk 8 (step S25).
  • the right eye side binarized image and the left eye side binarized image are in the image storage area 16 of the hard disk 8 in FIFO (First / In / First / Out) format.
  • the frame is stored for tens of frames.
  • the right eye side is selected from the binarized images of several frames to several tens of frames stored in the image storage area 16 of the hard disk 13 by the CPU 9.
  • Several consecutive frames of binarized images including the latest binarized image corresponding to are sequentially read (step S7).
  • step S9 the inter-frame difference / histogram creation process is started (step S9). That is, as shown in the flowchart of FIG. 4, the inter-frame difference processing is performed on the binarized images for two consecutive frames among the respective binarized images (steps S31 and S32), and the inter-frame difference is performed. Each difference image obtained by the process is cumulatively added for each divided area to create a right-eye histogram and stored in the image storage area 16 of the hard disk 8 (steps S33 and S34).
  • step S10 the latest 2 corresponding to the left eye side among the binarized images of several frames to several tens of frames stored in the image storage area 16 of the hard disk 13 by the CPU 9.
  • Several consecutive frames of binarized images including the binarized image are sequentially read out (step S10).
  • step S11 the number of frames of the binarized image that can be read out by the CPU 9 is checked. If the number of frames is equal to or larger than the predetermined number (step S11), the interframe difference / histogram creation process is started (step S12). That is, as shown in the flowchart of FIG. 4, the inter-frame difference processing is performed on the binarized images for two consecutive frames among the respective binarized images (steps S31 and S32), and the inter-frame difference is performed. Each difference image obtained by the process is cumulatively added for each divided area to create a left eye side histogram and stored in the image storage area 16 of the hard disk 8 (steps S33 and S34).
  • Step S14 statistical processing is performed on the density value of each divided area of the histogram, and an average value, density variance value, maximum value, deviation ( ⁇ 1 ⁇ , ⁇ 2 ⁇ ), and the like are calculated ( Step S41).
  • the CPU 9 extracts each divided area having a density value larger than the threshold value for extracting the change area rectangle (for example, average value ⁇ 1 ⁇ ) from each divided area of the histogram.
  • a rectangular change area rectangle 65 (see FIG. 31) is determined so as to include (activity division area) and stored in the image storage area 16 of the hard disk 8.
  • the CPU 9 extracts the virtual cursor rectangle from the divided areas 20 (see FIGS. 9 and 10) constituting the histogram as shown in the three-dimensional density distribution diagram of FIG.
  • a divided area (activity divided area 21) having a density value larger than the threshold value (for example, maximum value ⁇ 1 ⁇ ) is extracted.
  • the histogram shows the frequency distribution of 20 changes in each divided area.
  • the active divided area 21 in which the change in the action is significant is extracted by capturing the user's operation.
  • a rectangular activity rectangular area 26 is determined so as to include each activity division area 21 as shown in FIG. 9, and based on the determination result, as shown in FIG.
  • a right-eye virtual cursor activity area screen 27 is created and stored in the image storage area 16 of the hard disk 8.
  • a rectangular activity rectangular area 26 is determined so as to include each activity division area 21 as shown in FIG. 10, and the virtual cursor activity on the right eye side as shown in FIG.
  • An area image 27 is created and stored in the image storage area 16 of the hard disk 8 (step S42).
  • the CPU 9 reads the right-eye histogram stored in the image storage area 16 of the hard disk 8 and extracts a threshold value for extracting a virtual button click rectangle in each divided area 20 (for example, the maximum value ⁇ 2)) divided areas (activity divided areas) having a density value greater than 2 ⁇ ) are extracted, and a rectangular activity rectangular area is determined so as to include each of these activity divided areas, and a virtual button on the right eye side is determined.
  • a click activity area image (not shown) is created and stored in the image storage area 16 of the hard disk 8 (step S43).
  • the virtual cursor click area extraction threshold value 27 for the right eye side virtual cursor active region image 27 obtained by the CPU 9 using the threshold value for virtual cursor rectangle extraction (for example, the maximum value ⁇ 1 ⁇ ) is obtained.
  • the threshold value for virtual cursor rectangle extraction for example, the maximum value ⁇ 1 ⁇
  • the horizontal center point “A” of the active rectangular area 26 is obtained, and an inactive area on the left side of the horizontal center point “A”, an active area (for example, an active division area 21), Boundary point “B”, a non-active area on the right side of the horizontal center point “A”, and a boundary point “C” between the active area (for example, the active division area 21) are detected.
  • B ”,“ C ” Are determined as each active rectangular area 26, and other active areas are determined as unnecessary active areas due to the shadows of the user and are invalidated (two-point extraction processing).
  • the CPU 9 checks whether or not each of the activity rectangular areas 26 for which the two-point extraction processing has been completed can be divided vertically, and if it can be divided vertically, as shown in FIG.
  • the vertical center point “A” of the area 26 is obtained, and the boundary point “B” between the inactive area above the vertical center point “A” and the active area (for example, the active division area 21) is detected.
  • the area including these boundary points “B” is determined as the active rectangular area 26, and the lower active area is determined as an unnecessary active area due to the shadow of the user, and is invalidated (minimization process) (step S44). ).
  • the CPU 9 performs a virtual cursor activity region image 27 on the right eye side including the activity rectangular region 26 obtained by the multi-stage rectangular object extraction process constituted by the two-point extraction process and the minimization process, and the right-eye virtual image.
  • the button click activity area image is stored in the image storage area 16 of the hard disk 8 (step S45).
  • Step S15 the CPU 9 reads the left eye side histogram stored in the image storage area 16 of the hard disk 8 (step S15), and starts the active rectangular area extraction process.
  • Step S16 statistical processing is performed on the density value of each divided area of the histogram, and an average value, density variance value, maximum value, deviation ( ⁇ 1 ⁇ , ⁇ 2 ⁇ ), and the like are calculated ( Step S41).
  • the CPU 9 extracts each divided area 20 having a density value larger than the threshold for extracting the change area rectangle (for example, the average value ⁇ 1 ⁇ ), and includes these divided areas (activity divided areas).
  • a rectangular change area rectangle 65 is determined and stored in the image storage area 16 of the hard disk 8.
  • the CPU 9 extracts the virtual cursor rectangle from the divided areas 20 (see FIGS. 9 and 10) constituting the histogram as shown in the three-dimensional density distribution diagram of FIG. A divided area (activity divided area 21) having a density value larger than the threshold value (for example, maximum value ⁇ 1 ⁇ ) is extracted.
  • the threshold value for example, maximum value ⁇ 1 ⁇
  • a rectangular activity rectangular area 26 is determined so as to include each activity division area 21 as shown in FIG. 9, and based on the determination result, as shown in FIG.
  • a virtual cursor activity area screen 27 on the left eye side is created and stored in the image storage area 16 of the hard disk 8.
  • a rectangular activity rectangular area 26 is determined so as to include each activity division area 21 as shown in FIG. 10, and the virtual cursor activity on the left eye side as shown in FIG.
  • An area image 27 is created and stored in the image storage area 16 of the hard disk 8 (step S42).
  • ⁇ Determine activity area for virtual button click ⁇ the left-eye histogram stored in the image storage area 16 of the hard disk 8 is read by the CPU 9, and a threshold value for extracting a virtual button click rectangle in each divided area 20 (for example, the maximum value ⁇ 2 ⁇ ) ) A divided area (activity divided area) having a larger density value is extracted, and a rectangular activity rectangular area is determined so as to include each of these activity divided areas, and a virtual button click on the left eye side is clicked
  • An active area image (not shown) is created and stored in the image storage area 16 of the hard disk 8 (step S43).
  • the horizontal center point “A” of the active rectangular area 26 is obtained, and an inactive area on the left side of the horizontal center point “A”, an active area (for example, an active division area 21), Boundary point “B”, a non-active area on the right side of the horizontal center point “A”, and a boundary point “C” between the active area (for example, the active division area 21) are detected.
  • B ”,“ C ” Are determined as each active rectangular area 26, and other active areas are determined as unnecessary active areas due to the shadows of the user and are invalidated (two-point extraction processing).
  • the CPU 9 checks whether or not each of the activity rectangular areas 26 for which the two-point extraction process has been completed can be divided vertically, and if it can be divided vertically, as shown in FIG. 26, the center point “A” in the vertical direction is obtained, and the boundary point “B” between the inactive region and the active region (for example, the active division area 21) above the vertical center point “A” is detected.
  • the area including the boundary point “B” is determined as the active rectangular area 26, and the lower active area is determined as an unnecessary active area due to the shadow of the user or the like, and is invalidated (minimization process) (step S44).
  • the click activity area image is stored in the image storage area 16 of the hard disk 8 (step S45).
  • the movement of the user's hand can be captured by the virtual area activity area image 27 and the virtual button click activity area image of the right eye and the left eye.
  • the input device 101 detects the movement of the hand not intended by the user. It would be even better if it could be detected.
  • the activity rectangular area 26 as a result of extracting the activity area in the form of a histogram in multiple stages is extracted on the histogram data even when the user's intention is not pointing or tapping, for example, when the hand is waving left and right. Is done. Therefore, if it can be determined that this is not due to the pointing operation intended by the user, the reliability of the input device 101 is increased.
  • step S41 of FIG. 5 the process of step S41 ′ shown in FIG. 34 is performed, and the CPU 9 extracts the active rectangular area 26 from the latest difference data among the multi-stage difference image data that is formed into a histogram. And hold for a certain time while tracking those points.
  • step S43 in FIG. 5 the process of step S43 ′ shown in FIG. 35 is performed, and the CPU 9 stores the virtual button click activity area (maximum value ⁇ 2 ⁇ ) in step S41 ′.
  • the corresponding tracking is obtained and the tracking result is active over an area of a specific size, it is determined that the activity rectangular area 26 is invalid.
  • FIG. 37 shows an active area extracted when the user moves his / her hand from the lower right to the upper left, further to the lower left, and then to the upper right.
  • the active area generated from the tracking data of the activity rectangular area 26 has a large movement. Ignore it as noise by judging what was generated.
  • the active rectangle area selection process is started by the CPU 9 (step S17). That is, as shown in the flowchart of FIG. 6, the right eye side virtual cursor activity region image 27 and the right eye side virtual button click activity region image stored in the image storage area 16 of the hard disk 8 are read out, and these right Position correction by the binocular parallax method is performed on each active rectangular area 26 included in the virtual cursor activity area image 27 on the eye side and the virtual button click activity area image on the right eye side as shown in the schematic diagram of FIG.
  • the CPU 9 reads the left eye side virtual cursor activity region image 27 and the left eye side virtual button click activity region image stored in the image storage area 16 of the hard disk 8, and these left eye side virtual cursor activity regions are read out.
  • Position correction by the binocular parallax method is performed on the area image 27 and each activity rectangular area 26 of the left eye side virtual button click activity area image as shown in the schematic diagram of FIG.
  • the coordinates “P L (X L , Y L )” of each active rectangular area 26 so as to correspond to the mounting position (horizontal distance “B”, vertical distance, etc.), the focal length “f” of each web camera 4, 6, etc. Are corrected, numbers are added in order of size (steps S53 and S54).
  • the user's hand is at a focus position corresponding to the focal length of each of the webcams 4 and 6, for example, “0.3 m” to “0.8 m” away from each of the webcams 4 and 6.
  • the binocular parallax method is used. As shown in the schematic diagram of FIG. 17, the right-side active rectangular area 26 and the left-eye active rectangular area 26 can be completely matched (or substantially matched). .
  • the center coordinates “X R , Y R ” of the activity rectangular area 26 corresponding to the right eye side to which the number “1” is assigned correspond to the left eye side.
  • the distance (center coordinate distance) with the center coordinates “X L , Y L ” of the active rectangular area 26 is calculated and stored in the memory 10 together with the number “1”.
  • the CPU 9 sequentially reads out the respective central coordinate distances stored in the memory 10 and compares them with a predetermined value.
  • an activity rectangle area 26 on the eye side and an activity rectangle area 26 on the left eye side that is, when the user's hand is at a position away from each web camera 4, 6 by “0.3 m” to “0.8 m”
  • the right-eye activity rectangle area 26 and the left-eye activity rectangle area 26 corresponding to the user's hand are the valid right-eye activity rectangle area 26 and the valid left-eye activity rectangle area 26.
  • the other activity rectangle area 26 on the right eye side and the activity rectangle area 26 on the left eye side are the invalid activity rectangle area 26 on the right eye side and the activity rectangle area 26 on the invalid left eye side.
  • the present invention further provides a method of correcting this and measuring a more accurate distance.
  • the distance is measured by calculating a correction value from the center coordinate distance in the active rectangular area 26 of each of the visual fields of the webcams 4 and 6 with the camera viewing angle measured in advance as a constant. Will be described.
  • FIG. 38 explains the conversion of distance from the viewing angle of one web camera 4A. From the definition of the trigonometric function, the relationship of Equation 1 is established. w: Width of the subject on the image [pixel] img_w: Number of horizontal pixels of the camera image [pixel] d: Distance between camera and subject [m] ⁇ : Subject width / 2 [m] ⁇ w: Camera viewing angle / 2 [rad] (measured in advance) However, w and (img_w) are “angles” as physical quantities because of their sizes in the image. When this is solved for d, Equation 2 is obtained. In Equation 2, since ⁇ w and img_w are constants, the distance d, which is a value to be obtained, depends on ⁇ and w. Here, since ⁇ varies depending on the subject 60 to be reflected, it is necessary to obtain the distance d without depending on ⁇ .
  • FIG. 39 explains the conversion of the distance from the viewing angle of the two main and sub webcams.
  • the distance d is expressed by Equation 3.
  • w ′ absolute value of the difference between the abscissa [pixel] of the center of the subject on the sub-camera image and the center abscissa [pixel] of the subject on the image of the main camera (
  • ⁇ ′ Distance [m] between two points at which the center line of sight of the sub camera and the center line of sight of the main camera intersect at the position of the subject (when both eyes are parallel, ⁇ ′ is always constant)
  • the distance d does not depend on ⁇ ′ because it is constant when the central line of sight of both eyes is parallel. Note that d is sufficiently larger than ⁇ ′ (d >> ⁇ ′), and even if the subject 60 exists at the edge of the image, it is approximately established.
  • the CPU 9 performs the process of step 57 ′ shown in FIG. 36 in the process of step S57, calculates the center coordinate distance w ′ (
  • the effective distance to the subject 60 is measured by comparing the distance to the subject 60 and it is determined whether the distance is within the range of “0.3 m” to “0.8 m”. This is a judgment.
  • each activity rectangular area 26 on the right eye side and each activity rectangle area 26 on the left eye side have the relationship shown in the schematic diagram of FIG. 18, the number “the center coordinate distance is equal to or greater than a predetermined value”.
  • the number “0” is determined to be invalid for the right-eye side active rectangle area (O R1 ) 26 and the left-eye side active rectangle area (O L1 ) 26 corresponding to “1”. It is determined that the right-eye active rectangular area (O R2 ) 26 and the left-eye active rectangular area (O L2 ) 26 corresponding to 2 ′′ are valid.
  • the right-side active rectangular area (O R2 ) 26 determined to be valid by the CPU 9 and the left-side active rectangular area (O L2 ) 26 determined to be valid are previously designated, For example, a virtual cursor activity region image 27 and a virtual button click activity region image in which the activity rectangle region ( OL2 ) on the left eye side determined to be valid is left and the other activity rectangle regions 26 are deleted are created, This is stored in the image storage area 16 of the hard disk 8 as a virtual cursor activity area image 27 and a virtual button click activity area image from which the movement of the person in front of the user and the movement of the person behind are removed by the binocular parallax method. (Step S57).
  • the virtual cursor control process / screen control process is started by the CPU 9 (step S18). That is, as shown in the flowchart of FIG. 7, each virtual cursor activity region image stored in the image storage area 16 of the hard disk 8 (the movement of the person in front of the user and the person behind the user by the binocular parallax method).
  • the virtual cursor activity region images 27 for several frames including the latest activity rectangle region 26 are read (step S61), and one or more adjacent ones are displayed. It is checked whether an active rectangle area group constituted by the active rectangle areas 26 exists in the virtual cursor active area image 27.
  • step S62 and S63 If there are active rectangular area groups in the latest virtual cursor active area image 27 and the number thereof is “1” and almost rectangular (steps S62 and S63), the CPU 9 determines the size of the active rectangular area group. Now, the moving direction is determined, and virtual cursor control is performed so as to correspond to the determination result (step S64).
  • the CPU 9 determines that it is a virtual cursor display instruction, and the display unit 3 has a large size and white virtual display as shown in FIG. A cursor 25 is displayed.
  • the CPU 9 performs virtual processing.
  • the cursor movement instruction is determined, and the large size, white virtual cursor 25 displayed on the display unit 3 is moved so as to correspond to the moving direction of the fingertip.
  • the CPU 9 determines that the movement of the virtual cursor is stopped, and the virtual cursor displayed on the display unit 3 as shown in FIG. The movement of 25 is stopped and the size is reduced.
  • the CPU 9 changes the color of the virtual cursor 25 to red and prohibits a large movement, and a cursor movement instruction is issued to the OS side, and the real cursor 28 is placed in the virtual cursor 25. Move.
  • the CPU 9 detects this, and after a certain time, the position of the virtual cursor 25 displayed on the display unit 3 is fixed and the virtual cursor 25 is displayed as shown in FIG. The color of is changed from red to gray to inform the user that it can be clicked.
  • step S62 and S63 If it is checked whether or not there is an active rectangular area group in the virtual cursor active area image 27 described above, if the number of active rectangular area groups is “1” and is long in the vertical direction (steps S62 and S63), the CPU 9 Is used to determine which direction is longer than the previous active rectangular area group, and an upward scroll instruction (or downward scroll instruction) corresponding to the longer direction is generated and passed to the application side, and the display unit 3 The application screen (operation target screen) displayed on the screen is scrolled upward (or downward) (step S64).
  • the CPU 9 checks whether the color of the virtual cursor 25 is gray. If the color of the virtual cursor 25 is gray, the latest virtual button click activity area image stored in the image storage area 16 of the hard disk 8 is checked. The virtual button click activity region images for several frames including the activity rectangle region are read out (step S66).
  • the CPU 9 checks whether there is an active rectangular area group composed of one or more active rectangular areas 26 in the virtual button click active area image, and whether the shape has changed. Is a change in which the active rectangular area group is set in advance, for example, as shown in FIG. 24 (a), the user extends his hand only once from the state of pointing, and the active rectangular area group is small once. If it has changed from “large” to “large” (step S67), it is determined that it is a single click, a single click instruction is issued to the OS side, and it is in the virtual cursor 25 as shown in FIG. An icon or the like is single-clicked by the real cursor 28 (step S68).
  • step S67 determines that it is a double click, and a double click instruction is issued to the OS side, and the icon or the like at the position of the real cursor 28 is double clicked (step S68).
  • step S65 the operation target screen that has been enlarged is reduced, reduced, rotated, or the like.
  • the user puts his right hand and left hand at the focus positions of the web cameras 4 and 6 and moves the right hand and left hand away from each other, and there are two groups of active rectangular areas shown in FIG. 25 corresponding thereto. Then, as shown in FIG. 26 (a), when the distance between each of the activity rectangle areas becomes longer than the previous time by moving in the direction wider than the previous time, it is determined that the screen enlargement instruction is input by the CPU 9, and the activity rectangle A screen enlargement instruction with an enlargement ratio corresponding to the distance change ratio of the region group is generated and passed to the application side, and the application screen (operation target screen) displayed on the display unit 3 is enlarged.
  • the user puts out his right hand and left hand at the focus positions of the web cameras 4 and 6 and moves the right hand and left hand in a direction approaching each other.
  • two active rectangular area groups are shown in FIG.
  • the CPU 9 determines that a screen reduction instruction has been input, and the distance change ratio of the active rectangular area groups is A screen reduction instruction with a corresponding reduction ratio is generated and passed to the application side, and the application screen (operation target screen) displayed on the display unit 3 is reduced.
  • the CPU 9 when the left and right distances of the respective active rectangular area groups are wide, when one of the active rectangular area groups is small and moves upward, the angle of the upper active rectangular area group with respect to the lower active rectangular area group is small.
  • the CPU 9 generates a screen rotation instruction with a small rotation angle and passes it to the application side, and the application screen (operation target screen) displayed on the display unit 3 is rotated small.
  • Color cursor processing, frame buffer processing, inter-frame difference processing, histogram processing, active rectangular region extraction processing, active rectangular region selection processing, virtual cursor control processing / screen control processing are performed to detect the movement of the user's hand, and virtual cursor 25 Size control, position control, color control, click control, operation target screen enlargement control, reduction control, rotation control, up / down scroll control, left / right scroll control, etc., so that the following effects can be obtained. it can.
  • the cost of the input device 1a can be kept low (effect of claim 1).
  • the input device 1 a can be configured even when the capacity of the hard disk 8 is small, and the cost of the entire device can be kept low (effect of claim 1).
  • low-resolution color images obtained by photographing the user with the web cameras 4 and 6 are subjected to a small number of image processing such as graying processing, image division / binarization processing, and color filtering processing. Since a binary image for one frame is obtained, it is possible to prevent the CPU 9 from being subjected to a large burden, and thus to respond to a user's movement almost in real time even when an inexpensive CPU 9 whose processing speed is not fast is used.
  • the size control, position control, color control, click control, enlargement control, reduction control, rotation control, up / down scroll control, left / right scroll control, and the like of the operation target screen can be performed. Can be kept low (effect of claim 1).
  • the center coordinate position is corrected by the binocular parallax method for each of the right-eye activity rectangle area 26 and the left-eye activity rectangle area 26 obtained by photographing the user with the web cameras 4 and 6. After that, numbers are assigned in order of size, and the center coordinate positions are compared. Based on the comparison result, the right-side active rectangular area 26 and the left-eye active rectangular area 26 corresponding to the focus position are selected. Therefore, even if there is something other than the user's hand at the focus position of each webcam 4, 6 such as a person behind the user and moving, only the movement of the user's hand is not affected by this. Can be extracted to perform size control, position control, color control, click control, enlargement control, reduction control, rotation control, up / down scroll control, left / right scroll control, and the like of the operation target screen. 1 Effect).
  • the user when the user moves only one hand, it is determined that it is a virtual cursor control instruction or scroll control of the operation target screen, and the size control, position control, and color of the virtual cursor 25 are determined. Since control, click control, scroll control of the operation target screen, and the like are performed, the size, position, color, click, scroll of the operation target screen, etc. of the virtual cursor 25 displayed on the display unit 3 with only one hand can be controlled. Remote control is possible (effect of claim 3).
  • the operation target screen enlargement / reduction control instruction or the operation target screen rotation control is detected. Since the instruction is determined to be an instruction, the application screen (operation target screen) displayed on the display unit 3 can be enlarged, reduced, and rotated only by the user moving the right hand and the left hand. Effect).
  • a virtual cursor activity region image 27 and a virtual button click activity region image are created from the histogram using the results obtained by statistically processing the histogram in the activity rectangular region extraction process. Therefore, a moving part such as a user's hand can be accurately detected, and stable virtual cursor control, click control, and operation target screen control can be performed (effect of claim 5).
  • the multi-step rectangular object extraction process is performed on the virtual cursor activity area image 27 and the virtual button click activity area image in the activity rectangular area extraction process. It is possible to prevent malfunction caused by a shadow or the like and perform stable virtual cursor control, click control, and operation target screen control (effect of claim 6).
  • FIG. 29 is a block diagram showing a second embodiment of the input device according to the present invention.
  • An input device 1b shown in this figure is composed of a plastic member or the like formed in a box shape, and is an input device housing (illustrated) disposed in the vicinity of a remote operation target device such as a personal computer, a television, an air conditioner, or a large screen screen device.
  • a remote operation target device such as a personal computer, a television, an air conditioner, or a large screen screen device.
  • the left eye side image processing board 33 to be generated, the right eye side virtual cursor activity area image and the right eye side virtual button click activity area image which are arranged in the input device housing and output from the right eye side image processing board 31
  • the left eye side virtual cursor activity region image and the left eye side virtual button click activity region image output from the left eye side image processing board 32 are image-processed to generate pointing data corresponding to the movement of the user's hand.
  • Common processing bases that are supplied to devices such as personal computers, televisions, air conditioners, and large screen devices via cables such as USB cables and signal connection cables. And a 34.
  • Pointing data is generated, and the pointing data is supplied to the remote operation target device through the path of the input device 1b ⁇ cable ⁇ remote operation target device, and the operation of the remote operation target device is controlled.
  • the right-eye video camera main body 30 is configured by a color camera having a resolution of about 320 pixels ⁇ 240 pixels, and captures a user when a power supply voltage, a clock signal, and the like are output from the right-eye image processing board 31. Then, the color video signal obtained thereby is supplied to the right eye side image processing board 31.
  • the right eye side image processing board 31 converts the color video signal output from the video camera body for the right eye into a color image in RGB format, and then is specified in advance by the HSV (Hue / Saturation / Brightness) method. Using a color mask necessary for extracting a color image of a color (for example, skin color), the skin color image extracting circuit 35 for extracting the skin color image in the color image and the video camera body 30 for the right eye are output. After the color video signal is converted into a color image in RGB format, a graying processing circuit 36 for converting the color video signal into a monochrome image having a preset gradation, and a monochrome image output from the graying processing circuit 36 are set in advance.
  • HSV Human / Saturation / Brightness
  • the screen is divided by the number of screen divisions (however, when screen division is not set, this screen division processing is skipped) and binarized by the maximum likelihood threshold method.
  • the right eye side image processing board 31 is stored in the frame buffer circuit 39 and the frame buffer circuit 39 for temporarily storing the binary image output from the color filtering processing circuit 38 for several frames to several tens of frames.
  • the inter-frame difference processing circuit 40 that performs the inter-frame difference processing while sequentially reading the binarized images, and generates the difference image, and the difference images output in units of frames from the inter-frame difference processing circuit 40 are divided into the divided areas.
  • the histogram processing circuit 41 that accumulates each time to generate a histogram, and performs statistical processing on the histogram output from the histogram processing circuit 41, and also uses the statistical processing result to determine virtual cursor activity area determination processing, virtual button click activity Performs area determination processing, multi-step rectangular object extraction processing, etc., and influences such as shadows Removing the right eye of the virtual cursor activity area image, and a work rectangular area extraction processing circuit 42 for generating a virtual button click activity area image of the right eye side.
  • the activity rectangular area extraction processing circuit 42 compares the activity rectangular area extracted from the latest difference image data of the multi-stage difference image data generated by the histogram processing circuit 41 with the virtual button click activity area image. When the extracted activity rectangle area exceeds the range of the virtual button click activity area, the extraction action rectangle area is judged to be invalid so that the movement that the user does not intend for the input operation is ignored as noise. can do.
  • graying processing, screen division / binarization processing, color filtering processing, frame buffer processing, inter-frame difference processing, histogram processing, active rectangular area Extraction processing is sequentially performed to generate a virtual cursor activity region image on the right eye side and a virtual button click activity region image on the right eye side, which are supplied to the common processing board 34.
  • the left-eye video camera main body 32 is composed of a color camera having a resolution of about 320 pixels ⁇ 240 pixels.
  • the power supply voltage, clock signal, etc. are output from the left-eye image processing board 33, the user And the color video signal obtained thereby is supplied to the left eye side image processing board 33.
  • the left-eye-side image processing board 33 converts the color video signal output from the left-eye video camera body 32 into a color image in RGB format, and is preset in the HSV (Hue / Saturation / Brightness) method.
  • a skin color image extraction circuit 43 that extracts a skin color image in a color image using a color mask necessary for extracting a color image of a specific color (for example, skin color) and an output from the left-eye video camera main body 32
  • a graying processing circuit 44 for converting to a monochrome image having a preset gradation, and a monochrome image output from the graying processing circuit 44 are set in advance.
  • the left eye side image processing board 33 is stored in the frame buffer circuit 47 and the frame buffer circuit 47 for temporarily storing the binarized image output from the color filtering processing circuit 46 for several frames to several tens of frames.
  • the inter-frame difference processing circuit 48 that performs the inter-frame difference processing while sequentially reading the binarized images, and generates the difference image, and the difference images output in units of frames from the inter-frame difference processing circuit 48 are divided into the divided areas.
  • a histogram processing circuit 49 that accumulates each time and generates a histogram, and performs statistical processing on the histogram output from the histogram processing circuit 49, and also uses the statistical processing result to determine virtual cursor activity area determination processing and virtual button click activity Performs area determination processing, multi-step rectangular object extraction processing, etc., and influences such as shadows It comprises removing the left eye side of the virtual cursor activity area image, and activities rectangular area extraction processing circuit 50 for generating a virtual button click activity area image of the left eye side.
  • the activity rectangular area extraction processing circuit 50 compares the activity rectangular area extracted from the latest difference image data of the multi-stage difference image data generated by the histogram processing circuit 49 with the virtual button click activity area image. As shown in FIG. 37 described above, when the extracted activity rectangular area exceeds the range of the virtual button click activity area, the user can perform an input operation by performing a process of determining that these extracted activity rectangle areas are invalid. Unintentional movement can be ignored as noise
  • graying processing For the color video signal output from the left-eye video camera main body 32, graying processing, screen division / binarization processing, color filtering processing, frame buffer processing, inter-frame difference processing, histogram processing, active rectangular area Extraction processing is sequentially performed to generate a virtual cursor activity region image on the left eye side and a virtual button click activity region image on the left eye side, and supply them to the common processing board 34.
  • the common processing board 34 includes right-eye web camera 30 and left-eye web camera 32 attachment position data (horizontal distance “B”, vertical distance, etc.) necessary for position correction by the binocular parallax method, right-eye web camera. 30, a shooting condition setting circuit 51 in which shooting condition information such as the focal length “f” of the left-eye web camera 32 is set, and a right-eye virtual cursor activity region image output from the right-eye image processing board 31
  • the right eye side virtual button click activity region image, the left eye side virtual cursor activity region image output from the left eye side image processing board 33, and the left eye side virtual button click activity region image include an activity rectangular region.
  • Process for correcting the position of each active rectangular area by the binocular parallax method using the shooting condition information set in the shooting condition setting circuit 51, and adding a number to each active rectangular area in order of size Processing to calculate the distance between the center coordinates (center coordinate distance) of each activity rectangle area to which the same number is added, processing to select the activity rectangle area corresponding to each center coordinate distance that is less than or equal to the predetermined value, and selected Perform a process such as creating a virtual cursor activity area image on the left eye side that contains only the activity rectangle area and not an unselected activity rectangle area, and a virtual button click activity area image on the left eye side.
  • the activity rectangular area selection processing circuit 52 uses the viewing angles of the web cameras 30 and 32 as constants to change the color according to the center coordinate distance in the horizontal direction between the center points of the right eye activity rectangular area and the left eye activity rectangle area. By correcting the distance from the camera to the subject, it is possible to more accurately remove the influence of the movement of the person in front of the user and the person behind.
  • the principle of correction and the calculation formula for correction are as shown in FIG. 39 and Formula 4 described above.
  • the active rectangular area selection processing circuit 52 can realize high-speed processing by holding the distance to the subject calculated in advance according to the center coordinate distance as table data.
  • the common processing board 34 when the active rectangle area group exists in the latest virtual cursor activity area image among the virtual cursor activity area images on the left eye side held in the activity rectangle area selection processing circuit 52, Virtual cursor position instruction, virtual cursor shape instruction, virtual cursor color instruction, operation target screen scroll instruction, operation target screen enlargement instruction, operation target screen reduction instruction based on the number of active rectangle areas, shape, movement presence / absence, movement direction, etc.
  • the virtual cursor control processing / display control processing circuit 53 for generating a pointing data such as a double-click instruction.
  • an active rectangle area is included in the cursor activity area image and the left eye side virtual button click activity area image, the user removes noise caused by the movement of the person in front of the user, the movement of the person behind the user, etc.
  • a virtual cursor position instruction, virtual cursor shape instruction, virtual cursor color instruction, operation target screen scroll instruction, operation target screen enlargement instruction, operation target screen Pointing data such as a reduction instruction, operation target screen rotation instruction, etc. is generated, and the personal computer, TV, air conditioner, large screen that is the target of remote operation Supplies such as clean equipment.
  • color filtering processing and graying are performed on a low-resolution color image obtained by photographing the user with the right-eye video camera body 30 and the left-eye video camera body 32.
  • User's hand movements including processing, image segmentation / binarization processing, frame buffer processing, inter-frame difference processing, histogram processing, active rectangular area extraction processing, active rectangular area selection processing, virtual cursor control processing / screen control processing, etc.
  • Generating pointing data such as a virtual cursor position instruction, virtual cursor shape instruction, virtual cursor color instruction, operation target screen scroll instruction, operation target screen enlargement instruction, operation target screen reduction instruction, operation target screen rotation instruction, Since it is supplied to the remote operation target device, the virtual cursor size and virtual Cursor position, the virtual cursor color, click, scroll up and down of the operation target screen, left and right scrolling, enlargement, reduction, rotation, etc., can be remotely operated (Effect of Claim 2).
  • the cost of the input device 1b can be kept low. Effect of 2).
  • the low-resolution color image obtained by photographing the user with the right-eye video camera main body 30 and the left-eye video camera main body 32 is subjected to graying processing, image division / 2. Since the binarized image obtained by performing the binarization process and the color filtering process is stored in the frame buffer circuits 39 and 47, the input device 1b can be used even when the storage capacity of the frame buffer circuits 39 and 47 is small. The cost of the entire apparatus can be kept low (effect of claim 2).
  • graying processing, image division / 2 are performed on a low-resolution color image obtained by photographing the user with the right-eye video camera body 30 and the left-eye video camera body 32.
  • Image processing with a small number of stages such as value processing, color filtering processing, frame buffer processing, inter-frame difference processing, histogram processing, active rectangular area extraction processing, active rectangular area selection processing, virtual cursor control processing / screen control processing, etc.
  • the right-eye activity rectangular area obtained by photographing the user with the right-eye video camera main body 30 and the left-eye video camera main body 32 and the left-eye activity After correcting the center coordinate position with the binocular parallax method for the rectangular area, the center coordinate position is compared in the order of size, and each active rectangle on the right eye side corresponding to the focus position based on the comparison result Since the area and each activity rectangle area on the left-eye side are selected, other than the user's hand at the focus position of the right-eye video camera main body 30 and the left-eye video camera main body 32, for example, the user Even if there is a person behind and moving, only the movement of the user's hand is extracted without being affected by this, size control of the virtual cursor, position control, color control, click control, enlargement control of the operation target screen , Reduction control, rotation Please, vertical scroll control can be performed such as the left and right scroll control (Effect of Claim 2).
  • the user when the user is moving only one hand, it is any one of the virtual cursor control instruction, the click control instruction, and the scroll control instruction. Since pointing data is generated to indicate and indicate virtual cursor size instructions, virtual cursor position instructions, virtual cursor color instructions, scroll control instructions, click instructions, etc., display on the display on the remote operation target device side with only one hand It is possible to remotely control the size, position, color, click operation, scroll of the operation target screen, and the like of the virtual cursor being operated (effect of claim 3).
  • the right hand movement and the left hand movement are detected, and control instructions on the operation target screen are displayed. Since it is determined that there is a pointing data that indicates an operation target screen enlargement instruction, an operation target screen reduction instruction, an operation target screen rotation instruction, etc., the user only needs to move the right hand and left hand to the remote operation target device side
  • the operation target screen displayed on the display can be enlarged, reduced, and rotated (effect of claim 4).
  • the active rectangular area extraction processing circuits 42 and 50 statistically process the histogram, and using the statistical processing result, the virtual cursor active area image on the right eye side from the histogram, the right eye
  • the virtual button click activity area image on the side, the virtual cursor activity area image on the left side, and the virtual button click activity area image on the left side are created so that moving parts such as the user's hand can be accurately detected.
  • stable virtual cursor control, click control, and operation target screen control can be performed (effect of claim 5).
  • the right eye side virtual cursor activity area image, the right eye side virtual button click activity area image, and the left virtual cursor activity area Multi-step rectangular object extraction processing is performed on the image and the virtual button click activity area image on the left side to prevent malfunctions caused by the shadow of the user, stable virtual cursor control, click control, operation target Screen control can be performed (effect of claim 6).
  • the entire area of the color image obtained by each web camera 4, 6, right-eye video camera body 30, and left-eye video camera body 32 is grayed out and binarized.
  • the histogram is statistically processed, and the change area rectangle (rectangle including the activity rectangle area) 65 obtained by the activity rectangle area extraction process is An enlarged / reduced rectangular mask 66 enlarged / reduced at a specified enlargement / reduction ratio (for example, enlargement ratio “10%”) is created (step S71), and the monochrome image obtained by graying out the entire color image area of the next frame. Only the portion corresponding to the enlarged / reduced rectangular mask 66 (the image 67 of the active region portion included in the monochrome image) may be extracted from the image and binarized. Step S72).
  • the skin color image is extracted by the color filtering processing by the CPU 9 or the color filtering processing circuits 38 and 46, but the user specifies When controlling the position, click, scrolling of the operation target screen, enlargement of the operation target screen, reduction of the operation target screen, rotation of the operation target screen, etc. using a color operation device such as a red pen Alternatively, a color mask for red extraction may be used, and a red color image may be extracted by the color filtering processing by the CPU 9 or the color filtering processing circuits 38 and 46.
  • color filtering processing moves such as the user's hand included in the color video signals output from the web cameras 4 and 6, the right-eye video camera body 30, and the left-eye video camera body 32. Since the process is performed to extract the image, the color filtering process is omitted when the lighting condition of the place where the user is located is good and the contrast between the image where the user's hand is moving and the background image is large. You may make it do.
  • the present invention is used by being connected to an information device such as an information terminal device or a personal computer, takes an operation image of an operator (user) by a camera, and controls cursor operation of the information device, selection and execution of an application program, and the like.
  • An input device in particular, video input that simplifies the algorithm and reduces the amount of processing data as much as possible to reduce the amount of computation and memory usage, as well as controlling the cursor of the personal computer in real time. It relates to a device and has industrial applicability.

Abstract

Disclosed is an input device by way of a video image for simplifying an algorithm when remotely operating an information terminal device or personal computer, lowering the amount of data to be processed to reduce the amount of computation and memory usage, greatly reducing the price of the device overall, and enabling remote operation of the personal computer in real time. For a low-resolution color image acquired by photographing a user by Web cameras (4 and 6), graying processing, image partitioning/binarization processing, inter-frame difference processing, histogram processing, active rectangular region extraction processing, active rectangular region selection processing, virtual cursor control processing/screen control processing are performed; movement of hands of an operator is detected without being affected by movement of a person in the background; and size control, position control, color control, and click control of a virtual cursor (25), as well as zoom-up control, zoom-down control, rotation control, vertical scroll control, and horizontal scroll control of a screen to be operated are performed.

Description

入力装置Input device
 本発明は、情報端末装置やパーソナルコンピュータなどの情報機器に接続されて使用され、カメラによって操作者(ユーザ)の動作画像を取り込み、情報機器のカーソル操作やアプリケーションプログラムの選択及び実行などを制御する入力装置であって、特に、アルゴリズムを簡素化すると共に処理データ量を極力少なくして演算量、メモリ使用量を低減すると共に、リアルタイムでパソコンのカーソルなどを制御するようにした、ビデオ映像による入力装置に関する。 The present invention is used by being connected to an information device such as an information terminal device or a personal computer, takes an operation image of an operator (user) by a camera, and controls cursor operation of the information device, selection and execution of an application program, and the like. An input device, in particular, video input that simplifies the algorithm and reduces the amount of processing data as much as possible to reduce the amount of computation and memory usage, as well as controlling the cursor of the personal computer in real time. Relates to the device.
 近年、カメラでユーザを撮影して画像解析し、この画像解析結果を使用してオーディオ装置、エアコンなどの機器を操作する入力装置が種々、提案されている。 In recent years, various input devices have been proposed in which a user is photographed with a camera, image analysis is performed, and an apparatus such as an audio device or an air conditioner is operated using the image analysis result.
 図32は、このような従来の映像による入力装置を説明するための操作入力装置のブロック図である(特許文献1を参照)。 FIG. 32 is a block diagram of an operation input device for explaining such a conventional image input device (see Patent Document 1).
 この図に示す操作入力装置101は、可視光カメラ106(図33参照)などを使用してユーザを撮影しカラー画像を出力する撮像手段102と、撮像手段102から出力されるカラー画像を解析し、ユーザの手の形状を検出する手領域検出手段103と、予め登録されている手形状と手領域検出手段103から出力される手形状とを比較し、操作指示内容を判定する手操作判定手段104と、手操作判定手段104の判定内容に基づき選択メニューを音声、又はプロジェクト画像などでユーザに知らせる選択メニュー表現手段105と、を備えている。 The operation input device 101 shown in this figure uses the visible light camera 106 (see FIG. 33) and the like to image the user and output a color image, and analyzes the color image output from the imaging unit 102. , Hand region detecting means 103 for detecting the shape of the user's hand, and hand operation determining means for comparing the hand shape registered in advance with the hand shape output from the hand region detecting means 103 to determine the operation instruction content 104, and a selection menu expression unit 105 that informs the user of the selection menu by voice or a project image based on the determination content of the manual operation determination unit 104.
 この操作入力装置101は、可視光カメラ106で撮影したカラー画像から手領域を抽出すると共に、その手の形状がどのような形状(例えば、手を傾ける、指を曲げるなど)になっているかを判定し、この判定内容に対応する手操作指示を音声、又はプロジェクト画像などでユーザに知らせるものである。 The operation input device 101 extracts a hand region from a color image photographed by the visible light camera 106 and determines what shape the hand is in (for example, tilting a hand or bending a finger). The judgment is made, and a manual operation instruction corresponding to the judgment content is notified to the user by voice or a project image.
特開2009-104297号JP 2009-104297 A
 ところで、このような従来の操作入力装置101では、ユーザの手形状を正確に検出する必要があることから、可視光カメラ106として手の部分を詳細に撮影できる解像度の高いカメラを使用しなければならず、操作入力装置101全体が高価になってしまうという問題があった。 By the way, in such a conventional operation input device 101, since it is necessary to accurately detect the hand shape of the user, a high-resolution camera capable of photographing the hand portion in detail must be used as the visible light camera 106. In other words, there is a problem that the entire operation input device 101 becomes expensive.
 また、従来の操作入力装置101では、手領域検出手段103によって処理される背景画像と現在の画像が共に解像度が高いため、ピクセル(画素)数、データ量が膨大だった。そのため、その分だけハードディスクの容量、メモリ容量などを大きくしなければならず、操作入力装置101全体が高価になってしまうという問題があった。 In the conventional operation input device 101, both the background image processed by the hand region detection means 103 and the current image have high resolution, so the number of pixels and the amount of data are enormous. Therefore, the capacity of the hard disk, the memory capacity, etc. have to be increased by that amount, and there is a problem that the entire operation input device 101 becomes expensive.
 さらに、従来の操作入力装置101では、手領域検出手段103として図33に示すように、差分領域抽出手段110、肌色領域抽出手段111、2値化補正手段112、距離算出手段113、中央重点補正手段114、手領域候補検出手段115、輪郭長/面積算出手段116、手領域決定手段117などを備え、複雑な演算を行わなければならないのみならず、かなり高速なCPU、又は専用回路を使用しなければリアルタイムで手の形状を検出することができないという問題があった。 Furthermore, in the conventional operation input device 101, as shown in FIG. 33 as the hand region detection unit 103, the difference region extraction unit 110, the skin color region extraction unit 111, the binarization correction unit 112, the distance calculation unit 113, the center weight correction Means 114, hand region candidate detection means 115, contour length / area calculation means 116, hand region determination means 117, etc., which not only have to perform complicated calculations, but also use a fairly fast CPU or dedicated circuit. Otherwise, there was a problem that the shape of the hand could not be detected in real time.
 このため、パソコンのカーソルなどを簡単に遠隔操作する入力装置として上記のような従来の操作入力装置101を使用することは難しく、安価なカメラが適用可能で、かつ少ない演算量、少ないメモリ量で遠隔操作入力できる入力装置の開発が強く望まれていた。 For this reason, it is difficult to use the above-described conventional operation input device 101 as an input device for easily remotely operating a cursor of a personal computer, etc., and an inexpensive camera can be applied, with a small amount of computation and a small amount of memory. The development of an input device capable of remote operation input has been strongly desired.
 また、従来の操作入力装置101では、可視光カメラ106の撮影範囲内にユーザ以外の人が居て手を動かした場合に、これを検出して誤動作してしまうという問題があった。 Further, the conventional operation input device 101 has a problem that when a person other than the user moves within the shooting range of the visible light camera 106 and moves his / her hand, this is detected and malfunctions.
 本発明は、上記の事情に鑑み、解像度が低く安価なカメラの使用を可能にすると共に、演算量、メモリ量を大幅に少なくすることにより装置全体のコストを大幅に低減させながら、ユーザの動きを検知してパソコンのカーソルなどを遠隔操作することができ、さらにカメラの撮影範囲内にユーザ以外の人が居て手を動かした場合でも誤動作しない入力装置を提供することを目的とする。 In view of the above circumstances, the present invention enables the use of a low-resolution and inexpensive camera, and greatly reduces the cost of the entire apparatus by greatly reducing the amount of calculation and memory, and the movement of the user. It is an object of the present invention to provide an input device that can remotely detect a cursor of a personal computer and can operate a personal computer cursor and the like, and does not malfunction even when a person other than the user moves within the shooting range of the camera.
 また、パソコンなどから分離された解像度が低く安価なカメラの使用を可能にすると共に、回路規模を大幅に小さくすることにより装置全体のコストを大幅に低減させながら、ユーザの動きを検知してパソコンのカーソルなどを遠隔操作することができ、さらにカメラの撮影範囲内にユーザ以外の人が居て手を動かした場合でも誤動作しない入力装置を提供することを目的とする。 In addition, it is possible to use an inexpensive camera with low resolution separated from a personal computer, etc., and by significantly reducing the cost of the entire device by significantly reducing the circuit scale, the user's movement is detected and the personal computer is It is an object of the present invention to provide an input device that can remotely control the cursor, etc., and that does not malfunction even when a person other than the user is in the shooting range of the camera and moves his / her hand.
 また、解像度が低く安価なカメラの使用を可能にすると共に、演算量、メモリ量、又は回路規模を大幅に小さくすることにより装置全体のコストを大幅に低減させながら、ユーザの片手動作を検知してパソコンのカーソル、操作対象画面のスクロールなどを遠隔操作することができ、さらにカメラの撮影範囲内にユーザ以外の人が居て手を動かした場合でも誤動作しない入力装置を提供することを目的とする。 It also enables the use of low-resolution and inexpensive cameras, and detects one-handed movement of the user while significantly reducing the overall cost of the device by significantly reducing the amount of computation, memory, or circuit scale. The purpose is to provide an input device that can remotely control the cursor of the computer, the scroll of the operation target screen, etc., and that does not malfunction even if a person other than the user is in the shooting range of the camera and moves his / her hand. To do.
 また、解像度が低く安価なカメラの使用を可能にすると共に、演算量、メモリ量、又は回路規模を大幅に小さくすることにより装置全体のコストを大幅に低減させながら、ユーザの両手動作を検知して操作対象画面の拡大/縮小、操作対象画面の回転などを遠隔操作することができ、さらにカメラの撮影範囲内にユーザ以外の人が居て手を動かした場合でも誤動作しない入力装置を提供することを目的とする。 In addition, it enables the use of low-resolution and inexpensive cameras, and detects the user's two-handed movement while significantly reducing the overall cost of the device by greatly reducing the amount of computation, memory, or circuit scale. Provide an input device that can remotely control operations such as enlargement / reduction of the operation target screen and rotation of the operation target screen, and that does not malfunction even if a person other than the user moves within the shooting range of the camera. For the purpose.
 また、解像度が低く安価なカメラの使用を可能にすると共に、演算量、メモリ量、又は回路規模を大幅に小さくすることにより装置全体のコストを大幅に低減させながら、ユーザの手など動いている部分のみを正確に検知し、安定した仮想カーソル制御、クリック制御、操作対象画面制御を行い、さらにカメラの撮影範囲内にユーザ以外の人が居て手を動かした場合でも誤動作しない入力装置を提供することを目的とする。 In addition to enabling the use of low-resolution and inexpensive cameras, and reducing the amount of computation, memory, or circuit scale significantly, the cost of the entire apparatus is greatly reduced, and the user's hand moves. Provides an input device that accurately detects only the part, performs stable virtual cursor control, click control, control of the operation target screen, and does not malfunction even if a person other than the user moves within the shooting range of the camera The purpose is to do.
 また、解像度が低く安価なカメラの使用を可能にすると共に、演算量、メモリ量、又は回路規模を大幅に小さくすることにより装置全体のコストを大幅に低減させながら、ユーザの影などに起因する誤動作を防止し、安定した仮想カーソル制御、クリック制御、操作対象画面制御を行い、さらにカメラの撮影範囲内にユーザ以外の人が居て手を動かした場合でも誤動作しない入力装置を提供することを目的とする。 In addition, it is possible to use an inexpensive camera with a low resolution, and due to the shadow of the user while greatly reducing the cost of the entire apparatus by significantly reducing the amount of computation, the amount of memory, or the circuit scale. To provide an input device that prevents malfunctions, performs stable virtual cursor control, click control, and operation target screen control, and does not malfunction even when a person other than the user moves within the shooting range of the camera. Objective.
 また、解像度が低く安価なカメラの使用を可能にすると共に、演算量、メモリ量、又は回路規模を大幅に小さくすることにより装置全体のコストを大幅に低減させながら、カメラで得られた画像の中からユーザの手画像より少し広い範囲に含まれる画像のみを有効にし、それ以外の画像を無効にして変化領域以外の部分に存在するノイズを除去することができ、さらにカメラの撮影範囲内にユーザ以外の人が居て手を動かした場合でも誤動作しない入力装置を提供することを目的とする。 In addition, it is possible to use an inexpensive camera with a low resolution, and to significantly reduce the cost of the entire apparatus by significantly reducing the amount of calculation, memory, or circuit scale. It is possible to enable only the image included in a slightly wider range than the user's hand image from inside, invalidate other images and remove noise existing in parts other than the change area, and within the shooting range of the camera An object of the present invention is to provide an input device that does not malfunction even when a person other than the user is present and moves his / her hand.
 また、解像度が低く安価なカメラの使用を可能にすると共に、演算量、メモリ量、又は回路規模を大幅に小さくすることにより装置全体のコストを大幅に低減させながら、ユーザの手など動いている部分のみを正確に検知し、安定した仮想カーソル制御、クリック制御、操作対象画面制御を行い、さらにユーザが意図していない手の動きによる誤動作を防止する入力装置を提供することを目的とする。 In addition to enabling the use of low-resolution and inexpensive cameras, and reducing the amount of computation, memory, or circuit scale significantly, the cost of the entire apparatus is greatly reduced, and the user's hand moves. It is an object of the present invention to provide an input device that accurately detects only a portion, performs stable virtual cursor control, click control, and operation target screen control, and prevents malfunction caused by hand movements that are not intended by the user.
 また、解像度が低く安価なカメラの使用を可能にすると共に、演算量、メモリ量、又は回路規模を大幅に小さくすることにより装置全体のコストを大幅に低減させながら、両眼視差法を使用してカメラから被写体までの計測距離を補正により精度を高めた入力装置を提供することを目的とする。 The binocular parallax method can be used while significantly reducing the cost of the entire device by significantly reducing the amount of computation, memory, or circuit scale while allowing the use of low-resolution and inexpensive cameras. An object of the present invention is to provide an input device with improved accuracy by correcting the measurement distance from the camera to the subject.
 また、解像度が低く安価なカメラの使用を可能にすると共に、演算量、メモリ量、又は回路規模を大幅に小さくすることにより装置全体のコストを大幅に低減させながら、両眼視差法を使用してカメラから被写体までの計測距離を補正する場合の処理を高速化させた入力装置を提供することを目的とする。 The binocular parallax method can be used while significantly reducing the cost of the entire device by significantly reducing the amount of computation, memory, or circuit scale while allowing the use of low-resolution and inexpensive cameras. An object of the present invention is to provide an input device that speeds up processing when correcting a measurement distance from a camera to a subject.
 上記の目的を達成するために、本発明の入力装置は、ビデオカメラで得られた操作者の画像を処理して、操作者の動作内容に応じた操作指示を生成する入力装置において、操作者を撮影する右眼用カラーカメラと、この右眼用カラーカメラから所定距離だけ離れた位置に、前記右眼用カラーカメラと並んで配置され、前記操作者を撮影する左眼用カラーカメラと、前記右眼用カラーカメラから出力されるカラー画像に対し、グレー化処理、画像分割/2値化処理、フレーム間差分処理、ヒストグラム処理、活動矩形領域抽出処理を行い、前記操作者の右眼側活動矩形領域を抽出する右眼側画像処理プログラムと、前記左眼用カラーカメラから出力されるカラー画像に対し、グレー化処理、画像分割/2値化処理、フレーム間差分処理、ヒストグラム処理、活動矩形領域抽出処理を行い、前記操作者の左眼側活動矩形領域を抽出する左眼側画像処理プログラムと、前記右眼側画像処理プログラムで得られた右眼側活動矩形領域、前記左眼側画像処理プログラムで得られた左眼側活動矩形領域に対し、両眼視差法を使用した活動矩形領域選択処理、仮想カーソル制御処理/画面制御処理を行って、前記操作者の手、又は指先の動きを検出し、この検出結果に応じた操作指示を生成する画像処理プログラムとを備えることを特徴としている。 In order to achieve the above object, an input device of the present invention is an input device that processes an operator's image obtained by a video camera and generates an operation instruction according to the operation content of the operator. A color camera for the right eye that shoots, and a color camera for the left eye that is arranged alongside the color camera for the right eye at a position away from the color camera for the right eye by a predetermined distance; The color image output from the right-eye color camera is subjected to graying processing, image division / binarization processing, interframe difference processing, histogram processing, and active rectangular area extraction processing, and the right eye side of the operator A right eye side image processing program for extracting an active rectangular area and a color image output from the left eye color camera, graying processing, image division / binarization processing, inter-frame difference processing, hysteresis Gram processing, activity rectangle area extraction processing, the left eye side image processing program for extracting the left eye side activity rectangle area of the operator, the right eye side activity rectangle area obtained by the right eye side image processing program, The left-eye side active rectangular area obtained by the left-eye side image processing program is subjected to an active rectangular area selection process using a binocular parallax method, a virtual cursor control process / screen control process, and the operator's hand Or an image processing program that detects the movement of the fingertip and generates an operation instruction according to the detection result.
 また、本発明の入力装置は、ビデオカメラで得られた操作者の画像を処理して、操作者の動作内容に応じた操作指示を生成し、遠隔操作対象機器の動作を制御する入力装置において、箱形に形成される入力装置筐体と、この入力装置筐体の前面左側に取り付けられ、操作者の画像を撮影する右眼用カラーカメラ本体と、前記入力装置筐体の前面右側に取り付けられ、前記操作者の画像を撮影する左眼用カラーカメラ本体と、前記入力装置筐体内に配置され、グレー化処理回路、画像分割/2値化処理回路、フレーム間差分処理回路、ヒストグラム処理回路、活動矩形領域抽出処理回路によって、前記右眼用カラーカメラ本体から出力されるカラー画像を処理して、前記操作者の右眼側活動矩形領域を抽出する右眼側画像処理基板と、前記入力装置筐体内に配置され、グレー化処理回路、画像分割/2値化処理回路、フレーム間差分処理回路、ヒストグラム処理回路、活動矩形領域抽出処理回路によって、前記左眼用カラーカメラ本体から出力されるカラー画像を処理して、前記操作者の左眼側活動矩形領域を抽出する左眼側画像処理基板と、前記入力装置筐体内に配置され、活動矩形領域選択処理回路、仮想カーソル制御処理/画面制御処理回路によって、前記右眼側画像処理基板で得られた右眼側活動矩形領域、前記左眼側画像処理基板で得られた左眼側活動矩形領域に、両眼視差法を使用した活動矩形領域選択処理、仮想カーソル制御処理/画面制御処理を行って、前記操作者の手、又は指先の動きを検出し、この検出結果に応じたポインティングデータを生成し、遠隔操作対象機器の動作を制御する共通処理基板とを備えることを特徴としている。 Further, the input device of the present invention is an input device that processes an image of an operator obtained by a video camera, generates an operation instruction according to the operation content of the operator, and controls the operation of the remote operation target device. An input device case formed in a box shape, a color camera body for a right eye that is attached to the front left side of the input device case and takes an image of an operator, and attached to the front right side of the input device case A color camera body for the left eye that captures an image of the operator, and a graying processing circuit, an image division / binarization processing circuit, an inter-frame difference processing circuit, and a histogram processing circuit, which are disposed in the input device casing. A right-eye side image processing board for processing a color image output from the right-eye color camera main body by the activity rectangular area extraction processing circuit and extracting the right-eye side activity rectangular area of the operator; Arranged in the apparatus casing and output from the color camera body for the left eye by a graying processing circuit, an image division / binarization processing circuit, an inter-frame difference processing circuit, a histogram processing circuit, and an active rectangular area extraction processing circuit A left-eye side image processing board for processing a color image and extracting the left-eye side active rectangular area of the operator, and an active rectangular area selection processing circuit, a virtual cursor control process / screen disposed in the input device casing Activity using the binocular parallax method on the right-eye side active rectangular area obtained on the right-eye side image processing board and the left-eye side active rectangular area obtained on the left-eye side image processing board by the control processing circuit Performs rectangular area selection processing, virtual cursor control processing / screen control processing, detects the movement of the operator's hand or fingertip, generates pointing data according to the detection result, and performs remote operation It is characterized in that it comprises a common substrate for controlling the operation of the elephant equipment.
 また、本発明の入力装置は、前記仮想カーソル制御処理/画面制御処理、又は前記仮想カーソル制御処理/画面制御処理回路は、仮想カーソル活動領域画像上に活動矩形領域群が1つあるとき、その形状、移動有無に基づき、カールソル制御指示、又は画面スクロール指示を生成することを特徴としている。 In the input device of the present invention, when the virtual cursor control process / screen control process or the virtual cursor control process / screen control processing circuit has one active rectangular area group on the virtual cursor active area image, Based on the shape and presence / absence of movement, a Carlsol control instruction or a screen scroll instruction is generated.
 また、本発明の入力装置は、前記仮想カーソル制御処理/画面制御処理、又は前記仮想カーソル制御処理/画面制御処理回路は、仮想カーソル活動領域画像上に活動矩形領域群が2つあるとき、その移動方向に基づき、画面回転指示、画面拡大指示、画面縮小指示の何れかを生成することを特徴としている。 In the input device of the present invention, when the virtual cursor control process / screen control process or the virtual cursor control process / screen control processing circuit has two active rectangular area groups on the virtual cursor active area image, Based on the moving direction, one of a screen rotation instruction, a screen enlargement instruction, and a screen reduction instruction is generated.
 また、本発明の入力装置は、前記活動矩形領域抽出処理、又は前記活動矩形領域抽出処理回路は、ヒストグラムの統計処理結果を使用して、前記ヒストグラムから仮想カーソル活動領域画像、仮想ボタンクリック活動領域画像を作成することを特徴としている。 In the input device of the present invention, the activity rectangular area extraction process, or the activity rectangular area extraction processing circuit uses a histogram statistical processing result to generate a virtual cursor activity area image, a virtual button click activity area from the histogram. It is characterized by creating images.
 また、本発明の入力装置は、前記活動矩形領域抽出処理、又は前記活動矩形領域抽出処理回路は、前記仮想カーソル活動領域画像、又は前記仮想ボタンクリック活動領域画像に対し、多段階矩形オブジェクト抽出処理を行い、ノイズ成分を除去することを特徴としている。 In the input device of the present invention, the activity rectangle area extraction process, or the activity rectangle area extraction processing circuit is a multi-stage rectangle object extraction process for the virtual cursor activity area image or the virtual button click activity area image. And removing noise components.
 また、本発明の入力装置は、拡大/縮小矩形マスク作成処理、又は拡大/縮小矩形マスク作成処理回路を付加し、前記拡大/縮小矩形マスク作成処理、又は前記拡大/縮小矩形マスク作成処理回路によって、前記カラーカメラ、前記カラーカメラ本体で得られたカラー画像の中から、前記仮想カーソル活動領域画像上の変化領域矩形、又は前記仮想ボタンクリック活動領域画像の変化領域矩形に対応する画像を抽出し、それ以外の画像をカットして、ノイズ成分を除去することを特徴としている。 Further, the input device of the present invention adds an enlargement / reduction rectangle mask creation process or an enlargement / reduction rectangle mask creation processing circuit, and performs the enlargement / reduction rectangle mask creation process or the enlargement / reduction rectangle mask creation processing circuit. The image corresponding to the change area rectangle on the virtual cursor activity area image or the change area rectangle of the virtual button click activity area image is extracted from the color images obtained by the color camera and the color camera body. The other image is cut to remove noise components.
 また、本発明の入力装置は、前記活動矩形領域抽出処理、又は前記活動矩形領域抽出処理回路は、前記ヒストグラム処理、又は前記ヒストグラム処理回路にて生成した多段階の差分画像データの最新の差分画像データから抽出した活動矩形領域と、前記仮想ボタンクリック活動領域画像との比較に基づき無効な抽出活動矩形領域を判定することを特徴としている。 In the input device of the present invention, the active rectangular area extraction process or the active rectangular area extraction processing circuit is the latest difference image of the multi-stage difference image data generated by the histogram processing or the histogram processing circuit. An invalid activity rectangle area is determined based on a comparison between the activity rectangle area extracted from the data and the virtual button click activity area image.
 また、本発明の入力装置は、前記活動矩形領域選択処理では、前記カラーカメラの視野角を定数として前記右眼活動矩形領域本発明の入力装置は、と前記左眼活動矩形領域の中心点間の横方向での中心座標距離に応じて、前記カラーカメラから被写体までの距離を補正することを特徴としている。 In the activity rectangular area selection process, the input apparatus according to the present invention may be configured such that the viewing angle of the color camera is a constant and the right eye activity rectangular area is input between the center point of the left eye activity rectangular area. The distance from the color camera to the subject is corrected according to the center coordinate distance in the horizontal direction.
 また、本発明の入力装置は、前記中心座標距離に応じて補正された前記被写体までの前記距離が予めテーブルデータとして保持していることを特徴としている。 The input device according to the present invention is characterized in that the distance to the subject corrected in accordance with the center coordinate distance is stored in advance as table data.
 本発明によれば、解像度が低く安価なカメラの使用を可能にすると共に、演算量、メモリ量を大幅に少なくすることにより装置全体のコストを大幅に低減させながら、ユーザの動きを検知してパソコンのカーソルなどを遠隔操作することができ、さらにカメラの撮影範囲内にユーザ以外の人が居て手を動かした場合でも誤動作しない入力装置を実現することが可能となる。 According to the present invention, it is possible to use an inexpensive camera with low resolution, and to detect a user's movement while greatly reducing the cost of the entire apparatus by greatly reducing the amount of calculation and memory. An input device that can remotely control a cursor of a personal computer or the like and that does not malfunction even when a person other than the user moves within the shooting range of the camera and moves his / her hand can be realized.
本発明による入力装置の第1の実施形態を示すブロック図である。1 is a block diagram showing a first embodiment of an input device according to the present invention. 図1に示す入力装置の詳細な動作例を示すフローチャートである。3 is a flowchart illustrating a detailed operation example of the input device illustrated in FIG. 1. 図2に示すグレー化/2値化画像処理の詳細な動作例を示すフローチャートである。3 is a flowchart showing a detailed operation example of grayed / binarized image processing shown in FIG. 2. 図2に示すフレーム間差分/ヒストグラム作成処理の詳細な動作例を示すフローチャートである。3 is a flowchart showing a detailed operation example of inter-frame difference / histogram creation processing shown in FIG. 2. 図2に示す活動矩形領域抽出処理の詳細な動作例を示すフローチャートである。It is a flowchart which shows the detailed operation example of the activity rectangular area extraction process shown in FIG. 図2に示す活動矩形領域選択処理の詳細な動作例を示すフローチャートである。It is a flowchart which shows the detailed operation example of the active rectangle area | region selection process shown in FIG. 図2に示す仮想カーソル制御処理、画面制御処理の詳細な動作例を示すフローチャートである。3 is a flowchart showing detailed operation examples of a virtual cursor control process and a screen control process shown in FIG. 2. 図1に示す入力装置で使用される元のヒストグラム例と、変化領域抽出後のヒストグラム例とを示す模式図である。It is a schematic diagram which shows the example of an original histogram used with the input device shown in FIG. 1, and the example of a histogram after change area extraction. 図1に示す入力装置で使用される変化領域抽出後のヒストグラム例と、仮想カーソル活動領域画像例とを示す模式図である。It is a schematic diagram which shows the histogram example after the change area extraction used with the input device shown in FIG. 1, and a virtual cursor active area image example. 図1に示す入力装置で使用される変化領域抽出後のヒストグラム例と、仮想カーソル活動領域画像例とを示す模式図である。It is a schematic diagram which shows the histogram example after the change area extraction used with the input device shown in FIG. 1, and a virtual cursor active area image example. 図1に示す入力装置において、活動矩形領域選択処理で選択された活動矩形領域群の一例を示す模式図である。FIG. 3 is a schematic diagram showing an example of an active rectangular area group selected by an active rectangular area selection process in the input device shown in FIG. 1. 図1に示す入力装置において、活動矩形領域選択処理で選択された活動矩形領域群の一例を示す模式図である。FIG. 3 is a schematic diagram showing an example of an active rectangular area group selected by an active rectangular area selection process in the input device shown in FIG. 1. 図1に示す入力装置で使用される多段階オブジェクト抽出処理の一例を示す模式図である。It is a schematic diagram which shows an example of the multistage object extraction process used with the input device shown in FIG. 図1に示す入力装置で使用される多段階オブジェクト抽出処理の一例を示す模式図である。It is a schematic diagram which shows an example of the multistage object extraction process used with the input device shown in FIG. 図1に示す入力装置で使用される両眼視差法の概要を示す模式図である。It is a schematic diagram which shows the outline | summary of the binocular parallax method used with the input device shown in FIG. 図1に示す入力装置において、両眼視差法による補正を行う前の右活動矩形領域、左活動矩形領域の関係を示す模式図である。In the input device shown in FIG. 1, it is a schematic diagram which shows the relationship between the right active rectangular area and the left active rectangular area before performing correction by the binocular parallax method. 図1に示す入力装置において、両眼視差法による補正を行った後の右活動矩形領域、左活動矩形領域の関係を示す模式図である。FIG. 2 is a schematic diagram showing a relationship between a right active rectangular area and a left active rectangular area after correction by the binocular parallax method in the input device shown in FIG. 1. 図1に示す入力装置において、両眼視差法による補正を行った後の右活動矩形領域、各左活動矩形領域の実例を示す模式図である。In the input device shown in FIG. 1, it is a schematic diagram which shows the actual example of the right activity rectangular area after correcting by the binocular parallax method, and each left activity rectangular area. 図1に示す入力装置によって撮影されるユーザの手の動きと、仮想カーソルとの関係例を示す模式図である。It is a schematic diagram which shows the example of a relationship between the motion of the user's hand image | photographed with the input device shown in FIG. 1, and a virtual cursor. 図1に示す入力装置によって撮影されるユーザの手の動きと、仮想カーソルとの関係例を示す模式図である。It is a schematic diagram which shows the example of a relationship between the motion of the user's hand image | photographed with the input device shown in FIG. 1, and a virtual cursor. 図1に示す入力装置によって制御される実カーソルの微調整動作例を示す模式図である。It is a schematic diagram which shows the fine adjustment operation example of the real cursor controlled by the input device shown in FIG. 図1に示す入力装置によって制御される仮想カーソルの色制御例を示す模式図である。It is a schematic diagram which shows the color control example of the virtual cursor controlled by the input device shown in FIG. 図1に示す入力装置で使用される変化領域抽出後のヒストグラム例と、仮想カーソル活動領域画像例とを示す模式図である。It is a schematic diagram which shows the histogram example after the change area extraction used with the input device shown in FIG. 1, and a virtual cursor active area image example. 図1に示す入力装置によって撮影されるユーザの手の動きと、クリック動作との関係例を示す模式図である。It is a schematic diagram which shows the example of a relationship between the motion of the user's hand image | photographed with the input device shown in FIG. 1, and click operation | movement. 図1に示す入力装置によって得られる仮想カーソル活動領域画像上に2つの各活動矩形領域群があるときの一例を示す模式図である。It is a schematic diagram which shows an example when there exist two each active rectangular area groups on the virtual cursor active area image obtained by the input device shown in FIG. 図1に示す入力装置によって得られる仮想カーソル活動領域画像上にある各活動矩形領域群の位置関係例を示す模式図である。It is a schematic diagram which shows the positional relationship example of each active rectangular area group on the virtual cursor active area image obtained by the input device shown in FIG. 図1に示す入力装置によって得られる仮想カーソル活動領域画像上に2つの各活動矩形領域群があるときの一例を示す模式図である。It is a schematic diagram which shows an example when there exist two each active rectangular area groups on the virtual cursor active area image obtained by the input device shown in FIG. 図1に示す入力装置によって得られる仮想カーソル活動領域画像上にある各活動矩形領域群の位置関係例を示す模式図である。It is a schematic diagram which shows the positional relationship example of each active rectangular area group on the virtual cursor active area image obtained by the input device shown in FIG. 本発明による入力装置の第2の実施形態を示すブロック図である。It is a block diagram which shows 2nd Embodiment of the input device by this invention. 本発明による入力装置の他の実施形態を示すフローチャートである。It is a flowchart which shows other embodiment of the input device by this invention. 図30に示すフローチャートの動作例を示す模式図である。FIG. 31 is a schematic diagram illustrating an operation example of the flowchart illustrated in FIG. 30. 従来から知られている操作入力装置の一例を示すブロック図である。It is a block diagram which shows an example of the operation input apparatus known conventionally. 図32に示す手領域検出手段の詳細な回路構成例を示すブロック図である。It is a block diagram which shows the detailed circuit structural example of the hand area | region detection means shown in FIG. 図5に示すフローチャートのステップ41に代えて処理するステップ41’を示している。FIG. 6 shows a step 41 ′ to be processed instead of the step 41 in the flowchart shown in FIG. 5. 図5に示すフローチャートのステップ43に代えて処理するステップ43’を示している。 Step 43 ′ to be processed is shown instead of step 43 in the flowchart shown in FIG. 5. 図6に示すフローチャートのステップ57に代えて処理するステップ57’を示している。FIG. 7 shows a step 57 ′ to be processed instead of the step 57 in the flowchart shown in FIG. 6. 図1に示す入力装置において、抽出活動領域においてユーザの意図しない動作により発生したものをノイズとして除去する動作を説明する模式図である。FIG. 2 is a schematic diagram for explaining an operation for removing noise generated by an operation unintended by a user in an extraction activity region in the input device shown in FIG. 1. 1台のカメラでのカメラ視野角と被写体の幅と被写体までの距離の関係を説明する模式図である。It is a schematic diagram explaining the relationship between the camera viewing angle, the subject width, and the distance to the subject in one camera. 図1に示す入力装置の2台のカメラによる両眼視差法にて、カメラから被写体までの正確な距離の計測を説明する模式図である。It is a schematic diagram explaining the measurement of the exact distance from a camera to a to-be-photographed object by the binocular parallax method by two cameras of the input device shown in FIG.
1.第1の実施形態の説明
 図1は、本発明による入力装置の第1の実施の形態を示すブロック図を示す。
 この図に示す入力装置1aは、パソコン2のディスプレイ部3に設けられる内蔵型のウェブカメラ(請求項1の左眼用カラーカメラ)4と、パソコン2内に設けられるビデオキャプチャ5と、パソコン2のディスプレイ部3に設けられる外付け型のウェブカメラ(請求項1の右眼用カラーカメラ)6と、パソコン2内に設けられるUSBインタフェース7と、パソコン2内に設けられるハードディスク8と、パソコン2内に設けられるCPU9と、パソコン2内に設けられるメモリ10とによって構成されている。この入力装置1aは、各ウェブカメラ4、6で得られたカラー画像を解析して、各ウェブカメラ4、6の設置位置から所定距離範囲、例えば“0.3m”~“0.8m”の範囲内にいるユーザとそれ以外の距離にいる他の人とを区別しながら、ユーザの手、指先などの動きのみを検出し、パソコン2のディスプレイ部3に表示された仮想カーソル25〔図19の(b)参照〕、操作対象画面(OS画面、アプリケーション画面)などを制御し、現在、起動中のアプリケーションをコントロールする。
1. Description of First Embodiment FIG. 1 is a block diagram showing a first embodiment of an input device according to the present invention.
The input device 1a shown in this figure includes a built-in web camera (a left-eye color camera of claim 1) 4 provided in the display unit 3 of the personal computer 2, a video capture 5 provided in the personal computer 2, and a personal computer 2 An external web camera (a right-eye color camera according to claim 1) 6 provided in the display unit 3, a USB interface 7 provided in the personal computer 2, a hard disk 8 provided in the personal computer 2, and the personal computer 2. The CPU 9 is provided inside and the memory 10 is provided in the personal computer 2. The input device 1a analyzes the color images obtained by the web cameras 4 and 6, and has a predetermined distance range from the installation position of the web cameras 4 and 6, for example, “0.3 m” to “0.8 m”. A virtual cursor 25 [FIG. 19] displayed on the display unit 3 of the personal computer 2 is detected by detecting only movements of the user's hand, fingertips, etc. while distinguishing between a user within the range and another person at a distance other than that. (See (b))], the operation target screen (OS screen, application screen) and the like are controlled, and the currently running application is controlled.
 ウェブカメラ4は、320画素×240画素程度の解像度を持つカラーカメラであり、ビデオキャプチャ5から撮影指示が出されているとき、ユーザを撮影して得られたカラービデオ信号をビデオキャプチャ5に供給する。 The web camera 4 is a color camera having a resolution of about 320 pixels × 240 pixels, and supplies the video capture 5 with a color video signal obtained by photographing the user when a shooting instruction is issued from the video capture 5. To do.
 ビデオキャプチャ5は、システムバス12を介して、CPU9から撮影指示が出されているとき、ウェブカメラ4を制御してユーザを撮影すると共に、撮影動作で得られたカラービデオ信号を取り込み、RGB信号形式のカラー画像に変換してCPU9に供給する。 When a shooting instruction is issued from the CPU 9 via the system bus 12, the video capture 5 controls the web camera 4 to shoot a user, captures a color video signal obtained by the shooting operation, and receives an RGB signal. It is converted into a color image of the format and supplied to the CPU 9.
 また、ウェブカメラ6は、ウェブカメラ4から水平方向に所定距離だけ離れてディスプレイ部3の上縁などに取り付けられた、320画素×240画素程度の解像度を持つカラーカメラであり、USBインタフェース7から撮影指示が出されているとき、ユーザを撮影して得られたYUV信号をUSBインタフェース7に供給する。 The web camera 6 is a color camera having a resolution of about 320 pixels × 240 pixels attached to the upper edge of the display unit 3 at a predetermined distance in the horizontal direction from the web camera 4. When the photographing instruction is issued, the YUV signal obtained by photographing the user is supplied to the USB interface 7.
 USBインタフェース7は、システムバス12を介して、CPU9から撮影指示が出されているとき、ウェブカメラ6を制御してユーザの画像を撮影すると共に、撮影動作で得られたYUV信号を取り込んでCPU9に供給し、RGB信号形式のカラー画像に変換する。 When a shooting instruction is issued from the CPU 9 via the system bus 12, the USB interface 7 controls the web camera 6 to take a user's image and captures the YUV signal obtained by the shooting operation. Is converted into a color image in the RGB signal format.
 ハードディスク8は、OS(Operating System)、定数データなどが格納されるOS格納エリア13と、インターネットエクスプローラプログラム、ブラウザプログラムなどのアプリケーションプログラムが格納されるアプリケーション格納エリア14と、本発明で使用する画像処理プログラム(請求項1の右眼側画像処理プログラム、左眼側画像処理プログラム、画像処理プログラム)が格納される画像処理プログラム格納エリア15と、HSV(色相・彩度・明度)方式で、予め設定されている特定色(例えば、肌色)のカラー画像を抽出するのに必要なカラーマスク、2値化画像、ヒストグラム、仮想カーソル活動領域画像27(図9参照)、仮想ボタンクリック活動領域画像などが格納される画像格納エリア16とを備えている。そして、CPU9から読み出し指示が出力されたとき、システムバス12を介してこれを取り込み、指定されたエリアに格納されているOS、定数データ、アプリケーションプログラム、画像処理プログラム、2値化画像、ヒストグラム、仮想カーソル活動領域画像27、仮想ボタンクリック活動領域画像などを読み出し、システムバス12を介してCPU9に供給する。また、CPU9から書き込み指示、データが出力されたとき、システムバス12を介してこれらを取り込み、書き込み指示で指定されたエリア、例えば画像格納エリア16などにデータを記憶させる。 The hard disk 8 includes an OS storage area 13 in which an OS (Operating System) and constant data are stored, an application storage area 14 in which application programs such as an Internet Explorer program and a browser program are stored, and image processing used in the present invention. An image processing program storage area 15 for storing a program (a right eye side image processing program, a left eye side image processing program, an image processing program) according to claim 1 and an HSV (hue / saturation / lightness) method are set in advance. A color mask, a binarized image, a histogram, a virtual cursor active area image 27 (see FIG. 9), a virtual button click active area image, and the like necessary for extracting a specific color image (for example, skin color). And an image storage area 16 to be stored. When a read instruction is output from the CPU 9, the CPU 9 captures this via the system bus 12 and stores the OS, constant data, application program, image processing program, binary image, histogram, The virtual cursor activity area image 27, the virtual button click activity area image, and the like are read out and supplied to the CPU 9 via the system bus 12. Further, when a write instruction and data are output from the CPU 9, these are taken in via the system bus 12 and stored in an area designated by the write instruction, for example, the image storage area 16.
 CPU9は、ハードディスク8に格納されているOS、定数データ、アプリケーションプログラムなどで指定された表示データを生成してシステムバス12に接続された表示インタフェース11に供給し、ディスプレイ部3に操作対象画面を表示させる。また、右眼側画像処理プログラム、左眼側画像処理プログラム、画像処理プログラムなどで記述された画像処理を行い、操作対象画面に表示されている仮想カーソルのサイズ、位置などの制御、クリック制御、スクロール制御、画面回転制御、画面拡大制御、画面縮小制御などを行う。 The CPU 9 generates display data specified by the OS, constant data, application program and the like stored in the hard disk 8 and supplies the display data to the display interface 11 connected to the system bus 12, and displays the operation target screen on the display unit 3. Display. In addition, the image processing described in the right eye side image processing program, the left eye side image processing program, the image processing program, etc. is performed, the control of the size and position of the virtual cursor displayed on the operation target screen, click control, Scroll control, screen rotation control, screen enlargement control, screen reduction control, etc.
 メモリ10は、数百メガバイト~数ギガバイト程度の容量を持ち、CPU9がアプリケーションプログラム、右眼側画像処理プログラム、左眼側画像処理プログラム、画像処理プログラムなどで指定された処理を行うときの一時データ格納エリアとして使用される。 The memory 10 has a capacity of about several hundred megabytes to several gigabytes, and temporary data when the CPU 9 performs processing specified by an application program, a right eye side image processing program, a left eye side image processing program, an image processing program, or the like. Used as storage area.
 次に、図2~図7に示す各フローチャート、図8~図28に示す各模式図などを参照しながら、入力装置1aの画像処理動作、カーソル制御動作、画面制御動作などを説明する。 Next, the image processing operation, cursor control operation, screen control operation, etc. of the input device 1a will be described with reference to the flowcharts shown in FIGS. 2 to 7, the schematic diagrams shown in FIGS. 8 to 28, and the like.
《2値化画像生成、格納》
 まず、パソコン2の電源が投入されて、アプリケーションプログラム、右眼側画像処理プログラム、左眼側画像処理プログラム、画像処理プログラムが起動されると、図2のフローチャートに示すようにCPU9によって、ビデオキャプチャ5が制御されて、ウェブカメラ4の撮影動作で得られたカラービデオ信号が取り込まれると共に、RGB信号形式のカラー画像に変換されてメモリ10などに一時記憶される(ステップS1)。
<< Binary image generation and storage >>
First, when the personal computer 2 is turned on and an application program, a right eye side image processing program, a left eye side image processing program, and an image processing program are started, the CPU 9 performs video capture as shown in the flowchart of FIG. 5 is controlled, the color video signal obtained by the photographing operation of the web camera 4 is taken in, and is converted into a color image in the RGB signal format and temporarily stored in the memory 10 or the like (step S1).
 また、この動作と並行し、CPU9によってUSBインタフェース7が制御されて、ウェブカメラ6の撮影動作で得られたYUV信号が取り込まれると共に、RGB信号形式のカラー画像に変換されてメモリ10などに一時記憶される(ステップS2)。 In parallel with this operation, the USB interface 7 is controlled by the CPU 9 so that the YUV signal obtained by the photographing operation of the web camera 6 is taken in and converted into a color image in the RGB signal format and temporarily stored in the memory 10 or the like. Stored (step S2).
 また、これらの各動作と並行し、CPU9によってメモリ10などに一時記憶されている各カラー画像(各ウェブカメラ4、6の撮影動作で得られたカラー画像)のうち、パソコン2側からユーザを見たとき右眼に対応するウェブカメラ、例えば外付け型のウェブカメラ6で得られたカラー画像が読み取られる(ステップS3)。 In parallel with these operations, among the color images (color images obtained by the shooting operations of the web cameras 4 and 6) temporarily stored in the memory 10 or the like by the CPU 9, a user is selected from the personal computer 2 side. When viewed, a color image obtained by a web camera corresponding to the right eye, for example, an external web camera 6 is read (step S3).
 この後、CPU9によってグレー化/2値化画像処理が開始される(ステップS4)。即ち、図3のフローチャートに示すようにハードディスク8の画像格納リア16に格納されているカラーマスクで、外付け型のウェブカメラ6で得られたカラー画像がマスクされ、カラー画像の中から予め設定されている特定色(例えば、肌色)のカラー画像(肌色画像)が抽出されると共に(ステップS21)、外付け型のウェブカメラ6で得られたカラー画像がグレー処理化されて、予め設定されている階調のモノクロ画像に変換され、1フレーム分の画像容量を低減させる(ステップS22)。 Thereafter, the CPU 9 starts graying / binarized image processing (step S4). That is, as shown in the flow chart of FIG. 3, the color image obtained by the external web camera 6 is masked by the color mask stored in the image storage rear 16 of the hard disk 8, and preset from the color images. A color image (skin color image) of a specific color (for example, skin color) is extracted (step S21), and the color image obtained by the external web camera 6 is gray-processed and set in advance. The image is converted into a monochrome image of a certain gradation, and the image capacity for one frame is reduced (step S22).
 そして、CPU9によって画面分割指示が設定されているかどうかがチェックされ、画面分割指示があればモノクロ画像が複数のエリア(各エリアは各々、数個~数十個の画素によって構成される)に分割され、また画面分割指示がなければ分割処理がスキップされた後、最大尤度しきい値法でモノクロ画像が2値化され、2値化画像が作成される(ステップS23)。 Then, the CPU 9 checks whether or not a screen division instruction is set. If there is a screen division instruction, the monochrome image is divided into a plurality of areas (each area is composed of several to several tens of pixels). If there is no screen division instruction, the division process is skipped, and then the monochrome image is binarized by the maximum likelihood threshold method to create a binarized image (step S23).
 次いで、CPU9によって2値化画像と肌色画像との論理和が取られて、2値化画像中の肌色部分が抽出され(ステップS24)、これが1フレーム分の2値化画像(右眼側の2値化画像)としてハードディスク8の画像格納エリア16に格納される(ステップS25)。 Next, the logical sum of the binarized image and the skin color image is taken by the CPU 9 and the skin color portion in the binarized image is extracted (step S24), and this is the binarized image for one frame (on the right eye side). (Binary image) is stored in the image storage area 16 of the hard disk 8 (step S25).
 この後、CPU9によって図2のフローチャートに示すようにメモリ10などに一時記憶されている各カラー画像(各ウェブカメラ4、6の撮影動作で得られた各カラー画像)のうち、パソコン2側からユーザを見たとき左眼に対応するウェブカメラ、例えば内蔵型のウェブカメラ4で得られたカラー画像が読み取られる(ステップS5)。 After that, among the color images (color images obtained by the shooting operations of the web cameras 4 and 6) temporarily stored in the memory 10 or the like by the CPU 9 as shown in the flowchart of FIG. When the user is viewed, a color image obtained by the web camera corresponding to the left eye, for example, the built-in web camera 4 is read (step S5).
 次いで、CPU9によってグレー化/2値化画像処理が開始される(ステップS6)。即ち、図3のフローチャートに示すようにハードディスク8の画像格納エリア16に格納されているカラーマスクで内蔵型のウェブカメラ4で得られたカラー画像がマスクされ、カラー画像の中から予め設定されている特定色(例えば、肌色)のカラー画像(肌色画像)が抽出されると共に(ステップS21)、内蔵型のウェブカメラ4で得られたカラー画像がグレー処理化されて、予め設定されている階調のモノクロ画像に変換され、1フレーム分の画像容量を低減させる(ステップS22)。 Next, the CPU 9 starts graying / binarized image processing (step S6). That is, as shown in the flowchart of FIG. 3, the color image obtained by the built-in web camera 4 is masked by the color mask stored in the image storage area 16 of the hard disk 8 and preset from the color image. A color image (skin color image) of a specific color (for example, skin color) is extracted (step S21), and the color image obtained by the built-in web camera 4 is subjected to gray processing, and is set in advance. The image is converted into a tone monochrome image, and the image capacity for one frame is reduced (step S22).
 そして、CPU9によって画面分割指示が設定されているかどうかがチェックされ、画面分割指示があればモノクロ画像が複数のエリア(各エリアは各々、数個~数十個の画素によって構成される)に分割され、また画面分割指示がなければ分割処理がスキップされた後、最大尤度しきい値法でモノクロ画像が2値化され、2値化画像が作成される(ステップS23)。 Then, the CPU 9 checks whether or not a screen division instruction is set. If there is a screen division instruction, the monochrome image is divided into a plurality of areas (each area is composed of several to several tens of pixels). If there is no screen division instruction, the division process is skipped, and then the monochrome image is binarized by the maximum likelihood threshold method to create a binarized image (step S23).
 次いで、CPU9によって2値化画像と肌色画像との論理和が取られて、2値化画像中の肌色部分が抽出され(ステップS24)、これが1フレーム分の2値化画像(左眼側の2値化画像)としてハードディスク8の画像格納エリア16に格納される(ステップS25)。 Next, the logical sum of the binarized image and the skin color image is obtained by the CPU 9 and the skin color portion in the binarized image is extracted (step S24), and this is converted into a binarized image for one frame (on the left eye side). (Binary image) is stored in the image storage area 16 of the hard disk 8 (step S25).
 以下、上述した画像処理が繰り返され、FIFO(First・In・First・Out)形式でハードディスク8の画像格納エリア16に右眼側の2値化画像、左眼側の2値化画像が各々数フレーム分~数十フレーム分蓄積される。 Thereafter, the above-described image processing is repeated, and the right eye side binarized image and the left eye side binarized image are in the image storage area 16 of the hard disk 8 in FIFO (First / In / First / Out) format. The frame is stored for tens of frames.
《フレーム間差分、ヒストグラム作成》
 また、この動作と並行し図2のフローチャートに示すように、CPU9によってハードディスク13の画像格納エリア16に格納されている数フレーム分~数十フレーム分の2値化画像の中から、右眼側に対応する最新の2値化画像を含む連続する数フレームの2値化画像が順次読み出される(ステップS7)。
《Difference between frames, creation of histogram》
In parallel with this operation, as shown in the flowchart of FIG. 2, the right eye side is selected from the binarized images of several frames to several tens of frames stored in the image storage area 16 of the hard disk 13 by the CPU 9. Several consecutive frames of binarized images including the latest binarized image corresponding to are sequentially read (step S7).
 そして、CPU9によって読み出すことができた2値化画像のフレーム数がチェックされ、所定数以上のフレーム数であれば(ステップS8)、フレーム間差分/ヒストグラム作成処理が開始される(ステップS9)。即ち、図4のフローチャートに示すように、各2値化画像のうち、連続する2フレーム分の2値化画像に対しフレーム間差分処理が行われると共に(ステップS31、S32)、このフレーム間差分処理で得られた各差分画像が各分割エリア毎に累積加算されて右眼側のヒストグラムが作成され、ハードディスク8の画像格納エリア16に格納される(ステップS33、S34)。 Then, the number of frames of the binarized image that can be read out by the CPU 9 is checked. If the number of frames is equal to or larger than the predetermined number (step S8), the inter-frame difference / histogram creation process is started (step S9). That is, as shown in the flowchart of FIG. 4, the inter-frame difference processing is performed on the binarized images for two consecutive frames among the respective binarized images (steps S31 and S32), and the inter-frame difference is performed. Each difference image obtained by the process is cumulatively added for each divided area to create a right-eye histogram and stored in the image storage area 16 of the hard disk 8 (steps S33 and S34).
 次いで、図2のフローチャートに示すように、CPU9によってハードディスク13の画像格納エリア16に格納されている数フレーム分~数十フレーム分の2値化画像の中から左眼側に対応する最新の2値化画像を含む、連続する数フレームの2値化画像が順次読み出される(ステップS10)。 Next, as shown in the flowchart of FIG. 2, the latest 2 corresponding to the left eye side among the binarized images of several frames to several tens of frames stored in the image storage area 16 of the hard disk 13 by the CPU 9. Several consecutive frames of binarized images including the binarized image are sequentially read out (step S10).
 そして、CPU9によって読み出すことができた2値化画像のフレーム数がチェックされ、所定数以上のフレーム数であれば(ステップS11)、フレーム間差分/ヒストグラム作成処理が開始される(ステップS12)。即ち、図4のフローチャートに示すように、各2値化画像のうち、連続する2フレーム分の2値化画像に対しフレーム間差分処理が行われると共に(ステップS31、S32)、このフレーム間差分処理で得られた各差分画像が各分割エリア毎に累積加算されて左眼側のヒストグラムが作成され、ハードディスク8の画像格納エリア16に格納される(ステップS33、S34)。 Then, the number of frames of the binarized image that can be read out by the CPU 9 is checked. If the number of frames is equal to or larger than the predetermined number (step S11), the interframe difference / histogram creation process is started (step S12). That is, as shown in the flowchart of FIG. 4, the inter-frame difference processing is performed on the binarized images for two consecutive frames among the respective binarized images (steps S31 and S32), and the inter-frame difference is performed. Each difference image obtained by the process is cumulatively added for each divided area to create a left eye side histogram and stored in the image storage area 16 of the hard disk 8 (steps S33 and S34).
《統計処理、仮想カーソルの活動領域決定、変化領域抽出》
 この後、図2のフローチャートに示すように、CPU9によってハードディスク8の画像格納エリア16に格納されている右眼側のヒストグラムが読み出されて(ステップS13)、活動矩形領域抽出処理が開始される(ステップS14)。即ち、図5のフローチャートに示すように、ヒストグラムの各分割エリアの濃度値に対する統計処理が行われ、平均値、濃度分散値、最大値、偏差(±1σ、±2σ)などが演算される(ステップS41)。
《Statistical processing, virtual cursor activity area determination, change area extraction》
Thereafter, as shown in the flowchart of FIG. 2, the right-eye histogram stored in the image storage area 16 of the hard disk 8 is read by the CPU 9 (step S13), and the active rectangular area extraction process is started. (Step S14). That is, as shown in the flowchart of FIG. 5, statistical processing is performed on the density value of each divided area of the histogram, and an average value, density variance value, maximum value, deviation (± 1σ, ± 2σ), and the like are calculated ( Step S41).
 次いで、CPU9によってヒストグラムの各分割エリアの中から、変化領域矩形抽出用のしきい値(例えば、平均値-1σ)より大きい濃度値になっている各分割エリアが抽出され、これらの各分割エリア(活動分割エリア)を含むように矩形状の変化領域矩形65(図31参照)が決定されて、ハードディスク8の画像格納エリア16に格納される。 Next, the CPU 9 extracts each divided area having a density value larger than the threshold value for extracting the change area rectangle (for example, average value −1σ) from each divided area of the histogram. A rectangular change area rectangle 65 (see FIG. 31) is determined so as to include (activity division area) and stored in the image storage area 16 of the hard disk 8.
 また、この動作と並行し、CPU9によって、例えば図8の3次元濃度分布図に示すようにヒストグラムを構成している各分割エリア20(図9、図10参照)のうち、仮想カーソル矩形抽出用のしきい値(例えば、最大値-1σ)より大きい濃度値になっている分割エリア(活動分割エリア21)が抽出される。ヒストグラムは、各分割エリアの20の変化の度数分布を示しており、この処理にてユーザが操作しているところを捉えて動作の変化が激しい活動分割エリア21が抽出される。 In parallel with this operation, the CPU 9 extracts the virtual cursor rectangle from the divided areas 20 (see FIGS. 9 and 10) constituting the histogram as shown in the three-dimensional density distribution diagram of FIG. A divided area (activity divided area 21) having a density value larger than the threshold value (for example, maximum value −1σ) is extracted. The histogram shows the frequency distribution of 20 changes in each divided area. In this process, the active divided area 21 in which the change in the action is significant is extracted by capturing the user's operation.
 これにより、ユーザが指先を大きく回しているとき、図9に示す通り各活動分割エリア21を含むように矩形状の活動矩形領域26が決定されると共に、この決定結果に基づき図11に示すような右眼側の仮想カーソル活動領域画面27が作成されて、ハードディスク8の画像格納エリア16に格納される。 As a result, when the user turns the fingertip greatly, a rectangular activity rectangular area 26 is determined so as to include each activity division area 21 as shown in FIG. 9, and based on the determination result, as shown in FIG. A right-eye virtual cursor activity area screen 27 is created and stored in the image storage area 16 of the hard disk 8.
 また、ユーザが両手を動かしているときには、図10に示す通り各活動分割エリア21を含むように矩形状の活動矩形領域26が決定されて、図12に示すような右眼側の仮想カーソル活動領域画像27が作成され、ハードディスク8の画像格納エリア16に格納される(ステップS42)。 Further, when the user moves both hands, a rectangular activity rectangular area 26 is determined so as to include each activity division area 21 as shown in FIG. 10, and the virtual cursor activity on the right eye side as shown in FIG. An area image 27 is created and stored in the image storage area 16 of the hard disk 8 (step S42).
《仮想ボタンクリックの活動領域決定》
 この後、CPU9によってハードディスク8の画像格納エリア16に格納されている右眼側のヒストグラムが読み出されて、各分割エリア20のうち仮想ボタンクリック矩形抽出用のしきい値(例えば、最大値-2σ)より大きい濃度値になっている分割エリア(活動分割エリア)が抽出されると共に、これらの各活動分割エリアを含むように矩形状の活動矩形領域が決定されて、右眼側の仮想ボタンクリック活動領域画像(図示は省略する)が作成され、ハードディスク8の画像格納エリア16に格納される(ステップS43)。
《Determine activity area for virtual button click》
Thereafter, the CPU 9 reads the right-eye histogram stored in the image storage area 16 of the hard disk 8 and extracts a threshold value for extracting a virtual button click rectangle in each divided area 20 (for example, the maximum value − 2)) divided areas (activity divided areas) having a density value greater than 2σ) are extracted, and a rectangular activity rectangular area is determined so as to include each of these activity divided areas, and a virtual button on the right eye side is determined. A click activity area image (not shown) is created and stored in the image storage area 16 of the hard disk 8 (step S43).
《多段階矩形オブジェクト抽出処理、影の影響除去》
 次いで、CPU9によって仮想カーソル矩形抽出用のしきい値(例えば、最大値-1σ)を使用して得られた右眼側の仮想カーソル活動領域画像27、仮想ボタンクリック矩形抽出用のしきい値(例えば、最大値-2σ)を使用して得られた右眼側の仮想ボタンクリック活動領域画像に対し、各々活動矩形領域26が左右に分割可能かどうかチェックされ、左右に分割可能であれば、図13に示すように活動矩形領域26の水平方向中心点“A”が求められると共に、水平方向中心点“A”から左側にある非活動領域と、活動領域(例えば、活動分割エリア21)との境界点“B”と、水平方向中心点“A”から右側にある非活動領域と、活動領域(例えば、活動分割エリア21)との境界点“C”とが検知され、これら境界点“B”、“C”を含む領域が各活動矩形領域26と判定され、それ以外の活動領域がユーザの影などによる不要な活動領域と判定され、無効とされる(2点抽出処理)。
《Multi-stage rectangular object extraction processing, shadow effect removal》
Next, the virtual cursor click area extraction threshold value 27 for the right eye side virtual cursor active region image 27 obtained by the CPU 9 using the threshold value for virtual cursor rectangle extraction (for example, the maximum value −1σ) is obtained. For example, with respect to the virtual button click activity region image on the right eye side obtained using the maximum value −2σ), it is checked whether or not each of the activity rectangular regions 26 can be divided into left and right. As shown in FIG. 13, the horizontal center point “A” of the active rectangular area 26 is obtained, and an inactive area on the left side of the horizontal center point “A”, an active area (for example, an active division area 21), Boundary point “B”, a non-active area on the right side of the horizontal center point “A”, and a boundary point “C” between the active area (for example, the active division area 21) are detected. B ”,“ C ” Are determined as each active rectangular area 26, and other active areas are determined as unnecessary active areas due to the shadows of the user and are invalidated (two-point extraction processing).
 この後、CPU9によって、2点抽出処理が終了した各活動矩形領域26に対し、各々活動矩形領域26が上下に分割可能かどうかチェックされ、上下分割可能であれば図14に示すように活動矩形領域26の上下方向中心点“A”が求められると共に、上下方向中心点“A”から上側にある非活動領域と、活動領域(例えば、活動分割エリア21)との境界点“B”が検知され、これら境界点“B”を含む領域が活動矩形領域26と判定され、下の活動領域がユーザの影などによる不要な活動領域と判定され、無効とされる(最小化処理)(ステップS44)。 Thereafter, the CPU 9 checks whether or not each of the activity rectangular areas 26 for which the two-point extraction processing has been completed can be divided vertically, and if it can be divided vertically, as shown in FIG. The vertical center point “A” of the area 26 is obtained, and the boundary point “B” between the inactive area above the vertical center point “A” and the active area (for example, the active division area 21) is detected. Then, the area including these boundary points “B” is determined as the active rectangular area 26, and the lower active area is determined as an unnecessary active area due to the shadow of the user, and is invalidated (minimization process) (step S44). ).
 次いで、CPU9によって、これら2点抽出処理、最小化処理によって構成される多段階矩形オブジェクト抽出処理で得られた活動矩形領域26を含む右眼側の仮想カーソル活動領域画像27、右眼側の仮想ボタンクリック活動領域画像が、ハードディスク8の画像格納エリア16に格納される(ステップS45)。 Subsequently, the CPU 9 performs a virtual cursor activity region image 27 on the right eye side including the activity rectangular region 26 obtained by the multi-stage rectangular object extraction process constituted by the two-point extraction process and the minimization process, and the right-eye virtual image. The button click activity area image is stored in the image storage area 16 of the hard disk 8 (step S45).
《統計処理、仮想カーソルの活動領域決定、変化領域抽出》
 この後、図2のフローチャートに示すようにCPU9によって、ハードディスク8の画像格納エリア16に格納されている左眼側のヒストグラムが読み出されて(ステップS15)、活動矩形領域抽出処理が開始される(ステップS16)。即ち、図5のフローチャートに示すように、ヒストグラムの各分割エリアの濃度値に対する統計処理が行われ、平均値、濃度分散値、最大値、偏差(±1σ、±2σ)などが演算される(ステップS41)。
《Statistical processing, virtual cursor activity area determination, change area extraction》
Thereafter, as shown in the flowchart of FIG. 2, the CPU 9 reads the left eye side histogram stored in the image storage area 16 of the hard disk 8 (step S15), and starts the active rectangular area extraction process. (Step S16). That is, as shown in the flowchart of FIG. 5, statistical processing is performed on the density value of each divided area of the histogram, and an average value, density variance value, maximum value, deviation (± 1σ, ± 2σ), and the like are calculated ( Step S41).
 次いで、CPU9によって変化領域矩形抽出用のしきい値(例えば、平均値-1σ)より大きい濃度値になっている各分割エリア20が抽出されて、これらの各分割エリア(活動分割エリア)を含むように矩形状の変化領域矩形65(図31参照)が決定され、ハードディスク8の画像格納エリア16に格納される。 Next, the CPU 9 extracts each divided area 20 having a density value larger than the threshold for extracting the change area rectangle (for example, the average value −1σ), and includes these divided areas (activity divided areas). Thus, a rectangular change area rectangle 65 (see FIG. 31) is determined and stored in the image storage area 16 of the hard disk 8.
 また、この動作と並行し、CPU9によって、例えば図8の3次元濃度分布図に示すようにヒストグラムを構成している各分割エリア20(図9、図10参照)のうち、仮想カーソル矩形抽出用のしきい値(例えば、最大値-1σ)より大きい濃度値になっている分割エリア(活動分割エリア21)が抽出される。 In parallel with this operation, the CPU 9 extracts the virtual cursor rectangle from the divided areas 20 (see FIGS. 9 and 10) constituting the histogram as shown in the three-dimensional density distribution diagram of FIG. A divided area (activity divided area 21) having a density value larger than the threshold value (for example, maximum value −1σ) is extracted.
 これにより、ユーザが指先を大きく回しているとき、図9に示す通り各活動分割エリア21を含むように矩形状の活動矩形領域26が決定されると共に、この決定結果に基づき図11に示すような左眼側の仮想カーソル活動領域画面27が作成されて、ハードディスク8の画像格納エリア16に格納される。 As a result, when the user turns the fingertip greatly, a rectangular activity rectangular area 26 is determined so as to include each activity division area 21 as shown in FIG. 9, and based on the determination result, as shown in FIG. A virtual cursor activity area screen 27 on the left eye side is created and stored in the image storage area 16 of the hard disk 8.
 また、ユーザが両手を動かしているときには、図10に示す通り各活動分割エリア21を含むように矩形状の活動矩形領域26が決定されて、図12に示すような左眼側の仮想カーソル活動領域画像27が作成され、ハードディスク8の画像格納エリア16に格納される(ステップS42)。 Further, when the user moves both hands, a rectangular activity rectangular area 26 is determined so as to include each activity division area 21 as shown in FIG. 10, and the virtual cursor activity on the left eye side as shown in FIG. An area image 27 is created and stored in the image storage area 16 of the hard disk 8 (step S42).
《仮想ボタンクリックの活動領域決定》
 次いで、CPU9によってハードディスク8の画像格納エリア16に格納されている左眼側のヒストグラムが読み出されて、各分割エリア20のうち仮想ボタンクリック矩形抽出用のしきい値(例えば、最大値-2σ)より大きい濃度値になっている分割エリア(活動分割エリア)が抽出されると共に、これらの各活動分割エリアを含むように矩形状の活動矩形領域が決定されて、左眼側の仮想ボタンクリック活動領域画像(図示は省略する)が作成され、ハードディスク8の画像格納エリア16に格納される(ステップS43)。
《Determine activity area for virtual button click》
Next, the left-eye histogram stored in the image storage area 16 of the hard disk 8 is read by the CPU 9, and a threshold value for extracting a virtual button click rectangle in each divided area 20 (for example, the maximum value −2σ) ) A divided area (activity divided area) having a larger density value is extracted, and a rectangular activity rectangular area is determined so as to include each of these activity divided areas, and a virtual button click on the left eye side is clicked An active area image (not shown) is created and stored in the image storage area 16 of the hard disk 8 (step S43).
《多段階矩形オブジェクト抽出処理、影の影響除去》
 次いで、CPU9によって仮想カーソル矩形抽出用のしきい値(例えば、最大値-1σ)を使用して得られた左眼側の仮想カーソル活動領域画像27、仮想ボタンクリック矩形抽出用のしきい値(例えば、最大値-2σ)を使用して得られた左眼側の仮想ボタンクリック活動領域画像に対し、各々活動矩形領域26が左右に分割可能かどうかチェックされ、左右に分割可能であれば、図13に示すように活動矩形領域26の水平方向中心点“A”が求められると共に、水平方向中心点“A”から左側にある非活動領域と、活動領域(例えば、活動分割エリア21)との境界点“B”と、水平方向中心点“A”から右側にある非活動領域と、活動領域(例えば、活動分割エリア21)との境界点“C”とが検知され、これら境界点“B”、“C”を含む領域が各活動矩形領域26と判定され、それ以外の活動領域がユーザの影などによる不要な活動領域と判定され、無効とされる(2点抽出処理)。
《Multi-stage rectangular object extraction processing, shadow effect removal》
Next, the virtual cursor active area image 27 on the left eye side obtained by the CPU 9 using the threshold for virtual cursor rectangle extraction (for example, the maximum value −1σ), the threshold for virtual button click rectangle extraction ( For example, with respect to the virtual button click activity region image on the left eye side obtained using the maximum value −2σ), it is checked whether each of the activity rectangular regions 26 can be divided into left and right. As shown in FIG. 13, the horizontal center point “A” of the active rectangular area 26 is obtained, and an inactive area on the left side of the horizontal center point “A”, an active area (for example, an active division area 21), Boundary point “B”, a non-active area on the right side of the horizontal center point “A”, and a boundary point “C” between the active area (for example, the active division area 21) are detected. B ”,“ C ” Are determined as each active rectangular area 26, and other active areas are determined as unnecessary active areas due to the shadows of the user and are invalidated (two-point extraction processing).
 この後、CPU9によって、2点抽出処理が終了した各活動矩形領域26に対し各々活動矩形領域26が上下に分割可能かどうかチェックされ、上下分割可能であれば図14に示すように活動矩形領域26の上下方向中心点“A”が求められると共に、上下方向中心点“A”から上側にある非活動領域と活動領域(例えば、活動分割エリア21)との境界点“B”が検知され、これら境界点“B”を含む領域が活動矩形領域26、下の活動領域がユーザの影などによる不要な活動領域と判定され、無効とされる(最小化処理)(ステップS44)。 Thereafter, the CPU 9 checks whether or not each of the activity rectangular areas 26 for which the two-point extraction process has been completed can be divided vertically, and if it can be divided vertically, as shown in FIG. 26, the center point “A” in the vertical direction is obtained, and the boundary point “B” between the inactive region and the active region (for example, the active division area 21) above the vertical center point “A” is detected. The area including the boundary point “B” is determined as the active rectangular area 26, and the lower active area is determined as an unnecessary active area due to the shadow of the user or the like, and is invalidated (minimization process) (step S44).
 次いで、CPU9によってこれら2点抽出処理、最小化処理によって構成される多段階矩形オブジェクト抽出処理で得られた活動矩形領域26を含む左眼側の仮想カーソル活動領域画像27、左眼側の仮想ボタンクリック活動領域画像が、ハードディスク8の画像格納エリア16に格納される(ステップS45)。 Next, the left eye-side virtual cursor activity region image 27 including the activity rectangle region 26 obtained by the multi-step rectangular object extraction process constituted by the two-point extraction process and the minimization process by the CPU 9, and the left-eye virtual button The click activity area image is stored in the image storage area 16 of the hard disk 8 (step S45).
 上記した如くユーザの手による動きを右眼と左眼の各仮想領域活動領域画像27及び仮想ボタンクリック活動領域画像で捉えることができるが、このとき入力装置101がユーザの意図しない手の動きを検知できるようにするとさらに良い。ヒストグラム化した活動領域を多段階で抽出した結果の活動矩形領域26がユーザの意図したポインティング、またはタッピングでない動作、例えば手を左右に振っている場合もヒストグラム化されたデータ上に活動領域が抽出される。よって、これがユーザの意図したポインティング動作によるものではないことを判断できれば、入力装置101としての信頼性が高まる。 As described above, the movement of the user's hand can be captured by the virtual area activity area image 27 and the virtual button click activity area image of the right eye and the left eye. At this time, the input device 101 detects the movement of the hand not intended by the user. It would be even better if it could be detected. The activity rectangular area 26 as a result of extracting the activity area in the form of a histogram in multiple stages is extracted on the histogram data even when the user's intention is not pointing or tapping, for example, when the hand is waving left and right. Is done. Therefore, if it can be determined that this is not due to the pointing operation intended by the user, the reliability of the input device 101 is increased.
 抽出された活動矩形領域が有効かどうかを判断するためには、以下の処理を行う。すなわち、図5のステップS41の処理においては、図34に示すステップS41’の処理を行い、CPU9はヒストグラム化された多段階の差分画像データのうち、最新の差分データから活動矩形領域26を抽出し、それらの点を追跡しながら一定時間保持する。 ∙ To determine whether the extracted activity rectangle area is valid, the following processing is performed. That is, in the process of step S41 of FIG. 5, the process of step S41 ′ shown in FIG. 34 is performed, and the CPU 9 extracts the active rectangular area 26 from the latest difference data among the multi-stage difference image data that is formed into a histogram. And hold for a certain time while tracking those points.
 次に、図5のステップS43の処理においては、図35に示すステップS43’の処理を行い、CPU9は仮想ボタンクリックの活動領域(最大値-2σ)が抽出されたら、ステップS41’で記憶している活動矩形領域26と比較し、該当する追跡を求めてその追跡結果が特定の大きさの領域に亘って活動している場合には、無効な活動矩形領域26と判断する。 Next, in the process of step S43 in FIG. 5, the process of step S43 ′ shown in FIG. 35 is performed, and the CPU 9 stores the virtual button click activity area (maximum value −2σ) in step S41 ′. When the corresponding tracking is obtained and the tracking result is active over an area of a specific size, it is determined that the activity rectangular area 26 is invalid.
 図37はユーザが手を右下から左上、さらに左下、次に右上に移動した場合に抽出される活動領域を示しているが、活動矩形領域26の追跡データから生成された活動領域は大きな動きの中で生成されたもの判断してノイズとして無視する。 FIG. 37 shows an active area extracted when the user moves his / her hand from the lower right to the upper left, further to the lower left, and then to the upper right. However, the active area generated from the tracking data of the activity rectangular area 26 has a large movement. Ignore it as noise by judging what was generated.
 CPU9による上記処理にて抽出活動領域がユーザの意図しない動作により発生したものはノイズとして除去することが可能となり、操作性が向上する。 In the above processing by the CPU 9, it is possible to remove the extraction activity area caused by the operation not intended by the user as noise, and the operability is improved.
《活動矩形領域選択》
 この後、図2のフローチャートに示すように、CPU9によって活動矩形領域選択処理が開始される(ステップS17)。即ち、図6のフローチャートに示すようにハードディスク8の画像格納エリア16に格納されている右眼側の仮想カーソル活動領域画像27、右眼側の仮想ボタンクリック活動領域画像が読み出され、これら右眼側の仮想カーソル活動領域画像27、右眼側の仮想ボタンクリック活動領域画像に含まれている各活動矩形領域26に対し、図15の模式図に示すように両眼視差法による位置補正が行われ、各ウェブカメラ4、6の取り付け位置(水平距離“B”、上下距離など)、各ウェブカメラ4、6の焦点距離“f”などと対応するように各活動矩形領域26の中心座標“P(X、Y)”が補正された後、大きさ順に番号が付加される(ステップS51、S52)。
《Activity rectangular area selection》
Thereafter, as shown in the flowchart of FIG. 2, the active rectangle area selection process is started by the CPU 9 (step S17). That is, as shown in the flowchart of FIG. 6, the right eye side virtual cursor activity region image 27 and the right eye side virtual button click activity region image stored in the image storage area 16 of the hard disk 8 are read out, and these right Position correction by the binocular parallax method is performed on each active rectangular area 26 included in the virtual cursor activity area image 27 on the eye side and the virtual button click activity area image on the right eye side as shown in the schematic diagram of FIG. Center coordinates of each active rectangular area 26 so as to correspond to the attachment position (horizontal distance “B”, vertical distance, etc.) of each webcam 4, 6, the focal length “f” of each webcam 4, 6, etc. After “P R (X R , Y R )” is corrected, numbers are added in order of magnitude (steps S51 and S52).
 次いで、CPU9によってハードディスク8の画像格納エリア16に格納されている左眼側の仮想カーソル活動領域画像27、左眼側の仮想ボタンクリック活動領域画像が読み出され、これら左眼側の仮想カーソル活動領域画像27、左眼側の仮想ボタンクリック活動領域画像の各活動矩形領域26に対し、図15の模式図に示すように両眼視差法による位置補正が行われ、各ウェブカメラ4、6の取り付け位置(水平距離“B”、上下距離など)、各ウェブカメラ4、6の焦点距離“f”などと対応するように各活動矩形領域26の座標“P(X、Y)”が補正された後、大きさ順に番号が付加される(ステップS53、S54)。 Next, the CPU 9 reads the left eye side virtual cursor activity region image 27 and the left eye side virtual button click activity region image stored in the image storage area 16 of the hard disk 8, and these left eye side virtual cursor activity regions are read out. Position correction by the binocular parallax method is performed on the area image 27 and each activity rectangular area 26 of the left eye side virtual button click activity area image as shown in the schematic diagram of FIG. The coordinates “P L (X L , Y L )” of each active rectangular area 26 so as to correspond to the mounting position (horizontal distance “B”, vertical distance, etc.), the focal length “f” of each web camera 4, 6, etc. Are corrected, numbers are added in order of size (steps S53 and S54).
 これにより、各ウェブカメラ4、6の焦点距離などと対応したピント位置、例えば各ウェブカメラ4、6から“0.3m”~“0.8m”離れた位置にユーザの手があり、両眼視差法による位置補正を行う前、図16の模式図に示すように右眼側の各活動矩形領域26と左眼側の各活動矩形領域26とが位置がずれていても、両眼視差法による位置補正を行うことにより、図17の模式図に示すように右眼側の各活動矩形領域26と左眼側の各活動矩形領域26とを完全一致(又は、ほぼ一致)させることができる。 As a result, the user's hand is at a focus position corresponding to the focal length of each of the webcams 4 and 6, for example, “0.3 m” to “0.8 m” away from each of the webcams 4 and 6. Before performing the position correction by the parallax method, as shown in the schematic diagram of FIG. 16, even if the positions of the activity rectangle areas 26 on the right eye side and the activity rectangle areas 26 on the left eye side are shifted, the binocular parallax method is used. As shown in the schematic diagram of FIG. 17, the right-side active rectangular area 26 and the left-eye active rectangular area 26 can be completely matched (or substantially matched). .
 この後、CPU9によって図15の模式図に示すように、番号“1”が付与された右眼側に対応する活動矩形領域26の中心座標“X、Y”と、左眼側に対応する活動矩形領域26の中心座標“X、Y”との距離(中心座標距離)が演算されて、これが番号“1”と共に、メモリ10に記憶される。 Thereafter, as shown in the schematic diagram of FIG. 15 by the CPU 9, the center coordinates “X R , Y R ” of the activity rectangular area 26 corresponding to the right eye side to which the number “1” is assigned correspond to the left eye side. The distance (center coordinate distance) with the center coordinates “X L , Y L ” of the active rectangular area 26 is calculated and stored in the memory 10 together with the number “1”.
 以下、CPU9によって次の番号“2”~最後の番号“N”が付与された右眼側に対応する活動矩形領域26の中心座標“X、Y”と、左眼側に対応する活動矩形領域26の中心座標“X、Y”との距離(中心座標距離)が順次演算されて、これらが次の番号“2”~最後の番号“N”と共にメモリ10に記憶される(ステップS55、S56)。 Hereinafter, the central coordinates “X R , Y R ” of the activity rectangular area 26 corresponding to the right eye side to which the next number “2” to the last number “N” are assigned by the CPU 9 and the activity corresponding to the left eye side will be described. The distance (center coordinate distance) with respect to the center coordinates “X L , Y L ” of the rectangular area 26 is sequentially calculated and stored in the memory 10 together with the next number “2” to the last number “N” ( Steps S55 and S56).
 そして、これらの処理が終了したとき、CPU9によって、メモリ10に記憶されている各中心座標距離が順次読み出されて所定値と比較され、所定値以下になっている各中心座標距離を持つ右眼側の活動矩形領域26、左眼側の活動矩形領域26があるとき、即ち各ウェブカメラ4、6から“0.3m”~“0.8m”離れた位置にユーザの手があるとき、ユーザの手に対応する右眼側の活動矩形領域26、左眼側の活動矩形領域26が、有効な右眼側の活動矩形領域26、有効な左眼側の活動矩形領域26であると判定され、それ以外の右眼側の活動矩形領域26、左眼側の活動矩形領域26が無効な右眼側の活動矩形領域26、無効な左眼側の活動矩形領域26であると判定される。 When these processes are completed, the CPU 9 sequentially reads out the respective central coordinate distances stored in the memory 10 and compares them with a predetermined value. When there is an activity rectangle area 26 on the eye side and an activity rectangle area 26 on the left eye side, that is, when the user's hand is at a position away from each web camera 4, 6 by “0.3 m” to “0.8 m”, It is determined that the right-eye activity rectangle area 26 and the left-eye activity rectangle area 26 corresponding to the user's hand are the valid right-eye activity rectangle area 26 and the valid left-eye activity rectangle area 26. It is determined that the other activity rectangle area 26 on the right eye side and the activity rectangle area 26 on the left eye side are the invalid activity rectangle area 26 on the right eye side and the activity rectangle area 26 on the invalid left eye side. .
 このように両眼視差法を用いて、活動矩形領域の位置補正を行ないながら右眼及び左眼の各活動矩形領域26の中心点間の横方向での座標距離(中心座標距離)から被写体までの距離を測定することにより活動領域26の有効性判定を精度良く行うことができるのである。しかしながら、ウェブカメラ4,6の視野で被写体の中心部分と端の部分は誤差が発生する。よって、本発明は、これを補正してより正確な距離を計測する方法を更に提供するものである。 Thus, using the binocular parallax method, while correcting the position of the active rectangular area, the coordinate distance in the horizontal direction (center coordinate distance) between the center points of the active rectangular areas 26 of the right eye and the left eye to the subject. Therefore, the effectiveness of the active area 26 can be determined with high accuracy. However, an error occurs in the center portion and the end portion of the subject in the field of view of the web cameras 4 and 6. Therefore, the present invention further provides a method of correcting this and measuring a more accurate distance.
 距離の計測は、予め計測したカメラ視野角を定数として活動領域がウェブカメラ4,6の視野のそれぞれの活動矩形領域26における中心座標距離から補正値を計算するものであるが、以下、この原理について説明する。 The distance is measured by calculating a correction value from the center coordinate distance in the active rectangular area 26 of each of the visual fields of the webcams 4 and 6 with the camera viewing angle measured in advance as a constant. Will be described.
 図38は1台のウェブカメラ4Aの視野角から距離の変換について説明するもので、三角関数の定義から、数式1の関係が成立する。
Figure JPOXMLDOC01-appb-M000001
 w:画像上での被写体の幅[pixel]
 img_w:カメラ画像の横画素数[pixel]
 d:カメラと被写体までの距離[m]
 δ:被写体の幅/2[m]
 θw:カメラの視野角/2[rad](予め計測)
但し、wと(img_w)は、それぞれ画像中の大きさのため物理量としては「角度」となる。
 これをdについて解くと数式2となる。
Figure JPOXMLDOC01-appb-M000002
数式2において、θwとimg_wは定数であるので、求めたい値である距離dはδとwに依存する。ここでδは映る被写体60によって変化するため、δに依存せずに距離dを求める必要がある。
FIG. 38 explains the conversion of distance from the viewing angle of one web camera 4A. From the definition of the trigonometric function, the relationship of Equation 1 is established.
Figure JPOXMLDOC01-appb-M000001
w: Width of the subject on the image [pixel]
img_w: Number of horizontal pixels of the camera image [pixel]
d: Distance between camera and subject [m]
δ: Subject width / 2 [m]
θw: Camera viewing angle / 2 [rad] (measured in advance)
However, w and (img_w) are “angles” as physical quantities because of their sizes in the image.
When this is solved for d, Equation 2 is obtained.
Figure JPOXMLDOC01-appb-M000002
In Equation 2, since θw and img_w are constants, the distance d, which is a value to be obtained, depends on δ and w. Here, since δ varies depending on the subject 60 to be reflected, it is necessary to obtain the distance d without depending on δ.
 そこで2台のウェブカメラ4、6を用いる本発明では、δに依存せずに距離dを求めることが可能である。図39はメインとサブの2台のウェブカメラの視野角から距離の変換について説明するもので、距離dは数式3にて示される。
Figure JPOXMLDOC01-appb-M000003
 w’:サブカメラの画像上での被写体の中心の横座標[pixel]と、メインカメラの画像上での被写体の中心横座標[pixel]の差の絶対値(|X-X|)
 δ’:被写体の位置で、サブカメラの中心視線とメインカメラの中心視線が交差する2点間の距離[m](両眼の視線が平行ならば、δ’は常に一定)
 数式3において、両眼の中心視線が平行である場合は一定であるから距離dはδ’に依存しなくなる。尚、dはδ’より十分大きく(d>>δ’)、被写体60が画像の端に存在していても近似的に成り立つ。
Therefore, in the present invention using two web cameras 4 and 6, the distance d can be obtained without depending on δ. FIG. 39 explains the conversion of the distance from the viewing angle of the two main and sub webcams. The distance d is expressed by Equation 3.
Figure JPOXMLDOC01-appb-M000003
w ′: absolute value of the difference between the abscissa [pixel] of the center of the subject on the sub-camera image and the center abscissa [pixel] of the subject on the image of the main camera (| X L −X R |)
δ ′: Distance [m] between two points at which the center line of sight of the sub camera and the center line of sight of the main camera intersect at the position of the subject (when both eyes are parallel, δ ′ is always constant)
In Expression 3, the distance d does not depend on δ ′ because it is constant when the central line of sight of both eyes is parallel. Note that d is sufficiently larger than δ ′ (d >> δ ′), and even if the subject 60 exists at the edge of the image, it is approximately established.
 上記から2台のカメラを用いれば数式4にて表すように、距離dはw’のみに依存することになる。よって、dは画素値から計算され、離散値であるため各w’のd(w’)を予め計算しておきテーブルに保存することで被写体60までの距離(d)の補正を高速にて処理することができる。
Figure JPOXMLDOC01-appb-M000004
From the above, if two cameras are used, the distance d depends only on w ′, as shown in Equation 4. Therefore, since d is calculated from the pixel value and is a discrete value, d (w ′) of each w ′ is calculated in advance and stored in a table, thereby correcting the distance (d) to the subject 60 at high speed. Can be processed.
Figure JPOXMLDOC01-appb-M000004
 以上の原理をCPU9による活動領域26の有効性の判断に取り込めば判定の精度を更に高めることができる。この場合、CPU9はステップS57の処理において、図36に示すステップ57’の処理を行い、中心座標距離w’(|X-X|)を演算してメモリ10に格納しているテーブルデータと照合することで被写体60までの正確な距離を計測し、この距離が“0.3m”~“0.8m”の範囲内であるかを判定することで、活動矩形領域26の有効・無効の判断を行うものである。 If the above principle is incorporated in the determination of the effectiveness of the activity area 26 by the CPU 9, the accuracy of the determination can be further improved. In this case, the CPU 9 performs the process of step 57 ′ shown in FIG. 36 in the process of step S57, calculates the center coordinate distance w ′ (| X L −X R |), and stores the table data in the memory 10. The effective distance to the subject 60 is measured by comparing the distance to the subject 60 and it is determined whether the distance is within the range of “0.3 m” to “0.8 m”. This is a judgment.
 これにより、右眼側の各活動矩形領域26、左眼側の各活動矩形領域26が図18の模式図に示す関係になっていれば、中心座標距離が所定値以上になっている番号“1”に対応した右眼側の活動矩形領域(OR1)26、左眼側の活動矩形領域(OL1)26が無効と判定され、中心座標距離がほぼ“0”になっている番号“2”に対応した右眼側の活動矩形領域(OR2)26、左眼側の活動矩形領域(OL2)26が有効と判定される。 As a result, if each activity rectangular area 26 on the right eye side and each activity rectangle area 26 on the left eye side have the relationship shown in the schematic diagram of FIG. 18, the number “the center coordinate distance is equal to or greater than a predetermined value”. The number “0” is determined to be invalid for the right-eye side active rectangle area (O R1 ) 26 and the left-eye side active rectangle area (O L1 ) 26 corresponding to “1”. It is determined that the right-eye active rectangular area (O R2 ) 26 and the left-eye active rectangular area (O L2 ) 26 corresponding to 2 ″ are valid.
 次いで、CPU9によって、有効と判定された右眼側の活動矩形領域(OR2)26、有効と判定された左眼側の活動矩形領域(OL2)26のうち、予め指定されている方、例えば有効と判定された左眼側の活動矩形領域(OL2)26が残され、それ以外の活動矩形領域26が削除された仮想カーソル活動領域画像27、仮想ボタンクリック活動領域画像が作成され、これが両眼視差法によってユーザの手前にいる人の動き、及び背後にいる人の動きが除去された仮想カーソル活動領域画像27、仮想ボタンクリック活動領域画像としてハードディスク8の画像格納エリア16に格納される(ステップS57)。 Next, the right-side active rectangular area (O R2 ) 26 determined to be valid by the CPU 9 and the left-side active rectangular area (O L2 ) 26 determined to be valid are previously designated, For example, a virtual cursor activity region image 27 and a virtual button click activity region image in which the activity rectangle region ( OL2 ) on the left eye side determined to be valid is left and the other activity rectangle regions 26 are deleted are created, This is stored in the image storage area 16 of the hard disk 8 as a virtual cursor activity area image 27 and a virtual button click activity area image from which the movement of the person in front of the user and the movement of the person behind are removed by the binocular parallax method. (Step S57).
《片手ジェスチャによる仮想カーソルの位置、大きさ、色制御》
 この後、図2のフローチャートに示すように、CPU9によって仮想カーソル制御処理/画面制御処理が開始される(ステップS18)。即ち、図7のフローチャートに示すように、ハードディスク8の画像格納エリア16に格納されている各仮想カーソル活動領域画像(両眼視差法によって、ユーザの手前にいる人の動き、及び背後にいる人の動きが除去された各仮想カーソル活動領域画像)27のうち、最新の活動矩形領域26を含む数フレーム分の仮想カーソル活動領域画像27が読み出され(ステップS61)、近接した1つ以上の活動矩形領域26によって構成される活動矩形領域群が仮想カーソル活動領域画像27内に存在しているかどうかチェックされる。
《Position, size, and color control of virtual cursor by one hand gesture》
Thereafter, as shown in the flowchart of FIG. 2, the virtual cursor control process / screen control process is started by the CPU 9 (step S18). That is, as shown in the flowchart of FIG. 7, each virtual cursor activity region image stored in the image storage area 16 of the hard disk 8 (the movement of the person in front of the user and the person behind the user by the binocular parallax method). Among the virtual cursor activity region images 27), the virtual cursor activity region images 27 for several frames including the latest activity rectangle region 26 are read (step S61), and one or more adjacent ones are displayed. It is checked whether an active rectangle area group constituted by the active rectangle areas 26 exists in the virtual cursor active area image 27.
 そして、最新の仮想カーソル活動領域画像27内に活動矩形領域群が存在し、その数が“1”であり、かつほぼ矩形であれば(ステップS62、S63)、CPU9によって活動矩形領域群の大きさ、移動方向が判定され、判定結果に対応するように仮想カーソル制御が行われる(ステップS64)。 If there are active rectangular area groups in the latest virtual cursor active area image 27 and the number thereof is “1” and almost rectangular (steps S62 and S63), the CPU 9 determines the size of the active rectangular area group. Now, the moving direction is determined, and virtual cursor control is performed so as to correspond to the determination result (step S64).
 例えば、図19(a)に示すように、前回とほぼ同じ、高さ、左右位置で、ユーザが指先を大きく回し、これに対応して前回の処理で得られた大きな活動矩形領域群の位置と同じ位置で、大きな活動矩形領域群が得られているとき、CPU9によって仮想カーソルの表示指示であると判定されて、図19(b)に示すようにディスプレイ部3に大きなサイズ、白色の仮想カーソル25が表示される。 For example, as shown in FIG. 19 (a), the position of the large active rectangular area group obtained by the previous processing corresponding to the user turning the fingertip at the same height, left and right positions as in the previous time. When a large group of active rectangular areas is obtained at the same position, the CPU 9 determines that it is a virtual cursor display instruction, and the display unit 3 has a large size and white virtual display as shown in FIG. A cursor 25 is displayed.
 また、ユーザが指先を大きく回しながら上下方向、又は左右方向に移動させ、これに対応して前回の処理で得られた位置から移動する大きな活動矩形領域群が得られているとき、CPU9によって仮想カーソルの移動指示であると判定されて、指先の移動方向に対応するようにディスプレイ部3に表示されている大きなサイズ、白色の仮想カーソル25を移動させる。 Further, when the user moves the fingertip up and down or left and right, and a large group of active rectangular areas moving from the position obtained in the previous process is obtained correspondingly, the CPU 9 performs virtual processing. The cursor movement instruction is determined, and the large size, white virtual cursor 25 displayed on the display unit 3 is moved so as to correspond to the moving direction of the fingertip.
 また、図20(a)に示すように、前回とほぼ同じ、高さ、左右位置で、ユーザが指先を小さく回し、これに対応して前回の処理で得られた大きな活動矩形領域群の位置と同じ位置で、小さな活動矩形領域群が得られているとき、CPU9によって仮想カーソルの移動停止であると判定されて、図20(b)に示すようにディスプレイ部3に表示されている仮想カーソル25の移動を停止させると共に、サイズを小さくする。 Also, as shown in FIG. 20 (a), the position of the large active rectangular area group obtained by the previous processing corresponding to the user turning the fingertip small at the same height and left and right positions as in the previous time. When the small active rectangle area group is obtained at the same position as the virtual cursor, the CPU 9 determines that the movement of the virtual cursor is stopped, and the virtual cursor displayed on the display unit 3 as shown in FIG. The movement of 25 is stopped and the size is reduced.
 この状態で一定時間が経過すれば、CPU9によって仮想カーソル25の色が赤色に変更され大きな移動が禁止されると共に、OS側にカーソル移動指示が出されて、仮想カーソル25内に実カーソル28を移動させる。 If a certain time has passed in this state, the CPU 9 changes the color of the virtual cursor 25 to red and prohibits a large movement, and a cursor movement instruction is issued to the OS side, and the real cursor 28 is placed in the virtual cursor 25. Move.
 この後、ユーザが指先を少し移動させれば、CPU9によってこれが検知されて、ディスプレイ部3に表示されている仮想カーソル25の位置が微調整されると共に、OS側にカーソル位置調整指示が出されて、図21に示すように実カーソル28の位置が微調整される。 Thereafter, when the user moves the fingertip slightly, this is detected by the CPU 9 and the position of the virtual cursor 25 displayed on the display unit 3 is finely adjusted, and a cursor position adjustment instruction is issued to the OS side. Thus, the position of the actual cursor 28 is finely adjusted as shown in FIG.
 次いで、ユーザが指先を動かすのを止めれば、CPU9によってこれが検知され、一定時間後に、図22に示すようにディスプレイ部3に表示されている仮想カーソル25の位置が固定されると共に、仮想カーソル25の色が赤色からグレー色に変更され、ユーザにクリック可能になったことを知らせる。 Next, if the user stops moving the fingertip, the CPU 9 detects this, and after a certain time, the position of the virtual cursor 25 displayed on the display unit 3 is fixed and the virtual cursor 25 is displayed as shown in FIG. The color of is changed from red to gray to inform the user that it can be clicked.
 この状態でも、ユーザが指先を再度大きく回せば、CPU9によってこれが検知されて、ディスプレイ部3に表示されている仮想カーソル25の色が白色に戻され、仮想カーソル25が移動可能な状態に戻される。 Even in this state, if the user turns the fingertip again largely, this is detected by the CPU 9, the color of the virtual cursor 25 displayed on the display unit 3 is returned to white, and the virtual cursor 25 is returned to a movable state. .
《片手ジェスチャによるスクロール制御》
 また、上述した仮想カーソル活動領域画像27に活動矩形領域群が存在するかどうかなどをチェックしたとき、図23に示すように活動矩形領域群の数が“1”で水平方向に長ければ(ステップS62、S63)、CPU9によって前回の活動矩形領域群に対しどちらの方向に長くなったか判定すると共に、長くなった方向に応じた右スクロール指示(又は、左スクロール指示)が生成されてアプリケーション側に渡され、ディスプレイ部3に表示されているアプリケーション画面(操作対象画面)が右方向(又は、左方向)にスクロールされる(ステップS64)。
《Scroll control with one hand gesture》
When it is checked whether or not there is an active rectangular area group in the virtual cursor active area image 27 described above, as shown in FIG. 23, if the number of active rectangular area groups is “1” and is long in the horizontal direction (step In S62, S63), the CPU 9 determines which direction is longer than the previous active rectangular area group, and generates a right scroll instruction (or left scroll instruction) corresponding to the longer direction to the application side. The application screen (operation target screen) displayed on the display unit 3 is scrolled rightward (or leftward) (step S64).
 また、上述した仮想カーソル活動領域画像27に活動矩形領域群が存在するかどうかなどをチェックしたとき、活動矩形領域群の数が“1”で上下方向に長ければ(ステップS62、S63)、CPU9によって前回の活動矩形領域群に対しどちらの方向に長くなったか判定すると共に、長くなった方向に応じた上スクロール指示(又は、下スクロール指示)が生成されてアプリケーション側に渡され、ディスプレイ部3に表示されているアプリケーション画面(操作対象画面)が上方向(又は、下方向)にスクロールされる(ステップS64)。 If it is checked whether or not there is an active rectangular area group in the virtual cursor active area image 27 described above, if the number of active rectangular area groups is “1” and is long in the vertical direction (steps S62 and S63), the CPU 9 Is used to determine which direction is longer than the previous active rectangular area group, and an upward scroll instruction (or downward scroll instruction) corresponding to the longer direction is generated and passed to the application side, and the display unit 3 The application screen (operation target screen) displayed on the screen is scrolled upward (or downward) (step S64).
《片手ジェスチャによる実カーソルのクリック制御》
 この後、CPU9によって仮想カーソル25の色がグレーかどうかチェックされ、仮想カーソル25の色がグレーであれば、ハードディスク8の画像格納エリア16に格納されている仮想ボタンクリック活動領域画像のうち、最新の活動矩形領域を含む数フレーム分の仮想ボタンクリック活動領域画像が読み出される(ステップS66)。
《Real cursor click control by one hand gesture》
Thereafter, the CPU 9 checks whether the color of the virtual cursor 25 is gray. If the color of the virtual cursor 25 is gray, the latest virtual button click activity area image stored in the image storage area 16 of the hard disk 8 is checked. The virtual button click activity region images for several frames including the activity rectangle region are read out (step S66).
 次いで、CPU9によって仮想ボタンクリック活動領域画像内に近接した1つ以上の活動矩形領域26によって構成される活動矩形領域群が存在し、形状が変化しているかどうかチェックされ、活動矩形領域群の数が“1”であり活動矩形領域群が予め設定された変化、例えば図24(a)に示すようにユーザが指さし状態から、1回だけ手を広げ、活動矩形領域群が1回だけ“小”から“大”に変化していれば(ステップS67)、シングルクリックであると判定され、OS側にシングルクリック指示が出されて、図24(b)に示すように仮想カーソル25内にある実カーソル28によってアイコンなどがシングルクリックされる(ステップS68)。 Next, the CPU 9 checks whether there is an active rectangular area group composed of one or more active rectangular areas 26 in the virtual button click active area image, and whether the shape has changed. Is a change in which the active rectangular area group is set in advance, for example, as shown in FIG. 24 (a), the user extends his hand only once from the state of pointing, and the active rectangular area group is small once. If it has changed from “large” to “large” (step S67), it is determined that it is a single click, a single click instruction is issued to the OS side, and it is in the virtual cursor 25 as shown in FIG. An icon or the like is single-clicked by the real cursor 28 (step S68).
 また、ユーザが指さし状態から2回以上手を広げたり縮めたりし、活動矩形領域群が複数回“大”から“小”、“小”から“大”に変化していれば(ステップS67)、CPU9によってダブルクリックであると判定され、OS側にダブルクリック指示が出されて、実カーソル28の位置にあるアイコンなどがダブルクリックされる(ステップS68)。 Also, if the user extends or shrinks his / her hand twice or more from the pointing state, and the activity rectangular area group has changed from “large” to “small” and “small” to “large” several times (step S67). The CPU 9 determines that it is a double click, and a double click instruction is issued to the OS side, and the icon or the like at the position of the real cursor 28 is double clicked (step S68).
《両手ジェスチャによる画面拡大、縮小制御》
 また、各ウェブカメラ4、6のピント位置にユーザが右手、左手を出して、これら右手、左手を動かし、上述した仮想カーソル活動領域画像27に活動矩形領域群が存在するかどうかなどをチェックしたとき、活動矩形領域群の数が“2”になり、各々矩形になっていれば(ステップS63)、CPU9によってこれが検知されて、これら各活動矩形領域群の動きに応じてディスプレイ部3に表示されている操作対象画面の拡大、縮小、回転などが行われる(ステップS65)。
《Screen enlargement and reduction control using two-hand gestures》
In addition, the user puts his right hand and left hand at the focus position of each of the webcams 4 and 6 and moves the right hand and left hand to check whether or not there is an active rectangular area group in the virtual cursor active area image 27 described above. If the number of active rectangular area groups is “2” and each is rectangular (step S63), this is detected by the CPU 9 and displayed on the display unit 3 in accordance with the movement of each active rectangular area group. The operation target screen that has been enlarged is reduced, reduced, rotated, or the like (step S65).
 例えば、各ウェブカメラ4、6のピント位置にユーザが右手、左手を出して、その右手、左手を互いに離れる方向に移動させ、これに対応して図25に示す2つの活動矩形領域群が存在し、図26(a)に示すように前回より広くなる方向に移動してこれら各活動矩形領域群の距離が前回より長くなったとき、CPU9によって画面拡大指示が入力されたと判定され、活動矩形領域群の距離変化比に応じた拡大率の画面拡大指示が生成されてアプリケーション側に渡され、ディスプレイ部3に表示されているアプリケーション画面(操作対象画面)が拡大される。 For example, the user puts his right hand and left hand at the focus positions of the web cameras 4 and 6 and moves the right hand and left hand away from each other, and there are two groups of active rectangular areas shown in FIG. 25 corresponding thereto. Then, as shown in FIG. 26 (a), when the distance between each of the activity rectangle areas becomes longer than the previous time by moving in the direction wider than the previous time, it is determined that the screen enlargement instruction is input by the CPU 9, and the activity rectangle A screen enlargement instruction with an enlargement ratio corresponding to the distance change ratio of the region group is generated and passed to the application side, and the application screen (operation target screen) displayed on the display unit 3 is enlarged.
 また、各ウェブカメラ4、6のピント位置にユーザが右手、左手を出して、その右手、左手を互いに近づく方向に移動させ、これに対応して2つの活動矩形領域群が図26(b)に示すように前回より狭くなる方向に移動し、これら各活動矩形領域群の距離が前回より短くなったとき、CPU9によって画面縮小指示が入力されたと判定され、活動矩形領域群の距離変化比に応じた縮小率の画面縮小指示が生成されてアプリケーション側に渡され、ディスプレイ部3に表示されているアプリケーション画面(操作対象画面)が縮小される。 In addition, the user puts out his right hand and left hand at the focus positions of the web cameras 4 and 6 and moves the right hand and left hand in a direction approaching each other. Correspondingly, two active rectangular area groups are shown in FIG. When the distance between each of the active rectangular area groups is shorter than the previous time, the CPU 9 determines that a screen reduction instruction has been input, and the distance change ratio of the active rectangular area groups is A screen reduction instruction with a corresponding reduction ratio is generated and passed to the application side, and the application screen (operation target screen) displayed on the display unit 3 is reduced.
《両手ジェスチャによる画面回転制御》
 また、各ウェブカメラ4、6のピント位置にユーザが右手、左手を出して、これら右手、左手の一方を上に他方を下に移動させ、これに対応して図27に示す2つの活動矩形領域群のうち少なくとも一方が上方向(又は、下方向)に移動したとき、CPU9によって画面回転指示が入力されたと判定され、下側の活動矩形領域群に対する上側の活動矩形領域群の角度に応じた回転角度の画面回転指示が生成されてアプリケーション側に渡され、ディスプレイ部3に表示されているアプリケーション画面(操作対象画面)が回転される。
《Screen rotation control using two-hand gestures》
Also, the user puts out his right hand and left hand at the focus position of each of the webcams 4 and 6 and moves one of the right hand and left hand up and the other down, corresponding to the two activity rectangles shown in FIG. When at least one of the area groups moves upward (or downward), the CPU 9 determines that a screen rotation instruction has been input, and depends on the angle of the upper active rectangular area group with respect to the lower active rectangular area group A screen rotation instruction with the rotation angle is generated and passed to the application side, and the application screen (operation target screen) displayed on the display unit 3 is rotated.
 この際、図28(a)に示すように各活動矩形領域群の左右距離が狭い状態で、一方が大きく上方に移動し下側の活動矩形領域群に対する上側の活動矩形領域群の角度が大きいとき、CPU9によって大きな回転角度の画面回転指示が生成されてアプリケーション側に渡され、ディスプレイ部3に表示されているアプリケーション画面(操作対象画面)が大きく回転される。 At this time, as shown in FIG. 28 (a), in a state where the left and right distances of the respective active rectangular area groups are narrow, one of the active rectangular area groups moves upward and the angle of the upper active rectangular area group with respect to the lower active rectangular area group is large. At this time, a screen rotation instruction with a large rotation angle is generated by the CPU 9 and passed to the application side, and the application screen (operation target screen) displayed on the display unit 3 is largely rotated.
 また、図28(b)に示すように各活動矩形領域群の左右距離が広い状態で、一方が小さく上方に移動し下側の活動矩形領域群に対する上側の活動矩形領域群の角度が小さいとき、CPU9によって小さな回転角度の画面回転指示が生成されてアプリケーション側に渡され、ディスプレイ部3に表示されているアプリケーション画面(操作対象画面)が小さく回転される。 Also, as shown in FIG. 28B, when the left and right distances of the respective active rectangular area groups are wide, when one of the active rectangular area groups is small and moves upward, the angle of the upper active rectangular area group with respect to the lower active rectangular area group is small. The CPU 9 generates a screen rotation instruction with a small rotation angle and passes it to the application side, and the application screen (operation target screen) displayed on the display unit 3 is rotated small.
 このように、本発明の第1の実施形態においては、各ウェブカメラ4、6によってユーザを撮影して得られた低解像度のカラー画像に対し、グレー化処理、画像分割/2値化処理、色フィルタリング処理、フレームバッファ処理、フレーム間差分処理、ヒストグラム処理、活動矩形領域抽出処理、活動矩形領域選択処理、仮想カーソル制御処理/画面制御処理を行いユーザの手の動きを検知し、仮想カーソル25のサイズ制御、位置制御、色制御、クリック制御、操作対象画面の拡大制御、縮小制御、回転制御、上下スクロール制御、左右スクロール制御などを行うようにしているので、次に述べる効果を得ることができる。 As described above, in the first embodiment of the present invention, a graying process, an image division / binarization process, and the like for a low-resolution color image obtained by photographing a user with each of the web cameras 4 and 6, Color cursor processing, frame buffer processing, inter-frame difference processing, histogram processing, active rectangular region extraction processing, active rectangular region selection processing, virtual cursor control processing / screen control processing are performed to detect the movement of the user's hand, and virtual cursor 25 Size control, position control, color control, click control, operation target screen enlargement control, reduction control, rotation control, up / down scroll control, left / right scroll control, etc., so that the following effects can be obtained. it can.
 まず、解像度が高くない安価なウェブカメラ4、6を使用できることから、入力装置1aのコストを低く抑えることができる(請求項1の効果)。 First, since the inexpensive web cameras 4 and 6 that do not have high resolution can be used, the cost of the input device 1a can be kept low (effect of claim 1).
 また、各ウェブカメラ4、6によってユーザを撮影して得られた低解像度のカラー画像に対し、グレー化処理、画像分割/2値化処理、色フィルタリング処理を施して得られた2値化画像を画像格納エリア16に格納するため、ハードディスク8の容量が小さい場合にも入力装置1aを構成することができ、装置全体のコストを低く抑えることができる(請求項1の効果)。 In addition, a binarized image obtained by performing graying processing, image division / binarization processing, and color filtering processing on a low-resolution color image obtained by photographing the user with the web cameras 4 and 6. Is stored in the image storage area 16, the input device 1 a can be configured even when the capacity of the hard disk 8 is small, and the cost of the entire device can be kept low (effect of claim 1).
 また、各ウェブカメラ4、6によってユーザを撮影して得られた低解像度のカラー画像に対し、グレー化処理、画像分割/2値化処理、色フィルタリング処理など、少ない段数の画像処理を施して1フレーム分の2値化画像を得るようにしているので、CPU9に大きな負担をかけることを防ぎ、これによって処理速度が速くない安価なCPU9を使用した場合でもほぼリアルタイムでユーザの動きに対応するように、仮想カーソル25のサイズ制御、位置制御、色制御、クリック制御、操作対象画面の拡大制御、縮小制御、回転制御、上下スクロール制御、左右スクロール制御などを行うことができ、装置全体のコストを低く抑えることができる(請求項1の効果)。 Further, low-resolution color images obtained by photographing the user with the web cameras 4 and 6 are subjected to a small number of image processing such as graying processing, image division / binarization processing, and color filtering processing. Since a binary image for one frame is obtained, it is possible to prevent the CPU 9 from being subjected to a large burden, and thus to respond to a user's movement almost in real time even when an inexpensive CPU 9 whose processing speed is not fast is used. As described above, the size control, position control, color control, click control, enlargement control, reduction control, rotation control, up / down scroll control, left / right scroll control, and the like of the operation target screen can be performed. Can be kept low (effect of claim 1).
 また、各ウェブカメラ4、6によってユーザを撮影して得られた右眼側の各活動矩形領域26と左眼側の各活動矩形領域26とに対し、両眼視差法で中心座標位置を補正した後、大きさ順に番号を付けて中心座標位置を比較し、この比較結果に基づきピント位置に対応する右眼側の各活動矩形領域26と左眼側の各活動矩形領域26とを選択するようにしているので、各ウェブカメラ4、6のピント位置にあるユーザの手以外のもの、例えばユーザの後ろに人がいて動いていても、これに影響されることなくユーザの手の動きのみを抽出して、仮想カーソル25のサイズ制御、位置制御、色制御、クリック制御、操作対象画面の拡大制御、縮小制御、回転制御、上下スクロール制御、左右スクロール制御などを行うことができる(請求項1の効果)。 In addition, the center coordinate position is corrected by the binocular parallax method for each of the right-eye activity rectangle area 26 and the left-eye activity rectangle area 26 obtained by photographing the user with the web cameras 4 and 6. After that, numbers are assigned in order of size, and the center coordinate positions are compared. Based on the comparison result, the right-side active rectangular area 26 and the left-eye active rectangular area 26 corresponding to the focus position are selected. Therefore, even if there is something other than the user's hand at the focus position of each webcam 4, 6 such as a person behind the user and moving, only the movement of the user's hand is not affected by this. Can be extracted to perform size control, position control, color control, click control, enlargement control, reduction control, rotation control, up / down scroll control, left / right scroll control, and the like of the operation target screen. 1 Effect).
 また、本発明の第1の実施形態では、ユーザが片手だけを動かしているとき仮想カーソル制御指示、又は操作対象画面のスクロール制御であると判定し、仮想カーソル25のサイズ制御、位置制御、色制御、クリック制御、操作対象画面のスクロール制御などを行うようにしているので、片手だけでディスプレイ部3に表示されている仮想カーソル25のサイズ、位置、色、クリック、操作対象画面のスクロールなどを遠隔操作することができる(請求項3の効果)。 In the first embodiment of the present invention, when the user moves only one hand, it is determined that it is a virtual cursor control instruction or scroll control of the operation target screen, and the size control, position control, and color of the virtual cursor 25 are determined. Since control, click control, scroll control of the operation target screen, and the like are performed, the size, position, color, click, scroll of the operation target screen, etc. of the virtual cursor 25 displayed on the display unit 3 with only one hand can be controlled. Remote control is possible (effect of claim 3).
 また、本発明の第1の実施形態では、ユーザが両手を動かしているとき、右手の動き、左手の動きを各々検出し、操作対象画面の拡大/縮小制御指示、又は操作対象画面の回転制御指示であると判定するようにしているので、ユーザが右手、左手を動かすだけでディスプレイ部3に表示されているアプリケーション画面(操作対象画面)を拡大、縮小、回転させることができる(請求項4の効果)。 Further, in the first embodiment of the present invention, when the user moves both hands, the right hand movement and the left hand movement are detected, and the operation target screen enlargement / reduction control instruction or the operation target screen rotation control is detected. Since the instruction is determined to be an instruction, the application screen (operation target screen) displayed on the display unit 3 can be enlarged, reduced, and rotated only by the user moving the right hand and the left hand. Effect).
 また、本発明の第1の実施形態では、活動矩形領域抽出処理においてヒストグラムを統計処理して得られた結果を使用し、ヒストグラムから仮想カーソル活動領域画像27、仮想ボタンクリック活動領域画像を作成するようにしているので、ユーザの手など動いている部分を正確に検知することができ、安定した仮想カーソル制御、クリック制御、操作対象画面制御を行うことができる(請求項5の効果)。 In the first embodiment of the present invention, a virtual cursor activity region image 27 and a virtual button click activity region image are created from the histogram using the results obtained by statistically processing the histogram in the activity rectangular region extraction process. Therefore, a moving part such as a user's hand can be accurately detected, and stable virtual cursor control, click control, and operation target screen control can be performed (effect of claim 5).
 また、本発明の第1の実施形態では、活動矩形領域抽出処理において仮想カーソル活動領域画像27、仮想ボタンクリック活動領域画像に対し、多段階矩形オブジェクト抽出処理を行うようにしているので、ユーザの影などに起因する誤動作を防止し、安定した仮想カーソル制御、クリック制御、操作対象画面制御を行うことができる(請求項6の効果)。 In the first embodiment of the present invention, the multi-step rectangular object extraction process is performed on the virtual cursor activity area image 27 and the virtual button click activity area image in the activity rectangular area extraction process. It is possible to prevent malfunction caused by a shadow or the like and perform stable virtual cursor control, click control, and operation target screen control (effect of claim 6).
2.第2の実施形態の説明
 図29は、本発明による入力装置の第2の実施形態を示すブロック図である。
 この図に示す入力装置1bは、箱形に形成されたプラスチック部材などによって構成され、パソコン、テレビ、エアコン、大画面スクリーン装置などの遠隔操作対象機器の近傍に配置される入力装置筐体(図示は省略する)と、入力装置筐体の前面左側に取り付けられ、ユーザを撮影してカラー画像信号を出力する右眼用ビデオカメラ本体(請求項2のカラーカメラ本体)30と、入力装置筐体内に配置され、右眼用ビデオカメラ本体30で撮影されたカラー画像を処理して、右眼側の仮想カーソル活動領域画像、右眼側の仮想ボタンクリック活動領域画像を生成する右眼側画像処理基板31と、入力装置筐体の前面右側に取り付けられ、ユーザを撮影してカラー画像信号を出力する左目用ビデオカメラ本体(請求項2のカラーカメラ本体)32と、入力装置筐体内に配置され、左眼用ビデオカメラ本体32で撮影されたカラー画像を処理して、左眼側の仮想カーソル活動領域画像、左眼側の仮想ボタンクリック活動領域画像を生成する左眼側画像処理基板33と、入力装置筐体内に配置され、右眼側画像処理基板31から出力される右眼側の仮想カーソル活動領域画像、右眼側の仮想ボタンクリック活動領域画像、左眼側画像処理基板32から出力される左眼側の仮想カーソル活動領域画像、左眼側の仮想ボタンクリック活動領域画像を画像処理してユーザの手の動きに対応するポインティングデータを生成し、USBケーブル、信号接続ケーブルなどのケーブルを介して、パソコン、テレビ、エアコン、大画面スクリーン装置などの遠隔操作対象機器などに供給する共通処理基板34とを備えている。
2. Description of Second Embodiment FIG. 29 is a block diagram showing a second embodiment of the input device according to the present invention.
An input device 1b shown in this figure is composed of a plastic member or the like formed in a box shape, and is an input device housing (illustrated) disposed in the vicinity of a remote operation target device such as a personal computer, a television, an air conditioner, or a large screen screen device. Is attached to the left side of the front surface of the input device housing, and the right-eye video camera main body (color camera main body of claim 2) 30 that captures a user and outputs a color image signal, and the input device housing Right-eye image processing that generates a right-eye side virtual cursor activity region image and a right-eye side virtual button click activity region image by processing a color image captured by the right-eye video camera body 30 A left-eye video camera body that is attached to the substrate 31 and the front right side of the input device housing and that captures a user and outputs a color image signal (color camera body of claim 2) 2 and a color image placed in the input device casing and photographed by the video camera main body 32 for the left eye to process a virtual cursor activity area image on the left eye side and a virtual button click activity area image on the left eye side. The left eye side image processing board 33 to be generated, the right eye side virtual cursor activity area image and the right eye side virtual button click activity area image which are arranged in the input device housing and output from the right eye side image processing board 31 The left eye side virtual cursor activity region image and the left eye side virtual button click activity region image output from the left eye side image processing board 32 are image-processed to generate pointing data corresponding to the movement of the user's hand. Common processing bases that are supplied to devices such as personal computers, televisions, air conditioners, and large screen devices via cables such as USB cables and signal connection cables. And a 34.
 そして、ユーザを撮影して得られたカラー画像を解析して、影などの影響、ユーザの手前側にいる人、及び背後にいる人の影響などを取り除きながら、ユーザの手の動きに対応したポインティングデータを生成し、入力装置1b→ケーブル→遠隔操作対象機器なる経路でポインティングデータを遠隔操作対象機器に供給し、この遠隔操作対象機器の動作を制御する。 And, by analyzing the color image obtained by shooting the user, it responded to the movement of the user's hand while removing the effects of shadows, the effects of the people on the front side of the user and the people behind them Pointing data is generated, and the pointing data is supplied to the remote operation target device through the path of the input device 1b → cable → remote operation target device, and the operation of the remote operation target device is controlled.
 右眼用ビデオカメラ本体30は、320画素×240画素程度の解像度を持つカラーカメラによって構成されており、右眼側画像処理基板31から電源電圧、クロック信号などが出されているときユーザを撮影し、これによって得られたカラービデオ信号を右眼側画像処理基板31に供給する。 The right-eye video camera main body 30 is configured by a color camera having a resolution of about 320 pixels × 240 pixels, and captures a user when a power supply voltage, a clock signal, and the like are output from the right-eye image processing board 31. Then, the color video signal obtained thereby is supplied to the right eye side image processing board 31.
 右眼側画像処理基板31は、右眼用ビデオカメラ本体から出力されるカラービデオ信号をRGB形式のカラー画像に変換した後、HSV(色相・彩度・明度)方式で予め設定されている特定色(例えば、肌色)のカラー画像を抽出するのに必要なカラーマスクを使用して、カラー画像中の肌色画像を抽出する肌色画像抽出回路35と、右眼用ビデオカメラ本体30から出力されるカラービデオ信号をRGB形式のカラー画像に変換した後、予め設定されている階調のモノクロ画像に変換するグレー化処理回路36と、グレー化処理回路36から出力されるモノクロ画像を予め設定されている画面分割数で分割する(但し、画面分割設定されていないときこの画面分割処理はスキップされる)と共に、最大尤度しきい値法で2値化して2値化画像にする画像分割/2値化処理回路37と、画像分割/2値化処理回路37から出力される2値化画像と肌色画像抽出回路35から出力される肌色画像との論理和を取り、2値化画像中の肌色部分を抽出する色フィルタリング処理回路38とを備えている。 The right eye side image processing board 31 converts the color video signal output from the video camera body for the right eye into a color image in RGB format, and then is specified in advance by the HSV (Hue / Saturation / Brightness) method. Using a color mask necessary for extracting a color image of a color (for example, skin color), the skin color image extracting circuit 35 for extracting the skin color image in the color image and the video camera body 30 for the right eye are output. After the color video signal is converted into a color image in RGB format, a graying processing circuit 36 for converting the color video signal into a monochrome image having a preset gradation, and a monochrome image output from the graying processing circuit 36 are set in advance. The screen is divided by the number of screen divisions (however, when screen division is not set, this screen division processing is skipped) and binarized by the maximum likelihood threshold method. The logical division of the image division / binarization processing circuit 37 to be an image, the binarized image output from the image division / binarization processing circuit 37 and the skin color image output from the skin color image extraction circuit 35 is performed, And a color filtering processing circuit 38 for extracting a skin color portion in the binarized image.
 さらに、右眼側画像処理基板31は、色フィルタリング処理回路38から出力される2値化画像を数フレーム分~数十フレーム分、一時記憶するフレームバッファ回路39と、フレームバッファ回路39に記憶されている2値化画像を順次読み出しながらフレーム間差分処理を行い、差分画像を生成するフレーム間差分処理回路40と、フレーム間差分処理回路40からフレーム単位で出力される各差分画像を各分割エリア毎に積算して、ヒストグラムを生成するヒストグラム処理回路41と、ヒストグラム処理回路41から出力されるヒストグラムに対し統計処理を行うと共に、統計処理結果を用いて仮想カーソル活動領域判定処理、仮想ボタンクリック活動領域判定処理、多段階矩形オブジェクト抽出処理などを行って、影などの影響を取り除いた右眼側の仮想カーソル活動領域画像、右眼側の仮想ボタンクリック活動領域画像を生成する活動矩形領域抽出処理回路42とを備えている。 Further, the right eye side image processing board 31 is stored in the frame buffer circuit 39 and the frame buffer circuit 39 for temporarily storing the binary image output from the color filtering processing circuit 38 for several frames to several tens of frames. The inter-frame difference processing circuit 40 that performs the inter-frame difference processing while sequentially reading the binarized images, and generates the difference image, and the difference images output in units of frames from the inter-frame difference processing circuit 40 are divided into the divided areas. The histogram processing circuit 41 that accumulates each time to generate a histogram, and performs statistical processing on the histogram output from the histogram processing circuit 41, and also uses the statistical processing result to determine virtual cursor activity area determination processing, virtual button click activity Performs area determination processing, multi-step rectangular object extraction processing, etc., and influences such as shadows Removing the right eye of the virtual cursor activity area image, and a work rectangular area extraction processing circuit 42 for generating a virtual button click activity area image of the right eye side.
 さらに、活動矩形領域抽出処理回路42は、ヒストグラム処理回路41にて生成した多段階の差分画像データの最新の差分画像データから抽出した活動矩形領域と、前記仮想ボタンクリック活動領域画像とを比較し、抽出した活動矩形領域が仮想ボタンクリック活動領域の範囲を超えている場合は、これら抽出活動矩形領域を無効判定する処理を行うことにより、ユーザが入力操作に意図していない動きをノイズとして無視することができる。 Furthermore, the activity rectangular area extraction processing circuit 42 compares the activity rectangular area extracted from the latest difference image data of the multi-stage difference image data generated by the histogram processing circuit 41 with the virtual button click activity area image. When the extracted activity rectangle area exceeds the range of the virtual button click activity area, the extraction action rectangle area is judged to be invalid so that the movement that the user does not intend for the input operation is ignored as noise. can do.
 そして、右眼用ビデオカメラ本体30から出力されるカラービデオ信号に対し、グレー化処理、画面分割/2値化処理、色フィルタリング処理、フレームバッファ処理、フレーム間差分処理、ヒストグラム処理、活動矩形領域抽出処理を順次施して、右眼側の仮想カーソル活動領域画像、右眼側の仮想ボタンクリック活動領域画像を生成し、共通処理基板34に供給する。 For the color video signal output from the right-eye video camera body 30, graying processing, screen division / binarization processing, color filtering processing, frame buffer processing, inter-frame difference processing, histogram processing, active rectangular area Extraction processing is sequentially performed to generate a virtual cursor activity region image on the right eye side and a virtual button click activity region image on the right eye side, which are supplied to the common processing board 34.
 また、左眼用ビデオカメラ本体32は、320画素×240画素程度の解像度を持つカラーカメラによって構成されており、左眼側画像処理基板33から電源電圧、クロック信号などが出されているときユーザを撮影し、これによって得られたカラービデオ信号を左眼側画像処理基板33に供給する。 The left-eye video camera main body 32 is composed of a color camera having a resolution of about 320 pixels × 240 pixels. When the power supply voltage, clock signal, etc. are output from the left-eye image processing board 33, the user And the color video signal obtained thereby is supplied to the left eye side image processing board 33.
 左眼側画像処理基板33は、左眼用ビデオカメラ本体32から出力されるカラービデオ信号をRGB形式のカラー画像に変換した後、HSV(色相・彩度・明度)方式で、予め設定されている特定色(例えば、肌色)のカラー画像を抽出するのに必要なカラーマスクを使用して、カラー画像中の肌色画像を抽出する肌色画像抽出回路43と、左眼用ビデオカメラ本体32から出力されるカラービデオ信号をRGB形式のカラー画像に変換した後、予め設定されている階調のモノクロ画像に変換するグレー化処理回路44と、グレー化処理回路44から出力されるモノクロ画像を予め設定されている画面分割数で分割する(但し、画面分割設定されていないとき、この画面分割処理はスキップされる)と共に、最大尤度しきい値法で2値化して、2値化画像にする画像分割/2値化処理回路45と、画像分割/2値化処理回路45から出力される2値化画像と肌色画像抽出回路43から出力される肌色画像との論理和を取り、2値化画像中の肌色部分を抽出する色フィルタリング処理回路46とを備えている。 The left-eye-side image processing board 33 converts the color video signal output from the left-eye video camera body 32 into a color image in RGB format, and is preset in the HSV (Hue / Saturation / Brightness) method. A skin color image extraction circuit 43 that extracts a skin color image in a color image using a color mask necessary for extracting a color image of a specific color (for example, skin color) and an output from the left-eye video camera main body 32 After converting the color video signal to be converted into a color image in RGB format, a graying processing circuit 44 for converting to a monochrome image having a preset gradation, and a monochrome image output from the graying processing circuit 44 are set in advance. (However, when screen division is not set, this screen division process is skipped) and binarization is performed using the maximum likelihood threshold method. Logic of the image division / binarization processing circuit 45 to be a binarized image, the binarized image output from the image division / binarization processing circuit 45, and the skin color image output from the skin color image extraction circuit 43 And a color filtering processing circuit 46 that extracts the skin color portion in the binarized image.
 さらに、左眼側画像処理基板33は、色フィルタリング処理回路46から出力される2値化画像を数フレーム分~数十フレーム分、一時記憶するフレームバッファ回路47と、フレームバッファ回路47に記憶されている2値化画像を順次読み出しながらフレーム間差分処理を行い、差分画像を生成するフレーム間差分処理回路48と、フレーム間差分処理回路48からフレーム単位で出力される各差分画像を各分割エリア毎に積算して、ヒストグラムを生成するヒストグラム処理回路49と、ヒストグラム処理回路49から出力されるヒストグラムに対し統計処理を行うと共に、統計処理結果を用いて仮想カーソル活動領域判定処理、仮想ボタンクリック活動領域判定処理、多段階矩形オブジェクト抽出処理などを行って、影などの影響を取り除いた左眼側の仮想カーソル活動領域画像、左眼側の仮想ボタンクリック活動領域画像を生成する活動矩形領域抽出処理回路50とを備えている。 Further, the left eye side image processing board 33 is stored in the frame buffer circuit 47 and the frame buffer circuit 47 for temporarily storing the binarized image output from the color filtering processing circuit 46 for several frames to several tens of frames. The inter-frame difference processing circuit 48 that performs the inter-frame difference processing while sequentially reading the binarized images, and generates the difference image, and the difference images output in units of frames from the inter-frame difference processing circuit 48 are divided into the divided areas. A histogram processing circuit 49 that accumulates each time and generates a histogram, and performs statistical processing on the histogram output from the histogram processing circuit 49, and also uses the statistical processing result to determine virtual cursor activity area determination processing and virtual button click activity Performs area determination processing, multi-step rectangular object extraction processing, etc., and influences such as shadows It comprises removing the left eye side of the virtual cursor activity area image, and activities rectangular area extraction processing circuit 50 for generating a virtual button click activity area image of the left eye side.
 さらに、活動矩形領域抽出処理回路50は、ヒストグラム処理回路49にて生成した多段階の差分画像データの最新の差分画像データから抽出した活動矩形領域と、前記仮想ボタンクリック活動領域画像とを比較し、前述した図37に示すように、抽出した活動矩形領域が仮想ボタンクリック活動領域の範囲を超えている場合は、これら抽出活動矩形領域を無効判定する処理を行うことにより、ユーザが入力操作に意図していない動きをノイズとして無視することができる Further, the activity rectangular area extraction processing circuit 50 compares the activity rectangular area extracted from the latest difference image data of the multi-stage difference image data generated by the histogram processing circuit 49 with the virtual button click activity area image. As shown in FIG. 37 described above, when the extracted activity rectangular area exceeds the range of the virtual button click activity area, the user can perform an input operation by performing a process of determining that these extracted activity rectangle areas are invalid. Unintentional movement can be ignored as noise
 そして、左眼用ビデオカメラ本体32から出力されるカラービデオ信号に対し、グレー化処理、画面分割/2値化処理、色フィルタリング処理、フレームバッファ処理、フレーム間差分処理、ヒストグラム処理、活動矩形領域抽出処理を順次施して、左眼側の仮想カーソル活動領域画像、左眼側の仮想ボタンクリック活動領域画像を生成し、共通処理基板34に供給する。 For the color video signal output from the left-eye video camera main body 32, graying processing, screen division / binarization processing, color filtering processing, frame buffer processing, inter-frame difference processing, histogram processing, active rectangular area Extraction processing is sequentially performed to generate a virtual cursor activity region image on the left eye side and a virtual button click activity region image on the left eye side, and supply them to the common processing board 34.
 共通処理基板34は、両眼視差法による位置補正に必要な右眼用ウェブカメラ30、左眼用ウェブカメラ32の取り付け位置データ(水平距離“B”、上下距離など)、右眼用ウェブカメラ30、左眼用ウェブカメラ32の焦点距離“f”などの撮影条件情報が設定される撮影条件設定回路51と、右眼側画像処理基板31から出力される右眼側の仮想カーソル活動領域画像、右眼側の仮想ボタンクリック活動領域画像、左眼側画像処理基板33から出力される左眼側の仮想カーソル活動領域画像、左眼側の仮想ボタンクリック活動領域画像に活動矩形領域が含まれるとき、撮影条件設定回路51に設定されている撮影条件情報を用いて両眼視差法で各活動矩形領域の位置を補正する処理、大きさ順に各活動矩形領域に番号を付加する処理、同じ番号が付加された各活動矩形領域の中心座標間の距離(中心座標距離)を演算する処理、所定値以下になっている各中心座標距離に対応する活動矩形領域を選択する処理、選択した活動矩形領域のみを含み、選択していない活動矩形領域を含まない左眼側の仮想カーソル活動領域画像、左眼側の仮想ボタンクリック活動領域画像を作成する処理などを行って、右眼側の仮想カーソル活動領域画像、右眼側の仮想ボタンクリック活動領域画像、左眼側の仮想カーソル活動領域画像、左眼側の仮想ボタンクリック活動領域画像に含まれるユーザの手前にいる人、及び背後にいる人の動きなどの影響を取り除いた後、左眼側の仮想カーソル活動領域画像、左眼側の仮想ボタンクリック活動領域画像を数画像分~数十画像分保持する活動矩形領域選択処理回路52とを備えている。 The common processing board 34 includes right-eye web camera 30 and left-eye web camera 32 attachment position data (horizontal distance “B”, vertical distance, etc.) necessary for position correction by the binocular parallax method, right-eye web camera. 30, a shooting condition setting circuit 51 in which shooting condition information such as the focal length “f” of the left-eye web camera 32 is set, and a right-eye virtual cursor activity region image output from the right-eye image processing board 31 The right eye side virtual button click activity region image, the left eye side virtual cursor activity region image output from the left eye side image processing board 33, and the left eye side virtual button click activity region image include an activity rectangular region. Process for correcting the position of each active rectangular area by the binocular parallax method using the shooting condition information set in the shooting condition setting circuit 51, and adding a number to each active rectangular area in order of size Processing to calculate the distance between the center coordinates (center coordinate distance) of each activity rectangle area to which the same number is added, processing to select the activity rectangle area corresponding to each center coordinate distance that is less than or equal to the predetermined value, and selected Perform a process such as creating a virtual cursor activity area image on the left eye side that contains only the activity rectangle area and not an unselected activity rectangle area, and a virtual button click activity area image on the left eye side. Virtual cursor activity area image, right eye side virtual button click activity area image, left eye side virtual cursor activity area image, left eye side virtual button click activity area image, and a person in front of the user, and behind After removing the influence of the movement of the person, the activity rectangle that holds several to several tens of virtual cursor activity area images on the left eye side and virtual button click activity area images on the left eye side And a selection processing circuit 52.
 さらに、活動矩形領域選択処理回路52は、ウェブカメラ30、32の視野角を定数として右眼活動矩形領域と左眼活動矩形領域の中心点間の横方向での中心座標距離に応じて、カラーカメラから被写体までの距離を補正することでより正確にユーザの手前にいる人、及び背後にいる人の動きなどの影響を取り除くことができる。この補正の原理及び補正の演算式については、前述した図39及び数式4の通りである。活動矩形領域選択処理回路52は、中心座標距離に応じて予め演算した被写体までの距離をテーブルデータで保持しておくことで高速処理を実現できる。 Furthermore, the activity rectangular area selection processing circuit 52 uses the viewing angles of the web cameras 30 and 32 as constants to change the color according to the center coordinate distance in the horizontal direction between the center points of the right eye activity rectangular area and the left eye activity rectangle area. By correcting the distance from the camera to the subject, it is possible to more accurately remove the influence of the movement of the person in front of the user and the person behind. The principle of correction and the calculation formula for correction are as shown in FIG. 39 and Formula 4 described above. The active rectangular area selection processing circuit 52 can realize high-speed processing by holding the distance to the subject calculated in advance according to the center coordinate distance as table data.
 さらに、共通処理基板34は、活動矩形領域選択処理回路52に保持されている左眼側の各仮想カーソル活動領域画像のうち、最新の仮想カーソル活動領域画像に活動矩形領域群が存在するとき、活動矩形領域群の数、形状、移動有無、移動方向などに基づき、仮想カーソル位置指示、仮想カーソル形状指示、仮想カーソル色指示、操作対象画面スクロール指示、操作対象画面拡大指示、操作対象画面縮小指示、操作対象画面回転指示などのポインティングデータを生成すると共に、仮想カーソルがクリック可能な状態になっているとき、活動矩形領域選択処理回路52に保持されている各仮想ボタンクリック活動領域画像のうち、最新の仮想ボタンクリック活動領域画像に活動矩形領域群が存在しているかどうかをチェックし、活動矩形領域群が存在しているとき、活動矩形領域群の形状などに基づきシングルクリック指示、ダブルクリック指示などのポインティングデータを生成する仮想カーソル制御処理/画面制御処理回路53を備えている。 Further, the common processing board 34, when the active rectangle area group exists in the latest virtual cursor activity area image among the virtual cursor activity area images on the left eye side held in the activity rectangle area selection processing circuit 52, Virtual cursor position instruction, virtual cursor shape instruction, virtual cursor color instruction, operation target screen scroll instruction, operation target screen enlargement instruction, operation target screen reduction instruction based on the number of active rectangle areas, shape, movement presence / absence, movement direction, etc. In addition to generating pointing data such as an operation target screen rotation instruction, and when the virtual cursor is in a clickable state, among the virtual button click activity area images held in the activity rectangular area selection processing circuit 52, Check if the activity rectangle area group exists in the latest virtual button click activity area image, and the activity rectangle When the frequency group is present, a work such as a single click instruction based on the rectangular region group shape, the virtual cursor control processing / display control processing circuit 53 for generating a pointing data such as a double-click instruction.
 そして、右眼側画像処理基板31から出力される右眼側の仮想カーソル活動領域画像、右眼側の仮想ボタンクリック活動領域画像、左眼側画像処理基板32から出力される左眼側の仮想カーソル活動領域画像、左眼側の仮想ボタンクリック活動領域画像に活動矩形領域が含まれるとき、ユーザの前にいる人の動き、ユーザの背後にいる人の動きなどに起因するノイズを取り除きながらユーザの各手がどのように動いているかを判定し、この判定結果に応じて仮想カーソル位置指示、仮想カーソル形状指示、仮想カーソル色指示、操作対象画面スクロール指示、操作対象画面拡大指示、操作対象画面縮小指示、操作対象画面回転指示などのポインティングデータを生成し、遠隔操作対象機器となっているパソコン、テレビ、エアコン、大画面スクリーン装置などに供給する。 Then, the right eye side virtual cursor activity region image output from the right eye side image processing substrate 31, the right eye side virtual button click activity region image, and the left eye side virtual output from the left eye side image processing substrate 32. When an active rectangle area is included in the cursor activity area image and the left eye side virtual button click activity area image, the user removes noise caused by the movement of the person in front of the user, the movement of the person behind the user, etc. In accordance with the determination result, a virtual cursor position instruction, virtual cursor shape instruction, virtual cursor color instruction, operation target screen scroll instruction, operation target screen enlargement instruction, operation target screen Pointing data such as a reduction instruction, operation target screen rotation instruction, etc. is generated, and the personal computer, TV, air conditioner, large screen that is the target of remote operation Supplies such as clean equipment.
 このように、この第2の実施形態では、右眼用ビデオカメラ本体30、左眼用ビデオカメラ本体32によってユーザを撮影して得られた低解像度のカラー画像に対し、色フィルタリング処理、グレー化処理、画像分割/2値化処理、フレームバッファ処理、フレーム間差分処理、ヒストグラム処理、活動矩形領域抽出処理、活動矩形領域選択処理、仮想カーソル制御処理/画面制御処理などを行いユーザの手の動きを検知し、仮想カーソル位置指示、仮想カーソル形状指示、仮想カーソル色指示、操作対象画面スクロール指示、操作対象画面拡大指示、操作対象画面縮小指示、操作対象画面回転指示などのポインティングデータを生成し、遠隔操作対象機器に供給するようにしているので、遠隔操作対象機器側の仮想カーソルサイズ、仮想カーソル位置、仮想カーソル色、クリック、操作対象画面の上下スクロール、左右スクロール、拡大、縮小、回転などを遠隔操作することができる(請求項2の効果)。 As described above, in the second embodiment, color filtering processing and graying are performed on a low-resolution color image obtained by photographing the user with the right-eye video camera body 30 and the left-eye video camera body 32. User's hand movements including processing, image segmentation / binarization processing, frame buffer processing, inter-frame difference processing, histogram processing, active rectangular area extraction processing, active rectangular area selection processing, virtual cursor control processing / screen control processing, etc. Generating pointing data such as a virtual cursor position instruction, virtual cursor shape instruction, virtual cursor color instruction, operation target screen scroll instruction, operation target screen enlargement instruction, operation target screen reduction instruction, operation target screen rotation instruction, Since it is supplied to the remote operation target device, the virtual cursor size and virtual Cursor position, the virtual cursor color, click, scroll up and down of the operation target screen, left and right scrolling, enlargement, reduction, rotation, etc., can be remotely operated (Effect of Claim 2).
 また、この第2の実施形態では、解像度が高くない安価な右眼用ビデオカメラ本体30、左眼用ビデオカメラ本体32を使用できることから、入力装置1bのコストを低く抑えることができる(請求項2の効果)。 In the second embodiment, since the inexpensive right-eye video camera main body 30 and left-eye video camera main body 32 that do not have high resolution can be used, the cost of the input device 1b can be kept low. Effect of 2).
 また、この第2の実施形態では、右眼用ビデオカメラ本体30、左眼用ビデオカメラ本体32によってユーザを撮影して得られた低解像度のカラー画像に対し、グレー化処理、画像分割/2値化処理、色フィルタリング処理を施して得られた2値化画像をフレームバッファ回路39、47に格納するようにしているので、フレームバッファ回路39、47の記憶容量が小さい場合にも入力装置1bを構成することができ、装置全体のコストを低く抑えることができる(請求項2の効果)。 In the second embodiment, the low-resolution color image obtained by photographing the user with the right-eye video camera main body 30 and the left-eye video camera main body 32 is subjected to graying processing, image division / 2. Since the binarized image obtained by performing the binarization process and the color filtering process is stored in the frame buffer circuits 39 and 47, the input device 1b can be used even when the storage capacity of the frame buffer circuits 39 and 47 is small. The cost of the entire apparatus can be kept low (effect of claim 2).
 さらに、この第2の実施形態では、右眼用ビデオカメラ本体30、左眼用ビデオカメラ本体32によってユーザを撮影して得られた低解像度のカラー画像に対し、グレー化処理、画像分割/2値化処理、色フィルタリング処理、フレームバッファ処理、フレーム間差分処理、ヒストグラム処理、活動矩形領域抽出処理、活動矩形領域選択処理、仮想カーソル制御処理/画面制御処理など、少ない段数の画像処理を施してポインティングデータを生成するようにしているので、肌色画像抽出理回路35、43、グレー化処理回路36、44、画像分割/2値化処理回路37、45、色フィルタリング処理回路38、46、フレームバッファ処理回路36、47、フレーム間差分処理回路40、48、ヒストグラム処理回路41、49、活動矩形領域抽出処理回路42、50、活動矩形領域選択回路52、仮想カーソル制御処理/画面制御処理回路53として、処理速度があまり速くない素子の使用を可能にして装置全体のコストを低く抑えながら、ほぼリアルタイムでユーザの動きを検知し、遠隔操作対象機器を制御することができる(請求項2の効果)。 Furthermore, in the second embodiment, graying processing, image division / 2 are performed on a low-resolution color image obtained by photographing the user with the right-eye video camera body 30 and the left-eye video camera body 32. Image processing with a small number of stages such as value processing, color filtering processing, frame buffer processing, inter-frame difference processing, histogram processing, active rectangular area extraction processing, active rectangular area selection processing, virtual cursor control processing / screen control processing, etc. Since pointing data is generated, skin color image extraction circuits 35 and 43, graying processing circuits 36 and 44, image division / binarization processing circuits 37 and 45, color filtering processing circuits 38 and 46, a frame buffer Processing circuits 36, 47, inter-frame difference processing circuits 40, 48, histogram processing circuits 41, 49, activity rectangle As the area extraction processing circuits 42 and 50, the active rectangular area selection circuit 52, and the virtual cursor control processing / screen control processing circuit 53, it is possible to use an element whose processing speed is not so fast and keep the cost of the entire apparatus low. It is possible to detect the movement of the user in real time and control the remote operation target device (effect of claim 2).
 また、この第2の実施形態では、右眼用ビデオカメラ本体30、左眼用ビデオカメラ本体32によってユーザを撮影して得られた右眼側の各活動矩形領域と、左眼側の各活動矩形領域とに対し、両眼視差法で中心座標位置を補正した後、大きさ順に番号を付けて中心座標位置を比較し、この比較結果に基づきピント位置に対応する右眼側の各活動矩形領域と、左眼側の各活動矩形領域とを選択するようにしているので、右眼用ビデオカメラ本体30、左眼用ビデオカメラ本体32のピント位置にあるユーザの手以外のもの、例えばユーザの後ろに人がいて動いていても、これに影響されることなくユーザの手の動きのみを抽出して、仮想カーソルのサイズ制御、位置制御、色制御、クリック制御、操作対象画面の拡大制御、縮小制御、回転制御、上下スクロール制御、左右スクロール制御などを行うことができる(請求項2の効果)。 In the second embodiment, the right-eye activity rectangular area obtained by photographing the user with the right-eye video camera main body 30 and the left-eye video camera main body 32 and the left-eye activity After correcting the center coordinate position with the binocular parallax method for the rectangular area, the center coordinate position is compared in the order of size, and each active rectangle on the right eye side corresponding to the focus position based on the comparison result Since the area and each activity rectangle area on the left-eye side are selected, other than the user's hand at the focus position of the right-eye video camera main body 30 and the left-eye video camera main body 32, for example, the user Even if there is a person behind and moving, only the movement of the user's hand is extracted without being affected by this, size control of the virtual cursor, position control, color control, click control, enlargement control of the operation target screen , Reduction control, rotation Please, vertical scroll control can be performed such as the left and right scroll control (Effect of Claim 2).
 そして、この第2の実施形態においても、上述した第1の実施形態と同様に、ユーザが片手だけを動かしているとき、仮想カーソル制御指示、クリック制御指示、スクロール制御指示の何れかであると判定し、仮想カーソルサイズ指示、仮想カーソル位置指示、仮想カーソル色指示、スクロール制御指示、クリック指示などを示すポインティングデータを生成するようにしているので、片手だけで遠隔操作対象機器側のディスプレイに表示されている仮想カーソルのサイズ、位置、色、クリック動作、操作対象画面のスクロールなどを遠隔操作することができる(請求項3の効果)。 Also in the second embodiment, as in the first embodiment described above, when the user is moving only one hand, it is any one of the virtual cursor control instruction, the click control instruction, and the scroll control instruction. Since pointing data is generated to indicate and indicate virtual cursor size instructions, virtual cursor position instructions, virtual cursor color instructions, scroll control instructions, click instructions, etc., display on the display on the remote operation target device side with only one hand It is possible to remotely control the size, position, color, click operation, scroll of the operation target screen, and the like of the virtual cursor being operated (effect of claim 3).
 また、この第2の実施形態においても、上述した第1の実施形態と同様に、ユーザが両手を動かしているとき、右手の動き、左手の動きを各々検出して操作対象画面の制御指示であると判定し、操作対象画面拡大指示、操作対象画面縮小指示、操作対象画面回転指示などを示すポインティングデータを生成するようにしているので、ユーザが右手、左手を動かすだけで遠隔操作対象機器側のディスプレイに表示されている操作対象画面を拡大、縮小、回転させることができる(請求項4の効果)。 Also in the second embodiment, as in the first embodiment described above, when the user is moving both hands, the right hand movement and the left hand movement are detected, and control instructions on the operation target screen are displayed. Since it is determined that there is a pointing data that indicates an operation target screen enlargement instruction, an operation target screen reduction instruction, an operation target screen rotation instruction, etc., the user only needs to move the right hand and left hand to the remote operation target device side The operation target screen displayed on the display can be enlarged, reduced, and rotated (effect of claim 4).
 また、本発明の第2の実施形態では、活動矩形領域抽出処理回路42、50においてヒストグラムを統計処理すると共に、統計処理結果を使用してヒストグラムから右眼側の仮想カーソル活動領域画像、右眼側の仮想ボタンクリック活動領域画像、左側の仮想カーソル活動領域画像、左側の仮想ボタンクリック活動領域画像を作成するようにしているので、ユーザの手など動いている部分を正確に検知することができ、安定した仮想カーソル制御、クリック制御、操作対象画面制御を行うことができる(請求項5の効果)。 Further, in the second embodiment of the present invention, the active rectangular area extraction processing circuits 42 and 50 statistically process the histogram, and using the statistical processing result, the virtual cursor active area image on the right eye side from the histogram, the right eye The virtual button click activity area image on the side, the virtual cursor activity area image on the left side, and the virtual button click activity area image on the left side are created so that moving parts such as the user's hand can be accurately detected. Thus, stable virtual cursor control, click control, and operation target screen control can be performed (effect of claim 5).
 また、本発明の第2の実施形態では、活動矩形領域抽出処理回路42、50において、右眼側の仮想カーソル活動領域画像、右眼側の仮想ボタンクリック活動領域画像、左側の仮想カーソル活動領域画像、左側の仮想ボタンクリック活動領域画像に対し、多段階矩形オブジェクト抽出処理を行うようにしているので、ユーザの影などに起因する誤動作を防止し、安定した仮想カーソル制御、クリック制御、操作対象画面制御を行うことができる(請求項6の効果)。 In the second embodiment of the present invention, in the active rectangle area extraction processing circuits 42 and 50, the right eye side virtual cursor activity area image, the right eye side virtual button click activity area image, and the left virtual cursor activity area Multi-step rectangular object extraction processing is performed on the image and the virtual button click activity area image on the left side to prevent malfunctions caused by the shadow of the user, stable virtual cursor control, click control, operation target Screen control can be performed (effect of claim 6).
3.他の実施形態の説明
 上述した各実施形態では、各ウェブカメラ4、6、右眼用ビデオカメラ本体30、左眼用ビデオカメラ本体32で得られたカラー画像の全領域をグレー化し、2値化するようにしているが、図30フローチャート、図31の模式図に示すようにヒストグラムを統計処理、活動矩形領域抽出処理で得られた変化領域矩形(活動矩形領域を含む矩形)65に対し、指定された拡大縮小率(例えば拡大率“10%”)で拡大/縮小した拡大/縮小矩形マスク66を作成すると共に(ステップS71)、次フレームのカラー画像全領域をグレー化して得られたモノクロ画像の中から拡大/縮小矩形マスク66に対応する部分(モノクロ画像に含まれる活動領域部分の画像67)だけを抽出して、2値化処理するようにしても良い(ステップS72)。
3. Description of Other Embodiments In each of the above-described embodiments, the entire area of the color image obtained by each web camera 4, 6, right-eye video camera body 30, and left-eye video camera body 32 is grayed out and binarized. As shown in the flowchart of FIG. 30 and the schematic diagram of FIG. 31, the histogram is statistically processed, and the change area rectangle (rectangle including the activity rectangle area) 65 obtained by the activity rectangle area extraction process is An enlarged / reduced rectangular mask 66 enlarged / reduced at a specified enlargement / reduction ratio (for example, enlargement ratio “10%”) is created (step S71), and the monochrome image obtained by graying out the entire color image area of the next frame. Only the portion corresponding to the enlarged / reduced rectangular mask 66 (the image 67 of the active region portion included in the monochrome image) may be extracted from the image and binarized. Step S72).
 このようにすれば、モノクロ画像などの中から活動領域より少し広い範囲に含まれる画像のみを有効にし、それ以外の領域にある画像を無効にして、変化領域以外の部分に存在するノイズを除去することができる(請求項7の効果)。 In this way, only the image included in the area slightly wider than the active area from the monochrome image etc. is enabled, the image in the other area is disabled, and the noise existing in the part other than the change area is removed. (Effect of claim 7).
 また、上述した第1の実施形態、及び第2の実施形態では、CPU9による色フィルタリング処理、又は色フィルタリング処理回路38、46によって、肌色のカラー画像を抽出するようにしているが、ユーザが特定色の操作器、例えば赤ペンなどを使用して仮想カーソルの位置、クリック、操作対象画面のスクロール、操作対象画面の拡大、操作対象画面の縮小、操作対象画面の回転などを制御する場合には、赤色抽出用のカラーマスクを使用し、CPU9による色フィルタリング処理、又は色フィルタリング処理回路38、46によって赤色のカラー画像を抽出するようにしても良い。 Further, in the first embodiment and the second embodiment described above, the skin color image is extracted by the color filtering processing by the CPU 9 or the color filtering processing circuits 38 and 46, but the user specifies When controlling the position, click, scrolling of the operation target screen, enlargement of the operation target screen, reduction of the operation target screen, rotation of the operation target screen, etc. using a color operation device such as a red pen Alternatively, a color mask for red extraction may be used, and a red color image may be extracted by the color filtering processing by the CPU 9 or the color filtering processing circuits 38 and 46.
 これにより、各ウェブカメラ4、6、右眼用ビデオカメラ本体30、左眼用ビデオカメラ本体32の撮影範囲に複数の人がいても、ユーザが持つ操作器の色に対応するカラー画像を抽出して仮想カーソルのサイズ制御、位置制御、クリック制御、操作対象画面のスクロール制御、拡大制御、縮小制御、回転制御などを行うことができる。 Thereby, even if there are a plurality of people in the shooting range of each of the web cameras 4 and 6, the right-eye video camera main body 30, and the left-eye video camera main body 32, a color image corresponding to the color of the operating device possessed by the user is extracted. Thus, size control, position control, click control, scroll control of the operation target screen, enlargement control, reduction control, rotation control, and the like can be performed.
 また、このような色フィルタリング処理は、各ウェブカメラ4、6、右眼用ビデオカメラ本体30、左眼用ビデオカメラ本体32から出力されるカラービデオ信号に含まれるユーザの手など、動いている画像を抽出するために行っている処理であることから、ユーザが居る場所の照明条件が良好で、ユーザの手などが動いている画像と背景画像とのコントラストが大きいとき、色フィルタリング処理を省略するようにしても良い。 In addition, such color filtering processing moves such as the user's hand included in the color video signals output from the web cameras 4 and 6, the right-eye video camera body 30, and the left-eye video camera body 32. Since the process is performed to extract the image, the color filtering process is omitted when the lighting condition of the place where the user is located is good and the contrast between the image where the user's hand is moving and the background image is large. You may make it do.
 本発明は、情報端末装置やパーソナルコンピュータなどの情報機器に接続されて使用され、カメラによって操作者(ユーザ)の動作画像を取り込み、情報機器のカーソル操作やアプリケーションプログラムの選択及び実行などを制御する入力装置であって、特に、アルゴリズムを簡素化すると共に処理データ量を極力少なくして演算量、メモリ使用量を低減すると共に、リアルタイムでパソコンのカーソルなどを制御するようにした、ビデオ映像による入力装置に関するものであり、産業上の利用可能性を有する。 The present invention is used by being connected to an information device such as an information terminal device or a personal computer, takes an operation image of an operator (user) by a camera, and controls cursor operation of the information device, selection and execution of an application program, and the like. An input device, in particular, video input that simplifies the algorithm and reduces the amount of processing data as much as possible to reduce the amount of computation and memory usage, as well as controlling the cursor of the personal computer in real time. It relates to a device and has industrial applicability.
 1a、1b:入力装置
 2:パソコン
 3:ディスプレイ部
 4:ウェブカメラ(左眼用カラーカメラ)
 5:ビデオキャプチャ
 6:ウェブカメラ(右眼用カラーカメラ)
 7:USBインタフェース
 8:ハードディスク
 9:CPU
 10:メモリ
 11:表示インタフェース
 12:システムバス
 13:OS格納エリア
 14:アプリケーション格納エリア
 15:画像処理プログラム格納エリア
 16:画像格納エリア
 20:分割エリア
 21:活動分割エリア
 25:仮想カーソル
 26:活動矩形領域
 27:仮想カーソル活動領域画像
 28:実カーソル
 30:右眼用ビデオカメラ本体(右眼用カラーカメラ本体)
 31:右眼側画像処理基板
 32:左眼用ビデオカメラ本体(左眼用カラーカメラ本体)
 33:左眼側画像処理基板
 34:共通処理基板
 35:肌色画像抽出回路
 36:グレー化処理回路
 37:画像分割/2値化処理回路
 38:色フィルタリング処理回路
 39:フレームバッファ回路
 40:フレーム間差分処理回路
 41:ヒストグラム処理回路
 42:活動矩形領域抽出処理回路
 43:肌色画像抽出回路
 44:グレー化処理回路
 45:画像分割/2値化処理回路
 46:色フィルタリング処理回路
 47:フレームバッファ回路
 48:フレーム間差分処理回路
 49:ヒストグラム処理回路
 50:活動矩形領域抽出処理回路
 51:撮影条件設定回路
 52:活動矩形領域選択処理回路
 53:仮想カーソル制御処理/画面制御処理回路
 65:変化領域矩形
 66:拡大/縮小矩形マスク
 67:マスク後の画像
1a, 1b: input device 2: personal computer 3: display unit 4: web camera (color camera for left eye)
5: Video capture 6: Web camera (color camera for right eye)
7: USB interface 8: Hard disk 9: CPU
10: Memory 11: Display interface 12: System bus 13: OS storage area 14: Application storage area 15: Image processing program storage area 16: Image storage area 20: Division area 21: Activity division area 25: Virtual cursor 26: Activity rectangle Area 27: Virtual cursor activity area image 28: Real cursor 30: Video camera body for right eye (color camera body for right eye)
31: Right-eye image processing board 32: Left-eye video camera body (left-eye color camera body)
33: Left eye side image processing board 34: Common processing board 35: Skin color image extraction circuit 36: Graying processing circuit 37: Image division / binarization processing circuit 38: Color filtering processing circuit 39: Frame buffer circuit 40: Between frames Difference processing circuit 41: Histogram processing circuit 42: Activity rectangular area extraction processing circuit 43: Skin color image extraction circuit 44: Graying processing circuit 45: Image division / binarization processing circuit 46: Color filtering processing circuit 47: Frame buffer circuit 48 : Frame difference processing circuit 49: Histogram processing circuit 50: Activity rectangular area extraction processing circuit 51: Shooting condition setting circuit 52: Activity rectangular area selection processing circuit 53: Virtual cursor control processing / screen control processing circuit 65: Change area rectangle 66 : Enlarged / reduced rectangular mask 67: Image after masking

Claims (10)

  1.  ビデオカメラで得られた操作者の画像を処理して、操作者の動作内容に応じた操作指示を生成する入力装置において、
     操作者を撮影する右眼用カラーカメラと、
     この右眼用カラーカメラから所定距離だけ離れた位置に、前記右眼用カラーカメラと並んで配置され、前記操作者を撮影する左眼用カラーカメラと、
     前記右眼用カラーカメラから出力されるカラー画像に対し、グレー化処理、画像分割/2値化処理、フレーム間差分処理、ヒストグラム処理、活動矩形領域抽出処理を行い、前記操作者の右眼側活動矩形領域を抽出する右眼側画像処理プログラムと、
     前記左眼用カラーカメラから出力されるカラー画像に対し、グレー化処理、画像分割/2値化処理、フレーム間差分処理、ヒストグラム処理、活動矩形領域抽出処理を行い、前記操作者の左眼側活動矩形領域を抽出する左眼側画像処理プログラムと、
     前記右眼側画像処理プログラムで得られた右眼側活動矩形領域、前記左眼側画像処理プログラムで得られた左眼側活動矩形領域に対し、両眼視差法を使用した活動矩形領域選択処理、仮想カーソル制御処理/画面制御処理を行って、前記操作者の手、又は指先の動きを検出し、この検出結果に応じた操作指示を生成する画像処理プログラムと、
     を備えることを特徴とする入力装置。
    In an input device that processes an operator image obtained by a video camera and generates an operation instruction according to the operation content of the operator,
    A color camera for the right eye to photograph the operator,
    A color camera for the left eye that is arranged side by side with the color camera for the right eye at a position away from the color camera for the right eye by a predetermined distance, and photographs the operator,
    The color image output from the right-eye color camera is subjected to graying processing, image division / binarization processing, interframe difference processing, histogram processing, and active rectangular area extraction processing, and the right eye side of the operator A right eye side image processing program for extracting an active rectangular area;
    The color image output from the left-eye color camera is subjected to graying processing, image division / binarization processing, inter-frame difference processing, histogram processing, and active rectangular area extraction processing, and the left eye side of the operator A left eye side image processing program for extracting an active rectangular area;
    Activity rectangle region selection processing using binocular parallax for the right eye side activity rectangular region obtained by the right eye side image processing program and the left eye side activity rectangle region obtained by the left eye side image processing program An image processing program that performs virtual cursor control processing / screen control processing to detect the movement of the operator's hand or fingertip, and generates an operation instruction according to the detection result;
    An input device comprising:
  2.  ビデオカメラで得られた操作者の画像を処理して、操作者の動作内容に応じた操作指示を生成し、遠隔操作対象機器の動作を制御する入力装置において、
     箱形に形成される入力装置筐体と、
     この入力装置筐体の前面左側に取り付けられ、操作者の画像を撮影する右眼用カラーカメラ本体と、
     前記入力装置筐体の前面右側に取り付けられ、前記操作者の画像を撮影する左眼用カラーカメラ本体と、
     前記入力装置筐体内に配置され、グレー化処理回路、画像分割/2値化処理回路、フレーム間差分処理回路、ヒストグラム処理回路、活動矩形領域抽出処理回路によって、前記右眼用カラーカメラ本体から出力されるカラー画像を処理して、前記操作者の右眼側活動矩形領域を抽出する右眼側画像処理基板と、
     前記入力装置筐体内に配置され、グレー化処理回路、画像分割/2値化処理回路、フレーム間差分処理回路、ヒストグラム処理回路、活動矩形領域抽出処理回路によって、前記左眼用カラーカメラ本体から出力されるカラー画像を処理して、前記操作者の左眼側活動矩形領域を抽出する左眼側画像処理基板と、
     前記入力装置筐体内に配置され、活動矩形領域選択処理回路、仮想カーソル制御処理/画面制御処理回路によって、前記右眼側画像処理基板で得られた右眼側活動矩形領域、前記左眼側画像処理基板で得られた左眼側活動矩形領域に、両眼視差法を使用した活動矩形領域選択処理、仮想カーソル制御処理/画面制御処理を行って、前記操作者の手、又は指先の動きを検出し、この検出結果に応じたポインティングデータを生成し、遠隔操作対象機器の動作を制御する共通処理基板と、
     を備えることを特徴とする入力装置。
    In the input device that processes the image of the operator obtained by the video camera, generates an operation instruction according to the operation content of the operator, and controls the operation of the remote operation target device.
    An input device housing formed in a box shape;
    A color camera body for the right eye that is attached to the left side of the front surface of the input device housing and captures an image of the operator,
    A color camera body for the left eye that is attached to the front right side of the input device housing and captures an image of the operator;
    Output from the right-eye color camera body by the graying processing circuit, the image segmentation / binarization processing circuit, the inter-frame difference processing circuit, the histogram processing circuit, and the active rectangular area extraction processing circuit which are arranged in the input device casing. A right eye side image processing board that processes the color image to be extracted and extracts the right eye side activity rectangular region of the operator;
    Output from the left-eye color camera body by the graying processing circuit, image segmentation / binarization processing circuit, inter-frame difference processing circuit, histogram processing circuit, and active rectangular area extraction processing circuit, which is arranged in the input device casing. A left eye side image processing board that processes the color image to be extracted and extracts the left eye side activity rectangular region of the operator;
    The right-eye side active rectangular area, the left-eye-side image, which is arranged in the input device casing and obtained on the right-eye side image processing board by the active rectangular area selection processing circuit and the virtual cursor control processing / screen control processing circuit. An activity rectangular area selection process using a binocular parallax method, a virtual cursor control process / a screen control process are performed on the left-eye activity rectangle area obtained on the processing board, and the movement of the operator's hand or fingertip is performed. A common processing board for detecting, generating pointing data according to the detection result, and controlling the operation of the remote operation target device;
    An input device comprising:
  3.  前記仮想カーソル制御処理/画面制御処理、又は前記仮想カーソル制御処理/画面制御処理回路は、仮想カーソル活動領域画像上に活動矩形領域群が1つあるとき、その形状、移動有無に基づき、カールソル制御指示、又は画面スクロール指示を生成することを特徴とする請求項1又は2に記載の入力装置。 The virtual cursor control process / screen control process, or the virtual cursor control process / screen control processing circuit, when there is one active rectangular area group on the virtual cursor active area image, based on the shape and the presence / absence of movement, 3. The input device according to claim 1, wherein an instruction or a screen scroll instruction is generated.
  4.  前記仮想カーソル制御処理/画面制御処理、又は前記仮想カーソル制御処理/画面制御処理回路は、仮想カーソル活動領域画像上に活動矩形領域群が2つあるとき、その移動方向に基づき、画面回転指示、画面拡大指示、画面縮小指示の何れかを生成することを特徴とする請求項1乃至3の何れかの項に記載の入力装置。 The virtual cursor control processing / screen control processing, or the virtual cursor control processing / screen control processing circuit, when there are two active rectangular area groups on the virtual cursor active area image, based on the moving direction, a screen rotation instruction, 4. The input device according to claim 1, wherein either an instruction for enlarging a screen or an instruction for reducing a screen is generated.
  5.  前記活動矩形領域抽出処理、又は前記活動矩形領域抽出処理回路は、ヒストグラムの統計処理結果を使用して、前記ヒストグラムから仮想カーソル活動領域画像、仮想ボタンクリック活動領域画像を作成することを特徴とする請求項1乃至4の何れかの項に記載の入力装置。 The activity rectangle area extraction process or the activity rectangle area extraction processing circuit creates a virtual cursor activity area image and a virtual button click activity area image from the histogram using a histogram statistical processing result. The input device according to claim 1.
  6.  前記活動矩形領域抽出処理、又は前記活動矩形領域抽出処理回路は、前記仮想カーソル活動領域画像、又は前記仮想ボタンクリック活動領域画像に対し、多段階矩形オブジェクト抽出処理を行い、ノイズ成分を除去することを特徴とする請求項1乃至5の何れかの項に記載の入力装置。 The active rectangular area extraction process or the active rectangular area extraction processing circuit performs a multi-stage rectangular object extraction process on the virtual cursor active area image or the virtual button click active area image to remove a noise component. The input device according to any one of claims 1 to 5, wherein:
  7.  拡大/縮小矩形マスク作成処理、又は拡大/縮小矩形マスク作成処理回路を付加し、前記拡大/縮小矩形マスク作成処理、又は前記拡大/縮小矩形マスク作成処理回路によって、前記カラーカメラ、前記カラーカメラ本体で得られたカラー画像の中から、前記仮想カーソル活動領域画像上の変化領域矩形、又は前記仮想ボタンクリック活動領域画像の変化領域矩形に対応する画像を抽出し、それ以外の画像をカットして、ノイズ成分を除去することを特徴とする請求項1乃至6の何れかの項に記載の入力装置。 An enlargement / reduction rectangle mask creation process or an enlargement / reduction rectangle mask creation processing circuit is added, and the color camera, the color camera main body is obtained by the enlargement / reduction rectangle mask creation process or the enlargement / reduction rectangle mask creation process circuit. Extract the image corresponding to the change area rectangle on the virtual cursor activity area image or the change area rectangle of the virtual button click activity area image from the color image obtained in the above, and cut the other images The input device according to claim 1, wherein noise components are removed.
  8.  前記活動矩形領域抽出処理、又は前記活動矩形領域抽出処理回路は、前記ヒストグラム処理、又は前記ヒストグラム処理回路にて生成した多段階の差分画像データの最新の差分画像データから抽出した活動矩形領域と、前記仮想ボタンクリック活動領域画像との比較に基づき無効な抽出活動矩形領域を判定することを特徴とする請求項5乃至7の何れかの項に記載の入力装置。 The activity rectangle area extraction process, or the activity rectangle area extraction processing circuit, the activity rectangle area extracted from the latest difference image data of the multi-stage difference image data generated by the histogram process or the histogram processing circuit, The input device according to claim 5, wherein an invalid extraction activity rectangular region is determined based on a comparison with the virtual button click activity region image.
  9.  前記活動矩形領域選択処理では、前記カラーカメラの視野角を定数として前記右眼活動矩形領域と前記左眼活動矩形領域の中心点間の横方向での中心座標距離に応じて、前記カラーカメラから被写体までの距離を補正することを特徴とする請求項1乃至8の何れかの項に記載の入力装置。 In the activity rectangular area selection process, from the color camera according to a center coordinate distance in a horizontal direction between the center points of the right eye activity rectangular area and the left eye activity rectangular area with a viewing angle of the color camera as a constant. The input device according to claim 1, wherein the distance to the subject is corrected.
  10.  前記中心座標距離に応じて補正された前記被写体までの前記距離が予めテーブルデータとして保持していることを特徴とする請求項9に記載の入力装置。 10. The input device according to claim 9, wherein the distance to the subject corrected according to the center coordinate distance is stored in advance as table data.
PCT/JP2011/052591 2010-02-08 2011-02-08 Input device WO2011096571A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2010-025291 2010-02-08
JP2010025291A JP2013080266A (en) 2010-02-08 2010-02-08 Input device

Publications (1)

Publication Number Publication Date
WO2011096571A1 true WO2011096571A1 (en) 2011-08-11

Family

ID=44355563

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2011/052591 WO2011096571A1 (en) 2010-02-08 2011-02-08 Input device

Country Status (2)

Country Link
JP (1) JP2013080266A (en)
WO (1) WO2011096571A1 (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9227662B2 (en) 2010-07-16 2016-01-05 Piaggio & C. S.P.A. Tiltable motorcycles with two front steering wheels
JP5936155B2 (en) * 2012-07-27 2016-06-15 Necソリューションイノベータ株式会社 3D user interface device and 3D operation method
CN111753771A (en) * 2020-06-29 2020-10-09 武汉虹信技术服务有限责任公司 Gesture event recognition method, system and medium

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107562351A (en) * 2017-09-27 2018-01-09 努比亚技术有限公司 A kind of method, terminal and computer-readable recording medium for controlling screen resolution
JP6935887B2 (en) * 2020-02-14 2021-09-15 知能技術株式会社 Terminal operation system and terminal operation program

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2001070293A (en) * 1999-09-06 2001-03-21 Toshiba Corp Radio-diagnostic device
JP2004265222A (en) * 2003-03-03 2004-09-24 Nippon Telegr & Teleph Corp <Ntt> Interface method, system, and program
JP2009151516A (en) * 2007-12-20 2009-07-09 Yaskawa Information Systems Co Ltd Information processor and operator designating point computing program for information processor
JP2010015553A (en) * 2008-06-03 2010-01-21 Shimane Pref Gov Image recognition device, manipulation determination method, and program

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2001070293A (en) * 1999-09-06 2001-03-21 Toshiba Corp Radio-diagnostic device
JP2004265222A (en) * 2003-03-03 2004-09-24 Nippon Telegr & Teleph Corp <Ntt> Interface method, system, and program
JP2009151516A (en) * 2007-12-20 2009-07-09 Yaskawa Information Systems Co Ltd Information processor and operator designating point computing program for information processor
JP2010015553A (en) * 2008-06-03 2010-01-21 Shimane Pref Gov Image recognition device, manipulation determination method, and program

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9227662B2 (en) 2010-07-16 2016-01-05 Piaggio & C. S.P.A. Tiltable motorcycles with two front steering wheels
US9688305B2 (en) 2010-07-16 2017-06-27 Piaggio & C. S.P.A. Tiltable motorcycles with two front steering wheels
JP5936155B2 (en) * 2012-07-27 2016-06-15 Necソリューションイノベータ株式会社 3D user interface device and 3D operation method
US9495068B2 (en) 2012-07-27 2016-11-15 Nec Solution Innovators, Ltd. Three-dimensional user interface apparatus and three-dimensional operation method
CN111753771A (en) * 2020-06-29 2020-10-09 武汉虹信技术服务有限责任公司 Gesture event recognition method, system and medium

Also Published As

Publication number Publication date
JP2013080266A (en) 2013-05-02

Similar Documents

Publication Publication Date Title
US8094204B2 (en) Image movement based device control method, program, and apparatus
TWI543610B (en) Electronic device and image selection method thereof
JP4575829B2 (en) Display screen position analysis device and display screen position analysis program
US9256324B2 (en) Interactive operation method of electronic apparatus
JP2012238293A (en) Input device
US20150035752A1 (en) Image processing apparatus and method, and program therefor
US20210390721A1 (en) Display control apparatus, display control method, and storage medium
US20120236180A1 (en) Image adjustment method and electronics system using the same
US10152137B2 (en) Using natural movements of a hand-held device to manipulate digital content
WO2017057106A1 (en) Input device, input method, and program
US11416078B2 (en) Method, system and computer program for remotely controlling a display device via head gestures
CN111527468A (en) Air-to-air interaction method, device and equipment
US10979700B2 (en) Display control apparatus and control method
WO2011096571A1 (en) Input device
EP1381947A2 (en) A system and method for robust foreground and background image data separation for location of objects in front of a controllable display within a camera view
Drab et al. Motion Detection as Interaction Technique for Games & Applications on Mobile Devices.
EP3617851B1 (en) Information processing device, information processing method, and recording medium
TWI465984B (en) Method and control apparatus for determining control output in control domain
CN110297545B (en) Gesture control method, gesture control device and system, and storage medium
JP2014029656A (en) Image processor and image processing method
CN111176425A (en) Multi-screen operation method and electronic system using same
JP2008203538A (en) Image display system
CN114201028B (en) Augmented reality system and method for anchoring display virtual object thereof
KR20170043202A (en) Image photographing apparatus and control method thereof
JP2011039594A (en) Input device

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 11739918

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 11739918

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: JP