KR20130133117A - Multi human interface devide having poiner location information input area including vertical area to text input area - Google Patents

Multi human interface devide having poiner location information input area including vertical area to text input area Download PDF

Info

Publication number
KR20130133117A
KR20130133117A KR1020120152644A KR20120152644A KR20130133117A KR 20130133117 A KR20130133117 A KR 20130133117A KR 1020120152644 A KR1020120152644 A KR 1020120152644A KR 20120152644 A KR20120152644 A KR 20120152644A KR 20130133117 A KR20130133117 A KR 20130133117A
Authority
KR
South Korea
Prior art keywords
pointer
information input
position information
input
mode
Prior art date
Application number
KR1020120152644A
Other languages
Korean (ko)
Inventor
조은형
Original Assignee
조은형
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from KR1020120056427A external-priority patent/KR20130053363A/en
Application filed by 조은형 filed Critical 조은형
Priority to US14/403,998 priority Critical patent/US9612666B2/en
Priority to PCT/KR2012/011822 priority patent/WO2013180369A1/en
Publication of KR20130133117A publication Critical patent/KR20130133117A/en
Priority to US14/718,107 priority patent/US9612667B2/en
Priority to US14/718,108 priority patent/US20150253869A1/en
Priority to US14/718,105 priority patent/US9639173B2/en
Priority to US14/794,825 priority patent/US9612668B2/en
Priority to US15/475,045 priority patent/US9880637B2/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/02Input arrangements using manually operated switches, e.g. using keyboards or dials
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/02Input arrangements using manually operated switches, e.g. using keyboards or dials
    • G06F3/023Arrangements for converting discrete items of information into a coded form, e.g. arrangements for interpreting keyboard generated codes as alphanumeric codes, operand codes or instruction codes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/033Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/14Digital output to display device ; Cooperation and interconnection of the display device with other functional units
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2203/00Indexing scheme relating to G06F3/00 - G06F3/048
    • G06F2203/038Indexing scheme relating to G06F3/038
    • G06F2203/0381Multimodal input, i.e. interface arrangements enabling the user to issue commands by simultaneous use of input devices of different nature, e.g. voice plus gesture on digitizer

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • General Engineering & Computer Science (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Position Input By Displaying (AREA)

Abstract

The present invention relates to a multi-human interface device for forming a pointer location information input area including a vertical area and a text input area comprising: a pointer location information input unit for receiving information related to a pointer location from a user; a text input unit comprised of a plurality of physical buttons; a first mode conversion unit for releasing or operating a pointer location information input mode; a pointer execution command reception unit for executing one or more functions in a location in which the pointer is located. The pointer location information input unit is located over the text input unit. The pointer execution command reception unit is located under the text input unit. The pointer location information input area formed by the pointer location information input unit includes the text input area formed in the text input unit and the vertical area.

Description

BACKGROUND OF THE INVENTION 1. Field of the Invention [0001] The present invention relates to a composite human interface device, which forms a pointer position information area including a text input area and a vertical plane,

The present invention relates to a human interface for receiving text information or pointing position information from a user in a digital device capable of receiving text or pointing position related information such as a computer, a notebook, a tablet PC, a mobile phone, and the like, and transmitting the text information or the pointing position information to the digital device.

A text input device such as a keyboard for text input has been disclosed in personal computers and portable digital devices. Also, a pointing device such as a mouse device for controlling the pointing position of the pointer for controlling the digital device and performing the function has been disclosed.

US 5734375 B US 5864334 B US 20110006991 A1

The conventional text input device and the pointing device are provided as separate devices or the pointing input area is formed at a separate position different from the text input area of the text input device. Therefore, the movement of the user's hand is more than necessary in the work environment in which the text input, the pointing position information input, and the pointer execution command input are frequently changed, thereby lowering the work efficiency.

In the present invention, a pointing position information input area of the pointing device is provided on the upper side of the text input area of the text input device, and a switching part for switching the text input mode and the pointing position information input mode is provided. The user can perform the pointing input while minimizing the movement of the user's hand, thereby improving the work efficiency.

By integrating the separately provided text input device and the pointing device into one human interface device, it is possible to reduce the unit price and the size of the product, and eliminate the unnecessary operation of the user operation, thereby improving the work efficiency.

1 is an exemplary view of a composite human interface device having a pointing device and a text input device.
2 is a flowchart for explaining an example of the operation sequence according to the mode switching of the pointing device and the text device.
3 is an example of a configuration of a text input device and a pointing device.
Fig. 4 shows examples of the pointer execution command unit integral mode switching unit.
Fig. 5 shows an example of the structure of the pointer position information input device.
6 is a configuration diagram of an infrared laser sensor module.
Fig. 7 shows an embodiment for displaying the pointer position information input area.
8 is a first embodiment of a pointer position information input device which can be separated from a text input device.
9 is a second embodiment of a pointer position information input device which can be separated from the text input device.
10 is a third embodiment of a pointer position information input device which can be separated from a text input device.
11 is a front view and a perspective view of a human interface apparatus to which the lower-positioned pointer position information input device is applied.
12 is an embodiment of the pointer execution command unit integral mode switching unit.
13 is an embodiment utilizing a cover of a human interface device to which a lower-positioned pointer positional information input device is applied.
14 is an embodiment of a human interface device applied to a portable electronic device.

The present invention relates to a composite human interface for inputting text information or pointing position information from a user to a digital device capable of receiving text or pointing position related information such as a computer, a notebook, a tablet PC, a mobile phone, and the like.

As a conventional text information input device, a keyboard composed of an elastic body and a plurality of physical buttons connected to a switch is widely used.

In a digital device having a touch interface, a virtual keyboard is displayed on a display. When a part of a user's body is touched with a virtual keyboard displayed on the display, a gesture or an electrical signal of the user is sensed, A method of inputting text on a virtual keyboard displayed in a portion is also used.

The touch interface recognizes the movement of a part of the user's body or recognizes the contact with a specific contact surface, senses the flow of current through the user's body, or intercepts or interferes with light or sound waves by a part of the user's body The touch of the user can be recognized by sensing or the like.

Examples of touch interfaces include a pressure sensitive touch screen, an electrostatic touch screen, an optical touch screen, and an ultrasonic touch screen.

A resistive touch screen or a pressure sensitive touch screen is a pressure sensitive touch screen.

Resistive touchscreen is cheap, and it is advantageous to write in handwriting or small cell with stylus pen. However, because it uses pressure, it is not recognized by pressing hard, and touch feeling is slightly duller than electrostatic touch method. It is known as a disadvantage.

The resistive touchscreen consists of several layers.

Among these several layers are conductive layers, with the two layers facing the air layer.

Pressing the outer screen will cause the two conductive layers to touch each other, which causes the resistance and current to change and recognize the touch.

Capacitive sensing or electrostatic touching is a type of sensing operation using a capacitive coupling effect.

The electrostatic touchscreen is made of highly conductive glass, indium tin oxide, unlike pressure-sensitive touchscreens that use pressure.

By attaching the sensor to the four corners of the glass, current flows through the glass surface.

It is the electrostatic principle to recognize the change of current through this.

At this moment, the moment the finger moves on the screen, the electrons that flowed through the glass flow into the body through the finger, and the sensor senses the position where the change occurred.

The electrostatic touch method is known to have a feeling of operation and smoothness compared with a pressure sensitive touch screen because the touch is recognized by touching the screen rather than pressing the screen by force.

In addition, the electrostatic touch method enables multi-touch that can touch various places.

Since the electrostatic touch method is a method of operating by using a change in electric current, it may not be used with a leather glove which does not communicate with electric current, or may not be operated with a nail or a stylus pen.

However, you can touch the screen with a separate electrostatic stylus pen.

Sensors are sensitive and may be affected by peripheral devices.

The optical touch screen operates by measuring coordinates with the infrared camera mounted on the apex of the touch screen and the infrared light illuminates with the shadow of the object to be touched.

The ultrasonic touch screen emits ultrasonic waves on the screen, detects the interference effect by the user's touch, and operates by measuring the coordinates.

Various touch input techniques that can be used to detect positional information of a part of a user's body by sensing contact or movement of another user and controlling position information of the pointer can be used in the present invention.

1 is an exemplary diagram of a composite human interface device having a pointing device and a text input device.

The pointing device-integrated text input device may have a housing (100) for supporting a text input device and a pointer location information input area.

The housing has a strength enough to withstand the pressure of a user input and receives text input information and pointer position information input information and transmits the information to a digital device connected to the pointing device integrated text input device by wire or wireless A control unit, a memory unit, a battery unit, an encoding unit, and a transmission unit.

The pointing device-integrated text input device may include a plurality of buttons 109 for receiving text input information from a user.

The plurality of buttons 109 may be configured as physical buttons or virtual buttons.

The physical button may include a button connected to the elastic body, or may have a feature that the button itself is elastic and moves to the position when the user receives the input, and returns to the original position when the pressure of the user is removed.

The physical button is connected to the electrical switch, and the user's pressure is applied to the physical button, so that the position of the button is shifted and the phase of the switch is changed, so that the text input value of the corresponding button is generated.

The physical button is structured to have elasticity without an electrical switch, and moves to a position when the user's pressure is removed. The user's text input information is input by the touch input device The user's pressure or gesture can generate textual input information based on the recognized location information.

The virtual button may be configured by displaying a text input button on a display device.

The virtual button can display an arbitrary button by projecting light onto a transparent, semi-transparent or opaque object.

The virtual button is not perceived by the user's eyes but can generate corresponding text input information based on user's pressure or gesture information with unique position information per text.

The complex human interface device may have a pointing location information input area 108a or 108b having a text input area 107 and at least a portion thereof as a common area.

The pointing position information input areas 108a and 108b are areas located at the upper or lower end of the button for text input. The pointing position information input area 108a and 108b may have at least a part of the pointing position information input area common to the text input area have.

The pointing location information input area may include the text input area or the text input area may include a point location information input area.

The pointing position information input area and the text input area are at least partially common and at least partially used as the pointing position information input area but not used as the text input area and at least partially used as the text input area, May not be used.

The complex human interface device includes a pointer location information input device 105 for text input or for pointing location information input or for forming a virtual text input area or a virtual pointing location information input area for text input and pointing location information input. It may be configured on the composite human interface device or external to the composite human interface device.

The pointer location information input device 105 is composed of a device such as an infrared generator and an infrared receiver or an RGB camera or an ultrasonic generator and an ultrasonic receiver or an infrared generator and an infrared camera, and the surface or the housing 100 of the housing 100. A part of the user's body may be sensed in the air to receive location information and gesture information.

The pointer location information input device may be modularized and configured as a plurality of modules to expand the pointer location information input area or to improve accuracy and sensitivity.

For example, a pointer position information input device 105b for configuring a text input or pointing position information input area 108b of the right hand.

Or a pointer position information input device 105a for configuring the left hand text input or pointing position information input area 108a.

The text input area or the pointing position information input area 108b for the right hand may include an area of the J button of the English standard keyboard.

The text input area or the pointing position information input area 108a for the left hand may include an area of an F (F) button of an English standard keyboard.

The pointer location information input device may include both the J and F buttons of the English standard keyboard when there is no division for the left hand and the right hand.

The pointing location information input area may be a pointer position movement on the plurality of display parts when the complex human interface device is connected to a plurality of digital devices having a display part.

For example, the pointer position information input region may be divided into regions so as to be matched with the respective display portions, or a separate button indicating the display portion may be operated to transmit pointer position information on the display portion, or a plurality of display portions may be referred to as a virtual single The pointer position information can be transmitted so that the pointer can be moved on the virtual single display unit.

The complex human interface device includes a pointing location information input device that configures a pointing input area and receives pointer location information from a user, and inputs a button or an image, a space, an icon, or a text where a pointer moved by the pointing location information input device is located. It may have a pointer execution command unit (101, 102) for executing at least one function on the window.

The pointer execution command section can be composed of one or two buttons, which are responsible for the first function and the second function, respectively, and can be located on the left side of the housing, in the mail, or in the center.

The first function may function as a left click of a computer mouse and the second function may function as a right click of a computer mouse, for example.

The pointer execution command unit including one or two buttons may be provided at both the left side and the post of the housing.

The pointer execution command unit may be configured to be operated by touching or light interception of a part of the user's body, interference of ultrasonic waves, or shadow of a part of the user's body using the above-described touch technology.

The pointer execution command unit may be configured as an elastic physical button.

The pointer execution command unit may operate using at least one button of a text button located in the text input area in addition to the pointing position information input area.

The pointer execution command unit can operate by selecting a physical or virtual text button on the pointing position information input area.

For example, when using a virtual input device that can input location information in space as a pointing location information input device and using a physical button as a text input device, the position of the pointer on the virtual pointing location information input area in the pointing location information input mode. You can issue a pointer execution command by entering information and pressing the physical text button at that location.

The pointer execution command unit may receive pointer position information by a user first gesture in a pointing position information input area and generate a pointer execution command by a user second gesture at the same position.

The pointer execution command unit may be configured to perform a first function by a first gesture of the user's body, a first voice, a first blink, a first mouth shape, or the like.

The pointer execution command unit may be configured to perform a second function by a second gesture of the user's body, a second voice, a second blink, a second mouth shape, or the like.

The complex human interface device may include a text input mode for receiving text information through the text input device and a pointing location information input mode for receiving pointing location information through the pointing location information input device.

The text input mode and the pointing position information input mode can be switched by the mode switching unit 103.

The mode switching unit 103 may be a switch located separately on the housing.

The mode switching unit 103 may detect the simultaneous reception of at least one text input button or a plurality of text inputs included in the text input device and perform mode switching.

The mode switching unit 103 may switch the mode by receiving control information from a digital device connected to the composite human interface device by wire or wirelessly.

The mode switching unit 103 may be integrated with the pointer execution command unit 102.

For example, by a first gesture such as contact of a part of the user's body on a first button that first reacts to a first touch or first pressure and to a second response to a second touch or second pressure. It may be configured to generate a pointer execution command in response to the second touch or the second pressure by a second gesture such as sensing a touch or the first pressure to switch modes and pressing a button.

The mode switching unit 103 may be configured as a temporary mode switching mode and a permanent mode switching mode.

For example, in response to a first touch or a first pressure, a mode is changed from a text input mode to a pointing position information input mode, and when a first touch or a first pressure is removed, have.

Alternatively, in the pointing position information input mode in response to the first touch or the first pressure, the mode may be switched to the text input mode, and when the first touch or the first pressure is removed, the mode may be set to return to the pointing position information input mode. .

Switch from text input mode to pointing position information input mode in response to the second touch or second pressure or toggle switch, and configure the permanent switching mode to maintain the pointing position information input mode even if the second touch or second pressure is removed. Can be.

Temporary mode switching may be performed by first control information received from a digital device connected to the composite human interface device by wire or wirelessly.

Permanent mode switching may be performed by the second control information received from the digital device connected to the composite human interface device by wire or wirelessly.

The mode switching unit 103 may be integrated with the pointer execution command unit.

For example, when the first touch or first pressure is detected on the pointer execution command unit, the pointer mode is switched from the text input mode to the pointer position information input mode. When the second touch or the second pressure is detected, And when the third touch or the third pressure is applied, it is set to the permanent switching mode that operates in the pointer position information input mode even if the third touch or the third pressure is removed.

At this time, it is possible to configure to be able to input a pointer execution command even in the permanent switching mode.

The mode switching unit may be formed on the left or right side of the housing (106a, 106b)

The mode switching units 106a and 106b formed on the left side, right side, left side, and right side of the housing may be configured to include a virtual button or a physical button, and configured to operate by sensing a user's touch input or pressure.

The mode switching units 106a and 106b formed on the left, right, left, and right sides of the housing may have an input area of 3 centimeters or more and less than 15 centimeters along the side surface of the housing.

When the length of the mode switch is less than 5 centimeters, it is preferable to set the position of the mode switch to include the left space of the left tab key of the window standard keyboard.

The composite human interface device may have a transmitter 104 for transmitting data by wire or wirelessly to a digital device including an external or composite human interface device.

The digital device can receive text input or receive pointer location information.

Fig. 2 is a flowchart for explaining an example of the operation sequence according to mode switching of the pointing device and the text device.

The complex human interface device may have a separate power supply unit or may receive power from the outside in a wired or wireless manner and may have a separate switch for controlling the power supply unit.

When power is supplied to the composite human interface device by the switch controlling the power supply unit, it may be determined whether the composite human interface device is currently in the text input mode or the pointing location information input mode (200).

As a result of the determination, if the text input mode is selected, the text input device may be activated and a text input may be received from the user (201).

The text input of the input user can be transmitted to a digital device connected by wire or wireless (202).

When the mode is switched to the pointing position information input mode during the text input mode, the text input mode can be switched to the pointing position information input mode.

Or the pointing position information can be input simultaneously with the text input.

When the mode is switched to the pointing position information input mode, the pointing position information may be received by the user (204).

The input pointer location information may be transmitted to a digital device connected by wire or wireless connection (205).

When the composite human interface device receives the pointer first execution command 206, the composite human interface device may transmit the pointer first execution command to a digital device that is connected by wire or wirelessly (207).

When the composite human interface device receives the pointer second execution command 208, the composite human interface device may transmit the pointer first execution command to a digital device connected by wire or wirelessly (209).

When the temporary mode switch is released or the permanent mode switch is released, the composite human interface device may switch to the text input mode.

The power switch of the power supply unit may be terminated when the power supply of the complex human interface device is disconnected or there is no connected digital device, or when the connection is lost, or when there is no user input for a predetermined time or by control of a digital device connected by wire or wirelessly.

3 is an example of a configuration of a text input device and a pointing device.

The complex human interface device may include a first housing 301 including a power supply unit, a controller, or a communication unit, and a second housing 302 for configuring a text input area and a pointing location information input area.

The composite human interface device may have a text input device and a text input area 303 that receive text input from a user.

At this time, the text input area 303 may be composed of a virtual button or a physical button.

The pointing location information input device may be configured as a pressure sensitive or capacitive touch pad type and may be located 304 above the physical buttons of the text input device.

At this time, if the text input device is constituted by a physical touch pad type such as a pressure sensitive type or capacitive type, the touch pad can be used as a text input device and a pointing position information input device.

At this time, a physical touch pad such as a pressure sensitive or capacitive touch panel may be configured as a large-area touch pad type complex human interface device 300 including a touch pad 304 including a plurality of text button regions of a text input unit.

Alternatively, a physical touch pad such as a pressure sensitive or capacitive touch panel may be configured as a plurality of touch pad type human interface devices 310 including a plurality of touch pads 311 including one text button area of the text input unit.

Or the upper camera type human interface device 320 constituting the pointer location information input area 321 by an infrared camera or an RGB camera 222 positioned at an upper end of the pointing location information input area. It can be configured as.

In this case, the upper camera type complex human interface device 320 may configure the virtual text input button 303 using the upper camera.

The camera may be configured as a bottom camera 332 type complex human interface device 330 configured to be positioned at the bottom of the second housing to form a pointing position information input area 331.

In this case, the lower camera type composite human interface device 330 may configure a virtual text input button 303 to replace the physical text button using the lower camera.

Or a virtual pointing location information input area composed of a pair of infrared receivers or ultrasound receivers for receiving the information that the infrared or ultrasonic waves transmitted by the infrared transmitter or the ultrasonic transmitter are blocked or interfered with by a part of the user's body. 341 may be configured as a transmit / receive composite human interface device 340.

At this time, the virtual pointing position information input area formed as a pair can be used as a virtual text button input means in place of the physical text button by the mode switching unit.

Fig. 4 shows examples of the pointer execution command unit integral mode switching unit.

An elastic body 403 for returning the position to the original position when the pressure is removed, a button 402 for receiving the pressure of the user, and a pointer operation command A mode switching unit 401 for switching between a text mode and a pointing position information input mode by detecting a user's touch input at an upper portion of a button in a button-type pointer execution command unit having switches 404 and 405 for generating a touch input The pointer execution command unit integrated mode switching unit 400 can be configured.

An elastic body 403 for returning to the original position when the pressure is removed, a button 402 for receiving the pressure of the user, and a pressure sensor A second switch unit 411 which is connected by a user second pressure smaller than the user first pressure to generate a mode switching execution command in a button-type pointer execution instruction unit having first switches 404 and 405 for causing execution commands to be generated, 412), and a pressure discriminating pointer execution command unit integrated mode switching unit 410.

 The pressure difference type pointer execution command unit integrated mode switching unit 410 moves the button unit 402 and fixes the button unit 402 so that the elastic unit 403 does not return to the original position.

In this case, the position where the button is fixed can be configured to be set to the position where the second switches 411 and 412 are joined and the first switches 404 and 405 are not joined, and to operate in the permanent mode switching mode when the button is fixed .

 The pointer execution instruction unit integrated mode switching unit 400 using the touch input can be configured to operate the permanent mode changeover switch 414 through position shifting such as slipping the button unit and the permanent mode changeover switch 414 The pointer execution command switches 404 and 405 can be configured to operate when additional pressure is applied in the operated state.

When the contact area of a part of the user body which is composed of the touch pad 421 and touches the touch pad is in the first predetermined range, the pointer execution command instruction part integrated mode switching part 410 operates as the mode switching part, It can be operated as a pointer execution command section.

Fig. 5 shows an example of the structure of the pointer position information input device.

The pointer position information input device of the human interface device of the present invention may include a sensor module 501 including a light emitter and a camera.

The light emitter of the sensor module 501 is blocked by an obstacle such as a user's finger on the pointing position information input area 108 formed on the light plane composed of light emitted from the light emitter And serves to construct the light plane to be reflected.

 The camera of the sensor module includes an optical sensor that detects light blocked or reflected by the obstacle.

The optical sensor may use a line camera capable of recognizing a line image.

It is preferable that the line camera is configured such that the image sensor is composed of 400 to 1000 elements in a line form, and preferably 500 to 700 elements.

The line camera may include a plurality of line image sensors that are smaller than one five fifth to one seventh one of the 500 to 700 elements.

The light emitter is constituted by a light emitting element for constituting a light plane on the upper end close to the letter input button 303 by about 1 millimeter.

The pointing position information input area 108 is composed of a part of the light plane.

As the light emitting element, for example, an infrared emitter, an infrared laser emitter, a laser emitter, an ultraviolet emitter, a visible light emitter, or the like can be used.

Ideally, the light emitting device should use an infrared laser emitter that does not require a reflector to be attached to the rim.

The infrared laser emitter is suitable for driving a laser with a laser wavelength of 800 to 850 nanometers at 0.3 to 1 milliwatts.

The light emitter allows the light from the infrared emitter to pass through the line lens so that it can spread in a flat plane.

The light emitter may further include a line lens in the light emitting portion so that the light emitted from the light emitting element can constitute the light plane.

The line lens is positioned so that the light emitted from the light emitting element may form a light plane substantially parallel to the character input unit.

The pointer position information input device may further configure the optical filter 504 to transmit light of the light emitter only through light of a specific wavelength band.

The optical filter 504 is installed so as to allow light emitted from the light emitter to pass therethrough and to prevent the light emitted from the light emitter from being received when light reflected or blocked by the obstacle is received by the camera Can also be installed.

The light emitter and the optical filter for the camera may be separately configured, but it is preferable to design the light emitter and the camera in close proximity so that the light transmitted to one optical filter and the received light can be filtered .

The optical filter may be arranged to be perpendicular to the light traveling direction of the light emitter.

The optical filter may be installed in a range of 30 to 60 degrees or 120 to 150 degrees in the traveling direction of the light emitter in consideration of design.

The sensor module 501 may be installed such that the light of the light emitter is radiated directly to the light plane.

The sensor module 501 may further include one reflector 503 to refract the light of the light emitter once and radiate to the light plane.

At this time, light blocked or reflected by the obstacle may be refracted once by the reflection plate 503 and received by the camera.

The sensor module 501 may include two reflectors 503a and 503b to refract the light of the light emitter twice and radiate to the light plane.

At this time, light blocked or reflected by the obstacle can be refracted twice by the two reflectors 503a and 503b and received by the camera.

In this way, the sensor module can be installed inside the first housing 301.

At this time, an optical filter may be coated on the reflection plate 503 to replace the optical filter 504.

When the pointing position information input area 108 is activated by the sensor module 501, the user moves his / her finger over the plurality of text input buttons 303 to enter pointing position information.

In this case, the interval between the text input buttons 303 is ideally 0.1 millimeter to 1 millimeter so that the movement of the finger is not disturbed due to the interval of the text input button 303 of the user's finger.

In addition, the plurality of text input buttons 303 positioned below the pointing location information input region 108 may be positioned on a plane where the top surfaces of the plurality of text input buttons are flat.

A light source, which is emitted from the sensor module 501, is reflected by a reflector or an obstacle and is received by the camera, passes through the light tunnel 502.

At this time, if the upper surface of the light tunnel is replaced with a portion of the first housing 301 can reduce the thickness of the housing.

The lower surface and the side surface of the light tunnel can be designed as a separate structure from the first housing.

6 is a configuration diagram of an infrared laser sensor module.

The infrared sensor module basically emits light covering the pointing position information area using the light source 601 and analyzes the light reflected or blocked by the obstacle to detect the position of the obstacle by one or two cameras 603 It is the principle of calculation.

At this time, a line lens 602 that scatters light so that the light emitted from the light source 601 can be diverted to a desired area can be used.

It is preferable to use a line camera for the camera to recognize the state where the light source is reflected or blocked by the obstacle.

The camera may install a guide light tunnel in front of the camera to help the incoming light to receive light in a desired area.

In the present invention, the light source may be an infrared ray, an ultraviolet ray, a laser, or the like.

Light sources and cameras can be replaced by ultrasonic and ultrasonic receivers.

In the present invention, it is preferable to use an infrared laser as a light source in order not to provide a reflector on the edge of the keyboard.

If you want to use infrared light, you can install a reflector on the keyboard's edge, or you can increase the light output more than twice as much as an optical touch module for a regular screen, or install an auxiliary light source.

The sensor module may have an auxiliary control board 604 that controls the light source and the camera and transmits a received signal of the camera to the main control board.

The pointer position information input unit may be positioned to include a part of the extended surface of the text input unit plane.

Wherein the text input unit comprises a plurality of buttons for receiving at least one text and the button moves from a first position to a second position by a pressure applied by a user to generate a text selection signal, Lt; RTI ID = 0.0 > a < / RTI > physical button.

The text input unit includes F4 and F5 buttons on a Windows standard keyboard, and two cameras of the pointer position information input unit are provided. The camera on the left of the two cameras and the text input unit are provided. One of the buttons can be designed to be the F4 or F5 button at the closest distance.

Alternatively, the camera on the left side may be positioned below the space bar of the Windows standard keyboard, above the zero space where the space between the F4 and F5 buttons is moved downward.

In this way, when the position of the index finger of the right hand is at the position of the F (F) of the text input button, it is possible to reduce the phenomenon that the index finger of the left hand obscures the camera when the pointer position information is inputted with the right hand.

If the pointer execution command unit is set to a button having the largest area of the text input unit, for example, a space bar, it is convenient to execute the pointer execution command unit with the left hand while inputting the mouse position information with the right hand, for example.

At this time, since the left hand may be operating the mode switching unit, it is preferable that the shortest distance between the mode switching unit and the pointer execution command unit is 8 to 15 centimeters.

The mode switching unit may be provided separately from the button of the text input unit, and recognizes that a part of the user's body is touched on a part of the human interface body, and switches the mode to the text input mode and the pointer position information input mode. It operates in the pointer position information mode and can operate in the text input mode when it is not in touch.

The pointer execution command unit is composed of two buttons and performs a first function when the first button is pressed, for example, a left click of the mouse, and a second function when the second button is pressed, for example, And the first button is a button having the largest area of the text input unit buttons, for example, a space bar button. When the second button is set as the first button located on the left or right of the first button, It's easy to control with the thumb of the thumb and provides an easy user experience.

When the mode switching unit is switched to the text input mode or the pointer position information input mode, the text input mode or the pointer position information input mode can be displayed.

Wherein the pointer position information input mode is displayed by temporarily displaying the pointer position information input mode at a moment when the touch of the mode switching unit occurs or until the touch point of the mode switching unit is released, lighting signal.

By doing so, the user can recognize which mode the keyboard is currently in, and can easily know at which position the pointer position information should be input in the pointer position information input mode.

The text input area of the text input unit may be the same as the pointer location information input area of the pointer location information input unit or may include the pointer location information input area of the pointer location information input unit as shown in FIG.

In an embodiment constituting the pointer position information input unit, one or two optical emitters and a camera may be used.

The optical signal received by the camera may pass through a light tunnel to block an external noise source.

The light tunnel may be configured before the optical signal is emitted from the optical emitter to the pointer location information input area, and may be installed on the front surface of the camera receiving the user input signal from the pointer location information input area.

At this time, an optical signal emitted from the optical emitter or a reflector for refracting an optical signal received by the camera can be disposed.

A degree of freedom is provided to the installation position and direction of the pointer position information input unit through the reflection plate, thereby achieving space utilization and design benefit in design.

The pointer position information input unit may configure the pointer position information input area at least 0.1 to 5 millimeters above the text input unit by the optical emitter and the camera.

Thus, the pointer position information can be input only when there is a clear touch intention of the user.

When the optical emitter is designed to operate at 0.3 to 1 milliwatt using an infrared laser having a wavelength of 800 to 850 nanometers, it exhibits excellent power consumption, cost, stability, safety, and visual characteristics.

The optical emitter may further include a line lens that scatters the optical signal so that the optical signal constitutes a pointer position information input area parallel to the text input part.

The pointer position information input unit may be designed to block optical external noise using an optical filter that transmits only the optical signal in the infrared region.

If the interval between the physical buttons is designed to be 0.1 millimeter to 1 millimeter, the text input unit can provide the user with a feeling of sufficiently smooth and flat feeling during the input of the pointer position information.

Fig. 7 shows an embodiment for displaying the pointer position information input area.

The human interface apparatus of the present invention may be designed to further include pointer position information input area display units 701 and 702 for visually displaying the pointer position information input area in the pointer position information input mode.

The mode switching unit is provided separately from the button of the text input unit and recognizes that a part of the user's body is touched by a part of the complex human interface housing to switch to the text input mode and the pointer position information input mode, but when the touch is performed, the pointer is touched. It can be designed to operate in location information input mode and to operate in text input mode during non-touch.

At this time, when a part of the body used for the mode switching is the right hand (106b), the hand which performs the pointer position information input 108a can be designed to be a left hand.

In this case, it is desirable that the pointer position information input region is designed to be closer to the opposite corner of the corner where the mode switching unit is located, and the same principle can be also applied when the left and right hands are reversed. (106a, 108b)

On the other hand, a part of the body used for the mode switching may be a right hand, and a hand for inputting the pointer position information may be designed to be a right hand.

In this case, it is preferable that the pointer position information input area is designed to be closer to the corner where the mode switching part is located than the opposite corner of the corner where the mode switching part is located, and the left hand can be designed on the same principle.

The mode switching unit may determine the text input mode and the pointer location information input mode based on the number or positions of user fingers recognized by the pointer location information input unit.

The number of the recognized user fingers when determining the pointer position information input mode is smaller than the number of the recognized user fingers when determining the text input mode.

For example, while controlling the mode switching unit with the left hand, the finger of the left hand may be untouched from the text input button, and the index of the right hand may be touched for inputting the pointer position information, Is one.

On the other hand, when the finger of the left hand and the right hand are touched on the keyboard for text input, the number of the fingers of two to eight recognized users is recognized.

The pointer position information input mode display unit displays the pointer position information input area from the point of time when the mode switching unit is switched to the pointer position information input mode to the time point when the mode is switched to the pointer position information input mode, It can be displayed visually.

The pointer position information input mode display unit may be displayed by the visible light generator from the text input button or the reflected light by the text input button or through the space between the text input buttons.

At this time, the input mode may be displayed even in the text input mode, and the text input mode and the pointer location information input mode may be displayed to be distinguished.

The pointer position information input mode display unit includes invisible light generators 701a and 701b or surfaces 702a and 702b coated with dyes that optically respond to the invisible light and emit visible light, and the dye is the text. It can be applied to the interval between the input button or the text input button.

The mode switching unit may further include a hand side determining unit for determining whether the user inputs the pointer position information with a left hand or a right hand.

The display of the pointer location information input display unit may be changed according to the handside determination unit.

It is possible to flexibly switch the button arrangement and operation of the pointer execution command unit according to the hand side judgment unit.

For example, when the pointer execution command includes a right-click and a left-click, the hand-side determination unit may switch the button to perform the right-click and left-click operations.

The pointer position information input area display unit may display (702a, 702b) different pointer position information input areas according to the determination result of the hand side determination unit.

The mode switching unit may be designed to automatically switch to the text input mode when the pointer position information input from the user is not input for a predetermined time in the pointer position information input mode or when the input is received by the text input button.

The pointer position information input area display unit may apply a dye to a part of the text input button or the text input area so that the pointer input area is visually displayed irrespective of the mode of the mode switching unit so that the pointer input area can be recognized even in the text input mode .

8 is a first embodiment of a pointer position information input device which can be separated from a text input device.

The pointer location information input device 810 of the present invention is designed to be separated from the text input device 800 and may be combined 801 or separated from the text input device 802.

At this time, the power connection unit 820 included in the pointer location information input device can receive power from the power source of the text input device 800.

The pointer position information input device 801 includes at least one pointer position information input device 105a and 105b and may further include a control unit and an optical filter 504. [

The detachable pointer location information input device 810 may be combined 801 to constitute a pointing location information input area on a plane parallel to the text input area and may be coupled to the left 803 or right 804 of the text input device And the pointing position information input area can be configured on the bottom surface of the text input device.

At this time, the pointing position information input device and the power connection part 820 are designed to be flexibly deformed according to the positional change of the pointing position information input device.

9 is a second embodiment of a pointer position information input device which can be separated from the text input device.

The pointer location information input device 900 includes a built-in power supply to receive power from an external or its own battery.

When the pointer position information input device 900 is in the first position information input mode 902, the pointer position information input area is parallel to the text input area of the text input device, The housing can be designed.

Alternatively, the apparatus may further include a first position information input mode adjusting device that adjusts the height and tilt of the housing so that the user may arbitrarily adjust the height and angle of the position information input area.

  When the pointer position information input device 900 is in the second position information input mode 904, the pointer position information input area is input with pointer position information at a predetermined height and angle so that the pointer position information input device 900 is parallel to the bottom surface on which the pointer position information input device 900 is placed The housing of the device 900 can be designed.

At this time, preferably, the height of the optical filter 504 in the first position information input mode is higher than the height of the optical filter 504 in the second position information input mode.

When the pointer location information input device 900 is in the third location information input mode 906, the pointer location information input area is formed to have a constant height so as to include an area perpendicular to the bottom surface on which the pointer location information input device 900 is placed. The housing of the pointer position information input device 900 can be designed at an angle with and.

The pointer position information input device 900 may include a position information input mode, a second position information input mode, and a third position information input mode so as to distinguish at least two of the first position information input mode, And a sensor unit.

The location information input mode sensor unit may include a sensor utilizing gravity, a sensor detecting a surface placed on the floor, or a switch.

The pointer location information input apparatus 900 may have at least one mode of the first to third location information input modes.

In particular, in the third location information input mode 906, mode switching for switching between the text input mode and the pointer location information input mode may be replaced by whether the user's hand or finger is detected in the location information input apparatus 900. Can be.

That is, when the user is inputting text, the user's hand or finger is not detected by the location information input device 900 in the third location information input mode.

When the user places a hand or finger on the top of the location information input device in the third location information input mode 906 to input the location information of the pointer, the location information input device 900 inputs the pointer location information. Switch to the mode.

Alternatively, a text input mode based on a state in which a digital device connected to the text input unit 800 and the location information input device 900 by wire or wirelessly detects the user's hand or finger. And pointer location information input mode can be switched.

Alternatively, the text input unit 800 to which the location information input device 900 is connected by wire or wirelessly may receive mode information of the location information input device 900 to change modes or the digital device may change modes. Information can also be sent.

At this time, when the text input is in progress, the location information input apparatus 900 may be switched to the first power saving mode which consumes less power than in the pointer location information input mode.

After the text input is interrupted, the first power saving mode may be switched from the first power saving mode to the active mode during the first first interval.

When the user's hand or finger is not detected during the first first interval after the text input is interrupted, the user may enter the second power saving mode which consumes less power than the first power saving mode.

In the first power saving mode, the sensor may be periodically controlled with less power than the active mode to determine the presence of the user's hand or finger.

When the user's hand or finger is detected in the first power saving mode, the user may switch to the activation mode.

 In the activation mode, it is possible to switch to the pointer position information input mode.

The pointer execution command unit for a pointer execution command in the pointer position information input mode may generate a pointer execution command using a button provided in the text input device or based on a position or shape change of the detected user's hand or finger. Can be.

The location information input device 900 may be activated by a mode switching unit on the text input device 800.

10 is a third embodiment of a pointer position information input device which can be separated from a text input device.

The pointer location information input apparatus 1000 may have a housing 1001 that encloses at least two sides of the text input apparatus so that the text input apparatus 800 can be mounted.

The pointer location information input device 1000 includes a pointer location information input device 105a, 105b, a mode switching unit 106, a controller 1001, a housing 1002, and an optical filter 907 (1010).

After the text input device 800 is mounted on the pointer location information input device 1000 (1020), an optical filter 907 may be used to configure a pointing location information input area parallel to an upper end of the text input device 800 input area. The optical filter is positioned higher than the text input device 800 so that the light source can pass therethrough.

At least one surface 1003 surrounding the text input device of the housing 1002 may be a surface such that a light source radiated from the pointer location information input device 105 may be reflected, transmitted, or absorbed according to the type of the pointer location information input device. Each material can be treated with a mirror, transparent glass or plastic, or a black light absorber.

8, 9 and 10 illustrate the following components in an embodiment of a pointing-position-information inputting apparatus that can be detached from a text inputting apparatus.

And a pointer location information input unit for receiving information related to the pointer location from the user.

The pointer position information input unit receives the pointing position information of the user input through the pointing position information input area formed by the pointer position information input device.

And a pointer execution command receiving unit that receives a signal of a pointer execution command unit that receives a user command to perform at least one function in a place where the pointer is located.

The pointer execution command unit may be constituted by a button or a touch switch included in the pointer position information input unit or a control unit for switching to a pointer execution instruction when a predetermined input of the user input received from the text input device is received.

And a mode switching instruction receiving unit for receiving a signal of a mode switching unit for switching to a pointer position information input mode.

The mode switching unit may be configured in a pointing position information input device, or configured in a text input device, or may be configured as a separate device.

And a power unit for transmitting power to the pointer position information input unit, the pointer execution command receiving unit, and the mode switching command receiving unit.

The power source unit may be a power source unit having a self-powered battery or a power source supplied from the outside.

And a pointer position information transmitter for transmitting information related to the pointer position input to the pointer position information input unit to a digital device connected to the human interface device by wire or wirelessly.

Wherein the pointer position information input area of the pointer position information input unit is positioned at a predetermined angle with the predetermined height so that the pointer position information input area can be positioned to be parallel to the upper end of at least a part of a text input area of a text input device composed of a plurality of physical buttons Is set.

Wherein the mode switching unit can be fixed to at least one side of the text input device when the mode switching unit is provided separately from the text input button of the text input device and recognizes that the user input is received in the mode switching unit, And can be transmitted to the mode switching command receiving unit by wire or wirelessly.

The pointer position information input unit may include an optical emitter and a camera, and the optical signal received by the camera may be designed to pass through an optical filter through a light tunnel.

A position state recognition unit for recognizing the position state information about the gravity direction of the human interface or the position state information about the floor on which the human interface is placed or the position information of the human interface through the input of the user.

The position state is composed of at least two kinds, and when the first position state is selected, the mode is switched to the first pointer position information input mode and when the second position state is entered, the mode is switched to the second pointer position information input mode

And the pointer position information input area in the first position state is located in parallel with at least a part of the text input area of the text input device.

The pointer position information input area in the second position state is located so as to be parallel to the floor surface on which the pointing position information input device is placed.

The upper surface of the pointing position information input device in the first pointer position information input mode is positioned to face the bottom surface. When switching to the second pointer position information input mode, the calculation of the ordinate of the pointing position information input is the same, Calculate by changing direction.

The pointer execution command unit may use a space bar among text input buttons of the text input device and may further use another button adjacent to the space bar as a pointer execution command unit.

The power supply unit may receive power from the text input device.

The predefined height and angle of the pointer location input area in the first positional state may be defined by a controller that allows adjustment within a predefined range by the user.

The adjuster can adjust the angle by adjusting the joining angle of the at least one surface of the housing of the pointing position information input device with the adjacent surface.

The controller adjusts the height of at least one surface of the housing of the pointing position information input device.

The controller adjusts the height and angle of the pointing position information input unit by adjusting the height and angle of the optical module of the pointing position information input unit.

11 is a front view of a human interface device to which a lower-positioned pointer positional information input device is applied.

The human interface apparatus includes a text input unit configured by a plurality of physical buttons, a pointer position information input unit for inputting information related to a pointer position from a user, a pointer execution command for inputting a user command to perform at least one function, A mode switching instruction receiving unit for receiving a signal of a mode switching unit for switching to a pointer position information input mode; a power supply unit for supplying power to the pointer position information input unit, the pointer execution command receiving unit, and the mode switching instruction receiving unit; A pointer position information transmitting unit for transmitting information related to a pointer position input to the pointer position information inputting unit to a digital device connected to the human interface device by wire or wirelessly, The information input area is positioned so as to be parallel to the upper end of at least a part of the text input area of the text input part composed of the plurality of physical buttons and the pointer position information input part comprises at least two sensor modules, And the module is located on the lower left 1102 and the right 1103 of the human interface device.

Wherein the pointer position information input unit comprises an optical emitter and a camera, and the optical signal received by the camera passes through a light tunnel.

And the mode switching unit is provided separately from the text input button of the text input device.

The mode switching unit is located on the left side of the text input area of the text input device and recognizes that the user input is received in the mode switching unit and transmits the pointer position information input mode to the mode switching instruction receiving unit.

The physical button is composed of an upper portion of a button receiving pressure from the user's finger and an elastic member applying a physical force to return the upper portion of the button to the original position when the user's finger pressure is removed.

The complex human interface device may include a reflector or an absorber 1101 that reflects a light source generated by the pointer location information input unit at left, right, and top sides of the text input area, but not at least partially below. It is done.

The reflector is used when the light source generated from the pointer position information input unit is an infrared ray, and when an infrared ray laser is used, an absorber is used.

The at least two sensor modules 1102 and 1103 of the pointer location information input unit are located at least partially in the left outer area of the text input area and the lower right area of the right outer area. That is, the left sensor includes at least a part of the left side of the control button of the standard window keyboard and is further lowered below the control button, that is, toward the housing corner without the key button. In this way, the sensor module enables the configuration of a minimal volume composite human interface device to maximize the aesthetic effect of the composite human interface.

Also, by orienting the sensor in the diagonal direction of the text input area by locating the sensor near the vertex of the text input area (1110), it is better to position the sensor in the corner area and face the transverse direction of the text input area The pointer position information input area having a wider area can be constituted. In addition, it is possible to maximize the pointer position information input area at the upper part of the text input area by moving the position of the portion where the position information detection accuracy of the pointer position information input area between the sensor and the sensor is relatively low.

In addition, the position of the sensor can be positioned 1120 below the space bar of the standard keyboard to be narrower than the width of the keyboard. The width between the sensor and the sensor determines the width of the pointer position information input area and thereby obtains relatively high pointer position information descriptive resolution although the area of the pointer position information input area is reduced rather than positioned at the corner (1110).

Between the sensor and the sensor, the reflector or absorber is removed, and the height of the housing is lowered to prevent the thumb from being caught by the housing when the user presses the space bar of the standard keyboard with his / her thumb.

12 is an embodiment of the pointer execution command unit integral mode switching unit.

When the input operation in which the text input mode and the pointer position information input mode are frequently changed is performed, the number of times the mode switching and the pointer execution command are input is increased.

For example, in case of right-handed person, when you want to input pointer position information while text input mode is being performed in text input mode, you need to input the pointer position information after inputting to pointer real name command with left hand after switching mode with left or right hand do.

In general, while performing a quick operation, the user is required to input text and pointer position information and a pointer execution command while keeping a gaze on the monitor. If a mode switching unit and a pointer execution command input unit are provided separately, hand movement frequently occurs It is not possible to find the exact position, so it may not be able to perform properly, or it may be troublesome to look at the keyboard with eyes.

In order to solve this problem, for example, in the case of a right-handed person, when a hand is placed on the mode switching unit with the left hand after inputting text in the text input mode, a mode switching command is generated and the mode is changed to the pointer input mode. The pointer execution command unit may receive the user input by inputting the pointer position information with the right hand with the hand on the mode switching unit and applying pressure to the mode switching unit.

Thereby, the user can share the position of the mode switching unit for mode switching and the position of the pointer execution command unit for the pointer execution command.

For example, the mode switching unit may include a touch switch capable of accepting a touch input of a user's hand, and the pointer execution command unit may be disposed under the mode switching unit, which includes a switch responsive to a pressure such as a tact switch.

In this case, the first pointer execution command unit 1104 and the second pointer execution command unit 1105 are connected to each other by a conductor material capable of recognizing a touch from the user's hand or commonly connected to the mode switching unit 1201 So that the user can perform the same mode switching regardless of whether the user touches the first pointer execution command unit or touches the second pointer execution command unit.

At this time, the mode switching is operated in the pointer position information input mode while the touch is being performed, and in the text input mode when the touch is released.

The mode switching by the touch may be composed of another switch operating at a pressure smaller than the pressure for executing the pointer execution command, or a sensor capable of sensing the position of the user's finger.

In this case, the mode switching and pointer execution command unit may be located in an area distinct from the text input area.

Preferably, in the case of right handedness, it is located in the left outer area of the text input area.

If the mode switching unit and the pointer execution command unit are located outside the text input area, the possibility that the pointer position information input unit is confused with the pointer position information input disappears and the pointer position information input area can be expanded.

The hybrid human interface apparatus of the present invention may include a second mode switching unit 1202. The second mode switching unit switches to a text input mode and a pointer position information input mode each time an input is received from a user with a switch operating as a toggle switch. Thus, even if the user does not touch the mode switching unit with the left hand, the pointer position information can be input with only the right hand.

At this time, when the compound human interface apparatus is in the text mode by the second mode switching unit, when the mode switching command by the touch is received, the compound human interface apparatus of the present invention operates in the pointer position information input mode. In addition, when the compound human interface apparatus by the second mode switching unit is in the pointer position information input mode, when the mode switching command by touch is received, the pointer position information input mode is maintained but the mode switching command by touch is released When a text input is received by the text input unit, the mode is switched to the text input mode.

In this case, when the text input mode is switched by the text input, at least the first text input may be ignored, and when at least two text inputs are received, the text input mode may be switched.

In this case, when at least two or more text inputs are received and the mode is switched to the text input mode, the composite human interface device transmits the text input including at least the first text, which has been ignored, to the digital device, Lt; / RTI >

13 is an embodiment utilizing a lid of a composite human interface device.

The composite human interface device requires space of at least 1 millimeter to 2 millimeters on the top of the text input device plane in which the pointer position information input device is constituted by the physical device to form the pointer position information input area, Or a border for locating the reflector may be required at the edges of at least three sides surrounding the text entry area.

At this time, a difference of at least 1 to 2 millimeters occurs between the height of the edge of the at least three sides of the edge and the height of the text input area plane, and a substantially rectangular parallelepiped type space is formed at the top of the text input area.

The complex human interface apparatus of the present invention may further include a multipurpose cover 1310 for protecting the text input area from an external impact in the space of the substantially rectangular parallelepiped type.

The multi-purpose lid 1322 is separated from the composite human interface device 1320 and can be coupled and separated by a magnet or a physical structure.

The multipurpose cover 1322 may be folded several times. Preferably, the width of at least one area of the divided area of the multipurpose cover may be less than the width of the other area (1341).

This can further reduce the slope of the composite human interface device when the multipurpose cover is folded down and then mounted to adjust the inclination of the composite human interface device.

When the multipurpose cover is folded and then mounted on the bottom surface of the composite human interface device, a part of the folded surface that is in contact with the ground surface may be provided with an inclined surface 1342 so that an area contacting with the ground surface can be widened.

The slope of the composite human interface device desired by the user can be adjusted by the number of times of folding of the multipurpose cover 1322.

The multi-purpose lid may have a built-in rechargeable battery 1323 inside the lid.

An electrode of a power source of the rechargeable battery 1323 may be positioned at a lower end of the multipurpose cover, and may be located at a portion of the protruding region rather than a portion covering the text input area (1331) and the multipurpose cover may cover the text input device. If present (1310) is connected to the electrode provided on the top of the composite human interface device.

Also, when the multipurpose lid is folded and mounted on the lower end to adjust the inclination of the composite human interface device, it can be connected to the electrode provided on the bottom surface of the composite human interface device.

The composite human interface device may be connected to an external power source to supply power to the rechargeable battery 1323 inside the multipurpose lid or separately charge the multipurpose lid.

At this time, the bottom surface of the composite human interface device has a region 1332 covering a protruding edge area where the electrode 1331 is located and a text input area having a height difference, and a magnet or a physical coupling device And a groove into which the cover is inserted.

The rim of the multi-purpose lid is coated with a substance that can remove minute dust, oil, and moisture, and can remove foreign substances from the absorption plate, the reflector, the optical emitter, and the front surface of the camera have.

If the versatile cover covers the text input area, it can detect the text input area and turn off the power of the composite human interface device.

When the multipurpose cover is mounted on the bottom of the composite human interface device, the composite human interface device may be turned on.

Also, if there is no input by the user for a certain period of time after the power is turned on, the power can be cut off or the apparatus can be switched to the standby mode.

When the multipurpose cover covers the text input area (1310) The composite human interface device 1321 has a substantially rectangular parallelepiped shape and has no inclination when placed on the ground, or when the multipurpose cover is mounted at the bottom. Designed to be smaller, it can maximize aesthetic effect and increase portability.

On the other hand, if the multipurpose cover is folded and mounted on the bottom of a composite human interface device, the inclination can be secured like a general keyboard.

The multipurpose lid can be constructed in a removable manner, but it can be designed in such a way that it is folded down from the upper surface to the lower surface of the composite human interface device by the hinge.

14 is an embodiment of a human interface device applied to a portable electronic device such as a portable electronic device or a tablet personal computer.

The composite human interface device 1420 of the present invention can be utilized as an input device of a portable electronic device.

The portable electronic device having the composite human interface apparatus of the present invention is configured such that the display unit composed of the display panel 1411 and the frame 1410 supporting the display panel is bound by the composite human interface apparatus of the present invention and the hinge 1427 And can be designed to be covered and opened.

At this time, the display unit is inserted into the text input area of the human interface device of the present invention and the substantial hexahedral space formed by the reflection plate or absorbing plate 1424. Accordingly, the thickness of the portable electronic device using the human interface device of the present invention can be minimized.

To this end, the display unit of the portable electronic device should be designed to have a width smaller than the width of the human interface device 1420 of the present invention by at least twice as wide as the thickness of the reflection plate or absorption plate 1423, 1424.

In other words, in Fig. 13, the display unit must be designed so that the multipurpose cover can be inserted into the place.

The display unit has a feature 1413 that allows both corners to be rounded or cut into oblique lines.

When the display unit is folded, the pointer location information input devices 1421 and 1422 may be positioned toward the outer edge of the rounded or diagonally cut portion.

This is to fold the portable electronic device so that it can be kept thin without being disturbed by the pointer position information input devices 1421 and 1422.

At this time, the mode switching unit and the pointer execution command unit are located outside the side text input area of the human interface device as described above, the mode switching unit is outside the text input area, and the pointer execution command unit is at least one of the text input buttons, or the mode switching. Both the negative and pointer command execution units may operate as at least one of the text input buttons.

(1425, 1426), and more preferably, can be located outside the text input area.

At this time, it is preferable that the mode switching unit and the pointer execution command unit are integrally formed as shown in FIG. 12, but they may be separately disposed.

At this time, when the mode switching unit or the pointer execution command unit is located outside the text input area, when the user places a finger to control the mode switching unit or the pointer execution command unit, the pointer location information input devices 1421 and 1422 The pointer location information input area should be set except the area where the mode switching unit or the pointer execution command unit is located so that the user does not malfunction by inputting the location information of the pointer.

For this purpose, it is possible to design an optical signal to be prevented from reaching the mode switching unit or the pointer execution command unit by adjusting an area covered by the optical signal generated from the pointer position information input device.

Or an optical signal receiving angle of the camera is adjusted so that the camera does not receive an optical signal generated by reflection, interference, or blocking between the user's finger and the optical signal located in the mode switching unit or the pointer execution command unit, The optical signal can be ignored.

The portable electronic device is designed to have at least three corner portions of an absorbing plate or a reflecting plate for absorbing or reflecting a light source generated by the pointer position information input device.

At this time, the two surfaces are located on the side of the human interface device (1423, 1424) and the other surface is the pointer position information input device (1421) when the display unit is open within a predetermined angle on the surface where the display unit is in contact with the human interface device. The reflector or absorber may be positioned in a constant region 1412 of the edge of the display unit so that the light source generated by the 1422 is sufficiently reflected or absorbed.

In FIG. 14, the display unit may be configured as a flexible display unit. The display unit may be curved to cover the text input area of the complex human interface device and be wound around the bottom surface of the complex human interface device.

In addition, in FIG. 14, the display unit may be replaced with the multipurpose cover described in FIG. 13, and the absorber or reflector may be positioned at a corner portion of the multipurpose cover bonded to the composite human interface device, and flipped to the rear by a hinge to allow the composite human It can also be used to adjust the height of the interface device.

1 to 14 illustrate an example of a complex human interface device to be described in the present invention. The type of text input device pointing position information input device and the technology used can be changed or replaced without changing the basic purpose by a person skilled in the art .

Claims (18)

A composite human interface device,
A pointer position information input unit for receiving information related to the pointer position from the user,
A text input unit configured by a plurality of physical buttons,
A first mode switching unit for activating or deactivating the pointer position information input mode,
And a pointer execution command receiving unit for executing at least one function at a position of the pointer,
Wherein the pointer position information input unit is located at an upper end of the text input unit, the pointer execution command receiving unit is located at a lower end of the text input unit, and the pointer position information input area formed by the pointer position information input unit is a text input area And a plane perpendicular to the first plane.
The apparatus of claim 1, wherein the first mode switching unit is operated by a first input of a user input to a first button and the pointer execution command receiving unit is operated by a second input of a user input to the first button A composite human interface device characterized by. The apparatus according to claim 1, further comprising a power source unit for controlling power of the pointer location information input apparatus, wherein the pointer location information input apparatus includes an activation mode for receiving pointer location information, A first power saving mode and a second power saving mode for minimizing power consumption, and the power consumption is reduced in the order of an activation mode and a first power saving mode second power saving mode. The text input device according to claim 1, wherein the pointer location information input device and the text input device are separate devices and communicate wirelessly, and the text input device includes the pointer execution command part. Interface device. The text input device according to claim 4, wherein the text input device wirelessly communicates with a digital device having a display unit to transmit text input information, and pointer position information input to the pointer location information input device is transmitted to the digital device Or is transferred from the pointer location information input device to the digital device and the mode switching information provided in the text input device is transmitted to the pointer location information input device through the digital device or from the text input device to the pointer location And the information is transmitted to an information input device. The composite human interface apparatus of claim 1, wherein the first mode switching unit activates a pointer position information input mode when a user's hand or finger is detected. The compound human interface apparatus according to claim 1, further comprising a communication unit for transmitting an input of a user input from the text input apparatus or the pointer position information input apparatus to an external digital device. [2] The apparatus of claim 1, wherein the pointer position information input unit includes at least two sensor modules including an optical signal emitter and an optical signal detector, Wherein the first and second positions are narrowly spaced from each other. The apparatus of claim 1, wherein the first input of the user is a switch operating at a pressure less than a pressure for inputting a pointer execution command or a switch for recognizing a touch of a user, And a switch. 2. The apparatus of claim 1, wherein the first mode switching unit operates in a pointer position information input mode while the first input is maintained and releases the pointer position information input mode when the first input of the user is released Complex human interface device. The hybrid human interface apparatus according to claim 1, further comprising a second mode switching unit for switching the pointer position information input mode from the operation to the release or release dynamically each time the user's input is received once. 12. The method of claim 11, wherein when the input is received and then released from the first mode switching unit during operation in the pointer position information input mode by the second mode switching unit, the pointer position information input mode is released. Interface device. 12. The compound human interface apparatus according to claim 11, wherein when the text input is received in the pointer position information input mode by the second mode switching unit, the pointer position information input mode is canceled. The method as claimed in claim 13, wherein if the text entry is received while the second mode switching unit is operated in the pointer position information input mode, at least the first text entry is ignored and if at least two text entries are received, To the human interface device. 15. The method of claim 14, wherein when the at least two text inputs are received and switched to the text input mode, the composite human interface device transmits the text input of the user including the ignored at least first text to the display unit, And transmits the text input information to the display unit. The method as claimed in claim 11, wherein the second mode is switched by the third input of the user inputted to the second button, and the pointer execution command is received by the fourth input of the user inputted to the second button. Device. 2. The compound human interface apparatus of claim 1, wherein the user second input is a switch operated by a pressing pressure. Receiving information related to a pointer position from a pointer position information input unit from a user,
Receiving text input from a text input comprising a plurality of physical buttons,
Activating or deactivating the pointer position information input mode from the first mode switching unit,
Receiving a pointer execution command for executing at least one function at a location of a pointer from a pointer execution command receiving unit,
Wherein the pointer position information input unit is located at the upper end of the text input unit, the first mode switching unit and the pointer execution command receiving unit are located at the lower end of the text input unit, and the pointer position information input area formed by the pointer position information input unit And a plane perpendicular to the text input area formed in the text input unit.
KR1020120152644A 2012-05-28 2012-12-24 Multi human interface devide having poiner location information input area including vertical area to text input area KR20130133117A (en)

Priority Applications (7)

Application Number Priority Date Filing Date Title
US14/403,998 US9612666B2 (en) 2012-05-28 2012-12-29 Human interface apparatus having input unit for pointer location information and pointer command execution unit
PCT/KR2012/011822 WO2013180369A1 (en) 2012-05-28 2012-12-29 Human interface apparatus having input unit for pointer location information and pointer command execution unit
US14/718,107 US9612667B2 (en) 2012-05-28 2015-05-21 Human interface apparatus having input unit for pointer location information and pointer command execution unit
US14/718,108 US20150253869A1 (en) 2012-05-28 2015-05-21 Human interface apparatus having input unit for pointer location information and pointer command execution unit
US14/718,105 US9639173B2 (en) 2012-05-28 2015-05-21 Human interface apparatus having input unit for pointer location information and pointer command execution unit
US14/794,825 US9612668B2 (en) 2012-05-28 2015-07-09 Human interface apparatus having input unit for pointer location information and pointer command execution unit
US15/475,045 US9880637B2 (en) 2012-05-28 2017-03-30 Human interface apparatus having input unit for pointer location information and pointer command execution unit

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
KR1020120056427 2012-05-28
KR1020120056427A KR20130053363A (en) 2011-11-15 2012-05-28 Apparatus and method for multi human interface devide

Related Child Applications (1)

Application Number Title Priority Date Filing Date
KR1020130148173A Division KR20140082920A (en) 2013-12-02 2013-12-02 Multi human interface devide having poiner location information input area including vertical area to text input area

Publications (1)

Publication Number Publication Date
KR20130133117A true KR20130133117A (en) 2013-12-06

Family

ID=48699448

Family Applications (2)

Application Number Title Priority Date Filing Date
KR1020120152644A KR20130133117A (en) 2012-05-28 2012-12-24 Multi human interface devide having poiner location information input area including vertical area to text input area
KR1020120156853A KR101268209B1 (en) 2012-05-28 2012-12-28 Multi human interface devide having poiner location information input device and pointer excution device

Family Applications After (1)

Application Number Title Priority Date Filing Date
KR1020120156853A KR101268209B1 (en) 2012-05-28 2012-12-28 Multi human interface devide having poiner location information input device and pointer excution device

Country Status (1)

Country Link
KR (2) KR20130133117A (en)

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9921661B2 (en) 2012-10-14 2018-03-20 Neonode Inc. Optical proximity sensor and associated user interface
US10282034B2 (en) 2012-10-14 2019-05-07 Neonode Inc. Touch sensitive curved and flexible displays
US10324565B2 (en) 2013-05-30 2019-06-18 Neonode Inc. Optical proximity sensor
US10585530B2 (en) 2014-09-23 2020-03-10 Neonode Inc. Optical proximity sensor
US11073948B2 (en) 2012-10-14 2021-07-27 Neonode Inc. Optical proximity sensors
US11842014B2 (en) 2019-12-31 2023-12-12 Neonode Inc. Contactless touch input system

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2003280791A (en) 2002-03-20 2003-10-02 Hitachi Ltd Keyboard with pointing device

Cited By (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9921661B2 (en) 2012-10-14 2018-03-20 Neonode Inc. Optical proximity sensor and associated user interface
US10282034B2 (en) 2012-10-14 2019-05-07 Neonode Inc. Touch sensitive curved and flexible displays
US10802601B2 (en) 2012-10-14 2020-10-13 Neonode Inc. Optical proximity sensor and associated user interface
US10928957B2 (en) 2012-10-14 2021-02-23 Neonode Inc. Optical proximity sensor
US10949027B2 (en) 2012-10-14 2021-03-16 Neonode Inc. Interactive virtual display
US11073948B2 (en) 2012-10-14 2021-07-27 Neonode Inc. Optical proximity sensors
US11379048B2 (en) 2012-10-14 2022-07-05 Neonode Inc. Contactless control panel
US11714509B2 (en) 2012-10-14 2023-08-01 Neonode Inc. Multi-plane reflective sensor
US11733808B2 (en) 2012-10-14 2023-08-22 Neonode, Inc. Object detector based on reflected light
US10324565B2 (en) 2013-05-30 2019-06-18 Neonode Inc. Optical proximity sensor
US10585530B2 (en) 2014-09-23 2020-03-10 Neonode Inc. Optical proximity sensor
US11842014B2 (en) 2019-12-31 2023-12-12 Neonode Inc. Contactless touch input system

Also Published As

Publication number Publication date
KR101268209B1 (en) 2013-05-29

Similar Documents

Publication Publication Date Title
US9880637B2 (en) Human interface apparatus having input unit for pointer location information and pointer command execution unit
KR102367253B1 (en) Electrical device having multi-functional human interface
KR102052752B1 (en) Multi human interface devide having text input unit and pointer location information input unit
KR101268209B1 (en) Multi human interface devide having poiner location information input device and pointer excution device
KR20140075651A (en) Multi human interface devide having display unit
KR20150032950A (en) Digital device having multi human interface devide
KR20150050546A (en) Multi functional human interface apparatus
KR102189744B1 (en) Multi human interface devide having text input unit and pointer location information input unit
KR20140082920A (en) Multi human interface devide having poiner location information input area including vertical area to text input area
KR20140063489A (en) Multi human interface devide having display unit
KR20140063488A (en) Multi human interface devide having display unit
KR20140063485A (en) Multi human interface devide having display unit
KR20140063484A (en) Multi human interface devide having display unit
KR20140063490A (en) Multi human interface devide having display unit
KR20140063486A (en) Multi human interface devide having display unit
KR20140063487A (en) Multi human interface devide having display unit
KR20140063483A (en) Multi human interface devide having display unit

Legal Events

Date Code Title Description
A107 Divisional application of patent
WITN Withdrawal due to no request for examination