KR20130133117A - Multi human interface devide having poiner location information input area including vertical area to text input area - Google Patents
Multi human interface devide having poiner location information input area including vertical area to text input area Download PDFInfo
- Publication number
- KR20130133117A KR20130133117A KR1020120152644A KR20120152644A KR20130133117A KR 20130133117 A KR20130133117 A KR 20130133117A KR 1020120152644 A KR1020120152644 A KR 1020120152644A KR 20120152644 A KR20120152644 A KR 20120152644A KR 20130133117 A KR20130133117 A KR 20130133117A
- Authority
- KR
- South Korea
- Prior art keywords
- pointer
- information input
- position information
- input
- mode
- Prior art date
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/02—Input arrangements using manually operated switches, e.g. using keyboards or dials
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/02—Input arrangements using manually operated switches, e.g. using keyboards or dials
- G06F3/023—Arrangements for converting discrete items of information into a coded form, e.g. arrangements for interpreting keyboard generated codes as alphanumeric codes, operand codes or instruction codes
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/033—Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/14—Digital output to display device ; Cooperation and interconnection of the display device with other functional units
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2203/00—Indexing scheme relating to G06F3/00 - G06F3/048
- G06F2203/038—Indexing scheme relating to G06F3/038
- G06F2203/0381—Multimodal input, i.e. interface arrangements enabling the user to issue commands by simultaneous use of input devices of different nature, e.g. voice plus gesture on digitizer
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- General Engineering & Computer Science (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Position Input By Displaying (AREA)
Abstract
Description
The present invention relates to a human interface for receiving text information or pointing position information from a user in a digital device capable of receiving text or pointing position related information such as a computer, a notebook, a tablet PC, a mobile phone, and the like, and transmitting the text information or the pointing position information to the digital device.
A text input device such as a keyboard for text input has been disclosed in personal computers and portable digital devices. Also, a pointing device such as a mouse device for controlling the pointing position of the pointer for controlling the digital device and performing the function has been disclosed.
The conventional text input device and the pointing device are provided as separate devices or the pointing input area is formed at a separate position different from the text input area of the text input device. Therefore, the movement of the user's hand is more than necessary in the work environment in which the text input, the pointing position information input, and the pointer execution command input are frequently changed, thereby lowering the work efficiency.
In the present invention, a pointing position information input area of the pointing device is provided on the upper side of the text input area of the text input device, and a switching part for switching the text input mode and the pointing position information input mode is provided. The user can perform the pointing input while minimizing the movement of the user's hand, thereby improving the work efficiency.
By integrating the separately provided text input device and the pointing device into one human interface device, it is possible to reduce the unit price and the size of the product, and eliminate the unnecessary operation of the user operation, thereby improving the work efficiency.
1 is an exemplary view of a composite human interface device having a pointing device and a text input device.
2 is a flowchart for explaining an example of the operation sequence according to the mode switching of the pointing device and the text device.
3 is an example of a configuration of a text input device and a pointing device.
Fig. 4 shows examples of the pointer execution command unit integral mode switching unit.
Fig. 5 shows an example of the structure of the pointer position information input device.
6 is a configuration diagram of an infrared laser sensor module.
Fig. 7 shows an embodiment for displaying the pointer position information input area.
8 is a first embodiment of a pointer position information input device which can be separated from a text input device.
9 is a second embodiment of a pointer position information input device which can be separated from the text input device.
10 is a third embodiment of a pointer position information input device which can be separated from a text input device.
11 is a front view and a perspective view of a human interface apparatus to which the lower-positioned pointer position information input device is applied.
12 is an embodiment of the pointer execution command unit integral mode switching unit.
13 is an embodiment utilizing a cover of a human interface device to which a lower-positioned pointer positional information input device is applied.
14 is an embodiment of a human interface device applied to a portable electronic device.
The present invention relates to a composite human interface for inputting text information or pointing position information from a user to a digital device capable of receiving text or pointing position related information such as a computer, a notebook, a tablet PC, a mobile phone, and the like.
As a conventional text information input device, a keyboard composed of an elastic body and a plurality of physical buttons connected to a switch is widely used.
In a digital device having a touch interface, a virtual keyboard is displayed on a display. When a part of a user's body is touched with a virtual keyboard displayed on the display, a gesture or an electrical signal of the user is sensed, A method of inputting text on a virtual keyboard displayed in a portion is also used.
The touch interface recognizes the movement of a part of the user's body or recognizes the contact with a specific contact surface, senses the flow of current through the user's body, or intercepts or interferes with light or sound waves by a part of the user's body The touch of the user can be recognized by sensing or the like.
Examples of touch interfaces include a pressure sensitive touch screen, an electrostatic touch screen, an optical touch screen, and an ultrasonic touch screen.
A resistive touch screen or a pressure sensitive touch screen is a pressure sensitive touch screen.
Resistive touchscreen is cheap, and it is advantageous to write in handwriting or small cell with stylus pen. However, because it uses pressure, it is not recognized by pressing hard, and touch feeling is slightly duller than electrostatic touch method. It is known as a disadvantage.
The resistive touchscreen consists of several layers.
Among these several layers are conductive layers, with the two layers facing the air layer.
Pressing the outer screen will cause the two conductive layers to touch each other, which causes the resistance and current to change and recognize the touch.
Capacitive sensing or electrostatic touching is a type of sensing operation using a capacitive coupling effect.
The electrostatic touchscreen is made of highly conductive glass, indium tin oxide, unlike pressure-sensitive touchscreens that use pressure.
By attaching the sensor to the four corners of the glass, current flows through the glass surface.
It is the electrostatic principle to recognize the change of current through this.
At this moment, the moment the finger moves on the screen, the electrons that flowed through the glass flow into the body through the finger, and the sensor senses the position where the change occurred.
The electrostatic touch method is known to have a feeling of operation and smoothness compared with a pressure sensitive touch screen because the touch is recognized by touching the screen rather than pressing the screen by force.
In addition, the electrostatic touch method enables multi-touch that can touch various places.
Since the electrostatic touch method is a method of operating by using a change in electric current, it may not be used with a leather glove which does not communicate with electric current, or may not be operated with a nail or a stylus pen.
However, you can touch the screen with a separate electrostatic stylus pen.
Sensors are sensitive and may be affected by peripheral devices.
The optical touch screen operates by measuring coordinates with the infrared camera mounted on the apex of the touch screen and the infrared light illuminates with the shadow of the object to be touched.
The ultrasonic touch screen emits ultrasonic waves on the screen, detects the interference effect by the user's touch, and operates by measuring the coordinates.
Various touch input techniques that can be used to detect positional information of a part of a user's body by sensing contact or movement of another user and controlling position information of the pointer can be used in the present invention.
1 is an exemplary diagram of a composite human interface device having a pointing device and a text input device.
The pointing device-integrated text input device may have a housing (100) for supporting a text input device and a pointer location information input area.
The housing has a strength enough to withstand the pressure of a user input and receives text input information and pointer position information input information and transmits the information to a digital device connected to the pointing device integrated text input device by wire or wireless A control unit, a memory unit, a battery unit, an encoding unit, and a transmission unit.
The pointing device-integrated text input device may include a plurality of
The plurality of
The physical button may include a button connected to the elastic body, or may have a feature that the button itself is elastic and moves to the position when the user receives the input, and returns to the original position when the pressure of the user is removed.
The physical button is connected to the electrical switch, and the user's pressure is applied to the physical button, so that the position of the button is shifted and the phase of the switch is changed, so that the text input value of the corresponding button is generated.
The physical button is structured to have elasticity without an electrical switch, and moves to a position when the user's pressure is removed. The user's text input information is input by the touch input device The user's pressure or gesture can generate textual input information based on the recognized location information.
The virtual button may be configured by displaying a text input button on a display device.
The virtual button can display an arbitrary button by projecting light onto a transparent, semi-transparent or opaque object.
The virtual button is not perceived by the user's eyes but can generate corresponding text input information based on user's pressure or gesture information with unique position information per text.
The complex human interface device may have a pointing location
The pointing position
The pointing location information input area may include the text input area or the text input area may include a point location information input area.
The pointing position information input area and the text input area are at least partially common and at least partially used as the pointing position information input area but not used as the text input area and at least partially used as the text input area, May not be used.
The complex human interface device includes a pointer location information input device 105 for text input or for pointing location information input or for forming a virtual text input area or a virtual pointing location information input area for text input and pointing location information input. It may be configured on the composite human interface device or external to the composite human interface device.
The pointer location information input device 105 is composed of a device such as an infrared generator and an infrared receiver or an RGB camera or an ultrasonic generator and an ultrasonic receiver or an infrared generator and an infrared camera, and the surface or the
The pointer location information input device may be modularized and configured as a plurality of modules to expand the pointer location information input area or to improve accuracy and sensitivity.
For example, a pointer position
Or a pointer position
The text input area or the pointing position
The text input area or the pointing position
The pointer location information input device may include both the J and F buttons of the English standard keyboard when there is no division for the left hand and the right hand.
The pointing location information input area may be a pointer position movement on the plurality of display parts when the complex human interface device is connected to a plurality of digital devices having a display part.
For example, the pointer position information input region may be divided into regions so as to be matched with the respective display portions, or a separate button indicating the display portion may be operated to transmit pointer position information on the display portion, or a plurality of display portions may be referred to as a virtual single The pointer position information can be transmitted so that the pointer can be moved on the virtual single display unit.
The complex human interface device includes a pointing location information input device that configures a pointing input area and receives pointer location information from a user, and inputs a button or an image, a space, an icon, or a text where a pointer moved by the pointing location information input device is located. It may have a pointer execution command unit (101, 102) for executing at least one function on the window.
The pointer execution command section can be composed of one or two buttons, which are responsible for the first function and the second function, respectively, and can be located on the left side of the housing, in the mail, or in the center.
The first function may function as a left click of a computer mouse and the second function may function as a right click of a computer mouse, for example.
The pointer execution command unit including one or two buttons may be provided at both the left side and the post of the housing.
The pointer execution command unit may be configured to be operated by touching or light interception of a part of the user's body, interference of ultrasonic waves, or shadow of a part of the user's body using the above-described touch technology.
The pointer execution command unit may be configured as an elastic physical button.
The pointer execution command unit may operate using at least one button of a text button located in the text input area in addition to the pointing position information input area.
The pointer execution command unit can operate by selecting a physical or virtual text button on the pointing position information input area.
For example, when using a virtual input device that can input location information in space as a pointing location information input device and using a physical button as a text input device, the position of the pointer on the virtual pointing location information input area in the pointing location information input mode. You can issue a pointer execution command by entering information and pressing the physical text button at that location.
The pointer execution command unit may receive pointer position information by a user first gesture in a pointing position information input area and generate a pointer execution command by a user second gesture at the same position.
The pointer execution command unit may be configured to perform a first function by a first gesture of the user's body, a first voice, a first blink, a first mouth shape, or the like.
The pointer execution command unit may be configured to perform a second function by a second gesture of the user's body, a second voice, a second blink, a second mouth shape, or the like.
The complex human interface device may include a text input mode for receiving text information through the text input device and a pointing location information input mode for receiving pointing location information through the pointing location information input device.
The text input mode and the pointing position information input mode can be switched by the
The
The
The
The
For example, by a first gesture such as contact of a part of the user's body on a first button that first reacts to a first touch or first pressure and to a second response to a second touch or second pressure. It may be configured to generate a pointer execution command in response to the second touch or the second pressure by a second gesture such as sensing a touch or the first pressure to switch modes and pressing a button.
The
For example, in response to a first touch or a first pressure, a mode is changed from a text input mode to a pointing position information input mode, and when a first touch or a first pressure is removed, have.
Alternatively, in the pointing position information input mode in response to the first touch or the first pressure, the mode may be switched to the text input mode, and when the first touch or the first pressure is removed, the mode may be set to return to the pointing position information input mode. .
Switch from text input mode to pointing position information input mode in response to the second touch or second pressure or toggle switch, and configure the permanent switching mode to maintain the pointing position information input mode even if the second touch or second pressure is removed. Can be.
Temporary mode switching may be performed by first control information received from a digital device connected to the composite human interface device by wire or wirelessly.
Permanent mode switching may be performed by the second control information received from the digital device connected to the composite human interface device by wire or wirelessly.
The
For example, when the first touch or first pressure is detected on the pointer execution command unit, the pointer mode is switched from the text input mode to the pointer position information input mode. When the second touch or the second pressure is detected, And when the third touch or the third pressure is applied, it is set to the permanent switching mode that operates in the pointer position information input mode even if the third touch or the third pressure is removed.
At this time, it is possible to configure to be able to input a pointer execution command even in the permanent switching mode.
The mode switching unit may be formed on the left or right side of the housing (106a, 106b)
The
The
When the length of the mode switch is less than 5 centimeters, it is preferable to set the position of the mode switch to include the left space of the left tab key of the window standard keyboard.
The composite human interface device may have a
The digital device can receive text input or receive pointer location information.
Fig. 2 is a flowchart for explaining an example of the operation sequence according to mode switching of the pointing device and the text device.
The complex human interface device may have a separate power supply unit or may receive power from the outside in a wired or wireless manner and may have a separate switch for controlling the power supply unit.
When power is supplied to the composite human interface device by the switch controlling the power supply unit, it may be determined whether the composite human interface device is currently in the text input mode or the pointing location information input mode (200).
As a result of the determination, if the text input mode is selected, the text input device may be activated and a text input may be received from the user (201).
The text input of the input user can be transmitted to a digital device connected by wire or wireless (202).
When the mode is switched to the pointing position information input mode during the text input mode, the text input mode can be switched to the pointing position information input mode.
Or the pointing position information can be input simultaneously with the text input.
When the mode is switched to the pointing position information input mode, the pointing position information may be received by the user (204).
The input pointer location information may be transmitted to a digital device connected by wire or wireless connection (205).
When the composite human interface device receives the pointer
When the composite human interface device receives the pointer
When the temporary mode switch is released or the permanent mode switch is released, the composite human interface device may switch to the text input mode.
The power switch of the power supply unit may be terminated when the power supply of the complex human interface device is disconnected or there is no connected digital device, or when the connection is lost, or when there is no user input for a predetermined time or by control of a digital device connected by wire or wirelessly.
3 is an example of a configuration of a text input device and a pointing device.
The complex human interface device may include a
The composite human interface device may have a text input device and a
At this time, the
The pointing location information input device may be configured as a pressure sensitive or capacitive touch pad type and may be located 304 above the physical buttons of the text input device.
At this time, if the text input device is constituted by a physical touch pad type such as a pressure sensitive type or capacitive type, the touch pad can be used as a text input device and a pointing position information input device.
At this time, a physical touch pad such as a pressure sensitive or capacitive touch panel may be configured as a large-area touch pad type complex
Alternatively, a physical touch pad such as a pressure sensitive or capacitive touch panel may be configured as a plurality of touch pad type
Or the upper camera type
In this case, the upper camera type complex
The camera may be configured as a
In this case, the lower camera type composite
Or a virtual pointing location information input area composed of a pair of infrared receivers or ultrasound receivers for receiving the information that the infrared or ultrasonic waves transmitted by the infrared transmitter or the ultrasonic transmitter are blocked or interfered with by a part of the user's body. 341 may be configured as a transmit / receive composite
At this time, the virtual pointing position information input area formed as a pair can be used as a virtual text button input means in place of the physical text button by the mode switching unit.
Fig. 4 shows examples of the pointer execution command unit integral mode switching unit.
An
An
The pressure difference type pointer execution command unit integrated
In this case, the position where the button is fixed can be configured to be set to the position where the
The pointer execution instruction unit integrated
When the contact area of a part of the user body which is composed of the
Fig. 5 shows an example of the structure of the pointer position information input device.
The pointer position information input device of the human interface device of the present invention may include a
The light emitter of the
The camera of the sensor module includes an optical sensor that detects light blocked or reflected by the obstacle.
The optical sensor may use a line camera capable of recognizing a line image.
It is preferable that the line camera is configured such that the image sensor is composed of 400 to 1000 elements in a line form, and preferably 500 to 700 elements.
The line camera may include a plurality of line image sensors that are smaller than one five fifth to one seventh one of the 500 to 700 elements.
The light emitter is constituted by a light emitting element for constituting a light plane on the upper end close to the
The pointing position information input area 108 is composed of a part of the light plane.
As the light emitting element, for example, an infrared emitter, an infrared laser emitter, a laser emitter, an ultraviolet emitter, a visible light emitter, or the like can be used.
Ideally, the light emitting device should use an infrared laser emitter that does not require a reflector to be attached to the rim.
The infrared laser emitter is suitable for driving a laser with a laser wavelength of 800 to 850 nanometers at 0.3 to 1 milliwatts.
The light emitter allows the light from the infrared emitter to pass through the line lens so that it can spread in a flat plane.
The light emitter may further include a line lens in the light emitting portion so that the light emitted from the light emitting element can constitute the light plane.
The line lens is positioned so that the light emitted from the light emitting element may form a light plane substantially parallel to the character input unit.
The pointer position information input device may further configure the
The
The light emitter and the optical filter for the camera may be separately configured, but it is preferable to design the light emitter and the camera in close proximity so that the light transmitted to one optical filter and the received light can be filtered .
The optical filter may be arranged to be perpendicular to the light traveling direction of the light emitter.
The optical filter may be installed in a range of 30 to 60 degrees or 120 to 150 degrees in the traveling direction of the light emitter in consideration of design.
The
The
At this time, light blocked or reflected by the obstacle may be refracted once by the
The
At this time, light blocked or reflected by the obstacle can be refracted twice by the two
In this way, the sensor module can be installed inside the
At this time, an optical filter may be coated on the
When the pointing position information input area 108 is activated by the
In this case, the interval between the
In addition, the plurality of
A light source, which is emitted from the
At this time, if the upper surface of the light tunnel is replaced with a portion of the
The lower surface and the side surface of the light tunnel can be designed as a separate structure from the first housing.
6 is a configuration diagram of an infrared laser sensor module.
The infrared sensor module basically emits light covering the pointing position information area using the
At this time, a
It is preferable to use a line camera for the camera to recognize the state where the light source is reflected or blocked by the obstacle.
The camera may install a guide light tunnel in front of the camera to help the incoming light to receive light in a desired area.
In the present invention, the light source may be an infrared ray, an ultraviolet ray, a laser, or the like.
Light sources and cameras can be replaced by ultrasonic and ultrasonic receivers.
In the present invention, it is preferable to use an infrared laser as a light source in order not to provide a reflector on the edge of the keyboard.
If you want to use infrared light, you can install a reflector on the keyboard's edge, or you can increase the light output more than twice as much as an optical touch module for a regular screen, or install an auxiliary light source.
The sensor module may have an
The pointer position information input unit may be positioned to include a part of the extended surface of the text input unit plane.
Wherein the text input unit comprises a plurality of buttons for receiving at least one text and the button moves from a first position to a second position by a pressure applied by a user to generate a text selection signal, Lt; RTI ID = 0.0 > a < / RTI > physical button.
The text input unit includes F4 and F5 buttons on a Windows standard keyboard, and two cameras of the pointer position information input unit are provided. The camera on the left of the two cameras and the text input unit are provided. One of the buttons can be designed to be the F4 or F5 button at the closest distance.
Alternatively, the camera on the left side may be positioned below the space bar of the Windows standard keyboard, above the zero space where the space between the F4 and F5 buttons is moved downward.
In this way, when the position of the index finger of the right hand is at the position of the F (F) of the text input button, it is possible to reduce the phenomenon that the index finger of the left hand obscures the camera when the pointer position information is inputted with the right hand.
If the pointer execution command unit is set to a button having the largest area of the text input unit, for example, a space bar, it is convenient to execute the pointer execution command unit with the left hand while inputting the mouse position information with the right hand, for example.
At this time, since the left hand may be operating the mode switching unit, it is preferable that the shortest distance between the mode switching unit and the pointer execution command unit is 8 to 15 centimeters.
The mode switching unit may be provided separately from the button of the text input unit, and recognizes that a part of the user's body is touched on a part of the human interface body, and switches the mode to the text input mode and the pointer position information input mode. It operates in the pointer position information mode and can operate in the text input mode when it is not in touch.
The pointer execution command unit is composed of two buttons and performs a first function when the first button is pressed, for example, a left click of the mouse, and a second function when the second button is pressed, for example, And the first button is a button having the largest area of the text input unit buttons, for example, a space bar button. When the second button is set as the first button located on the left or right of the first button, It's easy to control with the thumb of the thumb and provides an easy user experience.
When the mode switching unit is switched to the text input mode or the pointer position information input mode, the text input mode or the pointer position information input mode can be displayed.
Wherein the pointer position information input mode is displayed by temporarily displaying the pointer position information input mode at a moment when the touch of the mode switching unit occurs or until the touch point of the mode switching unit is released, lighting signal.
By doing so, the user can recognize which mode the keyboard is currently in, and can easily know at which position the pointer position information should be input in the pointer position information input mode.
The text input area of the text input unit may be the same as the pointer location information input area of the pointer location information input unit or may include the pointer location information input area of the pointer location information input unit as shown in FIG.
In an embodiment constituting the pointer position information input unit, one or two optical emitters and a camera may be used.
The optical signal received by the camera may pass through a light tunnel to block an external noise source.
The light tunnel may be configured before the optical signal is emitted from the optical emitter to the pointer location information input area, and may be installed on the front surface of the camera receiving the user input signal from the pointer location information input area.
At this time, an optical signal emitted from the optical emitter or a reflector for refracting an optical signal received by the camera can be disposed.
A degree of freedom is provided to the installation position and direction of the pointer position information input unit through the reflection plate, thereby achieving space utilization and design benefit in design.
The pointer position information input unit may configure the pointer position information input area at least 0.1 to 5 millimeters above the text input unit by the optical emitter and the camera.
Thus, the pointer position information can be input only when there is a clear touch intention of the user.
When the optical emitter is designed to operate at 0.3 to 1 milliwatt using an infrared laser having a wavelength of 800 to 850 nanometers, it exhibits excellent power consumption, cost, stability, safety, and visual characteristics.
The optical emitter may further include a line lens that scatters the optical signal so that the optical signal constitutes a pointer position information input area parallel to the text input part.
The pointer position information input unit may be designed to block optical external noise using an optical filter that transmits only the optical signal in the infrared region.
If the interval between the physical buttons is designed to be 0.1 millimeter to 1 millimeter, the text input unit can provide the user with a feeling of sufficiently smooth and flat feeling during the input of the pointer position information.
Fig. 7 shows an embodiment for displaying the pointer position information input area.
The human interface apparatus of the present invention may be designed to further include pointer position information input area display units 701 and 702 for visually displaying the pointer position information input area in the pointer position information input mode.
The mode switching unit is provided separately from the button of the text input unit and recognizes that a part of the user's body is touched by a part of the complex human interface housing to switch to the text input mode and the pointer position information input mode, but when the touch is performed, the pointer is touched. It can be designed to operate in location information input mode and to operate in text input mode during non-touch.
At this time, when a part of the body used for the mode switching is the right hand (106b), the hand which performs the pointer
In this case, it is desirable that the pointer position information input region is designed to be closer to the opposite corner of the corner where the mode switching unit is located, and the same principle can be also applied when the left and right hands are reversed. (106a, 108b)
On the other hand, a part of the body used for the mode switching may be a right hand, and a hand for inputting the pointer position information may be designed to be a right hand.
In this case, it is preferable that the pointer position information input area is designed to be closer to the corner where the mode switching part is located than the opposite corner of the corner where the mode switching part is located, and the left hand can be designed on the same principle.
The mode switching unit may determine the text input mode and the pointer location information input mode based on the number or positions of user fingers recognized by the pointer location information input unit.
The number of the recognized user fingers when determining the pointer position information input mode is smaller than the number of the recognized user fingers when determining the text input mode.
For example, while controlling the mode switching unit with the left hand, the finger of the left hand may be untouched from the text input button, and the index of the right hand may be touched for inputting the pointer position information, Is one.
On the other hand, when the finger of the left hand and the right hand are touched on the keyboard for text input, the number of the fingers of two to eight recognized users is recognized.
The pointer position information input mode display unit displays the pointer position information input area from the point of time when the mode switching unit is switched to the pointer position information input mode to the time point when the mode is switched to the pointer position information input mode, It can be displayed visually.
The pointer position information input mode display unit may be displayed by the visible light generator from the text input button or the reflected light by the text input button or through the space between the text input buttons.
At this time, the input mode may be displayed even in the text input mode, and the text input mode and the pointer location information input mode may be displayed to be distinguished.
The pointer position information input mode display unit includes
The mode switching unit may further include a hand side determining unit for determining whether the user inputs the pointer position information with a left hand or a right hand.
The display of the pointer location information input display unit may be changed according to the handside determination unit.
It is possible to flexibly switch the button arrangement and operation of the pointer execution command unit according to the hand side judgment unit.
For example, when the pointer execution command includes a right-click and a left-click, the hand-side determination unit may switch the button to perform the right-click and left-click operations.
The pointer position information input area display unit may display (702a, 702b) different pointer position information input areas according to the determination result of the hand side determination unit.
The mode switching unit may be designed to automatically switch to the text input mode when the pointer position information input from the user is not input for a predetermined time in the pointer position information input mode or when the input is received by the text input button.
The pointer position information input area display unit may apply a dye to a part of the text input button or the text input area so that the pointer input area is visually displayed irrespective of the mode of the mode switching unit so that the pointer input area can be recognized even in the text input mode .
8 is a first embodiment of a pointer position information input device which can be separated from a text input device.
The pointer location
At this time, the
The pointer position
The detachable pointer location
At this time, the pointing position information input device and the
9 is a second embodiment of a pointer position information input device which can be separated from the text input device.
The pointer location
When the pointer position
Alternatively, the apparatus may further include a first position information input mode adjusting device that adjusts the height and tilt of the housing so that the user may arbitrarily adjust the height and angle of the position information input area.
When the pointer position
At this time, preferably, the height of the
When the pointer location
The pointer position
The location information input mode sensor unit may include a sensor utilizing gravity, a sensor detecting a surface placed on the floor, or a switch.
The pointer location
In particular, in the third location
That is, when the user is inputting text, the user's hand or finger is not detected by the location
When the user places a hand or finger on the top of the location information input device in the third location
Alternatively, a text input mode based on a state in which a digital device connected to the
Alternatively, the
At this time, when the text input is in progress, the location
After the text input is interrupted, the first power saving mode may be switched from the first power saving mode to the active mode during the first first interval.
When the user's hand or finger is not detected during the first first interval after the text input is interrupted, the user may enter the second power saving mode which consumes less power than the first power saving mode.
In the first power saving mode, the sensor may be periodically controlled with less power than the active mode to determine the presence of the user's hand or finger.
When the user's hand or finger is detected in the first power saving mode, the user may switch to the activation mode.
In the activation mode, it is possible to switch to the pointer position information input mode.
The pointer execution command unit for a pointer execution command in the pointer position information input mode may generate a pointer execution command using a button provided in the text input device or based on a position or shape change of the detected user's hand or finger. Can be.
The location
10 is a third embodiment of a pointer position information input device which can be separated from a text input device.
The pointer location
The pointer location
After the
At least one
8, 9 and 10 illustrate the following components in an embodiment of a pointing-position-information inputting apparatus that can be detached from a text inputting apparatus.
And a pointer location information input unit for receiving information related to the pointer location from the user.
The pointer position information input unit receives the pointing position information of the user input through the pointing position information input area formed by the pointer position information input device.
And a pointer execution command receiving unit that receives a signal of a pointer execution command unit that receives a user command to perform at least one function in a place where the pointer is located.
The pointer execution command unit may be constituted by a button or a touch switch included in the pointer position information input unit or a control unit for switching to a pointer execution instruction when a predetermined input of the user input received from the text input device is received.
And a mode switching instruction receiving unit for receiving a signal of a mode switching unit for switching to a pointer position information input mode.
The mode switching unit may be configured in a pointing position information input device, or configured in a text input device, or may be configured as a separate device.
And a power unit for transmitting power to the pointer position information input unit, the pointer execution command receiving unit, and the mode switching command receiving unit.
The power source unit may be a power source unit having a self-powered battery or a power source supplied from the outside.
And a pointer position information transmitter for transmitting information related to the pointer position input to the pointer position information input unit to a digital device connected to the human interface device by wire or wirelessly.
Wherein the pointer position information input area of the pointer position information input unit is positioned at a predetermined angle with the predetermined height so that the pointer position information input area can be positioned to be parallel to the upper end of at least a part of a text input area of a text input device composed of a plurality of physical buttons Is set.
Wherein the mode switching unit can be fixed to at least one side of the text input device when the mode switching unit is provided separately from the text input button of the text input device and recognizes that the user input is received in the mode switching unit, And can be transmitted to the mode switching command receiving unit by wire or wirelessly.
The pointer position information input unit may include an optical emitter and a camera, and the optical signal received by the camera may be designed to pass through an optical filter through a light tunnel.
A position state recognition unit for recognizing the position state information about the gravity direction of the human interface or the position state information about the floor on which the human interface is placed or the position information of the human interface through the input of the user.
The position state is composed of at least two kinds, and when the first position state is selected, the mode is switched to the first pointer position information input mode and when the second position state is entered, the mode is switched to the second pointer position information input mode
And the pointer position information input area in the first position state is located in parallel with at least a part of the text input area of the text input device.
The pointer position information input area in the second position state is located so as to be parallel to the floor surface on which the pointing position information input device is placed.
The upper surface of the pointing position information input device in the first pointer position information input mode is positioned to face the bottom surface. When switching to the second pointer position information input mode, the calculation of the ordinate of the pointing position information input is the same, Calculate by changing direction.
The pointer execution command unit may use a space bar among text input buttons of the text input device and may further use another button adjacent to the space bar as a pointer execution command unit.
The power supply unit may receive power from the text input device.
The predefined height and angle of the pointer location input area in the first positional state may be defined by a controller that allows adjustment within a predefined range by the user.
The adjuster can adjust the angle by adjusting the joining angle of the at least one surface of the housing of the pointing position information input device with the adjacent surface.
The controller adjusts the height of at least one surface of the housing of the pointing position information input device.
The controller adjusts the height and angle of the pointing position information input unit by adjusting the height and angle of the optical module of the pointing position information input unit.
11 is a front view of a human interface device to which a lower-positioned pointer positional information input device is applied.
The human interface apparatus includes a text input unit configured by a plurality of physical buttons, a pointer position information input unit for inputting information related to a pointer position from a user, a pointer execution command for inputting a user command to perform at least one function, A mode switching instruction receiving unit for receiving a signal of a mode switching unit for switching to a pointer position information input mode; a power supply unit for supplying power to the pointer position information input unit, the pointer execution command receiving unit, and the mode switching instruction receiving unit; A pointer position information transmitting unit for transmitting information related to a pointer position input to the pointer position information inputting unit to a digital device connected to the human interface device by wire or wirelessly, The information input area is positioned so as to be parallel to the upper end of at least a part of the text input area of the text input part composed of the plurality of physical buttons and the pointer position information input part comprises at least two sensor modules, And the module is located on the
Wherein the pointer position information input unit comprises an optical emitter and a camera, and the optical signal received by the camera passes through a light tunnel.
And the mode switching unit is provided separately from the text input button of the text input device.
The mode switching unit is located on the left side of the text input area of the text input device and recognizes that the user input is received in the mode switching unit and transmits the pointer position information input mode to the mode switching instruction receiving unit.
The physical button is composed of an upper portion of a button receiving pressure from the user's finger and an elastic member applying a physical force to return the upper portion of the button to the original position when the user's finger pressure is removed.
The complex human interface device may include a reflector or an
The reflector is used when the light source generated from the pointer position information input unit is an infrared ray, and when an infrared ray laser is used, an absorber is used.
The at least two
Also, by orienting the sensor in the diagonal direction of the text input area by locating the sensor near the vertex of the text input area (1110), it is better to position the sensor in the corner area and face the transverse direction of the text input area The pointer position information input area having a wider area can be constituted. In addition, it is possible to maximize the pointer position information input area at the upper part of the text input area by moving the position of the portion where the position information detection accuracy of the pointer position information input area between the sensor and the sensor is relatively low.
In addition, the position of the sensor can be positioned 1120 below the space bar of the standard keyboard to be narrower than the width of the keyboard. The width between the sensor and the sensor determines the width of the pointer position information input area and thereby obtains relatively high pointer position information descriptive resolution although the area of the pointer position information input area is reduced rather than positioned at the corner (1110).
Between the sensor and the sensor, the reflector or absorber is removed, and the height of the housing is lowered to prevent the thumb from being caught by the housing when the user presses the space bar of the standard keyboard with his / her thumb.
12 is an embodiment of the pointer execution command unit integral mode switching unit.
When the input operation in which the text input mode and the pointer position information input mode are frequently changed is performed, the number of times the mode switching and the pointer execution command are input is increased.
For example, in case of right-handed person, when you want to input pointer position information while text input mode is being performed in text input mode, you need to input the pointer position information after inputting to pointer real name command with left hand after switching mode with left or right hand do.
In general, while performing a quick operation, the user is required to input text and pointer position information and a pointer execution command while keeping a gaze on the monitor. If a mode switching unit and a pointer execution command input unit are provided separately, hand movement frequently occurs It is not possible to find the exact position, so it may not be able to perform properly, or it may be troublesome to look at the keyboard with eyes.
In order to solve this problem, for example, in the case of a right-handed person, when a hand is placed on the mode switching unit with the left hand after inputting text in the text input mode, a mode switching command is generated and the mode is changed to the pointer input mode. The pointer execution command unit may receive the user input by inputting the pointer position information with the right hand with the hand on the mode switching unit and applying pressure to the mode switching unit.
Thereby, the user can share the position of the mode switching unit for mode switching and the position of the pointer execution command unit for the pointer execution command.
For example, the mode switching unit may include a touch switch capable of accepting a touch input of a user's hand, and the pointer execution command unit may be disposed under the mode switching unit, which includes a switch responsive to a pressure such as a tact switch.
In this case, the first pointer
At this time, the mode switching is operated in the pointer position information input mode while the touch is being performed, and in the text input mode when the touch is released.
The mode switching by the touch may be composed of another switch operating at a pressure smaller than the pressure for executing the pointer execution command, or a sensor capable of sensing the position of the user's finger.
In this case, the mode switching and pointer execution command unit may be located in an area distinct from the text input area.
Preferably, in the case of right handedness, it is located in the left outer area of the text input area.
If the mode switching unit and the pointer execution command unit are located outside the text input area, the possibility that the pointer position information input unit is confused with the pointer position information input disappears and the pointer position information input area can be expanded.
The hybrid human interface apparatus of the present invention may include a second
At this time, when the compound human interface apparatus is in the text mode by the second mode switching unit, when the mode switching command by the touch is received, the compound human interface apparatus of the present invention operates in the pointer position information input mode. In addition, when the compound human interface apparatus by the second mode switching unit is in the pointer position information input mode, when the mode switching command by touch is received, the pointer position information input mode is maintained but the mode switching command by touch is released When a text input is received by the text input unit, the mode is switched to the text input mode.
In this case, when the text input mode is switched by the text input, at least the first text input may be ignored, and when at least two text inputs are received, the text input mode may be switched.
In this case, when at least two or more text inputs are received and the mode is switched to the text input mode, the composite human interface device transmits the text input including at least the first text, which has been ignored, to the digital device, Lt; / RTI >
13 is an embodiment utilizing a lid of a composite human interface device.
The composite human interface device requires space of at least 1 millimeter to 2 millimeters on the top of the text input device plane in which the pointer position information input device is constituted by the physical device to form the pointer position information input area, Or a border for locating the reflector may be required at the edges of at least three sides surrounding the text entry area.
At this time, a difference of at least 1 to 2 millimeters occurs between the height of the edge of the at least three sides of the edge and the height of the text input area plane, and a substantially rectangular parallelepiped type space is formed at the top of the text input area.
The complex human interface apparatus of the present invention may further include a multipurpose cover 1310 for protecting the text input area from an external impact in the space of the substantially rectangular parallelepiped type.
The
The
This can further reduce the slope of the composite human interface device when the multipurpose cover is folded down and then mounted to adjust the inclination of the composite human interface device.
When the multipurpose cover is folded and then mounted on the bottom surface of the composite human interface device, a part of the folded surface that is in contact with the ground surface may be provided with an
The slope of the composite human interface device desired by the user can be adjusted by the number of times of folding of the
The multi-purpose lid may have a built-in
An electrode of a power source of the
Also, when the multipurpose lid is folded and mounted on the lower end to adjust the inclination of the composite human interface device, it can be connected to the electrode provided on the bottom surface of the composite human interface device.
The composite human interface device may be connected to an external power source to supply power to the
At this time, the bottom surface of the composite human interface device has a
The rim of the multi-purpose lid is coated with a substance that can remove minute dust, oil, and moisture, and can remove foreign substances from the absorption plate, the reflector, the optical emitter, and the front surface of the camera have.
If the versatile cover covers the text input area, it can detect the text input area and turn off the power of the composite human interface device.
When the multipurpose cover is mounted on the bottom of the composite human interface device, the composite human interface device may be turned on.
Also, if there is no input by the user for a certain period of time after the power is turned on, the power can be cut off or the apparatus can be switched to the standby mode.
When the multipurpose cover covers the text input area (1310) The composite
On the other hand, if the multipurpose cover is folded and mounted on the bottom of a composite human interface device, the inclination can be secured like a general keyboard.
The multipurpose lid can be constructed in a removable manner, but it can be designed in such a way that it is folded down from the upper surface to the lower surface of the composite human interface device by the hinge.
14 is an embodiment of a human interface device applied to a portable electronic device such as a portable electronic device or a tablet personal computer.
The composite
The portable electronic device having the composite human interface apparatus of the present invention is configured such that the display unit composed of the
At this time, the display unit is inserted into the text input area of the human interface device of the present invention and the substantial hexahedral space formed by the reflection plate or absorbing
To this end, the display unit of the portable electronic device should be designed to have a width smaller than the width of the
In other words, in Fig. 13, the display unit must be designed so that the multipurpose cover can be inserted into the place.
The display unit has a
When the display unit is folded, the pointer location
This is to fold the portable electronic device so that it can be kept thin without being disturbed by the pointer position
At this time, the mode switching unit and the pointer execution command unit are located outside the side text input area of the human interface device as described above, the mode switching unit is outside the text input area, and the pointer execution command unit is at least one of the text input buttons, or the mode switching. Both the negative and pointer command execution units may operate as at least one of the text input buttons.
(1425, 1426), and more preferably, can be located outside the text input area.
At this time, it is preferable that the mode switching unit and the pointer execution command unit are integrally formed as shown in FIG. 12, but they may be separately disposed.
At this time, when the mode switching unit or the pointer execution command unit is located outside the text input area, when the user places a finger to control the mode switching unit or the pointer execution command unit, the pointer location
For this purpose, it is possible to design an optical signal to be prevented from reaching the mode switching unit or the pointer execution command unit by adjusting an area covered by the optical signal generated from the pointer position information input device.
Or an optical signal receiving angle of the camera is adjusted so that the camera does not receive an optical signal generated by reflection, interference, or blocking between the user's finger and the optical signal located in the mode switching unit or the pointer execution command unit, The optical signal can be ignored.
The portable electronic device is designed to have at least three corner portions of an absorbing plate or a reflecting plate for absorbing or reflecting a light source generated by the pointer position information input device.
At this time, the two surfaces are located on the side of the human interface device (1423, 1424) and the other surface is the pointer position information input device (1421) when the display unit is open within a predetermined angle on the surface where the display unit is in contact with the human interface device. The reflector or absorber may be positioned in a
In FIG. 14, the display unit may be configured as a flexible display unit. The display unit may be curved to cover the text input area of the complex human interface device and be wound around the bottom surface of the complex human interface device.
In addition, in FIG. 14, the display unit may be replaced with the multipurpose cover described in FIG. 13, and the absorber or reflector may be positioned at a corner portion of the multipurpose cover bonded to the composite human interface device, and flipped to the rear by a hinge to allow the composite human It can also be used to adjust the height of the interface device.
1 to 14 illustrate an example of a complex human interface device to be described in the present invention. The type of text input device pointing position information input device and the technology used can be changed or replaced without changing the basic purpose by a person skilled in the art .
Claims (18)
A pointer position information input unit for receiving information related to the pointer position from the user,
A text input unit configured by a plurality of physical buttons,
A first mode switching unit for activating or deactivating the pointer position information input mode,
And a pointer execution command receiving unit for executing at least one function at a position of the pointer,
Wherein the pointer position information input unit is located at an upper end of the text input unit, the pointer execution command receiving unit is located at a lower end of the text input unit, and the pointer position information input area formed by the pointer position information input unit is a text input area And a plane perpendicular to the first plane.
Receiving text input from a text input comprising a plurality of physical buttons,
Activating or deactivating the pointer position information input mode from the first mode switching unit,
Receiving a pointer execution command for executing at least one function at a location of a pointer from a pointer execution command receiving unit,
Wherein the pointer position information input unit is located at the upper end of the text input unit, the first mode switching unit and the pointer execution command receiving unit are located at the lower end of the text input unit, and the pointer position information input area formed by the pointer position information input unit And a plane perpendicular to the text input area formed in the text input unit.
Priority Applications (7)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/403,998 US9612666B2 (en) | 2012-05-28 | 2012-12-29 | Human interface apparatus having input unit for pointer location information and pointer command execution unit |
PCT/KR2012/011822 WO2013180369A1 (en) | 2012-05-28 | 2012-12-29 | Human interface apparatus having input unit for pointer location information and pointer command execution unit |
US14/718,107 US9612667B2 (en) | 2012-05-28 | 2015-05-21 | Human interface apparatus having input unit for pointer location information and pointer command execution unit |
US14/718,108 US20150253869A1 (en) | 2012-05-28 | 2015-05-21 | Human interface apparatus having input unit for pointer location information and pointer command execution unit |
US14/718,105 US9639173B2 (en) | 2012-05-28 | 2015-05-21 | Human interface apparatus having input unit for pointer location information and pointer command execution unit |
US14/794,825 US9612668B2 (en) | 2012-05-28 | 2015-07-09 | Human interface apparatus having input unit for pointer location information and pointer command execution unit |
US15/475,045 US9880637B2 (en) | 2012-05-28 | 2017-03-30 | Human interface apparatus having input unit for pointer location information and pointer command execution unit |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR1020120056427 | 2012-05-28 | ||
KR1020120056427A KR20130053363A (en) | 2011-11-15 | 2012-05-28 | Apparatus and method for multi human interface devide |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
KR1020130148173A Division KR20140082920A (en) | 2013-12-02 | 2013-12-02 | Multi human interface devide having poiner location information input area including vertical area to text input area |
Publications (1)
Publication Number | Publication Date |
---|---|
KR20130133117A true KR20130133117A (en) | 2013-12-06 |
Family
ID=48699448
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
KR1020120152644A KR20130133117A (en) | 2012-05-28 | 2012-12-24 | Multi human interface devide having poiner location information input area including vertical area to text input area |
KR1020120156853A KR101268209B1 (en) | 2012-05-28 | 2012-12-28 | Multi human interface devide having poiner location information input device and pointer excution device |
Family Applications After (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
KR1020120156853A KR101268209B1 (en) | 2012-05-28 | 2012-12-28 | Multi human interface devide having poiner location information input device and pointer excution device |
Country Status (1)
Country | Link |
---|---|
KR (2) | KR20130133117A (en) |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9921661B2 (en) | 2012-10-14 | 2018-03-20 | Neonode Inc. | Optical proximity sensor and associated user interface |
US10282034B2 (en) | 2012-10-14 | 2019-05-07 | Neonode Inc. | Touch sensitive curved and flexible displays |
US10324565B2 (en) | 2013-05-30 | 2019-06-18 | Neonode Inc. | Optical proximity sensor |
US10585530B2 (en) | 2014-09-23 | 2020-03-10 | Neonode Inc. | Optical proximity sensor |
US11073948B2 (en) | 2012-10-14 | 2021-07-27 | Neonode Inc. | Optical proximity sensors |
US11842014B2 (en) | 2019-12-31 | 2023-12-12 | Neonode Inc. | Contactless touch input system |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2003280791A (en) | 2002-03-20 | 2003-10-02 | Hitachi Ltd | Keyboard with pointing device |
-
2012
- 2012-12-24 KR KR1020120152644A patent/KR20130133117A/en not_active Application Discontinuation
- 2012-12-28 KR KR1020120156853A patent/KR101268209B1/en active IP Right Grant
Cited By (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9921661B2 (en) | 2012-10-14 | 2018-03-20 | Neonode Inc. | Optical proximity sensor and associated user interface |
US10282034B2 (en) | 2012-10-14 | 2019-05-07 | Neonode Inc. | Touch sensitive curved and flexible displays |
US10802601B2 (en) | 2012-10-14 | 2020-10-13 | Neonode Inc. | Optical proximity sensor and associated user interface |
US10928957B2 (en) | 2012-10-14 | 2021-02-23 | Neonode Inc. | Optical proximity sensor |
US10949027B2 (en) | 2012-10-14 | 2021-03-16 | Neonode Inc. | Interactive virtual display |
US11073948B2 (en) | 2012-10-14 | 2021-07-27 | Neonode Inc. | Optical proximity sensors |
US11379048B2 (en) | 2012-10-14 | 2022-07-05 | Neonode Inc. | Contactless control panel |
US11714509B2 (en) | 2012-10-14 | 2023-08-01 | Neonode Inc. | Multi-plane reflective sensor |
US11733808B2 (en) | 2012-10-14 | 2023-08-22 | Neonode, Inc. | Object detector based on reflected light |
US10324565B2 (en) | 2013-05-30 | 2019-06-18 | Neonode Inc. | Optical proximity sensor |
US10585530B2 (en) | 2014-09-23 | 2020-03-10 | Neonode Inc. | Optical proximity sensor |
US11842014B2 (en) | 2019-12-31 | 2023-12-12 | Neonode Inc. | Contactless touch input system |
Also Published As
Publication number | Publication date |
---|---|
KR101268209B1 (en) | 2013-05-29 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US9880637B2 (en) | Human interface apparatus having input unit for pointer location information and pointer command execution unit | |
KR102367253B1 (en) | Electrical device having multi-functional human interface | |
KR102052752B1 (en) | Multi human interface devide having text input unit and pointer location information input unit | |
KR101268209B1 (en) | Multi human interface devide having poiner location information input device and pointer excution device | |
KR20140075651A (en) | Multi human interface devide having display unit | |
KR20150032950A (en) | Digital device having multi human interface devide | |
KR20150050546A (en) | Multi functional human interface apparatus | |
KR102189744B1 (en) | Multi human interface devide having text input unit and pointer location information input unit | |
KR20140082920A (en) | Multi human interface devide having poiner location information input area including vertical area to text input area | |
KR20140063489A (en) | Multi human interface devide having display unit | |
KR20140063488A (en) | Multi human interface devide having display unit | |
KR20140063485A (en) | Multi human interface devide having display unit | |
KR20140063484A (en) | Multi human interface devide having display unit | |
KR20140063490A (en) | Multi human interface devide having display unit | |
KR20140063486A (en) | Multi human interface devide having display unit | |
KR20140063487A (en) | Multi human interface devide having display unit | |
KR20140063483A (en) | Multi human interface devide having display unit |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
A107 | Divisional application of patent | ||
WITN | Withdrawal due to no request for examination |