CN113531423A - Interactive intelligent projection table lamp and method - Google Patents
Interactive intelligent projection table lamp and method Download PDFInfo
- Publication number
- CN113531423A CN113531423A CN202110790844.0A CN202110790844A CN113531423A CN 113531423 A CN113531423 A CN 113531423A CN 202110790844 A CN202110790844 A CN 202110790844A CN 113531423 A CN113531423 A CN 113531423A
- Authority
- CN
- China
- Prior art keywords
- projection
- interface
- touch
- touch message
- type
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000000034 method Methods 0.000 title claims abstract description 27
- 230000002452 interceptive effect Effects 0.000 title claims abstract description 11
- 230000009471 action Effects 0.000 claims abstract description 9
- 230000008569 process Effects 0.000 claims description 12
- 230000007547 defect Effects 0.000 abstract description 4
- 230000008901 benefit Effects 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 230000009286 beneficial effect Effects 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 230000004438 eyesight Effects 0.000 description 1
- 230000002349 favourable effect Effects 0.000 description 1
Classifications
-
- F—MECHANICAL ENGINEERING; LIGHTING; HEATING; WEAPONS; BLASTING
- F21—LIGHTING
- F21S—NON-PORTABLE LIGHTING DEVICES; SYSTEMS THEREOF; VEHICLE LIGHTING DEVICES SPECIALLY ADAPTED FOR VEHICLE EXTERIORS
- F21S6/00—Lighting devices intended to be free-standing
- F21S6/002—Table lamps, e.g. for ambient lighting
-
- F—MECHANICAL ENGINEERING; LIGHTING; HEATING; WEAPONS; BLASTING
- F21—LIGHTING
- F21V—FUNCTIONAL FEATURES OR DETAILS OF LIGHTING DEVICES OR SYSTEMS THEREOF; STRUCTURAL COMBINATIONS OF LIGHTING DEVICES WITH OTHER ARTICLES, NOT OTHERWISE PROVIDED FOR
- F21V33/00—Structural combinations of lighting devices with other articles, not otherwise provided for
- F21V33/0004—Personal or domestic articles
- F21V33/0052—Audio or video equipment, e.g. televisions, telephones, cameras or computers; Remote control devices therefor
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0487—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
- G06F3/0488—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
- G06F3/04883—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures for inputting data by handwriting, e.g. gesture or text
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02B—CLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO BUILDINGS, e.g. HOUSING, HOUSE APPLIANCES OR RELATED END-USER APPLICATIONS
- Y02B20/00—Energy efficient lighting technologies, e.g. halogen lamps or gas discharge lamps
- Y02B20/40—Control techniques providing energy savings, e.g. smart controller or presence detection
Abstract
The invention discloses an interactive intelligent projection table lamp and a method, wherein a camera and a projection device are arranged on a table lamp body, and the shooting direction of the camera and the projection direction of the projection device are both opposite to the same projection medium; the camera and the projection device are both in communication connection with a control module of the desk lamp body; the control module is connected with a projection switch, the projection switch is used for sending a switching signal of the projection device to the control module, and the control module controls the projection device to be switched on and off according to the switching signal; the control module runs an operating system and a system-level touch recognition service, and the touch recognition service is used for recognizing the fingertip action of a user on the projection interface through an image acquired by the camera, generating a touch message according to the fingertip action and transmitting the touch message to the operating system so that the operating system can execute corresponding control operation. By using the invention, a user can perform accurate touch interactive operation on the projected interface, and the defect that the touch operation cannot be performed on the projected interface in the conventional desk lamp can be overcome.
Description
Technical Field
The invention relates to the technical field of intelligent table lamps, in particular to an interactive intelligent projection table lamp and a method.
Background
The intelligent desk lamp has entered daily life of people, the intelligent desk lamp with a display function in the current market is mainly divided into two types, one type is the intelligent desk lamp with a display screen and a touch screen, the desk lamp has the advantages of strong interactivity and the defects that a display area is fixed and the damage of the display screen to eyes is generally considered to be large; one is the intelligent desk lamp that can carry out the content and throw, and this kind of intelligent desk lamp advantage is that the display area who throws can change, throws light and forms through reflection such as desktop or wall and throw the interface, generally thinks to be favorable to the protection of eyesight like this, and the shortcoming is interactive poor.
Disclosure of Invention
Aiming at the defects of the prior art, the invention aims to provide an interactive intelligent projection table lamp and a method.
In order to achieve the purpose, the invention adopts the following technical scheme:
an interactive intelligent projection table lamp comprises a table lamp body, wherein a camera and a projection device are arranged on the table lamp body, and the shooting direction of the camera and the projection direction of the projection device are both opposite to the same projection medium; the camera and the projection device are both in communication connection with a control module of the desk lamp body; the control module is connected with a projection switch, the projection switch is used for sending a switching signal of the projection device to the control module, and the control module controls the projection device to be switched on and off according to the switching signal;
the control module runs an operating system and a system-level touch recognition service, and the touch recognition service is used for recognizing the fingertip action of a user on a projection interface projected by the projection device through an image acquired by the camera, generating a touch message according to the fingertip action and transmitting the touch message to the operating system so that the operating system can execute corresponding control operation; the specific process is as follows:
the method comprises the steps that after a touch recognition service is started, a camera is opened, whether a projection interface appears on a projection medium or not is continuously recognized through an image shot by the camera, the projection interface is recognized and obtained, meanwhile, an area interface area (x, y, width, height) of the projection interface in the image shot by the camera is also recognized, the interface area is divided by the width of a preset display interface to obtain the magnification factor of the projection interface, whether a fingertip appears in the projection interface or not and whether the fingertip is in contact with the projection medium or not is recognized, if the fingertip is in contact with the projection medium to indicate that touch exists, the position (x ', y') of the fingertip relative to the projection interface is obtained, a touch message is generated, and the position coordinate of the touch message on the projection interface is (x ', y'); the type of the touch message is divided into down pressing, move moving and up lifting, if the type of the last touch message is up and the fingertip and the projection medium are in contact, the type of the current touch message is judged to be down, if the type of the last touch message is down or move and the fingertip and the projection medium are still in contact, the type of the current touch message is judged to be move, and if the type of the last touch message is down or move and the fingertip and the projection medium are not in contact any more, the type of the current touch message is judged to be up; after the type of the touch message and the position coordinate on the projection interface are obtained, dividing the position coordinate (x ', y') by the magnification factor of the projection interface to obtain the position coordinate of the touch message mapped on the display interface, and then forwarding the type of the touch message and the position coordinate of the touch message on the display interface to an operating system; and after receiving the touch message, the operating system executes corresponding operation according to the position of the touch message on the display interface and the type of the touch message.
Further, the process of identifying the projected interface by the touch recognition service is as follows: and comparing the image acquired by the camera with a preset display interface image, and if the preset display interface image appears on the image acquired by the camera, identifying the projection interface.
The invention also provides a method for intelligently projecting the desk lamp, which comprises the following specific processes:
when the display interface is required to be projected, the projection switch is turned on, the operating system detects that the projection switch is turned on, the touch identification service is started, and the display interface is projected onto a projection medium by using the projection device;
the method comprises the steps that a camera is opened through touch recognition service, whether a projection interface exists on a projection medium or not is judged by using an image shot by the camera, the projection interface is recognized and obtained, meanwhile, an area (x, y, width, height) of the projection interface in the image shot by the camera is also recognized, the width of a preset display interface is divided by the area to obtain the magnification factor of the projection interface, and then whether fingertips appear in the projection interface or not and whether the fingertips are in contact with the projection medium or not is recognized; if the fingertip is in contact with the projection medium to indicate touch, acquiring the position (x ', y') of the fingertip relative to the projection interface, and generating a touch message, wherein the position coordinate of the touch message on the projection interface is (x ', y');
judging the type of the touch message, wherein the type of the touch message is divided into down pressing, move moving and up lifting, if the type of the last touch message is up and the fingertip and the projection medium are in contact, the type of the current touch message is judged to be down, if the type of the last touch message is down or move and the fingertip and the projection medium are still in contact, the type of the current touch message is judged to be move, and if the type of the last touch message is down or move and the fingertip and the projection medium are not in contact any more, the type of the current touch message is judged to be up;
after the type of the touch message and the position coordinate on the projection interface are obtained, dividing the position coordinate (x ', y') by the magnification factor of the projection interface to obtain the position coordinate of the touch message mapped on the display interface, and then forwarding the type of the touch message and the position coordinate of the touch message on the display interface to an operating system;
and after receiving the touch message, the operating system executes corresponding operation according to the position of the touch message on the display interface and the type of the touch message.
Further, the process of identifying the projected interface by the touch recognition service is as follows: and comparing the image acquired by the camera with a preset display interface image, and if the preset display interface image appears on the image acquired by the camera, identifying the projection interface.
The invention has the beneficial effects that: by using the invention, a user can perform accurate touch interactive operation on the projected interface, and the defect that the touch operation cannot be performed on the projected interface in the conventional desk lamp can be overcome.
Detailed Description
The present invention will be further described below, and it should be noted that the present embodiment is based on the technical solution, and a detailed implementation manner and a specific operation process are provided, but the protection scope of the present invention is not limited to the present embodiment.
Example 1
The embodiment provides an interactive intelligent projection table lamp, which comprises a table lamp body, wherein a camera and a projection device are arranged on the table lamp body, and the shooting direction of the camera and the projection direction of the projection device are both right opposite to the same projection medium (such as a desktop, a wall surface and the like); the camera and the projection device are both in communication connection with a control module of the desk lamp body; the control module is connected with a projection switch, the projection switch is used for sending a switching signal of the projection device to the control module, and the control module controls the projection device to be switched on and off according to the switching signal;
the control module runs an operating system and a system-level touch recognition service, and the touch recognition service is used for recognizing the fingertip action of a user on a projection interface projected by the projection device through an image acquired by the camera, generating a touch message according to the fingertip action and transmitting the touch message to the operating system so that the operating system can execute corresponding control operation; the specific process is as follows:
the method comprises the steps that after a touch recognition service is started, a camera is opened, whether a projection interface appears on a projection medium or not is continuously recognized through an image shot by the camera, the projection interface is recognized and obtained, meanwhile, a rectangular area interfacial area (x, y, width, height) of the projection interface in the image shot by the camera is recognized (calculated according to picture pixels), the width represents the width of the rectangular area, the height represents the height of the rectangular area, and the (x, y) is the offset position of the rectangular area relative to the image shot by the camera, namely the position of which point in the image is the upper left corner of the rectangular area of the projection interface; dividing the width (pixel value) of a preset display interface by the width (pixel value) of the projection interface (even if no display screen exists, the display interface exists, and the touch recognition service projects the display interface through the projection device) to obtain the magnification of the projection interface, then identifying whether a fingertip exists in the projection interface and the fingertip is in contact with the projection medium, if the fingertip is in contact with the projection medium, acquiring the position (x ', y') of the fingertip relative to the projection interface, and generating a touch message, wherein the position coordinate of the touch message on the projection interface is (x ', y') (calculated according to picture pixels); the type of the touch message is divided into down pressing, move moving and up lifting, if the type of the last touch message is up and the fingertip and the projection medium are in contact, the type of the current touch message is judged to be down, if the type of the last touch message is down or move and the fingertip and the projection medium are still in contact, the type of the current touch message is judged to be move, and if the type of the last touch message is down or move and the fingertip and the projection medium are not in contact any more, the type of the current touch message is judged to be up; and after the type of the touch message and the position coordinate on the projection interface are obtained, dividing the position coordinate (x ', y') by the magnification of the projection interface to obtain the position coordinate of the touch message mapped on the display interface, and then forwarding the type of the touch message and the position coordinate of the touch message on the display interface to the operating system.
And after receiving the touch message, the operating system executes corresponding operation according to the position of the touch message on the display interface and the type of the touch message. For example, if a finger clicks an application icon of the projection interface, the application is entered, and the interface of the application is projected by the projection device; if the finger slides left and right on the projection interface, the operation of moving the display interface left and right is executed, the corresponding interface is switched to, and the image is projected through the projection device. It should be noted that, a complete touch generally starts from down to up, and the operating system determines the touch operation process of each user and makes corresponding control accordingly.
The process converts finger operation into touch operation, the recognition is faster than gesture recognition, and after all, the recognition gesture can be recognized only after a complete gesture action is finished. In addition, more accurate click-like operation can be supported.
Further, the camera and the projection device can be arranged to be adjustable in angle, so that the projection medium can be switched according to requirements. .
Further, the process of identifying the projected interface by the touch recognition service is as follows: and comparing the image acquired by the camera with a preset display interface image, and if the preset display interface image appears on the image acquired by the camera, identifying the projection interface.
Example 2
The embodiment provides a method for intelligently projecting a desk lamp according to the embodiment 1, which specifically comprises the following steps:
when the display interface is required to be projected, the projection switch is turned on, the operating system detects that the projection switch is turned on, the touch identification service is started, and the display interface is projected onto a projection medium by using the projection device;
the method comprises the steps that a camera is opened through touch recognition service, whether a projection interface exists on a projection medium or not is judged by using an image shot by the camera, the projection interface is recognized and obtained, meanwhile, an area (x, y, width, height) of the projection interface in the image shot by the camera is also recognized, the width of a preset display interface is divided by the area to obtain the magnification factor of the projection interface, and then whether a fingertip exists in the projection interface or not is recognized, and the fingertip is in contact with the projection medium; if the fingertip is in contact with the projection medium to indicate touch, acquiring the position (x ', y') of the fingertip relative to the projection interface, and generating a touch message, wherein the position coordinate of the touch message on the projection interface is (x ', y');
judging the type of the touch message, wherein the type of the touch message is divided into down pressing, move moving and up lifting, if the type of the last touch message is up and the fingertip and the projection medium are in contact, the type of the current touch message is judged to be down, if the type of the last touch message is down or move and the fingertip and the projection medium are still in contact, the type of the current touch message is judged to be move, and if the type of the last touch message is down or move and the fingertip and the projection medium are not in contact any more, the type of the current touch message is judged to be up;
after the type of the touch message and the position coordinate on the projection interface are obtained, dividing the position coordinate (x ', y') by the magnification factor of the projection interface to obtain the position coordinate of the touch message mapped on the display interface, and then forwarding the type of the touch message and the position coordinate of the touch message on the display interface to an operating system;
and after receiving the touch message, the operating system executes corresponding operation according to the position of the touch message on the display interface and the type of the touch message.
Various corresponding changes and modifications can be made by those skilled in the art based on the above technical solutions and concepts, and all such changes and modifications should be included in the protection scope of the present invention.
Claims (4)
1. An interactive intelligent projection table lamp is characterized by comprising a table lamp body, wherein a camera and a projection device are arranged on the table lamp body, and the shooting direction of the camera and the projection direction of the projection device are both opposite to the same projection medium; the camera and the projection device are both in communication connection with a control module of the desk lamp body; the control module is connected with a projection switch, the projection switch is used for sending a switching signal of the projection device to the control module, and the control module controls the projection device to be switched on and off according to the switching signal;
the control module runs an operating system and a system-level touch recognition service, and the touch recognition service is used for recognizing the fingertip action of a user on a projection interface projected by the projection device through an image acquired by the camera, generating a touch message according to the fingertip action and transmitting the touch message to the operating system so that the operating system can execute corresponding control operation; the specific process is as follows:
the method comprises the steps that after a touch recognition service is started, a camera is opened, whether a projection interface appears on a projection medium or not is continuously recognized through an image shot by the camera, the projection interface is recognized, meanwhile, an area interfacial area (x, y, width, height) of the projection interface in the image shot by the camera is recognized, the width represents the width of a rectangular area, the height represents the height of the rectangular area, and the (x, y) represents the offset position of the rectangular area relative to the image shot by the camera, namely the position of which point in the image is the upper left corner of the rectangular area of the projection interface; dividing the width of a preset display interface by the interface area to obtain the magnification of a projection interface, then identifying whether a fingertip appears in the projection interface and whether the fingertip is in contact with the projection medium, if the fingertip is in contact with the projection medium, indicating that a touch exists, acquiring the position (x ', y') of the fingertip relative to the projection interface, and generating a touch message, wherein the position coordinate of the touch message on the projection interface is (x ', y'); the type of the touch message is divided into down pressing, move moving and up lifting, if the type of the last touch message is up and the fingertip and the projection medium are in contact, the type of the current touch message is judged to be down, if the type of the last touch message is down or move and the fingertip and the projection medium are still in contact, the type of the current touch message is judged to be move, and if the type of the last touch message is down or move and the fingertip and the projection medium are not in contact any more, the type of the current touch message is judged to be up; after the type of the touch message and the position coordinate on the projection interface are obtained, dividing the position coordinate (x ', y') by the magnification factor of the projection interface to obtain the position coordinate of the touch message mapped on the display interface, and then forwarding the type of the touch message and the position coordinate of the touch message on the display interface to an operating system; and after receiving the touch message, the operating system executes corresponding operation according to the position of the touch message on the display interface and the type of the touch message.
2. The intelligent projection desk lamp of claim 1, wherein the process of identifying the projection interface by the touch recognition service is as follows: and comparing the image acquired by the camera with a preset display interface image, and if the preset display interface image appears on the image acquired by the camera, identifying the projection interface.
3. A method for using the intelligent projection table lamp as claimed in any one of claims 1-2, characterized in that the specific process is as follows:
when the display interface is required to be projected, the projection switch is turned on, the operating system detects that the projection switch is turned on, the touch identification service is started, and the display interface is projected onto a projection medium by using the projection device;
the method comprises the steps that a camera is opened through touch recognition service, whether a projection interface exists on a projection medium or not is judged by using an image shot by the camera, the projection interface is recognized and obtained, meanwhile, an area (x, y, width, height) of the projection interface in the image shot by the camera is also recognized, the width of a preset display interface is divided by the area to obtain the magnification factor of the projection interface, and then whether fingertips appear in the projection interface or not and whether the fingertips are in contact with the projection medium or not is recognized; if the fingertip is in contact with the projection medium to indicate touch, acquiring the position (x ', y') of the fingertip relative to the projection interface, and generating a touch message, wherein the position coordinate of the touch message on the projection interface is (x ', y');
judging the type of the touch message, wherein the type of the touch message is divided into down pressing, move moving and up lifting, if the type of the last touch message is up and the fingertip and the projection medium are in contact, the type of the current touch message is judged to be down, if the type of the last touch message is down or move and the fingertip and the projection medium are still in contact, the type of the current touch message is judged to be move, and if the type of the last touch message is down or move and the fingertip and the projection medium are not in contact any more, the type of the current touch message is judged to be up;
after the type of the touch message and the position coordinate on the projection interface are obtained, dividing the position coordinate (x ', y') by the magnification factor of the projection interface to obtain the position coordinate of the touch message mapped on the display interface, and then forwarding the type of the touch message and the position coordinate of the touch message on the display interface to an operating system;
and after receiving the touch message, the operating system executes corresponding operation according to the position of the touch message on the display interface and the type of the touch message.
4. The method of claim 3, wherein the touch recognition service recognizes the projected interface by: and comparing the image acquired by the camera with a preset display interface image, and if the preset display interface image appears on the image acquired by the camera, identifying the projection interface.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110790844.0A CN113531423A (en) | 2021-07-13 | 2021-07-13 | Interactive intelligent projection table lamp and method |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110790844.0A CN113531423A (en) | 2021-07-13 | 2021-07-13 | Interactive intelligent projection table lamp and method |
Publications (1)
Publication Number | Publication Date |
---|---|
CN113531423A true CN113531423A (en) | 2021-10-22 |
Family
ID=78127718
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202110790844.0A Pending CN113531423A (en) | 2021-07-13 | 2021-07-13 | Interactive intelligent projection table lamp and method |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN113531423A (en) |
Citations (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN1912816A (en) * | 2005-08-08 | 2007-02-14 | 北京理工大学 | Virtus touch screen system based on camera head |
CN106297211A (en) * | 2015-06-12 | 2017-01-04 | 上海渐华科技发展有限公司 | Sitting posture correcting system for prompting based on movable camera and projection light and application thereof |
CN205880834U (en) * | 2016-05-27 | 2017-01-11 | 北京小鸟看看科技有限公司 | Finger power input connector |
JP2017045332A (en) * | 2015-08-27 | 2017-03-02 | キヤノン株式会社 | Information processing device, control method thereof, program, and storage medium |
CN108874030A (en) * | 2018-04-27 | 2018-11-23 | 努比亚技术有限公司 | Wearable device operating method, wearable device and computer readable storage medium |
CN109032360A (en) * | 2018-08-30 | 2018-12-18 | 广东小天才科技有限公司 | A kind of method for controlling projection and intelligent desk lamp of intelligent desk lamp |
CN109829888A (en) * | 2018-12-27 | 2019-05-31 | 北京林业大学 | Wild animal monitoring analysis system and method based on depth convolutional neural networks |
CN110162173A (en) * | 2019-05-06 | 2019-08-23 | 上海翎腾智能科技有限公司 | A kind of gesture interaction method and intelligent desk lamp of the intelligent desk lamp based on AI |
CN110377250A (en) * | 2019-06-05 | 2019-10-25 | 华为技术有限公司 | A kind of touch control method and electronic equipment thrown under screen scene |
CN112286584A (en) * | 2020-10-22 | 2021-01-29 | 江西艾普若科技有限责任公司 | Mobile terminal screen control method |
-
2021
- 2021-07-13 CN CN202110790844.0A patent/CN113531423A/en active Pending
Patent Citations (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN1912816A (en) * | 2005-08-08 | 2007-02-14 | 北京理工大学 | Virtus touch screen system based on camera head |
CN106297211A (en) * | 2015-06-12 | 2017-01-04 | 上海渐华科技发展有限公司 | Sitting posture correcting system for prompting based on movable camera and projection light and application thereof |
JP2017045332A (en) * | 2015-08-27 | 2017-03-02 | キヤノン株式会社 | Information processing device, control method thereof, program, and storage medium |
CN205880834U (en) * | 2016-05-27 | 2017-01-11 | 北京小鸟看看科技有限公司 | Finger power input connector |
CN108874030A (en) * | 2018-04-27 | 2018-11-23 | 努比亚技术有限公司 | Wearable device operating method, wearable device and computer readable storage medium |
CN109032360A (en) * | 2018-08-30 | 2018-12-18 | 广东小天才科技有限公司 | A kind of method for controlling projection and intelligent desk lamp of intelligent desk lamp |
CN109829888A (en) * | 2018-12-27 | 2019-05-31 | 北京林业大学 | Wild animal monitoring analysis system and method based on depth convolutional neural networks |
CN110162173A (en) * | 2019-05-06 | 2019-08-23 | 上海翎腾智能科技有限公司 | A kind of gesture interaction method and intelligent desk lamp of the intelligent desk lamp based on AI |
CN110377250A (en) * | 2019-06-05 | 2019-10-25 | 华为技术有限公司 | A kind of touch control method and electronic equipment thrown under screen scene |
CN112286584A (en) * | 2020-10-22 | 2021-01-29 | 江西艾普若科技有限责任公司 | Mobile terminal screen control method |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US6512507B1 (en) | Pointing position detection device, presentation system, and method, and computer-readable medium | |
US9134814B2 (en) | Input device, display system and input method | |
CN103294280A (en) | Optical touch device, passive touch system and input detection method thereof | |
CN105260021A (en) | Intelligent interactive projection system | |
CN102945091B (en) | A kind of man-machine interaction method based on laser projection location and system | |
CN105677206A (en) | System and method for controlling head-up display based on vision | |
CN102436327A (en) | Screen input system and implementation method thereof | |
CN110488980B (en) | Human-computer interaction system of projection system | |
CN103049109B (en) | A kind of pointer and touch point identification method | |
CN102375564A (en) | Method for interacting by adopting optical indicator, optical indicator as well as presentation method and system | |
CN109871155A (en) | It is embedded into the radium-shine projection input scheme of mobile terminal device | |
CN113531423A (en) | Interactive intelligent projection table lamp and method | |
JP2017182109A (en) | Display system, information processing device, projector, and information processing method | |
CN112631445A (en) | Touch display system, touch display method, stylus and medium | |
US20200241721A1 (en) | Interactive display apparatus and method | |
CN202257528U (en) | Screen input system | |
CN2664064Y (en) | Pointer input device | |
CN201630600U (en) | Table with touch screen | |
CN109002198A (en) | projection device laser pen input system | |
CN105005434A (en) | Touch screen display method and system and applicable touch screen and touch screen device | |
CN105278760B (en) | Optical touch system | |
CN113531425B (en) | Intelligent projection desk lamp capable of achieving accurate interaction | |
CN216651666U (en) | Transparent desk of single table leg intelligence | |
CN106201020B (en) | A kind of method and system determining focusing by cursor flashing | |
CN107749928A (en) | The operating method and device of a kind of cell phone incoming call |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination |