US20190133081A1 - Remote interaction device with tracking of remote movement input - Google Patents
Remote interaction device with tracking of remote movement input Download PDFInfo
- Publication number
- US20190133081A1 US20190133081A1 US16/122,776 US201816122776A US2019133081A1 US 20190133081 A1 US20190133081 A1 US 20190133081A1 US 201816122776 A US201816122776 A US 201816122776A US 2019133081 A1 US2019133081 A1 US 2019133081A1
- Authority
- US
- United States
- Prior art keywords
- location
- remote
- movement input
- local device
- interaction device
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Images
Classifications
-
- A—HUMAN NECESSITIES
- A01—AGRICULTURE; FORESTRY; ANIMAL HUSBANDRY; HUNTING; TRAPPING; FISHING
- A01K—ANIMAL HUSBANDRY; CARE OF BIRDS, FISHES, INSECTS; FISHING; REARING OR BREEDING ANIMALS, NOT OTHERWISE PROVIDED FOR; NEW BREEDS OF ANIMALS
- A01K15/00—Devices for taming animals, e.g. nose-rings or hobbles; Devices for overturning animals in general; Training or exercising equipment; Covering boxes
- A01K15/02—Training or exercising equipment, e.g. mazes or labyrinths for animals ; Electric shock devices ; Toys specially adapted for animals
- A01K15/025—Toys specially adapted for animals
-
- A—HUMAN NECESSITIES
- A01—AGRICULTURE; FORESTRY; ANIMAL HUSBANDRY; HUNTING; TRAPPING; FISHING
- A01K—ANIMAL HUSBANDRY; CARE OF BIRDS, FISHES, INSECTS; FISHING; REARING OR BREEDING ANIMALS, NOT OTHERWISE PROVIDED FOR; NEW BREEDS OF ANIMALS
- A01K15/00—Devices for taming animals, e.g. nose-rings or hobbles; Devices for overturning animals in general; Training or exercising equipment; Covering boxes
- A01K15/02—Training or exercising equipment, e.g. mazes or labyrinths for animals ; Electric shock devices ; Toys specially adapted for animals
- A01K15/021—Electronic training devices specially adapted for dogs or cats
-
- F—MECHANICAL ENGINEERING; LIGHTING; HEATING; WEAPONS; BLASTING
- F21—LIGHTING
- F21V—FUNCTIONAL FEATURES OR DETAILS OF LIGHTING DEVICES OR SYSTEMS THEREOF; STRUCTURAL COMBINATIONS OF LIGHTING DEVICES WITH OTHER ARTICLES, NOT OTHERWISE PROVIDED FOR
- F21V14/00—Controlling the distribution of the light emitted by adjustment of elements
- F21V14/02—Controlling the distribution of the light emitted by adjustment of elements by movement of light sources
-
- F—MECHANICAL ENGINEERING; LIGHTING; HEATING; WEAPONS; BLASTING
- F21—LIGHTING
- F21V—FUNCTIONAL FEATURES OR DETAILS OF LIGHTING DEVICES OR SYSTEMS THEREOF; STRUCTURAL COMBINATIONS OF LIGHTING DEVICES WITH OTHER ARTICLES, NOT OTHERWISE PROVIDED FOR
- F21V21/00—Supporting, suspending, or attaching arrangements for lighting devices; Hand grips
- F21V21/14—Adjustable mountings
- F21V21/15—Adjustable mountings specially adapted for power operation, e.g. by remote control
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/76—Television signal recording
- H04N5/765—Interface circuits between an apparatus for recording and another apparatus
- H04N5/77—Interface circuits between an apparatus for recording and another apparatus between a recording apparatus and a television camera
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/76—Television signal recording
- H04N5/765—Interface circuits between an apparatus for recording and another apparatus
- H04N5/77—Interface circuits between an apparatus for recording and another apparatus between a recording apparatus and a television camera
- H04N5/772—Interface circuits between an apparatus for recording and another apparatus between a recording apparatus and a television camera the recording apparatus and the television camera being placed in the same enclosure
Definitions
- the subject matter described herein relates generally to a remote interaction device, and more particularly to a device at a first location which allows a user at a second, remote location to control a laser pointer at the first location with movement input.
- pet owners generally interact with their pets only when they are in the same general location, such as a home. Many pet owners are required to leave their pets alone and unsupervised for numerous hours every day when the pet owner goes to work, runs errands, or leaves town on trips or vacations. Some pets become bored, lethargic, or sedentary when left alone. This can lead to numerous health problems including obesity and depression. Alternatively, some pets become ornery and mischievous when left alone. This can lead to property damage, barking which irritates neighbors, and in extreme cases injury or death of the pet may occur.
- Pet sitters who may take care of pets while the pet owner is away. Pet sitters often charge an hourly fee and may do little more than feed the pet before leaving. In some cases the pet owner may never know that the pet sitter did not interact with the pet for more than a few minutes. Even in the case of a pet sitter who plays with the pet, the pet owner does not receive the direct benefit of interacting with the pet personally.
- the remote interaction device includes, among other components, a photonic emission device and photonic emission aiming device.
- the photonic emission device is generally a laser which can be controlled by a user at a remote location by issuing commands on a connected device. Accordingly, it would be desirable for the user to be able to control the laser using swipe and tap input at the connected device.
- the present invention is directed to a remote interaction device, and more particularly to a device at a first location which allows a user at a second, remote location to control a laser pointer at the first location with movement input.
- the movement input may include tap and swipe on a touch screen, movement of a computer mouse, pressing of keys on a keyboard, and so on.
- a remote interaction device In accordance with an example embodiment of the present invention, a remote interaction device is provided.
- the device generally has audio-visual recording and transmitting functionality to provide an operator at a remote location with an audio-visual feed of the environment near the device.
- the device also has a light emission component which the operator controls and which projects light onto a surface in the environment in the vicinity of the device.
- the systems, devices, and methods provide operators with the ability to control the positions of the light emission by tracking movement at a remote device at the remote location.
- a method of facilitating remote interaction includes recording visual data at a first location and transmitting the recorded visual data to a second location, receiving the visual data, that was recorded at the first location, at the second location and displaying the visual data on a local device at the second location, and tracking movement input positions on a display of the local device to control the aim of a photonic emission device at the first location.
- a system for facilitating remote interaction includes an interaction device located at a first location, a local device located at a second location, the local device connected to the interaction device over a network, and wherein the local device tracks movement input positions on a display of the local device to control the aim of a photonic emission device at the interaction device.
- FIG. 1 is a visual representation of an exemplary embodiment of a remote interaction system and subjects in accordance with an embodiment of the present invention.
- FIG. 2 is a diagram illustrating exemplary data transmission between a remote interaction device and a remote connected device in accordance with an embodiment of the present invention.
- FIG. 3 is a cutaway view of the front of a remote interaction device in accordance with an embodiment of the present invention.
- FIG. 4 is a schematic diagram of various modules of a remote interaction device and their relation to one another in accordance with an embodiment of the present invention.
- FIG. 5 is a visual portrayal of an exemplary user interface to interact with a remote interaction device in accordance with an embodiment of the present invention.
- FIG. 6 is another visual portrayal of an exemplary user interface to interact with a remote interaction device in accordance with an embodiment of the present invention.
- FIG. 7A is another visual portrayal of an exemplary user interface to interact with remote interaction device in accordance with an embodiment of the present invention.
- FIG. 7B is another visual portrayal of an exemplary user interface to interact with remote interaction device in accordance with an embodiment of the present invention.
- FIG. 8 is a diagram of various exemplary components of a remote connected device in accordance with an embodiment of the present invention.
- the term “and/or” placed between a first entity and a second entity means one of (1) the first entity, (2) the second entity, and (3) the first entity and the second entity.
- Multiple entities listed with “and/or” should be construed in the same manner, i.e., “one or more” of the entities so conjoined.
- Other entities may optionally be present other than the entities specifically identified by the “and/or” clause, whether related or unrelated to those entities specifically identified.
- a reference to “A and/or B”, when used in conjunction with open-ended language such as “comprising” can refer, in one embodiment, to A only (optionally including entities other than B); in another embodiment, to B only (optionally including entities other than A); in yet another embodiment, to both A and B (optionally including other entities).
- These entities may refer to elements, actions, structures, steps, operations, values, and the like.
- FIGS. 1-8 illustrate exemplary embodiments of an interaction device at a first location which allows a user at a second, remote location to control a laser pointer at the first location using movement input.
- the movement input may include tap and swipe on a touch screen, movement of a computer mouse, pressing of keys on a keyboard, and so on.
- the first location may be a geographic area where a pet is located, for example, in a room in a home.
- the interaction device may include data and video recording capability, and at least a pointer, for example, a laser pointer.
- the interaction device may record data at the first location and transmit data to a remote device at a second location which is geographically remote from the first location, for example, at an office remote from the home where the pet is located.
- a user at the second, remote location may use movement input at the remote device to remotely track and control the laser pointer in the interaction device.
- FIG. 1 is a visual representation of a remote interaction system and subjects in accordance with an example embodiment of the present invention.
- the subjects may be, for example, pets or people.
- FIG. 1 shows an operator 114 , a connected device 101 , pet(s) 111 , network 116 , and remote interaction device 100 .
- the network 116 may be, for example, the Internet and/or a cellular-based wireless network.
- the remote interaction device 100 may be wired or wirelessly connected to the network 116 , for example, via a wired or wireless local area network (LAN).
- LAN local area network
- the operator 114 is a pet owner or other human.
- the operator 114 may be able to interact with the pet 111 at a remote location, geographically remote from the pet 111 , by using the connected device 101 to monitor and control the remote interaction device 100 .
- Connected device 101 in the exemplary embodiment of FIG. 1 is shown as a tablet computer, however, in other embodiments, the connected device 101 may be a desktop, laptop, notebook computer, wearable computer such as a smart watch, glasses or contact lenses, a smart phone, or any other device with wireless network connectivity.
- the remote interaction device 100 and the connected device 101 may be connected to each other wirelessly over a wireless network connected to the network 116 , which may be the Internet and/or a cellular network.
- the network 116 may be a broadband wireless network or other high bandwidth packet switch network.
- Remote interaction device 100 in the example embodiment may be made of various modules and components which facilitate the operator 114 's interaction with the pet 111 .
- the remote interaction device 100 may connect to the network 116 using a wireless connection module 102 .
- the wireless connection module 102 and other modules and components of the remote interaction device 100 may receive power from a power module 103 .
- the power module 103 may receive power via a universal serial bus (USB) interface, although in other embodiments other interfaces may be used.
- the CPU module 104 is a central processing unit (CPU) which generally controls all systems and processes in the remote interaction device 100 .
- a microphone 108 and a camera module 107 provide for audio and visual data capture at the location of remote interaction device 100 and may transmit the captured data to allow the operator 114 to view and hear what is going on at the location of the remote interaction device 100 , using the connected device 101 .
- the remote interaction device 100 may record and store the captured data in a data storage (not shown).
- a laser positioning module 105 is operatively connected to laser beam 110 and controls its positioning.
- the casing 112 provides a protective housing for all components and modules of the remote interaction device 100 .
- the laser beam 110 and the speakers 106 may allow the operator 114 to have interaction with the location of remote interaction device 100 , thus providing visual stimulation and audio stimulation respectively for the pet 111 .
- FIG. 2 illustrates an exemplary communication exchange between a remote interaction device 100 and a connected device 101 .
- the remote interaction device 100 may record and transmit data and video of activities at the location of the remote interaction device 100 to the connected device 101 (e.g., through a server connected to the network 116 ).
- a user at the connected device 101 may use movement input which is transformed to relative coordinates and transmitted to the remote interaction device 100 to control a laser pointer of the remote interaction device 100 .
- FIG. 3 shows a cutaway view of the front of the remote interaction device 100 in accordance with an exemplary embodiment of the present invention.
- the remote interaction device 100 shown in FIG. 3 may generally include a camera module 107 , a laser pointer module 136 , a pan and tilt platform 138 , a microphone 108 , an RGB LED notifier 118 , and one or more speakers 106 .
- each of the components shown in FIG. 3 may be located at different locations within the device from those shown in the exemplary embodiment. But in general, the camera module 107 and laser pointer module 136 face outward from the same face of the device so as to allow the camera module 107 to record a video feed in the same direction that the laser pointer module 136 allows for interaction with the environment, such as shining on a surface.
- the camera module 107 in some embodiments may be a video recording device with a wide angle lens which allows for a video recording of the environment in front of camera module 107 .
- the camera module 107 may be a CMOS sensor and a camera lens as well as a printed circuit board (PCB).
- the camera module 107 may use other digital video recording devices or other appropriate video recording devices.
- a wide angle lens may be used in some embodiments to allow for video recording of the environment without the need to move the camera to follow particular subjects or specific locations within the field of view of the camera module 107 . In other embodiments, other appropriate lenses may be used.
- the camera module 107 may capture high definition (HD) video although in other embodiments, lower definition video may be captured.
- HD high definition
- the camera module 107 in some embodiments may have focusing capabilities which allow for focusing based on the distance of a subject from the camera module 107 .
- the focusing capability may be performed automatically by internal processing of a camera processor which is operable to process visual data signals from the camera module 107 .
- focusing may be performed manually by a user at a remote location by engaging an appropriate command on the connected device 101 .
- additional components may be provided in the camera module 107 such as camera aiming devices, alternate and/or changeable filters, and others which allow a user to view different areas of the room by positioning the direction of the camera and viewing through different filters.
- automatic motion-capture components may be used in order to direct the camera to capture movement in the environment such as movement of the pet 111 .
- the laser positioning module 105 in some embodiments may be made of a laser pointer module 136 which may be a laser pointer that emits light through optical amplification. Light emitted by the laser pointer module 136 may be directed to a specific location in the environment such as on a surface. Typical surfaces may be floors, furniture, walls, or other suitable surfaces. Many animals become interested in light such as lasers projected on surfaces. These animals will follow the light and try to catch it or capture it, providing entertainment for the animal. In some embodiments, the laser pointer module 136 uses a laser which is safe for use around humans and animals.
- the pan and tilt platform 138 in some embodiments may be a platform to which laser pointer module 136 is mounted.
- the pan and tilt platform 138 provides the mechanical support which controls the physical location that the laser module 136 is pointing a laser beam 110 .
- electromagnets may be used to control the panning and tilting of the pan and tilt platform 138 .
- the pan and tilt platform 138 will be described further herein and is also referred to as laser positioning device 600 .
- the microphone 108 in some embodiments may be a microphone which is operable to receive audio input signals from the environment such as barking from a dog, meowing from a cat, or other input signals.
- the microphone 108 may be coupled to a processor which is operable to recognize when a sound is made in the environment. In some embodiments, this may trigger processes within the remote interaction device 100 such as notifying the operator 114 via or at the connected device 101 that noise is being made near the remote interaction device 100 , beginning visual recording using the camera module 107 , or other processes.
- the RGB LED notifier 118 in some embodiment may be a light emitting diode (LED) which indicates the status of the remote interaction device 100 .
- status indications may include power, standby, transmit/receive, charging, or other suitable status indications.
- the RGB LED notifier 118 may indicate different device status in some embodiments by flashing, constant color display, alternating color display, or other suitable display methods.
- the RGB LED notifier 118 in some embodiments may be a single RGB LED. In other embodiments, the RGB LED notifier 118 may include multiple RGB LED's in various configurations.
- the speaker 106 in some embodiments may be a speaker device which outputs audio signals into the environment near the remote interaction device 100 .
- the speaker 106 in some embodiments may be operable to output audio signals such as a human voice, music, or other sounds received from the operator 114 via the connected device 101 over a wireless network connected to the Internet 116 and processed by an audio processor so as to communicate with the pet 111 near the remote interaction device 100 .
- multiple speakers may be used.
- FIG. 4 a diagram of the device modules 400 of the remote interaction device 100 is shown.
- a CPU module 402 controls a USB 404 , a Wi-Fi module 406 , a laser positioning module 408 , an RGB LED 410 , a laser 412 , a video encoder 414 , and an audio codec 418 .
- the audio codec 418 controls one or more speakers 420 and a microphone 422
- the video encoder controls the camera 416 .
- additional processing units may be used and additional other modules may be added to provide increased functionality.
- Modules as used herein should be understood to have broad functionality, such as the laser positioning module 408 being operable in different embodiments to control electromagnets, servo-motors, or other laser positioning devices as required by the particular embodiments. Additionally, modules may control more than one components, such as when multiple speakers are used.
- the operator 114 when the operator 114 first powers on the remote interaction device 100 , he/she must configure the device to communicate with a wireless network connected to the Internet 116 . This may be called first-time mode. In the first-time mode, the operator 114 may receive data about the network name and password, if required. After completion of the first-time mode process, the remote interaction device 100 may operate in normal operation mode. The operator 114 may not need to be a remote interaction device 100 owner, but the operator 114 may be any person who wishes to interact with the remote interaction device 100 s . User interfaces on the connected device 101 provide a way for the operator 114 to interact with and control the remote interaction device 100 .
- FIGS. 5-8 an exemplary interaction screen 500 user interface and movement input in accordance with some embodiments of the present invention are shown.
- several fields and buttons may be shown in a pop-up menu 510 .
- the pop-up menu 510 may be pulled out from the left side of the display on the connected device 101 , or by tapping on a short-cut (not shown).
- the fields and buttons may include, for example, back button 512 , tips button 514 , share button 516 , sound button 518 , and laser control 520 .
- the laser positioning module 105 may be made of the laser pointer module 136 which may be a laser pointer that emits light through optical amplification. Light emitted by the laser pointer module 136 may be directed to a specific location in the environment such as on a surface.
- the operator 114 may control the laser positioning module 105 of the remote device 100 and project the laser beam 110 to desired locations through input movement captured by a tracking source.
- the remote device 100 and the connected device 101 may be operated and function as follows.
- the operator 114 opens a software application provided on the connected device 101 to receive (e.g., through a server) a video stream (as shown in FIG. 5 ) from the remote device 100 , as captured by the camera module 107 .
- the operator 114 may tap on the laser control 520 .
- the operator 114 may touch the screen of the connected device 101 , for example, with a finger (see FIG. 7A ).
- the provided application reads the position of the finger in coordinates of the screen of the connected device 101 .
- the finger position is automatically transformed into video stream coordinates (e.g., at 1280 ⁇ 720 pixel).
- the position is then transformed to relative coordinates (e.g., [ ⁇ 1 . . . 1]).
- the position is then adjusted by selected calibration data, if any is available.
- the adjusted relative coordinates are then transmitted to the remote device 100 , for example, through the server.
- the remote device 100 receives and processes the coordinates, and points the laser beam 110 to the position in the platform corresponding to the position from the video (as shown in example in FIG. 6 ).
- the positions of the finger on the touch screen are tracked and processed as above, resulting in the laser beam 110 pointing to corresponding positions at the location of the remote device 100 (see, for example, FIG. 7B ).
- the connected device 101 may be a desktop, laptop, or notebook computer, wearable computer such as a smart watch, glasses or contact lenses, a smart phone, or any other device with wireless network connectivity.
- the connected device 101 may include corresponding tracking source, for example, to control the laser beam 110 .
- the tracking source may be a touch screen.
- the tracking source may be a mouse or a keyboard.
- Other suitable tracking sources are also contemplated and are not limited to the examples above.
- a tracking processor receives data from the tracking source to produce the tracking event.
- a video viewport converter and calibration converter may transform and adjust the tracking event to produce the calibrated coordinates as described above. The adjusted relative coordinates are then transmitted to the remote device 100 .
- the tracking processor, the video viewport converter, the calibration converter, and the network transmitter may be software, hardware, or a combination thereof.
Landscapes
- Life Sciences & Earth Sciences (AREA)
- Environmental Sciences (AREA)
- Engineering & Computer Science (AREA)
- Physical Education & Sports Medicine (AREA)
- Animal Behavior & Ethology (AREA)
- Zoology (AREA)
- Animal Husbandry (AREA)
- Biodiversity & Conservation Biology (AREA)
- General Health & Medical Sciences (AREA)
- Health & Medical Sciences (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- General Engineering & Computer Science (AREA)
- User Interface Of Digital Computer (AREA)
- Selective Calling Equipment (AREA)
Abstract
Systems, devices, and methods are provided for remote interaction with a subject in an environment. The device has audio-visual recording and transmitting functionality to provide an operator at a remote location with an audio-visual feed of the environment near the device. The device also has a light emission component which the operator controls and which projects light onto a surface in the environment in the vicinity of the device. The systems, devices, and methods provide operators with the ability to control the positions of the light emission by tracking movement at a remote device at the remote location.
Description
- The present application is a continuation of U.S. patent application Ser. No. 15/352,270, filed Nov. 15, 2016, which claims the benefit of priority to U.S. Provisional Application No. 62/257,436, filed Nov. 19, 2015, both of which are hereby incorporated by reference in their entireties for all purposes.
- The subject matter described herein relates generally to a remote interaction device, and more particularly to a device at a first location which allows a user at a second, remote location to control a laser pointer at the first location with movement input.
- Presently, pet owners generally interact with their pets only when they are in the same general location, such as a home. Many pet owners are required to leave their pets alone and unsupervised for numerous hours every day when the pet owner goes to work, runs errands, or leaves town on trips or vacations. Some pets become bored, lethargic, or sedentary when left alone. This can lead to numerous health problems including obesity and depression. Alternatively, some pets become ornery and mischievous when left alone. This can lead to property damage, barking which irritates neighbors, and in extreme cases injury or death of the pet may occur.
- One attempted solution to a lack of interaction and stimulation for pets has been to hire pet sitters who may take care of pets while the pet owner is away. Pet sitters often charge an hourly fee and may do little more than feed the pet before leaving. In some cases the pet owner may never know that the pet sitter did not interact with the pet for more than a few minutes. Even in the case of a pet sitter who plays with the pet, the pet owner does not receive the direct benefit of interacting with the pet personally.
- Other attempted solutions have included leaving televisions or radios on for the pet while the pet owner is away, attempting to use automatically controlled toys, electroshock punishment for misbehaving, and passive surveillance systems which provide one-directional monitoring of the pet. Each of these passive and active systems has its own drawbacks ranging from being inefficient to inhumane.
- Accordingly, to overcome the above and other problems, a remote interaction device for interacting with pets was proposed in U.S. patent application Ser. No. 14/186,793, Pub. No. US 2014/0233906 A1, to Neskin, et al, the entire content and disclosure of which are herein incorporated by reference. The remote interaction device includes, among other components, a photonic emission device and photonic emission aiming device. The photonic emission device is generally a laser which can be controlled by a user at a remote location by issuing commands on a connected device. Accordingly, it would be desirable for the user to be able to control the laser using swipe and tap input at the connected device.
- The present invention is directed to a remote interaction device, and more particularly to a device at a first location which allows a user at a second, remote location to control a laser pointer at the first location with movement input. The movement input may include tap and swipe on a touch screen, movement of a computer mouse, pressing of keys on a keyboard, and so on.
- In accordance with an example embodiment of the present invention, a remote interaction device is provided. The device generally has audio-visual recording and transmitting functionality to provide an operator at a remote location with an audio-visual feed of the environment near the device. The device also has a light emission component which the operator controls and which projects light onto a surface in the environment in the vicinity of the device. The systems, devices, and methods provide operators with the ability to control the positions of the light emission by tracking movement at a remote device at the remote location.
- In some embodiments, a method of facilitating remote interaction is provided. The method includes recording visual data at a first location and transmitting the recorded visual data to a second location, receiving the visual data, that was recorded at the first location, at the second location and displaying the visual data on a local device at the second location, and tracking movement input positions on a display of the local device to control the aim of a photonic emission device at the first location.
- In some embodiments, a system for facilitating remote interaction is provided. The system includes an interaction device located at a first location, a local device located at a second location, the local device connected to the interaction device over a network, and wherein the local device tracks movement input positions on a display of the local device to control the aim of a photonic emission device at the interaction device.
- Other systems, devices, methods, features and advantages of the subject matter described herein will be or will become apparent to one with skill in the art upon examination of the following figures and detailed description. It is intended that all such additional systems, devices, methods, features and advantages be included within this description, be within the scope of the subject matter described herein, and be protected by the accompanying claims. In no way should the features of the example embodiments be construed as limiting the appended claims, absent express recitation of those features in the claims.
- The details of the subject matter set forth herein, both as to its structure and operation, may be apparent by study of the accompanying figures, in which like reference numerals refer to like parts. The components in the figures are not necessarily to scale, emphasis instead being placed upon illustrating the principles of the subject matter. Moreover, all illustrations are intended to convey concepts, where relative sizes, shapes and other detailed attributes may be illustrated schematically rather than literally or precisely.
-
FIG. 1 is a visual representation of an exemplary embodiment of a remote interaction system and subjects in accordance with an embodiment of the present invention. -
FIG. 2 is a diagram illustrating exemplary data transmission between a remote interaction device and a remote connected device in accordance with an embodiment of the present invention. -
FIG. 3 is a cutaway view of the front of a remote interaction device in accordance with an embodiment of the present invention. -
FIG. 4 is a schematic diagram of various modules of a remote interaction device and their relation to one another in accordance with an embodiment of the present invention. -
FIG. 5 is a visual portrayal of an exemplary user interface to interact with a remote interaction device in accordance with an embodiment of the present invention. -
FIG. 6 is another visual portrayal of an exemplary user interface to interact with a remote interaction device in accordance with an embodiment of the present invention. -
FIG. 7A is another visual portrayal of an exemplary user interface to interact with remote interaction device in accordance with an embodiment of the present invention. -
FIG. 7B is another visual portrayal of an exemplary user interface to interact with remote interaction device in accordance with an embodiment of the present invention. -
FIG. 8 is a diagram of various exemplary components of a remote connected device in accordance with an embodiment of the present invention. - Before the present subject matter is described in detail, it is to be understood that this disclosure is not limited to the particular embodiments described, as such may, of course, vary. It is also to be understood that the terminology used herein is for the purpose of describing particular embodiments only, and is not intended to be limiting, since the scope of the present disclosure will be limited only by the appended claims.
- As used herein and in the appended claims, the singular forms “a”, “an”, and “the” include plural referents unless the context clearly dictates otherwise.
- In the following description and in the figures, like elements are identified with like reference numerals. The use of “e.g.,” “etc,” and “or” indicates non-exclusive alternatives without limitation, unless otherwise noted. The use of “including” or “includes” means “including, but not limited to,” or “includes, but not limited to,” unless otherwise noted.
- As used herein, the term “and/or” placed between a first entity and a second entity means one of (1) the first entity, (2) the second entity, and (3) the first entity and the second entity. Multiple entities listed with “and/or” should be construed in the same manner, i.e., “one or more” of the entities so conjoined. Other entities may optionally be present other than the entities specifically identified by the “and/or” clause, whether related or unrelated to those entities specifically identified. Thus, as a non-limiting example, a reference to “A and/or B”, when used in conjunction with open-ended language such as “comprising” can refer, in one embodiment, to A only (optionally including entities other than B); in another embodiment, to B only (optionally including entities other than A); in yet another embodiment, to both A and B (optionally including other entities). These entities may refer to elements, actions, structures, steps, operations, values, and the like.
- The publications discussed herein are provided solely for their disclosure prior to the filing date of the present application. Nothing herein is to be construed as an admission that the present disclosure is not entitled to antedate such publication by virtue of prior disclosure. Further, the dates of publication provided may be different from the actual publication dates which may need to be independently confirmed.
- It should be noted that all features, elements, components, functions, and steps described with respect to any embodiment provided herein are intended to be freely combinable and substitutable with those from any other embodiment. If a certain feature, element, component, function, or step is described with respect to only one embodiment, then it should be understood that that feature, element, component, function, or step can be used with every other embodiment described herein unless explicitly stated otherwise. This paragraph therefore serves as antecedent basis and written support for the introduction of claims, at any time, that combine features, elements, components, functions, and steps from different embodiments, or that substitute features, elements, components, functions, and steps from one embodiment with those of another, even if the following description does not explicitly state, in a particular instance, that such combinations or substitutions are possible. It is explicitly acknowledged that express recitation of every possible combination and substitution is overly burdensome, especially given that the permissibility of each and every such combination and substitution will be readily recognized by those of ordinary skill in the art.
- Turning now to the drawings,
FIGS. 1-8 illustrate exemplary embodiments of an interaction device at a first location which allows a user at a second, remote location to control a laser pointer at the first location using movement input. The movement input may include tap and swipe on a touch screen, movement of a computer mouse, pressing of keys on a keyboard, and so on. The first location may be a geographic area where a pet is located, for example, in a room in a home. The interaction device may include data and video recording capability, and at least a pointer, for example, a laser pointer. The interaction device may record data at the first location and transmit data to a remote device at a second location which is geographically remote from the first location, for example, at an office remote from the home where the pet is located. A user at the second, remote location may use movement input at the remote device to remotely track and control the laser pointer in the interaction device. -
FIG. 1 is a visual representation of a remote interaction system and subjects in accordance with an example embodiment of the present invention. The subjects may be, for example, pets or people. Generally,FIG. 1 shows anoperator 114, aconnected device 101, pet(s) 111,network 116, andremote interaction device 100. Thenetwork 116 may be, for example, the Internet and/or a cellular-based wireless network. Theremote interaction device 100 may be wired or wirelessly connected to thenetwork 116, for example, via a wired or wireless local area network (LAN). - In the exemplary embodiment shown in
FIG. 1 , theoperator 114 is a pet owner or other human. Theoperator 114 may be able to interact with thepet 111 at a remote location, geographically remote from thepet 111, by using the connecteddevice 101 to monitor and control theremote interaction device 100.Connected device 101 in the exemplary embodiment ofFIG. 1 is shown as a tablet computer, however, in other embodiments, theconnected device 101 may be a desktop, laptop, notebook computer, wearable computer such as a smart watch, glasses or contact lenses, a smart phone, or any other device with wireless network connectivity. Theremote interaction device 100 and theconnected device 101 may be connected to each other wirelessly over a wireless network connected to thenetwork 116, which may be the Internet and/or a cellular network. In some embodiments, thenetwork 116 may be a broadband wireless network or other high bandwidth packet switch network. -
Remote interaction device 100 in the example embodiment may be made of various modules and components which facilitate theoperator 114's interaction with thepet 111. In some embodiments, theremote interaction device 100 may connect to thenetwork 116 using awireless connection module 102. Thewireless connection module 102 and other modules and components of theremote interaction device 100 may receive power from apower module 103. In some embodiments, thepower module 103 may receive power via a universal serial bus (USB) interface, although in other embodiments other interfaces may be used. TheCPU module 104 is a central processing unit (CPU) which generally controls all systems and processes in theremote interaction device 100. - A
microphone 108 and acamera module 107 provide for audio and visual data capture at the location ofremote interaction device 100 and may transmit the captured data to allow theoperator 114 to view and hear what is going on at the location of theremote interaction device 100, using the connecteddevice 101. In some embodiments, theremote interaction device 100 may record and store the captured data in a data storage (not shown). Alaser positioning module 105 is operatively connected tolaser beam 110 and controls its positioning. Thecasing 112 provides a protective housing for all components and modules of theremote interaction device 100. In some exemplary operations, thelaser beam 110 and thespeakers 106 may allow theoperator 114 to have interaction with the location ofremote interaction device 100, thus providing visual stimulation and audio stimulation respectively for thepet 111. -
FIG. 2 illustrates an exemplary communication exchange between aremote interaction device 100 and aconnected device 101. As described herein, theremote interaction device 100 may record and transmit data and video of activities at the location of theremote interaction device 100 to the connected device 101 (e.g., through a server connected to the network 116). A user at theconnected device 101 may use movement input which is transformed to relative coordinates and transmitted to theremote interaction device 100 to control a laser pointer of theremote interaction device 100. -
FIG. 3 shows a cutaway view of the front of theremote interaction device 100 in accordance with an exemplary embodiment of the present invention. Theremote interaction device 100 shown inFIG. 3 may generally include acamera module 107, alaser pointer module 136, a pan andtilt platform 138, amicrophone 108, anRGB LED notifier 118, and one ormore speakers 106. - It should be noted that each of the components shown in
FIG. 3 may be located at different locations within the device from those shown in the exemplary embodiment. But in general, thecamera module 107 andlaser pointer module 136 face outward from the same face of the device so as to allow thecamera module 107 to record a video feed in the same direction that thelaser pointer module 136 allows for interaction with the environment, such as shining on a surface. - The
camera module 107 in some embodiments may be a video recording device with a wide angle lens which allows for a video recording of the environment in front ofcamera module 107. In some embodiments, thecamera module 107 may be a CMOS sensor and a camera lens as well as a printed circuit board (PCB). In other embodiments, thecamera module 107 may use other digital video recording devices or other appropriate video recording devices. A wide angle lens may be used in some embodiments to allow for video recording of the environment without the need to move the camera to follow particular subjects or specific locations within the field of view of thecamera module 107. In other embodiments, other appropriate lenses may be used. - In some embodiments, the
camera module 107 may capture high definition (HD) video although in other embodiments, lower definition video may be captured. - The
camera module 107 in some embodiments may have focusing capabilities which allow for focusing based on the distance of a subject from thecamera module 107. In some embodiments, the focusing capability may be performed automatically by internal processing of a camera processor which is operable to process visual data signals from thecamera module 107. In some embodiments, focusing may be performed manually by a user at a remote location by engaging an appropriate command on theconnected device 101. - In some embodiments, additional components may be provided in the
camera module 107 such as camera aiming devices, alternate and/or changeable filters, and others which allow a user to view different areas of the room by positioning the direction of the camera and viewing through different filters. In some embodiments, automatic motion-capture components may be used in order to direct the camera to capture movement in the environment such as movement of thepet 111. - The
laser positioning module 105 in some embodiments may be made of alaser pointer module 136 which may be a laser pointer that emits light through optical amplification. Light emitted by thelaser pointer module 136 may be directed to a specific location in the environment such as on a surface. Typical surfaces may be floors, furniture, walls, or other suitable surfaces. Many animals become interested in light such as lasers projected on surfaces. These animals will follow the light and try to catch it or capture it, providing entertainment for the animal. In some embodiments, thelaser pointer module 136 uses a laser which is safe for use around humans and animals. - The pan and
tilt platform 138 in some embodiments may be a platform to whichlaser pointer module 136 is mounted. The pan andtilt platform 138 provides the mechanical support which controls the physical location that thelaser module 136 is pointing alaser beam 110. In some embodiments, electromagnets may be used to control the panning and tilting of the pan andtilt platform 138. The pan andtilt platform 138 will be described further herein and is also referred to as laser positioning device 600. - The
microphone 108 in some embodiments may be a microphone which is operable to receive audio input signals from the environment such as barking from a dog, meowing from a cat, or other input signals. In some embodiments, themicrophone 108 may be coupled to a processor which is operable to recognize when a sound is made in the environment. In some embodiments, this may trigger processes within theremote interaction device 100 such as notifying theoperator 114 via or at theconnected device 101 that noise is being made near theremote interaction device 100, beginning visual recording using thecamera module 107, or other processes. - The
RGB LED notifier 118 in some embodiment may be a light emitting diode (LED) which indicates the status of theremote interaction device 100. In some embodiments, status indications may include power, standby, transmit/receive, charging, or other suitable status indications. TheRGB LED notifier 118 may indicate different device status in some embodiments by flashing, constant color display, alternating color display, or other suitable display methods. TheRGB LED notifier 118 in some embodiments may be a single RGB LED. In other embodiments, theRGB LED notifier 118 may include multiple RGB LED's in various configurations. - The
speaker 106 in some embodiments may be a speaker device which outputs audio signals into the environment near theremote interaction device 100. Thespeaker 106 in some embodiments may be operable to output audio signals such as a human voice, music, or other sounds received from theoperator 114 via theconnected device 101 over a wireless network connected to theInternet 116 and processed by an audio processor so as to communicate with thepet 111 near theremote interaction device 100. In some embodiments, multiple speakers may be used. - Turning to
FIG. 4 , in accordance with some embodiments of the present invention, a diagram of thedevice modules 400 of theremote interaction device 100 is shown. As shown, aCPU module 402 controls aUSB 404, a Wi-Fi module 406, alaser positioning module 408, anRGB LED 410, alaser 412, avideo encoder 414, and anaudio codec 418. In turn, theaudio codec 418 controls one ormore speakers 420 and amicrophone 422, and the video encoder controls thecamera 416. In some embodiments, additional processing units may be used and additional other modules may be added to provide increased functionality. Modules as used herein should be understood to have broad functionality, such as thelaser positioning module 408 being operable in different embodiments to control electromagnets, servo-motors, or other laser positioning devices as required by the particular embodiments. Additionally, modules may control more than one components, such as when multiple speakers are used. - Generally, in some exemplary operations, when the
operator 114 first powers on theremote interaction device 100, he/she must configure the device to communicate with a wireless network connected to theInternet 116. This may be called first-time mode. In the first-time mode, theoperator 114 may receive data about the network name and password, if required. After completion of the first-time mode process, theremote interaction device 100 may operate in normal operation mode. Theoperator 114 may not need to be aremote interaction device 100 owner, but theoperator 114 may be any person who wishes to interact with the remote interaction device 100 s. User interfaces on theconnected device 101 provide a way for theoperator 114 to interact with and control theremote interaction device 100. - Turning to
FIGS. 5-8 , anexemplary interaction screen 500 user interface and movement input in accordance with some embodiments of the present invention are shown. In some embodiments, several fields and buttons may be shown in a pop-upmenu 510. The pop-upmenu 510 may be pulled out from the left side of the display on theconnected device 101, or by tapping on a short-cut (not shown). The fields and buttons may include, for example,back button 512,tips button 514,share button 516,sound button 518, andlaser control 520. - As mentioned above in
FIGS. 1 and 3 , thelaser positioning module 105 may be made of thelaser pointer module 136 which may be a laser pointer that emits light through optical amplification. Light emitted by thelaser pointer module 136 may be directed to a specific location in the environment such as on a surface. In some embodiments of theinteraction screen 500 user interface, theoperator 114 may control thelaser positioning module 105 of theremote device 100 and project thelaser beam 110 to desired locations through input movement captured by a tracking source. - As shown in
FIGS. 5, 6, 7A and 7B , in some exemplary embodiments, theremote device 100 and theconnected device 101 may be operated and function as follows. Theoperator 114 opens a software application provided on theconnected device 101 to receive (e.g., through a server) a video stream (as shown inFIG. 5 ) from theremote device 100, as captured by thecamera module 107. To turn on/off the laser at theremote device 100, theoperator 114 may tap on thelaser control 520. When the laser at theremote device 100 has been turned on, theoperator 114 may touch the screen of theconnected device 101, for example, with a finger (seeFIG. 7A ). As theoperator 114 touches a position on the touch screen (e.g., location A), the provided application reads the position of the finger in coordinates of the screen of theconnected device 101. The finger position is automatically transformed into video stream coordinates (e.g., at 1280×720 pixel). The position is then transformed to relative coordinates (e.g., [−1 . . . 1]). The position is then adjusted by selected calibration data, if any is available. The adjusted relative coordinates are then transmitted to theremote device 100, for example, through the server. Theremote device 100 receives and processes the coordinates, and points thelaser beam 110 to the position in the platform corresponding to the position from the video (as shown in example inFIG. 6 ). As theoperator 114 performs a movement input, for example, moves (or swipes) his/her finger, the positions of the finger on the touch screen are tracked and processed as above, resulting in thelaser beam 110 pointing to corresponding positions at the location of the remote device 100 (see, for example,FIG. 7B ). - It should be noted that although the above example shows the
connected device 101 with a touch screen for tracking the positions of a finger, other tracking sources may be used when theconnected device 101 does not have a touch screen as discussed herein. - Turning to
FIG. 8 , a diagram 800 of theconnected device 101 in accordance with some embodiments of the present invention is shown. As mentioned herein, theconnected device 101 may be a desktop, laptop, or notebook computer, wearable computer such as a smart watch, glasses or contact lenses, a smart phone, or any other device with wireless network connectivity. Accordingly, theconnected device 101 may include corresponding tracking source, for example, to control thelaser beam 110. For example, when theconnected device 101 is a smart phone or a tablet, the tracking source may be a touch screen. When theconnected device 101 is a desktop or laptop, the tracking source may be a mouse or a keyboard. Other suitable tracking sources are also contemplated and are not limited to the examples above. A tracking processor receives data from the tracking source to produce the tracking event. A video viewport converter and calibration converter may transform and adjust the tracking event to produce the calibrated coordinates as described above. The adjusted relative coordinates are then transmitted to theremote device 100. It is noted that the tracking processor, the video viewport converter, the calibration converter, and the network transmitter may be software, hardware, or a combination thereof. - While embodiments of the present invention have been shown and described, various modifications may be made without departing from the spirit and scope of the present invention, and all such modifications and equivalents are intended to be covered.
- In many instances entities are described herein as being coupled to other entities. It should be understood that the terms “coupled” and “connected” (or any of their forms) are used interchangeably herein and, in both cases, are generic to the direct coupling of two entities (without any non-negligible (e.g., parasitic) intervening entities) and the indirect coupling of two entities (with one or more non-negligible intervening entities). Where entities are shown as being directly coupled together, or described as coupled together without description of any intervening entity, it should be understood that those entities can be indirectly coupled together as well unless the context clearly dictates otherwise.
- While the embodiments are susceptible to various modifications and alternative forms, specific examples thereof have been shown in the drawings and are herein described in detail. It should be understood, however, that these embodiments are not to be limited to the particular form disclosed, but to the contrary, these embodiments are to cover all modifications, equivalents, and alternatives falling within the spirit of the disclosure. Furthermore, any features, functions, steps, or elements of the embodiments may be recited in or added to the claims, as well as negative limitations that define the inventive scope of the claims by features, functions, steps, or elements that are not within that scope.
Claims (14)
1. A method of facilitating remote interaction comprising:
recording visual data at a first location and transmitting the recorded visual data to a second location;
receiving the visual data, that was recorded at the first location, at the second location and displaying the visual data on a local device at the second location; and
tracking movement input positions on a display of the local device to control the aim of a photonic emission device at the first location.
2. The method of claim 1 , wherein the movement input positions are automatically transformed, at the local device at the second location, into video stream coordinates.
3. The method of claim 1 , wherein the movement input positions are transformed, at the local device at the second location, to relative coordinates.
4. The method of claim 3 , wherein the relative coordinates are further adjusted, at the local device at the second location, by selected calibration data.
5. The method of claim 4 , wherein the adjusted relative coordinates are transmitted to the first location.
6. The method of claim 1 , wherein the photonic emission device is a laser positioning device.
7. The method of claim 1 , wherein the movement input positions include touch and swipe movement on a touch screen display.
8. A system for facilitating remote interaction comprising:
an interaction device located at a first location;
a local device located at a second location, the local device connected to the interaction device over a network; and
wherein the local device tracks movement input positions on a display of the local device to control the aim of a photonic emission device at the interaction device.
9. The system of claim 8 , wherein the photonic emission device is a laser positioning device.
10. The system of claim 8 , wherein the movement input positions are automatically transformed, at the local device at the second location, into video stream coordinates.
11. The system of claim 8 , wherein the movement input positions are transformed, at the local device at the second location, to relative coordinates.
12. The system of claim 11 , wherein the relative coordinates are further adjusted, at the local device at the second location, by selected calibration data.
13. The system of claim 12 , wherein the adjusted relative coordinates are transmitted to the first location.
14. The system of claim 8 , wherein the movement input positions include touch and swipe movement on a touch screen display.
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US16/122,776 US20190133081A1 (en) | 2015-11-19 | 2018-09-05 | Remote interaction device with tracking of remote movement input |
US16/664,691 US20200154675A1 (en) | 2015-11-19 | 2019-10-25 | Remote interaction device with tracking of remote movement input |
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201562257436P | 2015-11-19 | 2015-11-19 | |
US15/352,270 US10085423B2 (en) | 2015-11-19 | 2016-11-15 | Remote interaction device with tracking of remote movement input |
US16/122,776 US20190133081A1 (en) | 2015-11-19 | 2018-09-05 | Remote interaction device with tracking of remote movement input |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/352,270 Continuation US10085423B2 (en) | 2015-11-19 | 2016-11-15 | Remote interaction device with tracking of remote movement input |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US16/664,691 Continuation US20200154675A1 (en) | 2015-11-19 | 2019-10-25 | Remote interaction device with tracking of remote movement input |
Publications (1)
Publication Number | Publication Date |
---|---|
US20190133081A1 true US20190133081A1 (en) | 2019-05-09 |
Family
ID=58717741
Family Applications (3)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/352,270 Expired - Fee Related US10085423B2 (en) | 2015-11-19 | 2016-11-15 | Remote interaction device with tracking of remote movement input |
US16/122,776 Abandoned US20190133081A1 (en) | 2015-11-19 | 2018-09-05 | Remote interaction device with tracking of remote movement input |
US16/664,691 Abandoned US20200154675A1 (en) | 2015-11-19 | 2019-10-25 | Remote interaction device with tracking of remote movement input |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/352,270 Expired - Fee Related US10085423B2 (en) | 2015-11-19 | 2016-11-15 | Remote interaction device with tracking of remote movement input |
Family Applications After (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US16/664,691 Abandoned US20200154675A1 (en) | 2015-11-19 | 2019-10-25 | Remote interaction device with tracking of remote movement input |
Country Status (3)
Country | Link |
---|---|
US (3) | US10085423B2 (en) |
EP (1) | EP3376855A4 (en) |
WO (1) | WO2017087412A1 (en) |
Families Citing this family (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20220415476A1 (en) * | 2012-06-14 | 2022-12-29 | Medibotics Llc | Wearable Device and System for Nutritional Intake Monitoring and Management |
AU2018245330B2 (en) * | 2017-03-30 | 2020-04-02 | Komatsu Ltd. | Control system for work vehicle, method for setting trajectory of work implement, and work vehicle |
US10306362B1 (en) * | 2017-04-20 | 2019-05-28 | Dynamount, Llc | Microphone remote positioning, amplification, and distribution systems and methods |
RU2666517C1 (en) * | 2017-06-09 | 2018-09-07 | Общество с ограниченной ответственностью "АЭЛИНА ЛС" | Method of rehabilitation of patients with impaired motor function using an automated information system for the implementation of audiovisual biofeedback |
KR20190002172A (en) * | 2017-06-29 | 2019-01-08 | 홍영준 | Rotary apparatus for throwing the feed |
CN107455282B (en) * | 2017-09-18 | 2018-07-03 | 尹美川 | The method and apparatus of funny cat, household electrical appliance |
CN108965815A (en) * | 2018-07-27 | 2018-12-07 | 合肥阅辞科技有限公司 | A kind of remote indication system and indicating means |
Family Cites Families (17)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6099522A (en) * | 1989-02-06 | 2000-08-08 | Visx Inc. | Automated laser workstation for high precision surgical and industrial interventions |
US6094189A (en) * | 1998-04-17 | 2000-07-25 | Quillen; Wendell A. | Visual echo remote laser pointer |
JP2002171460A (en) | 2000-11-30 | 2002-06-14 | Sony Corp | Reproducing device |
US8046701B2 (en) * | 2003-08-07 | 2011-10-25 | Fuji Xerox Co., Ltd. | Peer to peer gesture based modular presentation system |
US7409924B2 (en) | 2004-07-15 | 2008-08-12 | Lawrence Kates | Training, management, and/or entertainment system for canines, felines, or other animals |
JP4339836B2 (en) | 2005-09-26 | 2009-10-07 | 株式会社コナミスポーツ&ライフ | Optical pet guidance device |
US20080180537A1 (en) | 2006-11-14 | 2008-07-31 | Uri Weinberg | Camera system and methods |
US7658694B2 (en) | 2007-04-30 | 2010-02-09 | Nike, Inc. | Adaptive training system |
JP5684577B2 (en) | 2008-02-27 | 2015-03-11 | ソニー コンピュータ エンタテインメント アメリカ リミテッド ライアビリテイ カンパニー | How to capture scene depth data and apply computer actions |
EP2281230A1 (en) | 2008-04-10 | 2011-02-09 | Karl Christopher Hansen | Simple-to-use optical wireless remote control |
US8644688B2 (en) | 2008-08-26 | 2014-02-04 | Opentv, Inc. | Community-based recommendation engine |
WO2010108186A1 (en) | 2009-03-20 | 2010-09-23 | Georgia Tech Research Corporation | Methods and apparatuses for using a mobile device to provide remote assistance |
GB201003933D0 (en) | 2010-03-09 | 2010-04-21 | Chamberlain David | Animal exercise apparatus |
US20130068173A1 (en) * | 2011-09-12 | 2013-03-21 | George R. Jamison | Internet communication system for pets |
US8866870B1 (en) * | 2011-10-20 | 2014-10-21 | Lockheed Martin Corporation | Methods, apparatus, and systems for controlling from a first location a laser at a second location |
CA2902194A1 (en) * | 2013-02-21 | 2014-08-28 | Petcube, Inc. | Remote interaction device |
US9848578B2 (en) * | 2013-03-15 | 2017-12-26 | Lee Miller | Toy and app for remotely viewing and playing with a pet |
-
2016
- 2016-11-15 US US15/352,270 patent/US10085423B2/en not_active Expired - Fee Related
- 2016-11-15 EP EP16866964.6A patent/EP3376855A4/en not_active Withdrawn
- 2016-11-15 WO PCT/US2016/062081 patent/WO2017087412A1/en active Application Filing
-
2018
- 2018-09-05 US US16/122,776 patent/US20190133081A1/en not_active Abandoned
-
2019
- 2019-10-25 US US16/664,691 patent/US20200154675A1/en not_active Abandoned
Also Published As
Publication number | Publication date |
---|---|
EP3376855A4 (en) | 2019-07-03 |
US20170142933A1 (en) | 2017-05-25 |
US20200154675A1 (en) | 2020-05-21 |
WO2017087412A1 (en) | 2017-05-26 |
US10085423B2 (en) | 2018-10-02 |
EP3376855A1 (en) | 2018-09-26 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US10085423B2 (en) | Remote interaction device with tracking of remote movement input | |
US10251370B2 (en) | Remote interaction device | |
US20210154850A1 (en) | Omni-Directional Mobile Robot for Remote Users | |
US9973846B2 (en) | Microphone array, monitoring system, and sound pickup setting method | |
US10574942B2 (en) | Systems and methods for virtual co-location | |
US8994776B2 (en) | Customizable robotic system | |
EP3776008B1 (en) | Method and electronic device for ultrasonic sensing in smart devices | |
US20140149546A1 (en) | Remote communication system for communication with companion animal and remote communication method with companion animal using the same | |
US20140267720A1 (en) | Toy and App for Remotely Viewing and Playing with a Pet | |
US20180054228A1 (en) | Teleoperated electronic device holder | |
JP2020170916A (en) | Information processor, and information processing method | |
WO2020116233A1 (en) | Information processing device, information processing method, and program | |
US11416002B1 (en) | Robotic vacuum with mobile security function | |
JP2019106908A (en) | Animal management system, program and computer readable memory medium | |
CN106305463A (en) | Intelligent feeder and intelligent feeding system | |
WO2020202353A1 (en) | Communication robot, method for controlling same, information processing server, and information processing method | |
JP2024054440A (en) | IMAGE OUTPUT CONTROL DEVICE, IMAGE OUTPUT CONTROL METHOD, PROGRAM, IMAGE OUTPUT CONTROL SYSTEM, AND ELECTRONIC DEVICE | |
JP2023130837A (en) | Apparatus system, loudness adjustment method, second apparatus, and first apparatus |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |