US20140169609A1 - Communication eyewear assembly with zone of safety capability - Google Patents
Communication eyewear assembly with zone of safety capability Download PDFInfo
- Publication number
- US20140169609A1 US20140169609A1 US13/713,789 US201213713789A US2014169609A1 US 20140169609 A1 US20140169609 A1 US 20140169609A1 US 201213713789 A US201213713789 A US 201213713789A US 2014169609 A1 US2014169609 A1 US 2014169609A1
- Authority
- US
- United States
- Prior art keywords
- wearer
- assembly
- structured
- data
- processor
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R1/00—Details of transducers, loudspeakers or microphones
- H04R1/10—Earpieces; Attachments therefor ; Earphones; Monophonic headphones
- H04R1/1058—Manufacture or assembly
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R1/00—Details of transducers, loudspeakers or microphones
- H04R1/02—Casings; Cabinets ; Supports therefor; Mountings therein
- H04R1/028—Casings; Cabinets ; Supports therefor; Mountings therein associated with devices performing functions other than acoustics, e.g. electric candles
-
- G—PHYSICS
- G02—OPTICS
- G02C—SPECTACLES; SUNGLASSES OR GOGGLES INSOFAR AS THEY HAVE THE SAME FEATURES AS SPECTACLES; CONTACT LENSES
- G02C11/00—Non-optical adjuncts; Attachment thereof
- G02C11/10—Electronic devices other than hearing aids
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R1/00—Details of transducers, loudspeakers or microphones
- H04R1/10—Earpieces; Attachments therefor ; Earphones; Monophonic headphones
- H04R1/1016—Earpieces of the intra-aural type
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R2420/00—Details of connection covered by H04R, not provided for in its groups
- H04R2420/07—Applications of wireless loudspeakers or wireless microphones
Definitions
- the present invention is directed to a communication eyewear assembly that is structured be worn on a person's head so as to protect the person's eyes and also to permit hands-free wireless multimedia communication.
- the communication eyewear assembly is additionally structured to identify a zone of safety about the wearer and to notify the wearer and/or others of the zone of safety.
- the assembly can process a variety of data, receivable from a variety of sources, to ascertain and/or identify the zone of safety.
- the assembly can also determine existence of extraordinary circumstances based on changes or lack of changes in the data, and notify the wearer and/or others of same, particularly with regard to the status of the wearer.
- eyewear particularly protective eyewear, such as sunglasses, sports eyewear, or safety glasses
- various types of communication devices such as radios, portable music players, smart phones, tablets, etc.
- attempts have been made to combine certain features of each of these types of equipment to address the needs of persons who require both eye protection and audiovisual communicability, such as people working in industrial, mining, construction, or military settings.
- some such technology has begun to expand into entertainment and leisure applications such as sports and interactive gaming.
- current technologies have failed to address many significant challenges.
- One problem with current devices is that they do not adequately combine audio, visual, and communicative features into a single, integrated device, particularly one that provides quality in-ear audio in a noisy environment.
- existing devices might typically include a visual display and remote communication features, but lack any audio capabilities.
- such devices might focus primarily on audio communications.
- some existing telecommunication eyewear might provide only rudimentary visual information such as a street name, a retail store location, etc., or perhaps even lack a visual display altogether.
- Another common drawback involves the typical speaker functionality which is not adequately structured to deliver quality sound to the wearer, especially in noisy environments.
- the present invention is directed to a communication eyewear assembly that is structured be worn on a person's head so as to protect the person's eyes and also to permit hands-free wireless multimedia communication.
- the communication eyewear assembly is additionally structured to identify a zone of safety about the wearer and to notify the wearer and/or others of the zone of safety.
- the assembly can process a variety of data, receivable from a variety of sources, to ascertain and/or identify the zone of safety.
- the assembly can also determine existence of extraordinary circumstances based on changes or lack of changes in the data, and notify the wearer and/or others of same, particularly with regard to the status of the wearer.
- the communication eyewear assembly comprises a frame assembly which is structured to be worn on a person's head. Accordingly, the frame assembly is structured to be supportably disposed on the wearer so as to maintain at least one substantially transparent eye panel in a generally shielding position in front of the wearer's eyes.
- the frame assembly comprises a front frame member and two eye panels coupled thereto.
- the frame assembly additionally comprises a pair of legs structured to effectively retain and position the front frame member in front of the wearer's eyes.
- the communication eyewear assembly provides for wireless communication between the wearer and other entities, such as, but not limited to, other wearers, control stations, broadcast stations, network servers, remote network peers, local and/or remote data sources, etc.
- the wireless communication could involve a wearer located on a factory floor communicating with a manager located in a control station.
- the wireless communication could involve a wearer participating in an investigation of a hazardous event while in contact with an emergency response vehicle and/or a related dispatch center.
- the wireless communication could involve a wearer participating in an interactive sport or game together with a plurality of other participants in real time and/or space.
- the communication eyewear assembly comprises at least one processor which is coupled to the frame assembly.
- the processor is structured and disposed to receive an incoming data signal, typically from a remote location, which can include, but is not limited to, audio, video, real-time voice data, environmental data, locational data, data on other objects or persons, etc.
- the processor is further structured and disposed to transmit an outbound data signal to another entity, such as, but not limited to, real-time voice data from the wearer to another person with whom the wearer is communicating. Further details of the multimedia communication capabilities of the processor, including incoming and outbound visual signals, environmental, and locational data, etc., are discussed in more detail below, particularly with regard to the zone of safety capability.
- the communication eyewear assembly additionally comprises at least one earpiece structured to be disposed in communication with the processor and further structured to be disposed at one of the wearer's ears.
- the processor retransmits the received incoming audio signal to at least one earpiece, which in turn communicates the incoming audio signal to the wearer via a speaker assembly.
- the communication eyewear assembly comprises two earpieces, each of which is structured to be disposed at a different one of the wearer's ears.
- the earpieces are additionally structured to substantially muffle ambient noise, further facilitating communication.
- the earpieces are structured to be removably secured to the frame assembly for storage.
- the communication eyewear assembly additionally comprises a camera assembly disposed and structured to pick up an outbound visual signal, which may comprise image data and/or video data. Because the assembly is worn on a wearer's head, the outbound visual signal will typically reflect the view as seen from the wearer's perspective, although other views may also be captured including, but not limited to, views directed behind the wearer.
- the camera assembly is further structured to transmit the outbound visual signal to the processor. Accordingly, the processor is structured to transmit to another entity the outbound visual signal received from the camera assembly.
- the processor is also structured to receive an incoming visual signal from a remote location.
- the incoming visual signal may comprise image data and/or video data, such as pictures or movies from a remotely stored and transmitted data file.
- the communication eyewear assembly further comprises a display assembly disposed and structured to display the incoming visual signal received by the processor. More specifically, the display assembly includes at least one display unit disposed and structured to display the incoming visual signal to the wearer.
- the communication eyewear assembly is additionally structured to identify a zone of safety about the wearer and to notify the wearer and/or others of the zone of safety.
- the processor is structured to receive, send, and process a variety of data associated with ascertaining and/or identifying the zone of safety. This includes multimedia communication capabilities associated various sources and recipients of information, preferably via a wireless communication system.
- the processor can communicate with remote data sources such as a remote environmental data source or a remote locational data source.
- Environmental data can include general as well as specific information about the environment around the eyewear wearer, such as dangerous temperatures, toxic gases, airborne chemical or biological agents, ionizing radiation, potentially damaging energy fields, atmospheric pressure, etc.
- Locational data can include information about the location or orientation of the wearer and/or other items of interest.
- the processor can also communicate with local data sources such as a local environmental data source or a local locational data source generally situated in the vicinity of the area in which the wearer is operating.
- local data sources such as a local environmental data source or a local locational data source generally situated in the vicinity of the area in which the wearer is operating.
- the environmental and locational data provided by the local sources can comprise similar types of information to that provided by the remote data sources as well as additional stored and/or real-time information based on direct local measurements, observations, inputs, etc.
- the processor can also communicate with data sources worn by the wearer such as a personal environmental data source, a personal locational data source, or a biometric data source, which may be embedded in the wearer's clothing or attached thereto, such as via arm band, leg band, belt, backpack, etc.
- data sources worn by the wearer such as a personal environmental data source, a personal locational data source, or a biometric data source, which may be embedded in the wearer's clothing or attached thereto, such as via arm band, leg band, belt, backpack, etc.
- the processor can also communicate with personal data sources worn by individuals other than the wearer, particularly those in proximity to the wearer. Information obtained and provided by such data sources can comprise similar types of information to that provided by the remote data sources, the local data sources, and the wearer's personal data sources.
- the communication eyewear assembly can include similar types of data sources integrated therewith, such as, but not limited to, environmental sensors, locational sensors, and/or biometric sensors indicating wearer's physical status.
- the above-described data can be utilized to compare the wearer's present physical status with predetermined criteria and to notify the wearer and/or others of the wearer's physical status.
- the above-described data can also be processed by the processor and/or a remote processor to identify a zone of safety about the wearer.
- the processors can provide a wide variety of associated functionality, such as in some cases identifying a predetermined zone of safety based on observed parameters, or in other cases, calculating a zone of safety based on pre-existing data received.
- the zone of safety can be determined for one, two, or three dimensions, as well as comprise a temporal aspect, such as in the case of a dynamic, changing situation.
- the communication eyewear assembly is capable of notifying the wearer of same, such as by displaying hazardous conditions information through a visual display, by communicating a pre-determined audio spoken message or tone to the eyewear's audio earpieces, or by vibrational means, electrical shock, heat, etc.
- zone of safety notification can be received on their personal communication device, as can the wearer if he/she is wearing a separate personal communication device.
- remote locations such as control stations, dispatch stations, etc., can receive the zone of safety notification and process the associated information accordingly.
- the present invention can further involve the assignment of a unique ID to the wearer's communication eyewear assembly.
- the unique ID can be transmitted to a remote location or as part of normal wireless transmissions to the wearer's personal communication device for purposes of confirming the identity of the wearer, and times of use.
- the unique ID can be contained in the processor and/or the locational sensors, either alone or in combination with processor, or by separate means such as those involving local RFID, for purposes of establishing location and movement from point to point.
- the communication eyewear assembly is structured to determine existence of extraordinary circumstances based on changes or lack of changes in the data, and to notify the wearer and/or others of same.
- Relevant data can comprise biometric or other data of the individual wearer and/or that of the environment surrounding the wearer, and can be utilized to assess the well-being of the wearer and to optionally notify remote locations of the wearer's health status, possible distress, etc.
- the communication eyewear assembly may comprise a user interface, such as a touch sensitive interface, brain wave monitor, motion sensor, optical recognition monitor, etc.
- the user interface is disposed in communication with the processor and is structured to relay received input to the processor so as to permit selective control of at least one type of input parameter that can be received and processed by the processor.
- the wearer can interact with various embodiments of the user interface by touching it (e.g., swiping, tapping, touching and holding), using forehead movement, nose movement, ear movement, eye movement, eyelid movement, or in some cases by simply thinking, etc., as well as ay combinations thereof.
- the user interface can be structured to facilitate selection or control of the visual display of incoming data.
- the user interface can be structured to facilitate control of audio volume.
- the user interface can be structured to facilitate input of an emergency distress signal, such as a pattern of successive taps, etc.
- the processor can be structured to process audio instructions received directly from the wearer via the microphone, such as for manipulation and adjustment of display and/or audio settings, or to initiate a distress signal.
- FIG. 1 is a perspective view of one embodiment of a communication eyewear assembly in accordance with the present invention.
- FIG. 2 is a schematic illustration of one embodiment of a transceiver configuration in accordance with the present invention.
- FIG. 3 is a schematic illustration of one embodiment of a wireless communication system in accordance with the present invention.
- FIG. 4 is a schematic illustration of one embodiment of an established zone of safety in accordance with the present invention.
- FIG. 4A is a schematic illustration of one embodiment of a display presentation in accordance with the established zone of safety of FIG. 4 .
- FIG. 5 is a cross section view of one embodiment of an earpiece in accordance with the present invention.
- FIG. 6 is a partial perspective view of another embodiment of a microphone assembly in accordance with the present invention.
- FIG. 7 is a partial perspective view of a communication eyewear assembly having a display assembly which includes two different types of display units.
- the present invention is directed towards a communication eyewear assembly generally indicated as 10 .
- the communication eyewear assembly 10 is structured to be worn on a person's head so as to protect the person's eyes and also to permit hands-free wireless multimedia communication.
- the communication eyewear assembly 10 is additionally structured to identify a zone of safety about the wearer, as distinguished from a hazardous object or zone, and to notify the wearer and/or others of the zone of safety, such as that of an industrial environment, or that found in a variety of other situations, such as, but not limited to, military operations, fire and rescue operations, public safety operations, sports activities, interactive gaming, etc.
- the assembly 10 can process a variety of data, receivable from a variety of sources, to ascertain and/or identify the zone of safety.
- the assembly 10 can also determine existence of extraordinary circumstances based on changes or lack of changes in the data, and notify the wearer and/or others of same, particularly with regard to the status of the wearer.
- the communication eyewear assembly 10 comprises a frame assembly 20 which is structured to be worn on a person's head. Accordingly, the frame assembly 20 is structured to be supportably disposed on the wearer so as to maintain at least one substantially transparent eye panel 25 in a generally shielding position in front of the wearer's eyes.
- the frame assembly 20 comprises a front frame member 21 and two eye panels 25 coupled thereto.
- each eye panel 25 is structured to substantially shield the front of a different corresponding eye of the wearer while also permitting clear vision therethrough.
- the eye panels 25 can be tinted.
- the eye panels 25 can include ultraviolet protection such as UV coatings.
- the eye panels 25 can also be polarized for protective purposes and/or for viewing enhancement purposes.
- the eye panels 25 can also comprise prescription lenses.
- the eye panels 25 can also comprise a shatterproof material of construction for additional protection in the event of significant impact to the eye panels 25 by a foreign object.
- the frame assembly 20 further comprises at least one leg 23 , 24 structured to effectively retain and position the front frame member 21 in front of the wearer's eyes during activity of the wearer.
- the frame assembly 20 comprises a first leg 23 and a second leg 24 , each leg 23 , 24 structured to extend from opposite sides of the front frame member 21 towards a rear thereof.
- the legs 23 , 24 are disposed in a generally perpendicular relation to the front frame member 21 so as to extend over and grip the wearer's ears when the front frame member 21 is operatively disposed in front of the wearer's eyes.
- the frame assembly 20 further comprises pair of side eye panels 27 .
- Each side eye panel 27 is disposed and structured to substantially shield a different eye of the wearer.
- the side eye panels 27 can be further structured to permit clear peripheral vision therethrough.
- the side eye panels 27 are each coupled to a different leg 23 , 24 .
- the side eye panels 27 can be tinted, UV coated, and/or polarized.
- the side eye panels 27 can also comprise a shatterproof material of construction.
- the communication eyewear assembly 10 provides for wireless communication between a wearer and other entities, such as, but not limited to, other wearers, control stations, broadcast stations, network servers, remote network peers, local and/or remote data sources, etc.
- the wireless communication could involve a wearer located on a factory floor communicating with a manager located in a control station.
- the wireless communication could involve a wearer participating in an investigation of a hazardous event while in contact with an emergency response vehicle and/or a related dispatch center.
- the wireless communication could involve a wearer participating in an interactive sport or game together with a plurality of other participants in real time and/or space.
- the communication eyewear assembly 10 comprises at least one processor 40 which is coupled to the frame assembly 20 .
- the processor 40 is structured and disposed to receive an incoming data signal, typically from a remote location.
- the incoming data signal comprises real-time voice data, such as from another person with whom the wearer is communicating.
- the incoming data signal comprises environmental data, such as from a remotely stored and transmitted database file.
- the processor 40 is further structured and disposed to transmit an outbound data signal to another entity, such as real-time voice data from the wearer to another person with whom the wearer is communicating. Further details of the multimedia communication capabilities of the processor 40 , including the incoming and outbound visual signals, environmental, and locational data, etc., are discussed in more detail below, particularly with regard to the zone of safety capability.
- the processor 40 is structured and disposed to fit inside of the frame assembly 20 .
- the processor 40 is disposed inside of the first leg 23 . It should be noted, however, that it is within the scope and intent of the present invention that the processor 40 can be structured and disposed to fit inside of other portions of the frame assembly 20 , including, but not limited to, the second leg 24 , the front frame member 21 , etc.
- the communication eyewear assembly 10 further comprises a power supply 90 structured and disposed to provide power to the processor 40 .
- the power supply 90 can also be structured to supply power to other components of the assembly 10 .
- the power supply 90 is structured and disposed to fit inside of the frame assembly 20 .
- the power supply 90 is disposable inside of the second leg 24 .
- the power supply 90 can be structured and disposed to fit inside of other portions of the frame assembly 20 , including, but not limited to, the first leg 23 , the front frame member 21 , etc.
- the power supply 90 can also be rechargeable.
- the power supply 90 comprises a rechargeable battery pack that can be charged by an external power source that has been removably coupled to the assembly 10 .
- the communication eyewear assembly 10 further comprises at least one earpiece 60 structured to be disposed in either wired or wireless communication with the processor 40 and further structured to be disposed at one of the wearer's ears.
- the processor 40 retransmits a received incoming audio signal to at least one earpiece 60 , which in turn communicates the incoming audio signal to the wearer.
- the communication eyewear assembly 10 comprises two earpieces 60 , each of which are structured to be disposed at a different one of the wearer's ears.
- the earpieces 60 are further structured to be at least partially inserted into the wearer's ears when disposed in their respective operative orientations.
- the processor 40 comprises at least one transceiver structured to facilitate wireless communication.
- the processor 40 can comprise a first transceiver 42 and a second transceiver 44 .
- the first transceiver 42 is structured to receive an incoming data signal from a remote source, and is also structured to transmit an outbound data signal to a remote destination.
- the data signal can include a wide variety of information, including, but not limited to, audio and/or visual information.
- the first transceiver utilizes radio frequency communication.
- the outbound data signal can comprise an audio signal corresponding to the wearer's voice which has been picked up by an outbound microphone assembly 70 , such as the one shown in FIG. 1 and described in greater detail below.
- an outbound visual signal will ordinarily correspond to an image captured by a camera assembly 105 , also described in greater detail below.
- the outbound data signal can also comprise information detected by various sensors, as discussed further below.
- the processor 40 can also comprise a local data storage capability or “memory” 43 , such as for use in scenarios that require local storage and convenient access to specific data.
- the second transceiver 44 of FIG. 2 is structured to wirelessly communicate with the earpieces 60 .
- the second transceiver utilizes radio frequency communication technology.
- the second transceiver utilizes magnetic induction communication technology. It is, however, within the scope and intent of the present invention that other suitable wireless technologies may also be utilized, including, but not limited to BLUETOOTH technology.
- the processor 40 comprises a single transceiver structured to communicate with remote entities as well as with the earpieces 60 .
- each earpiece 60 comprises an earpiece transceiver 64 structured to facilitate communication with the processor 40 .
- the earpiece transceivers 64 are each disposed in wireless communication with the second transceiver 44 .
- the specific communication technology of the earpiece transceivers 64 matches that of the second transceiver 44 , such as, but not limited to, radio frequency technology, magnetic induction technology, BLUETOOTH, etc.
- each earpiece transceiver 64 is uniquely identifiable by the processor 40 .
- each earpiece transceiver 64 comprises a unique signature, such as a radio, BLUETOOTH or magnetic identification, which is recognizable by the processor 40 . Accordingly, the processor 40 can transmit different content to each earpiece transceiver 64 , as may be desired.
- a unique signature such as a radio, BLUETOOTH or magnetic identification
- the communication eyewear assembly 10 is additionally structured to identify a zone of safety about the wearer and to notify the wearer and/or others of the zone of safety.
- the processor 40 is structured to receive, send, and process a variety of data associated with ascertaining and/or identifying the zone of safety. This includes multimedia communication capabilities associated various sources and recipients of information, preferably via a wireless communication system.
- a schematic illustration depicts one embodiment of a wireless communication system 110 in accordance with the present invention.
- the processor 40 and various electronic data sources are structured to communicate with one or more communication networks 120 and protocols, such as, but not limited to, the Internet, cellular, satellite, WIFI, WIMAX, WLAN, WAN, LAN, WIFI hotspot, Intranet, ZIGBEE, ENOCEAN, etc.
- the various data sources can also communicate directly with the processor 40 , using a variety of wireless technologies and protocols, such as via cellular, satellite, WIFI, WIMAX, ZIGBEE, ENOCEAN, BLUETOOTH, etc.
- a remote processor 140 can facilitate processing and organization of data and other information, and may be structured to enhance and/or supplement the computational capacity of the in-frame processor 40 , such as for determining the zone of safety.
- a variety of processing and data storage hardware and software, operating systems, and communication protocols may be implemented to carry out various embodiments of the present invention.
- the processor 40 can communicate with remote data sources such as a remote environmental data source 181 or a remote locational data source 183 .
- Environmental data can include general as well as specific information about the environment around the eyewear wearer.
- information provided by the remote environmental data source 181 can pertain to potentially dangerous temperatures, toxic gases, airborne chemical or biological agents, ionizing radiation, potentially damaging energy fields, atmospheric pressure, etc.
- Locational data can include information about the location of the wearer as well as the location of items of interest.
- the information provided by the remote locational data source 183 can include, but is not limited to, the wearer's position on the face of the earth, location of restricted areas, relative locations of potentially hazardous stationary or moving objects around the eyewear wearer, etc.
- the processor 40 can also communicate with local data sources such as a local environmental data source 281 or a local locational data source 283 generally situated in the vicinity of the area in which the wearer is operating.
- the environmental and locational data provided by the local sources 281 , 283 can comprise similar types of information to that provided by the remote data sources 181 , 183 , as discussed above. Moreover, they may comprise additional stored and/or real-time information based on direct local measurements, observations, inputs, etc.
- the local data sources can comprise a wide variety of active devices, such as thermometers, altimeters, locational beacons, alarms, motion sensors, and sensors of biological agents, chemical agents, and radiation, among others.
- the processor 40 can also communicate with data sources worn by the wearer such as a personal environmental data source 381 , a personal locational data source 383 , or a biometric data source 385 , which may be embedded in the wearer's clothing or attached thereto, such as via arm band, leg band, belt, backpack, etc.
- the environmental and locational data provided by the personal sources 381 , 382 can comprise similar types of information to that provided by the remote data sources 181 , 183 and the local data sources 281 , 283 , as discussed above.
- the biometric data sources 385 may comprise additional stored and/or real-time information based on biometric data of the wearer, such as, but not limited to, body temperature, heartbeat, and rate of respiration.
- the personal biometric data sources can comprise a wide variety of active devices such as body temperature monitors, heartbeat monitors, respiration monitors, perspiration monitors, muscle reflex monitors, blood sugar monitors, etc.
- the processor 40 can also communicate with personal data sources 250 worn by individuals other than the wearer, particularly those in proximity to the wearer.
- Information obtained and provided by such data sources 250 can comprise similar types of information to that provided by the remote data sources 181 , 183 , the local data sources 281 , 283 , and the personal sources 381 , 383 , 385 as discussed above.
- the communication eyewear assembly 10 can include similar types of data sources integrated therewith, which are each structured to communicate with the processor 40 .
- the communication eyewear assembly 10 comprises one or more environmental sensors 81 which are coupled to the frame assembly 20 .
- the environmental sensors 81 are structured and disposed to obtain specific information about the environment around the eyewear wearer.
- information provided by the environmental sensors 81 can pertain to potentially dangerous temperatures, toxic gases, airborne chemical or biological agents, ionizing radiation, potentially damaging energy fields, atmospheric pressure, ultrasonic waves, etc.
- the communication eyewear assembly 10 comprises one or more locational sensors 83 which are coupled to the frame assembly 20 and structured to communicate with the processor 40 .
- the locational sensors 83 are specifically disposed and structured to provide locational data which can include information about the location of the wearer as well as the location of items of interest.
- a locational sensor 83 is disposed inside of the first leg 23 . It should be noted, however, that it is within the scope and intent of the present invention that the locational sensors 83 can be structured and disposed to fit inside of other portions of the frame assembly 20 , including, but not limited to, the second leg 24 , the front frame member 21 , etc.
- the processor 40 is structured to utilize information from the locational sensor 83 and/or other aforementioned sources of locational data to determine the wearer's position on the face of the earth, location of restricted areas, relative locations of potentially hazardous stationary or moving objects around the eyewear wearer, etc.
- the locational sensor 83 can be structured to receive signals from the Global Positioning System (GPS) via a geomagnetic sensing capability.
- GPS Global Positioning System
- Such information can be used to determine the wearer's general location and also to determine situational awareness, such as whether the wearer is venturing into restricted areas or perhaps is immobile in excess of a predetermined amount of time, signaling possible incapacitation.
- the processor 40 can utilize the locational data to determine in real time the wearer's vertical and/or horizontal and/or angular orientation thus permitting precise characterization of the wearer's location and orientation in one, two, or three spatial dimensions.
- Such information can include gyroscopic orientation, speed, and/or acceleration.
- situational awareness such as whether the wearer is venturing into restricted areas or perhaps may be horizontally oriented for too long a period, thus may be incapacitated and require assistance.
- such information can be utilized to inform a company on the location and status of a worker at a work site, and to facilitate a decision on whether to send help as well as to notify first responders of the wearer's status and location.
- the communication eyewear assembly 10 also comprises one or more biometric sensors 85 which are coupled to the frame assembly 20 and structured to communicate with the processor 40 .
- the biometric sensors 85 are specifically disposed and structured to provide biometric data on a variety of bodily functions such as body temperature, heartbeat, rate of respiration, perspiration, muscle reflex, blood sugar, etc.
- a biometric sensor 85 is disposed on the nasal bridge portion of the front frame member 21 . It should be noted, however, that it is within the scope and intent of the present invention that the biometric sensors 85 can be structured and disposed to fit on or inside of other portions of the frame assembly 20 , including, but not limited to, the first leg 23 , the second leg 24 , etc.
- the processor 40 is structured to utilize information from the biometric sensor 85 and/or other aforementioned sources of biometric data to determine the wearer's physical status.
- the processor is further structured to compare the wearer's present physical status with predetermined criteria and to notify the wearer and/or others of the wearer's physical status.
- the above-described data can be processed by the processor 40 and/or remote processor 140 to identify a zone of safety about the wearer.
- the processors can provide a wide variety of associated functionality, such as in some cases identifying a predetermined zone of safety based on observed parameters, or in other cases, calculating a zone of safety based on data received.
- the processor 40 can be disposed in communication with the remote processor 140 to receive information on the zone of safety.
- the remote processor 140 can be structured to calculate and/or identify the zone of safety and to notify the processor 40 accordingly.
- the zone of safety 123 is calculated based on data involving hazardous objects 112 , such as contaminated pieces of equipment from a wreckage site, as well as a hazard area 113 , such as the affected area associated with a chemical spill, as well as locational data including the global position of the event and the relative locations of the wearer 9 and the hazard area or zone 113 .
- hazardous objects 112 such as contaminated pieces of equipment from a wreckage site
- a hazard area 113 such as the affected area associated with a chemical spill
- locational data including the global position of the event and the relative locations of the wearer 9 and the hazard area or zone 113 .
- the processor 40 is structured to process locational data as well as data on the hazardous objects and the hazardous environment provided by various local and/or remote sensors in order to distinguish safe areas, features, and objects from dangerous ones, and thus determine, calculate, and generate the zone of safety 123 accordingly. While the zone of safety 123 shown in FIG. 4 is two dimensional, as noted above, the zone of safety can be determined for one, two, or three dimensions, as well as comprise a temporal aspect, such as in the case of a dynamic, changing situation.
- the communication eyewear assembly 10 is capable of notifying the wearer 9 and/or others of same, such as in contrast to potentially hazardous conditions around the wearer.
- the eyewear assembly 10 worn by the wearer 9 can be structured to notify the wearer 9 as to the existence of the zone of safety 123 (and its boundaries) by displaying hazardous conditions information through a variety of video means, or, through a pre-determined audio spoken message or tone to the eyewear's audio earpieces 60 .
- the wearer 9 is notified by the eyewear assembly 10 via vibrational means, gentle electrical shock, heat, etc., such as via the physical notification mechanism 87 shown in FIG. 1 .
- the physical notification mechanism 87 comprises a vibration device.
- the physical notification mechanism 87 comprises an electro-shocking device.
- physical notification mechanism 87 comprises a heating device.
- the present invention can further involve the assignment of a unique ID 45 to the wearer's communication eyewear assembly 10 .
- the unique ID can be transmitted to a remote location or as part of normal wireless transmissions to the wearer's personal communication device 11 for purposes of confirming the identity of the wearer, and times of use.
- the unique ID 45 can constitute a singular ‘public’ identity, similar to that associated with a cellphone account.
- current telecommunication eyewear units might have a separate BLUETOOTH identity for use in connection with peripheral devices, such limited identity is not specifically tied to the eyewear or the wearer in the manner contemplated herein.
- each eyewear/user can have an identity.
- the individualized identity affiliated with the unique ID 45 can be used in large industrial plants to monitor conduct and productivity of corresponding workers/wearers.
- the unique ID can be contained in the processor 40 and/or the locational sensors 83 , either alone or in combination with processor 40 , or by separate means such as those involving local RFID, for purposes of establishing location and movement from point to point.
- FIG. 4A is a schematic illustration of one embodiment of a visual display presentation associated with a display unit 26 in accordance with the established zone of safety 123 of FIG. 4 .
- the zone of safety 123 ′ comprises three spatial dimensions and is also depicted in contrast to a zone of danger 124 ′.
- the zone of safety 123 ′ also comprises a temporal dimension.
- the visual display is primarily displayed in connection with the communication eyewear assembly 10 worn by the wearer 9 but can also be displayed on other devices as well, including personal communication devices 11 , 11 ′, or remote displays, such as those associated with a control center or dispatch.
- the three-dimensional zone of safety 123 ′ borders a corresponding danger zone 124 ′ which encompasses a spill zone or similar type of hazard area 113 .
- the zone of safety 123 ′ includes a two-dimensional base area, along with a height dimension, such as to avoid vapors associated with the spill zone 113 , hazardous objects 112 , etc.
- Other objects and items of FIG. 4 are also represented by the visual display presentation shown in FIG. 4A .
- the display unit 26 is structured to present additional information in graphical and/or script format, such as a communication reception gauge 151 , a danger scale 153 , and/or a text box 155 . Each of these can be customized to represent a variety of safety, environmental, personal, and/or communicative data.
- information received from the aforementioned sensor devices embedded in the communication eyewear assembly 10 , or other remote sources can be graphically and/or textually displayed via the display unit 26 , including, by way of example, locational information about potentially hazardous stationary or moving objects around the wearer.
- the processor 40 is structured to identify whether a potentially hazardous moving object 112 is increasing its risk to the wearer 9 , through movement of the object itself toward the wearer, or the wearer toward the object.
- the communication eyewear assembly 10 through use of the processor 40 and multiple microphones 76 (discussed in more detail, below), can cooperatively determine whether a potentially dangerous object 112 is approaching the wearer 9 , as well as the object's 122 position, via triangulation or other appropriate calculations, and inform the wearer accordingly.
- the assembly 10 is structured to determine existence of extraordinary circumstances based on changes or lack of changes in the data, and to notify the wearer and/or others of same.
- the data can comprise biometric or other data of the individual wearer and/or that of the environment surrounding the wearer. Examples of such data include, but are not limited to, potentially dangerous ambient temperatures, toxic gases, airborne chemical or biological agents, ionizing radiation, or other potentially damaging energy fields, atmospheric pressure, unusual body temperature, heartbeat, or rate of respiration, etc.
- the data can be used by the processor 40 alone or in conjunction with the remote processor 140 to assess the well-being of the wearer 9 and to optionally notify remote locations, of the wearer's health status, possible distress, etc.
- the processors 40 , 140 can compare received data with corresponding baseline data to determine the potential existence of a threat to the well-being of the wearer 9 , and autonomously and wirelessly report the information and condition to remote sources. Conditions determined to constitute an anomaly or other unusual occurrence, can thus signal a potential risk to the wearer, which in turn can trigger a corresponding notification, such as via an alarm feature. Additionally, the information could be used for purposes of security, insurance documentation, or other documentation useful to human resources at the employer of the wearer.
- the earpiece 60 further comprises a speaker assembly 68 structured and disposed to audibly communicate an incoming audio signal (audio data) to the wearer.
- the speaker assembly 68 is structured to be at least partially inserted into the ear of the wearer when disposed in an operative orientation.
- the earpiece 60 is additionally structured to substantially muffle ambient noise, further facilitating communication. As shown in the embodiment of FIG. 5 , at least a portion of the earpiece 60 is structured to be inserted into the wearer's ear so as to substantially muffle the ambient noise. More specifically, the earpiece 60 of FIG. 5 additionally comprises a disposable insert 62 structured to substantially muffle the ambient noise.
- the disposable insert 62 can be formed, for instance, from a sound-deadening, compressible material, such as high density foam, although a variety of other suitable materials may also be utilized, including materials which readily conform to fit inside of the ear canal. For at least sanitary purposes, the disposable insert 62 can be discarded and replaced after use.
- the earpiece transceiver 64 which is communicatively coupled to both the speaker assembly 68 and an optional earpiece microphone assembly 66 .
- the earpiece transceiver 64 is structured to facilitate communication between the respective earpiece components 66 , 68 and the processor 40 .
- the earpiece 60 further comprises an earpiece power supply 92 structured and disposed to provide power to the speaker assembly 68 and the earpiece microphone assembly 66 , as well as to the earpiece transceiver 64 .
- the earpiece power supply 92 can be rechargeable.
- the communication eyewear assembly 10 further comprises an outbound microphone assembly 70 disposed and structured to pick up an outbound audio signal from the wearer and transmit the outbound audio signal to the processor 40 .
- the outbound microphone assembly 70 comprises an adjustable elongate segment 71 which extends from the first leg 23 of the frame assembly 20 .
- the adjustable elongate segment 71 is structured to permit adjustment of the microphone assembly 70 in front of the wearer's mouth to better pick up the outbound audio signal, which, in at least one embodiment, comprises real-time voice data of the wearer's spoken voice.
- the outbound microphone assembly 70 is in disposed in wired communication with the processor 40 .
- the outbound microphone assembly 70 can be disposed in wireless communication with the processor 40 , such as with the second transceiver 44 .
- the communication eyewear assembly 10 can be further structured to enhance an incoming and/or outbound audio signal based on an ambient audio signal. This can include enhancing a voice component of the received audio data, such as that of the wearer or that of others in the vicinity of the wearer. This can also include enhancement of informational sounds such sirens, whistles, horns, gunfire, explosions, and other identifiable audio sounds that are distinct from the general background noise.
- the ambient audio signal will typically comprise the background sounds generated by the surrounding environment.
- one or more ambient microphone assemblies can be disposed and structured to pick up an ambient audio signal and transmit the ambient audio signal to the processor 40 for enhancement processing, filtering, etc.
- the earpiece microphone assembly 66 can serve as an ambient microphone assembly.
- one or more frame microphone assemblies 76 such as those depicted in FIG. 6 , can provide an ambient microphone capability.
- the communication eyewear assembly 10 further comprises a user interface 80 , such as a touch sensitive interface, brain wave monitor, motion sensor, optical recognition monitor, etc.
- the user interface is disposed in communication with the processor 40 and is structured to relay received input to the processor 40 so as to permit selective control of at least one type of input parameter that can be received and processed by the processor 40 .
- the wearer can interact with various embodiments of the user interface 80 by touching it (e.g., swiping, tapping, touching and holding), using forehead movement, nose movement, ear movement, eye movement, eyelid movement, or in some cases by simply thinking, etc., as well as ay combinations thereof.
- the user interface 80 can be structured to facilitate selection or control of the visual display of incoming data.
- the user interface 80 can be structured to facilitate control of audio volume.
- the user interface 80 can be structured to facilitate input of an emergency distress signal, such as a pattern of successive taps, etc.
- the user interface 80 can be disposed in a variety of suitable locations, such as on the frame assembly 20 or on at least one of the earpieces 60 . In the embodiment of FIG. 1 , for instance, the user interface 80 is disposed on the first leg 23 .
- the processor 40 can be structured to process audio instructions received directly from the wearer 9 via the microphone 70 , such as for manipulation and adjustment of display and/or audio settings, or to initiate a distress signal.
- two frame microphone assemblies 76 are each disposed on a different side of the front frame member 21 .
- the frame microphone assemblies 76 are disposed and structured to pick up an ambient audio signal and transmit the ambient audio signal to the processor 40 .
- a distance of several inches, preferably about four to five inches (4′′-5′′) which separates the frame microphone assemblies 76 from one another provides sufficient separation such that disparate sources of sound can be geometrically differentiated by the processor 40 based on the variations in the audio signals picked up by each individual frame microphone assembly 76 .
- the frame microphone assemblies 76 are in disposed in wired communication with the processor 40 .
- the frame microphone assemblies 76 can be disposed in wireless communication with the processor 40 , such as with the second transceiver 44 . It should also be noted that, other numbers of frame microphone assemblies 76 , could be used. Moreover, in other embodiments, the frame microphone assemblies 76 can also be disposed at other suitable locations, such as on the legs 23 , 24 , or elsewhere.
- the earpieces 60 are further structured to be removably secured to the frame assembly 20 for storage.
- each earpiece 60 comprises a clip 69 structured to removably secure the earpiece 60 to the frame assembly 20 when the earpiece 60 is not in use.
- the earpieces 60 can be removably secured to at least one of the legs 23 , 24 when disposed in a stored orientation.
- the earpieces 60 may also be removably secured to other portions of the frame assembly, including, but not limited to, the front frame member 21 .
- the earpiece 60 and the frame assembly 20 can be cooperatively structured to charge a rechargeable earpiece power supply 92 when the earpiece 60 is disposed in a stored orientation.
- the assembly 10 further includes a camera assembly 105 disposed and structured to pick up an outbound visual signal, which may comprise image data and/or video data.
- the camera assembly 105 comprises a single camera element disposed on a central area of the front frame member 21 and facing towards a field of view from the wearer's perspective. It should be noted, however, that it is within the scope and intent of the present invention that a plurality of cameras elements may be utilized, and further, that one or more camera assemblies 105 may be disposed on a variety of other suitable portions of the assembly 10 , including, but not limited to, other positions on the front frame member, on the legs 23 , 24 , etc.
- the outbound visual signal will typically reflect the view as seen from the wearer's perspective, although other views may also be captured including, but not limited to, views directed behind the wearer.
- the camera assembly 105 will typically pick up visual signals in the visible spectrum.
- the camera assembly 105 can pick up visual signals in other spectra, including, but not limited to, the infrared spectrum.
- the camera assembly 105 is further structured to transmit the outbound visual signal to the processor 40 .
- the processor 40 is structured to transmit to another entity the outbound visual signal or data received from the camera assembly 105 .
- the processor 40 is also structured to receive an incoming visual signal from a remote location.
- the incoming visual signal may comprise image data and/or video data, such as pictures or movies from a remotely stored and transmitted data file, or real-time feeds from remote sources such as hidden cameras, satellites, aerial vehicles, control towers, etc.
- the assembly 10 includes at least one display unit 26 disposed and structured to display the incoming visual signal to the wearer.
- the assembly 10 comprises a single display unit 26 which is at least partially disposed on an exterior surface of one of the eye panels 25 .
- the display unit 26 is further structured to project an image of the received incoming visual signal towards the wearer's eye.
- the display unit 26 is structured to project an image of the received incoming visual signal onto the eye panel 25 such that the wearer's eye can perceive the displayed visual signal as emanating from the interior and/or exterior surface of the eye panel 25 .
- FIG. 1 depicts the display unit 26 positioned on the exterior surface of the eye panel 25
- the display unit 26 may alternatively be disposed on the interior surface of the eye panel 25 , or within the eye panel 25 .
- the assembly 10 includes at least one display unit 26 corresponding to each eye panel 25 .
- the embodiment of FIG. 7 depicts two other variations of the display unit 26 ′, 26 ′′, each coupled to a respective eye panel 25 .
- a first display unit 26 ′ is disposed on the interior surface of one eye panel 25 , and is further structured to display the incoming visual signal in a partial field of vision defined by the eye panel 25 .
- the second display unit 26 ′′ is disposed within the other eye panel 25 ′, and is further structured to display the incoming visual signal in substantially the entire field of vision defined by the eye panel 25 ′.
- the display unit 26 , 26 ′, 26 ′′ may physically comprise a variety of suitable materials and structures, including active and passive electro-mechanical, opto-mechanical, and/or and electro-optical components.
- the display unit 26 , 26 ′, 26 ′′ may comprise flat panel display technology, such as, but not limited to, liquid crystal display (LCD) technology, organic light emitting diode (OLED) display technology, etc.
- the display unit 26 , 26 ′, 26 ′′ can be further structured to at least partially correspond to the contour of the eye panel 25 .
- the display unit 26 , 26 ′, 26 ′′ can be additionally structured to maintain a substantially transparent state until receipt of the incoming visual signal from the processor 40 , whereupon the display unit 26 , 26 ′, 26 ′′ will display the visual signal accordingly.
- the display of the incoming visual signal is physically triggered by an electrical input received by the display unit 26 , 26 ′, 26 ′′ from the processor 40 .
Landscapes
- Physics & Mathematics (AREA)
- Acoustics & Sound (AREA)
- Health & Medical Sciences (AREA)
- Engineering & Computer Science (AREA)
- Signal Processing (AREA)
- Otolaryngology (AREA)
- General Health & Medical Sciences (AREA)
- Optics & Photonics (AREA)
- Ophthalmology & Optometry (AREA)
- General Physics & Mathematics (AREA)
- Alarm Systems (AREA)
- Manufacturing & Machinery (AREA)
- Eyeglasses (AREA)
Abstract
A communication eyewear assembly is configured to identify a zone of safety about a wearer and to notify the wearer and/or others of same. A frame assembly is worn to maintain at least one transparent eye panel in front of the wearer's eyes. A processor coupled to the frame assembly receives incoming data from at least one data source. Utilizing the received data, the processor generates zone of safety information in least one spatial dimension relative to a hazardous area. The assembly can also determine existence of extraordinary circumstances based on changes or lack of changes in the data, and notify the wearer and/or others of same, particularly with regard to the status of the wearer. In accordance with the monitoring of the wearer and/or the communication eyewear itself, a unique ID is assigned to the communication eyewear assembly.
Description
- 1. Field of the Invention
- The present invention is directed to a communication eyewear assembly that is structured be worn on a person's head so as to protect the person's eyes and also to permit hands-free wireless multimedia communication. The communication eyewear assembly is additionally structured to identify a zone of safety about the wearer and to notify the wearer and/or others of the zone of safety. Moreover, the assembly can process a variety of data, receivable from a variety of sources, to ascertain and/or identify the zone of safety. The assembly can also determine existence of extraordinary circumstances based on changes or lack of changes in the data, and notify the wearer and/or others of same, particularly with regard to the status of the wearer.
- 2. Description of the Related Art
- The use of eyewear, particularly protective eyewear, such as sunglasses, sports eyewear, or safety glasses, has achieved near universal adoption by modern society. So too has the use of various types of communication devices, such as radios, portable music players, smart phones, tablets, etc. More recently, attempts have been made to combine certain features of each of these types of equipment to address the needs of persons who require both eye protection and audiovisual communicability, such as people working in industrial, mining, construction, or military settings. Indeed, some such technology has begun to expand into entertainment and leisure applications such as sports and interactive gaming. However, despite some progress in the art, current technologies have failed to address many significant challenges.
- One problem with current devices is that they do not adequately combine audio, visual, and communicative features into a single, integrated device, particularly one that provides quality in-ear audio in a noisy environment. In some cases, such as those involving the fledgling technology of “augmented reality” eyewear, existing devices might typically include a visual display and remote communication features, but lack any audio capabilities. In other cases, such devices might focus primarily on audio communications. Indeed, some existing telecommunication eyewear might provide only rudimentary visual information such as a street name, a retail store location, etc., or perhaps even lack a visual display altogether. Another common drawback involves the typical speaker functionality which is not adequately structured to deliver quality sound to the wearer, especially in noisy environments. Accordingly, there is a need to provide a communication eyewear assembly that is structured to permit hands-free wireless multimedia communication and which overcomes these shortcomings to combine audio, visual, and communicative features into a single, integrated device, particularly one that provides quality in-ear audio in a noisy environment.
- Another problem with current head-worn devices is that they lack adequate measures for dealing with the effects of operating in potentially hazardous environments, in that they are not sufficiently structured to distinguish ordinary, safe environments or situations from hazardous ones. For instance, when a person is operating in a hazardous environment that requires use of telecommunication eyewear, it can be critical for the person to have the ability to discern a safe situation from a dangerous one. Presently, this is left to either the wearer's own instincts and observations, or perhaps a combination of those plus observations from remote observers who are in radio contact with the wearer. It would therefore be advantageous to overcome these limitations by providing for a communication eyewear assembly that is structured to identify a zone of safety about the wearer and to notify the wearer and/or others of the zone of safety, as distinguished from hazardous zones. It would be a further benefit for such an assembly to have the capability of processing a variety of data, receivable from a variety of sources, to ascertain and/or identify the zone of safety.
- Further still, human beings operating in dangerous environments are susceptible to numerous dangers which can lead to immobility, injury, incapacity, or other sudden changes in status. Moreover, the general status of such environments can change rapidly, potentially affecting the wearer of communications equipment and/or other people or things in the vicinity. As such, it would also be beneficial to provide for a communication eyewear assembly that can determine the existence of extraordinary circumstances based on changes or lack of changes in observed data, and notify the wearer and/or others of same, particularly with regard to the identity, condition, location or other status of the wearer.
- The present invention is directed to a communication eyewear assembly that is structured be worn on a person's head so as to protect the person's eyes and also to permit hands-free wireless multimedia communication. The communication eyewear assembly is additionally structured to identify a zone of safety about the wearer and to notify the wearer and/or others of the zone of safety. Moreover, the assembly can process a variety of data, receivable from a variety of sources, to ascertain and/or identify the zone of safety. The assembly can also determine existence of extraordinary circumstances based on changes or lack of changes in the data, and notify the wearer and/or others of same, particularly with regard to the status of the wearer.
- In at least one embodiment, the communication eyewear assembly comprises a frame assembly which is structured to be worn on a person's head. Accordingly, the frame assembly is structured to be supportably disposed on the wearer so as to maintain at least one substantially transparent eye panel in a generally shielding position in front of the wearer's eyes. In one such embodiment, the frame assembly comprises a front frame member and two eye panels coupled thereto. In a further embodiment, the frame assembly additionally comprises a pair of legs structured to effectively retain and position the front frame member in front of the wearer's eyes.
- The communication eyewear assembly provides for wireless communication between the wearer and other entities, such as, but not limited to, other wearers, control stations, broadcast stations, network servers, remote network peers, local and/or remote data sources, etc. By way of example, the wireless communication could involve a wearer located on a factory floor communicating with a manager located in a control station. As another example, the wireless communication could involve a wearer participating in an investigation of a hazardous event while in contact with an emergency response vehicle and/or a related dispatch center. As yet another example, the wireless communication could involve a wearer participating in an interactive sport or game together with a plurality of other participants in real time and/or space. These examples are meant for illustrative purposes only, as the present invention can apply to a variety of other applications and circumstances.
- Accordingly, the communication eyewear assembly comprises at least one processor which is coupled to the frame assembly. The processor is structured and disposed to receive an incoming data signal, typically from a remote location, which can include, but is not limited to, audio, video, real-time voice data, environmental data, locational data, data on other objects or persons, etc.
- The processor is further structured and disposed to transmit an outbound data signal to another entity, such as, but not limited to, real-time voice data from the wearer to another person with whom the wearer is communicating. Further details of the multimedia communication capabilities of the processor, including incoming and outbound visual signals, environmental, and locational data, etc., are discussed in more detail below, particularly with regard to the zone of safety capability.
- The communication eyewear assembly additionally comprises at least one earpiece structured to be disposed in communication with the processor and further structured to be disposed at one of the wearer's ears. As such, the processor retransmits the received incoming audio signal to at least one earpiece, which in turn communicates the incoming audio signal to the wearer via a speaker assembly. In a further embodiment, the communication eyewear assembly comprises two earpieces, each of which is structured to be disposed at a different one of the wearer's ears. In a still further embodiment, the earpieces are additionally structured to substantially muffle ambient noise, further facilitating communication. In yet another embodiment, the earpieces are structured to be removably secured to the frame assembly for storage.
- The communication eyewear assembly additionally comprises a camera assembly disposed and structured to pick up an outbound visual signal, which may comprise image data and/or video data. Because the assembly is worn on a wearer's head, the outbound visual signal will typically reflect the view as seen from the wearer's perspective, although other views may also be captured including, but not limited to, views directed behind the wearer. As with the other aforementioned multimedia components, the camera assembly is further structured to transmit the outbound visual signal to the processor. Accordingly, the processor is structured to transmit to another entity the outbound visual signal received from the camera assembly.
- The processor is also structured to receive an incoming visual signal from a remote location. For instance, the incoming visual signal may comprise image data and/or video data, such as pictures or movies from a remotely stored and transmitted data file. As such, the communication eyewear assembly further comprises a display assembly disposed and structured to display the incoming visual signal received by the processor. More specifically, the display assembly includes at least one display unit disposed and structured to display the incoming visual signal to the wearer.
- As indicated, above, the communication eyewear assembly is additionally structured to identify a zone of safety about the wearer and to notify the wearer and/or others of the zone of safety. In particular, the processor is structured to receive, send, and process a variety of data associated with ascertaining and/or identifying the zone of safety. This includes multimedia communication capabilities associated various sources and recipients of information, preferably via a wireless communication system.
- Accordingly, the processor can communicate with remote data sources such as a remote environmental data source or a remote locational data source. Environmental data can include general as well as specific information about the environment around the eyewear wearer, such as dangerous temperatures, toxic gases, airborne chemical or biological agents, ionizing radiation, potentially damaging energy fields, atmospheric pressure, etc. Locational data can include information about the location or orientation of the wearer and/or other items of interest.
- The processor can also communicate with local data sources such as a local environmental data source or a local locational data source generally situated in the vicinity of the area in which the wearer is operating. The environmental and locational data provided by the local sources can comprise similar types of information to that provided by the remote data sources as well as additional stored and/or real-time information based on direct local measurements, observations, inputs, etc.
- The processor can also communicate with data sources worn by the wearer such as a personal environmental data source, a personal locational data source, or a biometric data source, which may be embedded in the wearer's clothing or attached thereto, such as via arm band, leg band, belt, backpack, etc.
- The processor can also communicate with personal data sources worn by individuals other than the wearer, particularly those in proximity to the wearer. Information obtained and provided by such data sources can comprise similar types of information to that provided by the remote data sources, the local data sources, and the wearer's personal data sources.
- In addition, the communication eyewear assembly, itself, can include similar types of data sources integrated therewith, such as, but not limited to, environmental sensors, locational sensors, and/or biometric sensors indicating wearer's physical status.
- The above-described data can be utilized to compare the wearer's present physical status with predetermined criteria and to notify the wearer and/or others of the wearer's physical status.
- The above-described data can also be processed by the processor and/or a remote processor to identify a zone of safety about the wearer. The processors can provide a wide variety of associated functionality, such as in some cases identifying a predetermined zone of safety based on observed parameters, or in other cases, calculating a zone of safety based on pre-existing data received. The zone of safety can be determined for one, two, or three dimensions, as well as comprise a temporal aspect, such as in the case of a dynamic, changing situation.
- Once the processor and/or remote processor determine the zone of safety, the communication eyewear assembly is capable of notifying the wearer of same, such as by displaying hazardous conditions information through a visual display, by communicating a pre-determined audio spoken message or tone to the eyewear's audio earpieces, or by vibrational means, electrical shock, heat, etc.
- Similarly, other individuals in the vicinity can receive appropriate zone of safety notification on their personal communication device, as can the wearer if he/she is wearing a separate personal communication device. Moreover, remote locations such as control stations, dispatch stations, etc., can receive the zone of safety notification and process the associated information accordingly.
- In accordance with the monitoring of the wearer and/or the communication eyewear worn by the wearer, the present invention can further involve the assignment of a unique ID to the wearer's communication eyewear assembly. Further, the unique ID can be transmitted to a remote location or as part of normal wireless transmissions to the wearer's personal communication device for purposes of confirming the identity of the wearer, and times of use. The unique ID can be contained in the processor and/or the locational sensors, either alone or in combination with processor, or by separate means such as those involving local RFID, for purposes of establishing location and movement from point to point.
- In a further embodiment, the communication eyewear assembly is structured to determine existence of extraordinary circumstances based on changes or lack of changes in the data, and to notify the wearer and/or others of same. Relevant data can comprise biometric or other data of the individual wearer and/or that of the environment surrounding the wearer, and can be utilized to assess the well-being of the wearer and to optionally notify remote locations of the wearer's health status, possible distress, etc.
- In a still further embodiment, the communication eyewear assembly may comprise a user interface, such as a touch sensitive interface, brain wave monitor, motion sensor, optical recognition monitor, etc. The user interface is disposed in communication with the processor and is structured to relay received input to the processor so as to permit selective control of at least one type of input parameter that can be received and processed by the processor. By way of further example, the wearer can interact with various embodiments of the user interface by touching it (e.g., swiping, tapping, touching and holding), using forehead movement, nose movement, ear movement, eye movement, eyelid movement, or in some cases by simply thinking, etc., as well as ay combinations thereof. For instance, in one embodiment, the user interface can be structured to facilitate selection or control of the visual display of incoming data. In another embodiment, the user interface can be structured to facilitate control of audio volume. In still another embodiment, the user interface can be structured to facilitate input of an emergency distress signal, such as a pattern of successive taps, etc.
- As an additional control interface feature, the processor can be structured to process audio instructions received directly from the wearer via the microphone, such as for manipulation and adjustment of display and/or audio settings, or to initiate a distress signal.
- These and other objects, features and advantages of the present invention will become clearer when the drawings as well as the detailed description are taken into consideration.
- For a fuller understanding of the nature of the present invention, reference should be had to the following detailed description taken in connection with the accompanying drawings in which:
-
FIG. 1 is a perspective view of one embodiment of a communication eyewear assembly in accordance with the present invention. -
FIG. 2 is a schematic illustration of one embodiment of a transceiver configuration in accordance with the present invention. -
FIG. 3 is a schematic illustration of one embodiment of a wireless communication system in accordance with the present invention. -
FIG. 4 is a schematic illustration of one embodiment of an established zone of safety in accordance with the present invention. -
FIG. 4A is a schematic illustration of one embodiment of a display presentation in accordance with the established zone of safety ofFIG. 4 . -
FIG. 5 is a cross section view of one embodiment of an earpiece in accordance with the present invention. -
FIG. 6 is a partial perspective view of another embodiment of a microphone assembly in accordance with the present invention. -
FIG. 7 is a partial perspective view of a communication eyewear assembly having a display assembly which includes two different types of display units. - Like reference numerals refer to like parts throughout the several views of the drawings.
- As shown throughout the accompanying figures, the present invention is directed towards a communication eyewear assembly generally indicated as 10.
- The
communication eyewear assembly 10 is structured to be worn on a person's head so as to protect the person's eyes and also to permit hands-free wireless multimedia communication. Thecommunication eyewear assembly 10 is additionally structured to identify a zone of safety about the wearer, as distinguished from a hazardous object or zone, and to notify the wearer and/or others of the zone of safety, such as that of an industrial environment, or that found in a variety of other situations, such as, but not limited to, military operations, fire and rescue operations, public safety operations, sports activities, interactive gaming, etc. Moreover, theassembly 10 can process a variety of data, receivable from a variety of sources, to ascertain and/or identify the zone of safety. Theassembly 10 can also determine existence of extraordinary circumstances based on changes or lack of changes in the data, and notify the wearer and/or others of same, particularly with regard to the status of the wearer. - As seen in
FIG. 1 , thecommunication eyewear assembly 10 comprises aframe assembly 20 which is structured to be worn on a person's head. Accordingly, theframe assembly 20 is structured to be supportably disposed on the wearer so as to maintain at least one substantiallytransparent eye panel 25 in a generally shielding position in front of the wearer's eyes. - In the embodiment of
FIG. 1 , theframe assembly 20 comprises afront frame member 21 and twoeye panels 25 coupled thereto. In this embodiment, eacheye panel 25 is structured to substantially shield the front of a different corresponding eye of the wearer while also permitting clear vision therethrough. - For applications where it is desirable to reduce the brightness of light perceived by the wearer's eyes, the
eye panels 25 can be tinted. For applications where it is desirable to reduce the amount of harmful wavelengths of light reaching the wearer's eyes, theeye panels 25 can include ultraviolet protection such as UV coatings. Theeye panels 25 can also be polarized for protective purposes and/or for viewing enhancement purposes. Theeye panels 25 can also comprise prescription lenses. In at least one embodiment, theeye panels 25 can also comprise a shatterproof material of construction for additional protection in the event of significant impact to theeye panels 25 by a foreign object. - The
frame assembly 20 further comprises at least oneleg front frame member 21 in front of the wearer's eyes during activity of the wearer. In the embodiment ofFIG. 1 , theframe assembly 20 comprises afirst leg 23 and asecond leg 24, eachleg front frame member 21 towards a rear thereof. Thelegs front frame member 21 so as to extend over and grip the wearer's ears when thefront frame member 21 is operatively disposed in front of the wearer's eyes. - In at least one embodiment, the
frame assembly 20 further comprises pair ofside eye panels 27. Eachside eye panel 27 is disposed and structured to substantially shield a different eye of the wearer. Moreover, theside eye panels 27 can be further structured to permit clear peripheral vision therethrough. As shown inFIG. 1 , in at least one embodiment, theside eye panels 27 are each coupled to adifferent leg eye panels 25, theside eye panels 27 can be tinted, UV coated, and/or polarized. Theside eye panels 27 can also comprise a shatterproof material of construction. - Turning now to the communication aspects of the present invention, the
communication eyewear assembly 10 provides for wireless communication between a wearer and other entities, such as, but not limited to, other wearers, control stations, broadcast stations, network servers, remote network peers, local and/or remote data sources, etc. By way of example, the wireless communication could involve a wearer located on a factory floor communicating with a manager located in a control station. As another example, the wireless communication could involve a wearer participating in an investigation of a hazardous event while in contact with an emergency response vehicle and/or a related dispatch center. As yet another example, the wireless communication could involve a wearer participating in an interactive sport or game together with a plurality of other participants in real time and/or space. These examples are meant for illustrative purposes only, as the present invention can apply to a variety of other applications and circumstances. - Accordingly, the
communication eyewear assembly 10 comprises at least oneprocessor 40 which is coupled to theframe assembly 20. Theprocessor 40 is structured and disposed to receive an incoming data signal, typically from a remote location. For example, in one embodiment, the incoming data signal comprises real-time voice data, such as from another person with whom the wearer is communicating. In another embodiment, the incoming data signal comprises environmental data, such as from a remotely stored and transmitted database file. Theprocessor 40 is further structured and disposed to transmit an outbound data signal to another entity, such as real-time voice data from the wearer to another person with whom the wearer is communicating. Further details of the multimedia communication capabilities of theprocessor 40, including the incoming and outbound visual signals, environmental, and locational data, etc., are discussed in more detail below, particularly with regard to the zone of safety capability. - In at least one embodiment, the
processor 40 is structured and disposed to fit inside of theframe assembly 20. For example, in the embodiment ofFIG. 1 , theprocessor 40 is disposed inside of thefirst leg 23. It should be noted, however, that it is within the scope and intent of the present invention that theprocessor 40 can be structured and disposed to fit inside of other portions of theframe assembly 20, including, but not limited to, thesecond leg 24, thefront frame member 21, etc. - The
communication eyewear assembly 10 further comprises apower supply 90 structured and disposed to provide power to theprocessor 40. Thepower supply 90 can also be structured to supply power to other components of theassembly 10. In at least one embodiment, thepower supply 90 is structured and disposed to fit inside of theframe assembly 20. For example, in the embodiment ofFIG. 1 , thepower supply 90 is disposable inside of thesecond leg 24. It should be noted, however, that it is within the scope and intent of the present invention that thepower supply 90 can be structured and disposed to fit inside of other portions of theframe assembly 20, including, but not limited to, thefirst leg 23, thefront frame member 21, etc. Thepower supply 90 can also be rechargeable. For example, in one embodiment, thepower supply 90 comprises a rechargeable battery pack that can be charged by an external power source that has been removably coupled to theassembly 10. - The
communication eyewear assembly 10 further comprises at least oneearpiece 60 structured to be disposed in either wired or wireless communication with theprocessor 40 and further structured to be disposed at one of the wearer's ears. As such, theprocessor 40 retransmits a received incoming audio signal to at least oneearpiece 60, which in turn communicates the incoming audio signal to the wearer. With specific reference to the embodiment ofFIG. 1 , thecommunication eyewear assembly 10 comprises twoearpieces 60, each of which are structured to be disposed at a different one of the wearer's ears. Moreover, in the embodiment ofFIG. 1 , theearpieces 60 are further structured to be at least partially inserted into the wearer's ears when disposed in their respective operative orientations. - With regard to inter-component connectivity as well as remote connectivity, the
processor 40 comprises at least one transceiver structured to facilitate wireless communication. For example, with reference to the embodiment ofFIG. 2 , theprocessor 40 can comprise afirst transceiver 42 and asecond transceiver 44. Thefirst transceiver 42 is structured to receive an incoming data signal from a remote source, and is also structured to transmit an outbound data signal to a remote destination. The data signal can include a wide variety of information, including, but not limited to, audio and/or visual information. In at least one embodiment, the first transceiver utilizes radio frequency communication. It is, however, within the scope and intent of the present invention that other suitable wireless technologies may also be utilized, including, but not limited to cellular, satellite, WIFI, WIMAX, BLUETOOTH, etc. The outbound data signal can comprise an audio signal corresponding to the wearer's voice which has been picked up by anoutbound microphone assembly 70, such as the one shown inFIG. 1 and described in greater detail below. Similarly, an outbound visual signal will ordinarily correspond to an image captured by acamera assembly 105, also described in greater detail below. The outbound data signal can also comprise information detected by various sensors, as discussed further below. - The
processor 40 can also comprise a local data storage capability or “memory” 43, such as for use in scenarios that require local storage and convenient access to specific data. - The
second transceiver 44 ofFIG. 2 is structured to wirelessly communicate with theearpieces 60. In one embodiment, the second transceiver utilizes radio frequency communication technology. In another embodiment, the second transceiver utilizes magnetic induction communication technology. It is, however, within the scope and intent of the present invention that other suitable wireless technologies may also be utilized, including, but not limited to BLUETOOTH technology. In an alternative embodiment, theprocessor 40 comprises a single transceiver structured to communicate with remote entities as well as with theearpieces 60. - In addition, each
earpiece 60 comprises anearpiece transceiver 64 structured to facilitate communication with theprocessor 40. For example, in the embodiment ofFIG. 2 , theearpiece transceivers 64 are each disposed in wireless communication with thesecond transceiver 44. In such an embodiment, the specific communication technology of theearpiece transceivers 64 matches that of thesecond transceiver 44, such as, but not limited to, radio frequency technology, magnetic induction technology, BLUETOOTH, etc. In a further embodiment, eachearpiece transceiver 64 is uniquely identifiable by theprocessor 40. For example, in one such embodiment, eachearpiece transceiver 64 comprises a unique signature, such as a radio, BLUETOOTH or magnetic identification, which is recognizable by theprocessor 40. Accordingly, theprocessor 40 can transmit different content to eachearpiece transceiver 64, as may be desired. - As indicated, above, the
communication eyewear assembly 10 is additionally structured to identify a zone of safety about the wearer and to notify the wearer and/or others of the zone of safety. In particular, as discussed next, theprocessor 40 is structured to receive, send, and process a variety of data associated with ascertaining and/or identifying the zone of safety. This includes multimedia communication capabilities associated various sources and recipients of information, preferably via a wireless communication system. - With reference to
FIG. 3 , a schematic illustration depicts one embodiment of awireless communication system 110 in accordance with the present invention. In particular, theprocessor 40 and various electronic data sources are structured to communicate with one ormore communication networks 120 and protocols, such as, but not limited to, the Internet, cellular, satellite, WIFI, WIMAX, WLAN, WAN, LAN, WIFI hotspot, Intranet, ZIGBEE, ENOCEAN, etc. The various data sources can also communicate directly with theprocessor 40, using a variety of wireless technologies and protocols, such as via cellular, satellite, WIFI, WIMAX, ZIGBEE, ENOCEAN, BLUETOOTH, etc. Further, aremote processor 140 can facilitate processing and organization of data and other information, and may be structured to enhance and/or supplement the computational capacity of the in-frame processor 40, such as for determining the zone of safety. For all of the foregoing, it is noted that a variety of processing and data storage hardware and software, operating systems, and communication protocols may be implemented to carry out various embodiments of the present invention. - Continuing with the representative embodiment of
FIG. 3 , theprocessor 40 can communicate with remote data sources such as a remoteenvironmental data source 181 or a remote locational data source 183. Environmental data can include general as well as specific information about the environment around the eyewear wearer. By way of example, only, information provided by the remoteenvironmental data source 181 can pertain to potentially dangerous temperatures, toxic gases, airborne chemical or biological agents, ionizing radiation, potentially damaging energy fields, atmospheric pressure, etc. Locational data can include information about the location of the wearer as well as the location of items of interest. For instance, the information provided by the remote locational data source 183 can include, but is not limited to, the wearer's position on the face of the earth, location of restricted areas, relative locations of potentially hazardous stationary or moving objects around the eyewear wearer, etc. - The
processor 40 can also communicate with local data sources such as a localenvironmental data source 281 or a locallocational data source 283 generally situated in the vicinity of the area in which the wearer is operating. The environmental and locational data provided by thelocal sources remote data sources 181,183, as discussed above. Moreover, they may comprise additional stored and/or real-time information based on direct local measurements, observations, inputs, etc. For instance, in addition to archived or stored information, the local data sources can comprise a wide variety of active devices, such as thermometers, altimeters, locational beacons, alarms, motion sensors, and sensors of biological agents, chemical agents, and radiation, among others. - The
processor 40 can also communicate with data sources worn by the wearer such as a personalenvironmental data source 381, a personallocational data source 383, or abiometric data source 385, which may be embedded in the wearer's clothing or attached thereto, such as via arm band, leg band, belt, backpack, etc. The environmental and locational data provided by thepersonal sources 381,382 can comprise similar types of information to that provided by theremote data sources 181,183 and thelocal data sources biometric data sources 385 may comprise additional stored and/or real-time information based on biometric data of the wearer, such as, but not limited to, body temperature, heartbeat, and rate of respiration. For instance, in addition to archived or stored information, the personal biometric data sources can comprise a wide variety of active devices such as body temperature monitors, heartbeat monitors, respiration monitors, perspiration monitors, muscle reflex monitors, blood sugar monitors, etc. - The
processor 40 can also communicate withpersonal data sources 250 worn by individuals other than the wearer, particularly those in proximity to the wearer. Information obtained and provided bysuch data sources 250 can comprise similar types of information to that provided by theremote data sources 181,183, thelocal data sources personal sources - In addition, the
communication eyewear assembly 10, itself, can include similar types of data sources integrated therewith, which are each structured to communicate with theprocessor 40. Accordingly, with reference again toFIGS. 1 and 3 , thecommunication eyewear assembly 10 comprises one or moreenvironmental sensors 81 which are coupled to theframe assembly 20. Theenvironmental sensors 81 are structured and disposed to obtain specific information about the environment around the eyewear wearer. By way of example, only, information provided by theenvironmental sensors 81 can pertain to potentially dangerous temperatures, toxic gases, airborne chemical or biological agents, ionizing radiation, potentially damaging energy fields, atmospheric pressure, ultrasonic waves, etc. - Similarly, the
communication eyewear assembly 10 comprises one or morelocational sensors 83 which are coupled to theframe assembly 20 and structured to communicate with theprocessor 40. Thelocational sensors 83 are specifically disposed and structured to provide locational data which can include information about the location of the wearer as well as the location of items of interest. As depicted in the embodiment ofFIG. 1 , alocational sensor 83 is disposed inside of thefirst leg 23. It should be noted, however, that it is within the scope and intent of the present invention that thelocational sensors 83 can be structured and disposed to fit inside of other portions of theframe assembly 20, including, but not limited to, thesecond leg 24, thefront frame member 21, etc. - The
processor 40 is structured to utilize information from thelocational sensor 83 and/or other aforementioned sources of locational data to determine the wearer's position on the face of the earth, location of restricted areas, relative locations of potentially hazardous stationary or moving objects around the eyewear wearer, etc. For instance, with regard to determining the wearer's global position, thelocational sensor 83 can be structured to receive signals from the Global Positioning System (GPS) via a geomagnetic sensing capability. Such information can be used to determine the wearer's general location and also to determine situational awareness, such as whether the wearer is venturing into restricted areas or perhaps is immobile in excess of a predetermined amount of time, signaling possible incapacitation. - In a further embodiment, the
processor 40 can utilize the locational data to determine in real time the wearer's vertical and/or horizontal and/or angular orientation thus permitting precise characterization of the wearer's location and orientation in one, two, or three spatial dimensions. Such information can include gyroscopic orientation, speed, and/or acceleration. Here too, such information can be used to determine situational awareness, such as whether the wearer is venturing into restricted areas or perhaps may be horizontally oriented for too long a period, thus may be incapacitated and require assistance. By way of practical example, such information can be utilized to inform a company on the location and status of a worker at a work site, and to facilitate a decision on whether to send help as well as to notify first responders of the wearer's status and location. - The
communication eyewear assembly 10 also comprises one or morebiometric sensors 85 which are coupled to theframe assembly 20 and structured to communicate with theprocessor 40. Thebiometric sensors 85 are specifically disposed and structured to provide biometric data on a variety of bodily functions such as body temperature, heartbeat, rate of respiration, perspiration, muscle reflex, blood sugar, etc. As depicted in the embodiment ofFIG. 1 , abiometric sensor 85 is disposed on the nasal bridge portion of thefront frame member 21. It should be noted, however, that it is within the scope and intent of the present invention that thebiometric sensors 85 can be structured and disposed to fit on or inside of other portions of theframe assembly 20, including, but not limited to, thefirst leg 23, thesecond leg 24, etc. - The
processor 40 is structured to utilize information from thebiometric sensor 85 and/or other aforementioned sources of biometric data to determine the wearer's physical status. The processor is further structured to compare the wearer's present physical status with predetermined criteria and to notify the wearer and/or others of the wearer's physical status. - With reference to
FIG. 4 , the above-described data can be processed by theprocessor 40 and/orremote processor 140 to identify a zone of safety about the wearer. The processors can provide a wide variety of associated functionality, such as in some cases identifying a predetermined zone of safety based on observed parameters, or in other cases, calculating a zone of safety based on data received. In a further embodiment, theprocessor 40 can be disposed in communication with theremote processor 140 to receive information on the zone of safety. Further still, in some embodiments, theremote processor 140 can be structured to calculate and/or identify the zone of safety and to notify theprocessor 40 accordingly. - By way of example only, one particular embodiment of an established zone of
safety 123 in a hazardous environment is schematically illustrated inFIG. 4 . In this embodiment, the zone ofsafety 123 is calculated based on data involvinghazardous objects 112, such as contaminated pieces of equipment from a wreckage site, as well as ahazard area 113, such as the affected area associated with a chemical spill, as well as locational data including the global position of the event and the relative locations of thewearer 9 and the hazard area orzone 113. Accordingly, theprocessor 40 is structured to process locational data as well as data on the hazardous objects and the hazardous environment provided by various local and/or remote sensors in order to distinguish safe areas, features, and objects from dangerous ones, and thus determine, calculate, and generate the zone ofsafety 123 accordingly. While the zone ofsafety 123 shown inFIG. 4 is two dimensional, as noted above, the zone of safety can be determined for one, two, or three dimensions, as well as comprise a temporal aspect, such as in the case of a dynamic, changing situation. - Once the
processor 40 and/orremote processor 140 determine the zone ofsafety 123, thecommunication eyewear assembly 10 is capable of notifying thewearer 9 and/or others of same, such as in contrast to potentially hazardous conditions around the wearer. For instance, theeyewear assembly 10 worn by thewearer 9 can be structured to notify thewearer 9 as to the existence of the zone of safety 123 (and its boundaries) by displaying hazardous conditions information through a variety of video means, or, through a pre-determined audio spoken message or tone to the eyewear'saudio earpieces 60. In a further embodiment, thewearer 9 is notified by theeyewear assembly 10 via vibrational means, gentle electrical shock, heat, etc., such as via thephysical notification mechanism 87 shown inFIG. 1 . In one embodiment, thephysical notification mechanism 87 comprises a vibration device. In another embodiment, thephysical notification mechanism 87 comprises an electro-shocking device. In yet another embodiment,physical notification mechanism 87 comprises a heating device. - Similarly, other individuals in the vicinity can receive appropriate zone of safety notification on their
personal communication device 11′, as can the wearer if he/she is wearing a separatepersonal communication device 11. Moreover, remote locations such as control stations, dispatch stations, etc., can receive the zone of safety notification and process the associated information accordingly. - In accordance with the monitoring of the wearer and/or the
communication eyewear 10 worn by the wearer, the present invention can further involve the assignment of aunique ID 45 to the wearer'scommunication eyewear assembly 10. Further, the unique ID can be transmitted to a remote location or as part of normal wireless transmissions to the wearer'spersonal communication device 11 for purposes of confirming the identity of the wearer, and times of use. Theunique ID 45 can constitute a singular ‘public’ identity, similar to that associated with a cellphone account. In contrast, while current telecommunication eyewear units might have a separate BLUETOOTH identity for use in connection with peripheral devices, such limited identity is not specifically tied to the eyewear or the wearer in the manner contemplated herein. In the present invention, each eyewear/user can have an identity. For instance, the individualized identity affiliated with theunique ID 45 can be used in large industrial plants to monitor conduct and productivity of corresponding workers/wearers. On the physical level, the unique ID can be contained in theprocessor 40 and/or thelocational sensors 83, either alone or in combination withprocessor 40, or by separate means such as those involving local RFID, for purposes of establishing location and movement from point to point. - Turning to some of the visual features of the present invention,
FIG. 4A is a schematic illustration of one embodiment of a visual display presentation associated with adisplay unit 26 in accordance with the established zone ofsafety 123 ofFIG. 4 . As depicted in this embodiment, the zone ofsafety 123′ comprises three spatial dimensions and is also depicted in contrast to a zone ofdanger 124′. In a further embodiment, wherein the display is updated in real-time, the zone ofsafety 123′ also comprises a temporal dimension. The visual display is primarily displayed in connection with thecommunication eyewear assembly 10 worn by thewearer 9 but can also be displayed on other devices as well, includingpersonal communication devices - As seen in
FIG. 4A , the three-dimensional zone ofsafety 123′ borders acorresponding danger zone 124′ which encompasses a spill zone or similar type ofhazard area 113. The zone ofsafety 123′ includes a two-dimensional base area, along with a height dimension, such as to avoid vapors associated with thespill zone 113,hazardous objects 112, etc. Other objects and items ofFIG. 4 are also represented by the visual display presentation shown inFIG. 4A . In a further embodiment, thedisplay unit 26 is structured to present additional information in graphical and/or script format, such as acommunication reception gauge 151, adanger scale 153, and/or atext box 155. Each of these can be customized to represent a variety of safety, environmental, personal, and/or communicative data. - For instance, information received from the aforementioned sensor devices embedded in the
communication eyewear assembly 10, or other remote sources can be graphically and/or textually displayed via thedisplay unit 26, including, by way of example, locational information about potentially hazardous stationary or moving objects around the wearer. Additionally, theprocessor 40 is structured to identify whether a potentially hazardous movingobject 112 is increasing its risk to thewearer 9, through movement of the object itself toward the wearer, or the wearer toward the object. For example, thecommunication eyewear assembly 10, through use of theprocessor 40 and multiple microphones 76 (discussed in more detail, below), can cooperatively determine whether a potentiallydangerous object 112 is approaching thewearer 9, as well as the object's 122 position, via triangulation or other appropriate calculations, and inform the wearer accordingly. - In a further embodiment, the
assembly 10 is structured to determine existence of extraordinary circumstances based on changes or lack of changes in the data, and to notify the wearer and/or others of same. The data can comprise biometric or other data of the individual wearer and/or that of the environment surrounding the wearer. Examples of such data include, but are not limited to, potentially dangerous ambient temperatures, toxic gases, airborne chemical or biological agents, ionizing radiation, or other potentially damaging energy fields, atmospheric pressure, unusual body temperature, heartbeat, or rate of respiration, etc. The data can be used by theprocessor 40 alone or in conjunction with theremote processor 140 to assess the well-being of thewearer 9 and to optionally notify remote locations, of the wearer's health status, possible distress, etc. - More specifically, the
processors wearer 9, and autonomously and wirelessly report the information and condition to remote sources. Conditions determined to constitute an anomaly or other unusual occurrence, can thus signal a potential risk to the wearer, which in turn can trigger a corresponding notification, such as via an alarm feature. Additionally, the information could be used for purposes of security, insurance documentation, or other documentation useful to human resources at the employer of the wearer. - Referring now to some audio capabilities of the
communication eyewear assembly 10, reference is made to the embodiment ofFIG. 5 , wherein theearpiece 60 further comprises aspeaker assembly 68 structured and disposed to audibly communicate an incoming audio signal (audio data) to the wearer. In this embodiment, thespeaker assembly 68 is structured to be at least partially inserted into the ear of the wearer when disposed in an operative orientation. - The
earpiece 60 is additionally structured to substantially muffle ambient noise, further facilitating communication. As shown in the embodiment ofFIG. 5 , at least a portion of theearpiece 60 is structured to be inserted into the wearer's ear so as to substantially muffle the ambient noise. More specifically, theearpiece 60 ofFIG. 5 additionally comprises adisposable insert 62 structured to substantially muffle the ambient noise. Thedisposable insert 62 can be formed, for instance, from a sound-deadening, compressible material, such as high density foam, although a variety of other suitable materials may also be utilized, including materials which readily conform to fit inside of the ear canal. For at least sanitary purposes, thedisposable insert 62 can be discarded and replaced after use. - Also depicted in
FIG. 5 is theearpiece transceiver 64, which is communicatively coupled to both thespeaker assembly 68 and an optionalearpiece microphone assembly 66. Theearpiece transceiver 64 is structured to facilitate communication between therespective earpiece components processor 40. Theearpiece 60 further comprises anearpiece power supply 92 structured and disposed to provide power to thespeaker assembly 68 and theearpiece microphone assembly 66, as well as to theearpiece transceiver 64. In at least one embodiment, theearpiece power supply 92 can be rechargeable. - Referring again to
FIG. 1 , thecommunication eyewear assembly 10 further comprises anoutbound microphone assembly 70 disposed and structured to pick up an outbound audio signal from the wearer and transmit the outbound audio signal to theprocessor 40. For instance, in the embodiment ofFIG. 1 , theoutbound microphone assembly 70 comprises an adjustableelongate segment 71 which extends from thefirst leg 23 of theframe assembly 20. The adjustableelongate segment 71 is structured to permit adjustment of themicrophone assembly 70 in front of the wearer's mouth to better pick up the outbound audio signal, which, in at least one embodiment, comprises real-time voice data of the wearer's spoken voice. In this embodiment, theoutbound microphone assembly 70 is in disposed in wired communication with theprocessor 40. However, in another embodiment, theoutbound microphone assembly 70 can be disposed in wireless communication with theprocessor 40, such as with thesecond transceiver 44. - The
communication eyewear assembly 10 can be further structured to enhance an incoming and/or outbound audio signal based on an ambient audio signal. This can include enhancing a voice component of the received audio data, such as that of the wearer or that of others in the vicinity of the wearer. This can also include enhancement of informational sounds such sirens, whistles, horns, gunfire, explosions, and other identifiable audio sounds that are distinct from the general background noise. By way of example, the ambient audio signal will typically comprise the background sounds generated by the surrounding environment. As such, one or more ambient microphone assemblies can be disposed and structured to pick up an ambient audio signal and transmit the ambient audio signal to theprocessor 40 for enhancement processing, filtering, etc. As seen in the embodiment ofFIG. 5 , theearpiece microphone assembly 66 can serve as an ambient microphone assembly. As another example, one or moreframe microphone assemblies 76, such as those depicted inFIG. 6 , can provide an ambient microphone capability. - Referring now to some additional operational capabilities of the
communication eyewear assembly 10, and with reference again toFIG. 1 , thecommunication eyewear assembly 10 further comprises auser interface 80, such as a touch sensitive interface, brain wave monitor, motion sensor, optical recognition monitor, etc. The user interface is disposed in communication with theprocessor 40 and is structured to relay received input to theprocessor 40 so as to permit selective control of at least one type of input parameter that can be received and processed by theprocessor 40. By way of further example, the wearer can interact with various embodiments of theuser interface 80 by touching it (e.g., swiping, tapping, touching and holding), using forehead movement, nose movement, ear movement, eye movement, eyelid movement, or in some cases by simply thinking, etc., as well as ay combinations thereof. For instance, in one embodiment, theuser interface 80 can be structured to facilitate selection or control of the visual display of incoming data. In another embodiment, theuser interface 80 can be structured to facilitate control of audio volume. In still another embodiment, theuser interface 80 can be structured to facilitate input of an emergency distress signal, such as a pattern of successive taps, etc. Theuser interface 80 can be disposed in a variety of suitable locations, such as on theframe assembly 20 or on at least one of theearpieces 60. In the embodiment ofFIG. 1 , for instance, theuser interface 80 is disposed on thefirst leg 23. - As an additional control interface feature, the
processor 40 can be structured to process audio instructions received directly from thewearer 9 via themicrophone 70, such as for manipulation and adjustment of display and/or audio settings, or to initiate a distress signal. - Turning again to
FIG. 6 , twoframe microphone assemblies 76 are each disposed on a different side of thefront frame member 21. As with theearpiece microphone assemblies 66, discussed above, theframe microphone assemblies 76 are disposed and structured to pick up an ambient audio signal and transmit the ambient audio signal to theprocessor 40. Further, in the embodiment ofFIG. 6 , a distance of several inches, preferably about four to five inches (4″-5″), which separates theframe microphone assemblies 76 from one another provides sufficient separation such that disparate sources of sound can be geometrically differentiated by theprocessor 40 based on the variations in the audio signals picked up by each individualframe microphone assembly 76. In this embodiment, theframe microphone assemblies 76 are in disposed in wired communication with theprocessor 40. However, in another embodiment, theframe microphone assemblies 76 can be disposed in wireless communication with theprocessor 40, such as with thesecond transceiver 44. It should also be noted that, other numbers offrame microphone assemblies 76, could be used. Moreover, in other embodiments, theframe microphone assemblies 76 can also be disposed at other suitable locations, such as on thelegs - With further reference to
FIGS. 5 , in at least one embodiment, theearpieces 60 are further structured to be removably secured to theframe assembly 20 for storage. As seen inFIG. 5 , eachearpiece 60 comprises aclip 69 structured to removably secure theearpiece 60 to theframe assembly 20 when theearpiece 60 is not in use. For instance, theearpieces 60 can be removably secured to at least one of thelegs earpieces 60 may also be removably secured to other portions of the frame assembly, including, but not limited to, thefront frame member 21. In a still further embodiment, theearpiece 60 and theframe assembly 20 can be cooperatively structured to charge a rechargeableearpiece power supply 92 when theearpiece 60 is disposed in a stored orientation. - Turning now to some additional aspects of the present invention involving visual communication, the
assembly 10 further includes acamera assembly 105 disposed and structured to pick up an outbound visual signal, which may comprise image data and/or video data. Referring again toFIG. 1 , in one embodiment, thecamera assembly 105 comprises a single camera element disposed on a central area of thefront frame member 21 and facing towards a field of view from the wearer's perspective. It should be noted, however, that it is within the scope and intent of the present invention that a plurality of cameras elements may be utilized, and further, that one ormore camera assemblies 105 may be disposed on a variety of other suitable portions of theassembly 10, including, but not limited to, other positions on the front frame member, on thelegs assembly 10 is worn on a wearer's head, the outbound visual signal will typically reflect the view as seen from the wearer's perspective, although other views may also be captured including, but not limited to, views directed behind the wearer. Further, thecamera assembly 105 will typically pick up visual signals in the visible spectrum. However, in a further embodiment, thecamera assembly 105 can pick up visual signals in other spectra, including, but not limited to, the infrared spectrum. - As with the other aforementioned multimedia components, the
camera assembly 105 is further structured to transmit the outbound visual signal to theprocessor 40. Accordingly, theprocessor 40 is structured to transmit to another entity the outbound visual signal or data received from thecamera assembly 105. - As noted, above, the
processor 40 is also structured to receive an incoming visual signal from a remote location. For instance, the incoming visual signal may comprise image data and/or video data, such as pictures or movies from a remotely stored and transmitted data file, or real-time feeds from remote sources such as hidden cameras, satellites, aerial vehicles, control towers, etc. As such, theassembly 10 includes at least onedisplay unit 26 disposed and structured to display the incoming visual signal to the wearer. For example, with reference to the embodiment ofFIG. 1 , theassembly 10 comprises asingle display unit 26 which is at least partially disposed on an exterior surface of one of theeye panels 25. Thedisplay unit 26 is further structured to project an image of the received incoming visual signal towards the wearer's eye. In another embodiment, thedisplay unit 26 is structured to project an image of the received incoming visual signal onto theeye panel 25 such that the wearer's eye can perceive the displayed visual signal as emanating from the interior and/or exterior surface of theeye panel 25. Although the embodiment ofFIG. 1 depicts thedisplay unit 26 positioned on the exterior surface of theeye panel 25, thedisplay unit 26 may alternatively be disposed on the interior surface of theeye panel 25, or within theeye panel 25. - In a further embodiment, the
assembly 10 includes at least onedisplay unit 26 corresponding to eacheye panel 25. For instance, the embodiment ofFIG. 7 depicts two other variations of thedisplay unit 26′, 26″, each coupled to arespective eye panel 25. As depicted, afirst display unit 26′ is disposed on the interior surface of oneeye panel 25, and is further structured to display the incoming visual signal in a partial field of vision defined by theeye panel 25. Thesecond display unit 26″ is disposed within theother eye panel 25′, and is further structured to display the incoming visual signal in substantially the entire field of vision defined by theeye panel 25′. - The
display unit display unit display unit eye panel 25. - The
display unit processor 40, whereupon thedisplay unit display unit processor 40. - It should also be noted that, although the above description focuses on certain specific uses of the present invention, particularly those in a hazardous industrial work environment, the present invention could be readily applied to other scenarios, such as to military operations, fire and rescue operations, public safety operations, sports activities, interactive gaming, etc.
- Since many modifications, variations and changes in detail can be made to the described preferred embodiment of the invention, it is intended that all matters in the foregoing description and shown in the accompanying drawings be interpreted as illustrative and not in a limiting sense. Thus, the scope of the invention should be determined by the appended claims and their legal equivalents.
- Now that the invention has been described,
Claims (32)
1. A communication eyewear assembly comprising:
a frame assembly;
at least one substantially transparent eye panel coupled to said frame assembly;
said frame assembly structured to be supportably disposed on a wearer so as to maintain said at least one eye panel disposed in front of a wearer's eye;
a processor coupled to said frame assembly, said processor structured and disposed to receive an incoming data signal from at least one source of data;
said processor structured to utilize said data to generate zone of safety information corresponding to a physical zone of safety, said physical zone of safety defining a physical boundary in least one spatial dimension relative to a hazardous area;
said processor further structured to communicate said zone of safety information to at least the wearer.
2. An assembly as recited in claim 1 wherein said zone of safety information comprises data pertaining to at least two spatial dimensions.
3. An assembly as recited in claim 2 wherein said zone of safety information comprises data pertaining to at least three spatial dimensions.
4. An assembly as recited in claim 1 wherein said zone of safety information comprises temporal data.
5. An assembly as recited in claim 1 wherein said processor coupled to said frame assembly is further structured to communicate with a remote processor structured to assist said processor coupled to said frame assembly.
6. An assembly as recited in claim 1 wherein said processor is further structured to determine existence of an extraordinary circumstance based on changes or lack of changes in said data, and further structured to notify the wearer of same.
7. An assembly as recited in claim 1 wherein said processor is further structured to determine existence of an extraordinary circumstance based on changes or lack of changes in said data, and further structured to notify an entity other than the wearer of same.
8. An assembly as recited in claim 1 further comprising a unique ID feature structured to communicate with a personal communication device of the wearer and/or remote devices.
9. An assembly as recited in claim 1 comprising at least one earpiece disposed in communication with said processor, and further structured to be disposed in operative proximity to the at least one of the wearer's ears, said at least one earpiece comprising a speaker assembly structured and disposed to audibly communicate an incoming audio signal to the wearer.
10. An assembly as recited in claim 9 wherein said at least one earpiece is further structured to substantially muffle ambient noise.
11. An assembly as recited in claim 10 wherein at least a portion of said at least one earpiece is structured to be inserted into the wearer's ear so as to substantially muffle ambient noise.
12. An assembly as recited in claim 11 wherein said at least one earpiece comprises a disposable insert structured to substantially muffle ambient noise.
13. An assembly as recited in claim 1 wherein said at least one data source is selected from a group consisting of a remote source, a local source, a device integrated with said communication eyewear worn by the wearer, a device worn by the wearer but not integrated with said communication eyewear, and a device worn by another individual in the vicinity of the wearer.
14. An assembly as recited in claim 1 wherein said data utilized to generate said zone of safety information is selected from a group consisting of environmental data, locational data of the wearer, locational data of potentially hazardous objects in the vicinity of the wearer, and biometric data of the wearer.
15. An assembly as recited in claim 1 wherein said assembly further comprises at least one display unit disposed in communication with said processor and structured to visually display a representation of said zone of safety to the wearer.
16. An assembly as recited in claim 1 wherein said processor is further structured to communicate said zone of safety information to at least one remote device.
17. A communication eyewear assembly comprising:
a frame assembly;
at least one substantially transparent eye panel coupled to said frame assembly;
said frame assembly structured to be supportably disposed on a wearer so as to maintain said at least one eye panel disposed in front of a wearer's eye;
a processor coupled to said frame assembly, said processor structured and disposed to receive an incoming data signal from at least one source of data;
said processor structured to utilize said data to generate zone of safety information corresponding to a physical zone of safety, said physical zone of safety defining a physical boundary in least one spatial dimension relative to a hazardous area;
said processor further structured to communicate said zone of safety information to the wearer and/or at least one remote device;
said processor further structured to determine existence of an extraordinary circumstance based on changes or lack of changes in said data, and further structured to notify the wearer and/or at least one remote device of same; and
a unique ID feature structured to communicate with a personal communication device of the wearer and/or remote devices.
18. An assembly as recited in claim 17 wherein said processor coupled to said frame assembly is structured to comprise said unique ID feature.
19. An assembly as recited in claim 17 wherein said unique ID feature is integrated with a locational sensor coupled to said frame.
20. An assembly as recited in claim 17 wherein said at least one data source is selected from a group consisting of a remote source, a local source, a device integrated with said communication eyewear worn by the wearer, a device worn by the wearer but not integrated with said communication eyewear, and a device worn by another individual in the vicinity of the wearer.
21. An assembly as recited in claim 17 wherein said data utilized to generate said zone of safety information is selected from a group consisting of environmental data, locational data of the wearer, locational data of potentially hazardous objects in the vicinity of the wearer, and biometric data of the wearer.
22. An assembly as recited in claim 17 wherein said assembly further comprises at least one display unit disposed in communication with said processor and structured to visually display a representation of said zone of safety to the wearer.
23. A communication eyewear assembly comprising:
a frame assembly;
at least one substantially transparent eye panel coupled to said frame assembly;
said frame assembly structured to be supportably disposed on a wearer so as to maintain said at least one eye panel disposed in front of a wearer's eye;
a processor coupled to said frame assembly, said processor structured and disposed to receive an incoming data signal from at least one source of data;
said processor structured to utilize said data to generate zone of safety information corresponding to a physical zone of safety, said physical zone of safety defining a physical boundary in least one spatial dimension relative to a hazardous area;
said processor further structured to communicate said zone of safety information to the wearer and/or at least one remote device;
at least one display unit disposed in communication with said processor and structured to visually display a representation of said zone of safety to the wearer;
said processor further structured to determine existence of an extraordinary circumstance based on changes or lack of changes in said data, and further structured to notify the wearer and/or at least one remote device of same;
a unique ID feature structured to communicate with a personal communication device of the wearer and/or remote devices; and
at least one earpiece disposed in communication with said processor, and further structured to be disposed in operative proximity to the at least one of the wearer's ears, said at least one earpiece comprising a speaker assembly structured and disposed to audibly communicate an incoming audio signal to the wearer, said at least one earpiece further structured to substantially muffle ambient noise, wherein at least a portion of said at least one earpiece is structured to be inserted into the wearer's ear so as to substantially muffle ambient noise.
24. An assembly as recited in claim 23 wherein said at least one data source is selected from a group consisting of a remote source, a local source, a device integrated with said communication eyewear worn by the wearer, a device worn by the wearer but not integrated with said communication eyewear, and a device worn by another individual in the vicinity of the wearer.
25. An assembly as recited in claim 23 wherein said data utilized to generate said zone of safety information is selected from a group consisting of environmental data, locational data of the wearer, locational data of potentially hazardous objects in the vicinity of the wearer, and biometric data of the wearer.
26. An assembly as recited in claim 23 , wherein a physical means for communicating said zone of safety information is selected from a group consisting of sounds, audio messages, vibrations, electro-shock, heat, graphic displays, and textual displays.
27. An assembly as recited in claim 23 wherein said processor coupled to said frame assembly is further structured to communicate with a remote processor structured to assist said processor coupled to said frame assembly.
28. An assembly as recited in claim 23 wherein said processor is further structured to determine said extraordinary circumstance based on a health related data of the wearer.
29. An assembly as recited in claim 23 wherein said processor is further structured to determine said extraordinary circumstance based on a safety related data of the wearer.
30. An assembly as recited in claim 23 further structured to detect moving objects, and wherein said processor is further structured to incorporate data pertaining to the moving objects into said zone of safety information.
31. An assembly as recited in claim 30 comprising a motion detection capability coupled to said frame assembly.
32. An assembly as recited in claim 31 wherein said motion detection capability is selected from a group consisting of camera assemblies, microphone assemblies, ultrasonic transducers, magnetic position sensors, and RFID sensors.
Priority Applications (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US13/713,789 US8744113B1 (en) | 2012-12-13 | 2012-12-13 | Communication eyewear assembly with zone of safety capability |
EP13863341.7A EP2932734A1 (en) | 2012-12-13 | 2013-11-14 | Communication eyewear assembly with safety capability zone |
JP2015547952A JP2016505889A (en) | 2012-12-13 | 2013-11-14 | Communication glasses with safety function zone |
PCT/US2013/069994 WO2014092921A1 (en) | 2012-12-13 | 2013-11-14 | Communication eyewear assembly with safety capability zone |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US13/713,789 US8744113B1 (en) | 2012-12-13 | 2012-12-13 | Communication eyewear assembly with zone of safety capability |
Publications (2)
Publication Number | Publication Date |
---|---|
US8744113B1 US8744113B1 (en) | 2014-06-03 |
US20140169609A1 true US20140169609A1 (en) | 2014-06-19 |
Family
ID=50781323
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/713,789 Expired - Fee Related US8744113B1 (en) | 2012-12-13 | 2012-12-13 | Communication eyewear assembly with zone of safety capability |
Country Status (4)
Country | Link |
---|---|
US (1) | US8744113B1 (en) |
EP (1) | EP2932734A1 (en) |
JP (1) | JP2016505889A (en) |
WO (1) | WO2014092921A1 (en) |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107636897A (en) * | 2015-01-12 | 2018-01-26 | 华为技术有限公司 | Multiband antenna for wearable glasses |
JP2019502964A (en) * | 2016-01-22 | 2019-01-31 | エシロール・アンテルナシオナル | Head mounted device including environmental sensing module |
US20190056707A1 (en) * | 2017-08-18 | 2019-02-21 | The Boeing Company | Methods and Systems for Controlling Operation of Equipment Based on Biometric Data |
US10330440B2 (en) * | 2014-11-26 | 2019-06-25 | Philip Lyren | Target analysis and recommendation |
USD933635S1 (en) * | 2020-04-17 | 2021-10-19 | Bose Corporation | Audio accessory |
US11837874B2 (en) * | 2018-11-20 | 2023-12-05 | Lg Electronics Inc. | Wireless charging device |
Families Citing this family (29)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8482488B2 (en) | 2004-12-22 | 2013-07-09 | Oakley, Inc. | Data input management system for wearable electronically enabled interface |
US20120105740A1 (en) | 2000-06-02 | 2012-05-03 | Oakley, Inc. | Eyewear with detachable adjustable electronics module |
US7013009B2 (en) | 2001-06-21 | 2006-03-14 | Oakley, Inc. | Eyeglasses with wireless communication features |
WO2008076774A2 (en) | 2006-12-14 | 2008-06-26 | Oakley, Inc. | Wearable high resolution audio visual interface |
JP5944134B2 (en) * | 2011-10-14 | 2016-07-05 | シャープ株式会社 | Wireless communication device |
US9016857B2 (en) * | 2012-12-06 | 2015-04-28 | Microsoft Technology Licensing, Llc | Multi-touch interactions on eyewear |
US10306389B2 (en) | 2013-03-13 | 2019-05-28 | Kopin Corporation | Head wearable acoustic system with noise canceling microphone geometry apparatuses and methods |
US9257952B2 (en) | 2013-03-13 | 2016-02-09 | Kopin Corporation | Apparatuses and methods for multi-channel signal compression during desired voice activity detection |
US20140268016A1 (en) | 2013-03-13 | 2014-09-18 | Kopin Corporation | Eyewear spectacle with audio speaker in the temple |
CN205177388U (en) | 2013-03-15 | 2016-04-20 | 奥克利有限公司 | Eyepiece system |
CN205691887U (en) | 2013-06-12 | 2016-11-16 | 奥克利有限公司 | Modular communication system and glasses communication system |
KR20150010229A (en) * | 2013-07-18 | 2015-01-28 | 삼성디스플레이 주식회사 | Eyeglasses including transparent display and controlling method thereof |
US9629120B2 (en) * | 2014-05-23 | 2017-04-18 | Samsung Electronics Co., Ltd. | Method and apparatus for providing notification |
US9622214B2 (en) * | 2014-05-23 | 2017-04-11 | Samsung Electronics Co., Ltd. | Method and apparatus for providing notification |
US10638452B2 (en) | 2014-05-23 | 2020-04-28 | Samsung Electronics Co., Ltd. | Method and apparatus for providing notification |
CN106465458B (en) | 2014-05-23 | 2019-12-13 | 三星电子株式会社 | method and apparatus for providing notification |
US20150358717A1 (en) * | 2014-06-06 | 2015-12-10 | Plantronics, Inc. | Audio Headset for Alerting User to Nearby People and Objects |
US9855171B2 (en) * | 2015-03-04 | 2018-01-02 | 3R Products, Llc | Safety glasses |
AU2016265703B2 (en) | 2015-05-15 | 2021-06-17 | Vertical Optics, LLC | Wearable vision redirecting devices |
IL244255A (en) | 2016-02-23 | 2017-04-30 | Vertical Optics Llc | Wearable vision redirecting devices |
US9690119B2 (en) | 2015-05-15 | 2017-06-27 | Vertical Optics, LLC | Wearable vision redirecting devices |
US11631421B2 (en) | 2015-10-18 | 2023-04-18 | Solos Technology Limited | Apparatuses and methods for enhanced speech recognition in variable environments |
US10184964B2 (en) | 2016-03-21 | 2019-01-22 | International Business Machines Corporation | Automatic measurement and notification of electrical level using smartphone sensors |
KR20170119922A (en) * | 2016-04-20 | 2017-10-30 | 엘지전자 주식회사 | Portable sound equipment |
US10074356B1 (en) | 2017-03-09 | 2018-09-11 | Plantronics, Inc. | Centralized control of multiple active noise cancellation devices |
USD875821S1 (en) * | 2018-01-26 | 2020-02-18 | Snail Innovation Institute | Fiber feeding display glasses |
TWI716828B (en) * | 2019-03-08 | 2021-01-21 | 美律實業股份有限公司 | System and earphone related to transforming sensed data |
US20220128841A1 (en) * | 2020-10-27 | 2022-04-28 | Skullcandy, Inc. | Eyeglasses with associated true wireless earbuds |
US20220357766A1 (en) * | 2021-05-10 | 2022-11-10 | Anjuma KARKERA | Mobile individual secure communications environment |
Family Cites Families (79)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US3169600A (en) | 1964-01-31 | 1965-02-16 | James E Thomas | Sound head set |
US3610841A (en) | 1970-03-06 | 1971-10-05 | Pacific Plantronics Inc | Self-supporting headset |
US3807526A (en) | 1972-10-26 | 1974-04-30 | Illinois Tool Works | Ear protector |
US3787643A (en) | 1972-11-07 | 1974-01-22 | American Danish Oticon | Hearing aid device |
US3825700A (en) | 1973-03-02 | 1974-07-23 | Sonotone Corp | Articulated hearing aid temple and behind-the-ear hearing aid element |
US3943925A (en) | 1973-04-02 | 1976-03-16 | Leight Howard S | Ear protector assembly |
US3946168A (en) | 1974-09-16 | 1976-03-23 | Maico Hearing Instruments Inc. | Directional hearing aids |
US4353364A (en) | 1979-01-18 | 1982-10-12 | Woods Thomas J | Ear acoustical attenuating device |
US4683587A (en) | 1985-06-11 | 1987-07-28 | Silverman Michael D | Submersible personal stereo |
US4819270A (en) | 1986-07-03 | 1989-04-04 | Leonard Lombardo | Stereo dimensional recording method and microphone apparatus |
US5002151A (en) | 1986-12-05 | 1991-03-26 | Minnesota Mining And Manufacturing Company | Ear piece having disposable, compressible polymeric foam sleeve |
US4972491A (en) | 1988-11-30 | 1990-11-20 | Wilcox Jr Edward R | Ear-mic headset/eardefender |
US5029216A (en) | 1989-06-09 | 1991-07-02 | The United States Of America As Represented By The Administrator Of The National Aeronautics & Space Administration | Visual aid for the hearing impaired |
US5020150A (en) | 1989-11-07 | 1991-05-28 | John Shannon | Combination radio and eyeglasses |
US5265165A (en) | 1990-03-16 | 1993-11-23 | Rauch Robert A | Multipurpose headwear |
US5298692A (en) | 1990-11-09 | 1994-03-29 | Kabushiki Kaisha Pilot | Earpiece for insertion in an ear canal, and an earphone, microphone, and earphone/microphone combination comprising the same |
CA2059597A1 (en) | 1991-01-22 | 1992-07-23 | Paul A. Vogt | Radio eyewear |
JP3235865B2 (en) | 1991-06-03 | 2001-12-04 | パイオニア株式会社 | Ear speakers |
CN2108942U (en) | 1991-09-05 | 1992-07-01 | 黎明超 | Hearing aid type glasses with radio |
US5367345A (en) | 1992-02-14 | 1994-11-22 | Da Silva Jean Pierre M | Audio-adapted eyeglass retainer |
JPH0593152U (en) | 1992-05-01 | 1993-12-17 | 邦彦 亀井 | Cordless handset device |
CA2071040A1 (en) | 1992-06-11 | 1993-12-12 | Jacob Gluz | Earphone-type speaker assembly |
JPH08505238A (en) | 1992-12-04 | 1996-06-04 | バーチャル ビジョン,インコーポレイティド | Head-mounted display system |
US6010216A (en) | 1993-01-19 | 2000-01-04 | Jesiek; Daniel Stephen | "Hear speak" two-way voice radio communications eyeglasses |
US5475449A (en) | 1993-09-22 | 1995-12-12 | Pyle; Nigel | Safety glass/ear plug combination |
JP4001643B2 (en) | 1993-10-05 | 2007-10-31 | スナップ−オン・テクノロジイズ・インク | Two-hand open type car maintenance equipment |
US5541677A (en) | 1994-12-28 | 1996-07-30 | Keith Huhtala | Spectacles retaining strap with connected earplugs |
JP2002515197A (en) | 1995-09-19 | 2002-05-21 | インターバル リサーチ コーポレイション | Earphone with eyeglass attachment |
US6005536A (en) | 1996-01-16 | 1999-12-21 | National Captioning Institute | Captioning glasses |
IT1288970B1 (en) | 1996-08-09 | 1998-09-25 | Fabrizio Marano | MULTIFUNCTION PERFECTED GLASSES |
US6091832A (en) | 1996-08-12 | 2000-07-18 | Interval Research Corporation | Wearable personal audio loop apparatus |
US6012812A (en) | 1996-09-06 | 2000-01-11 | The Energy Corp. | Industrial safety assembly |
US6950531B2 (en) | 1997-11-06 | 2005-09-27 | Energy Telecom, Inc. | Industrial hearing protection and communication assembly |
US5717479A (en) | 1996-09-06 | 1998-02-10 | Rickards; Tom | Industrial safety assembly including disposable ear protection and earphone |
US5715337A (en) | 1996-09-19 | 1998-02-03 | The Mirco Optical Corporation | Compact display system |
US5886822A (en) | 1996-10-08 | 1999-03-23 | The Microoptical Corporation | Image combining system for eyeglasses and face masks |
US6204974B1 (en) | 1996-10-08 | 2001-03-20 | The Microoptical Corporation | Compact image display system for eyeglasses or other head-borne frames |
US5703670A (en) | 1996-11-19 | 1997-12-30 | Callard; Shawn R. | Earplugs adapted to eyeglasses and combination thereof |
JPH10304030A (en) | 1997-04-28 | 1998-11-13 | Saitama Nippon Denki Kk | Portable telephone system |
US6176576B1 (en) | 1997-06-06 | 2001-01-23 | Radians, Inc. | Eyewear supported by a wearer's concha of an ear |
US5988812A (en) | 1997-10-27 | 1999-11-23 | Sony Corporation | Headphone eyeglasses |
WO1999023524A1 (en) | 1997-10-30 | 1999-05-14 | The Microoptical Corporation | Eyeglass interface system |
AU759466B2 (en) | 1997-11-24 | 2003-04-17 | Tom Rickards | Industrial safety assembly including disposable ear protection and ear phone |
US6233345B1 (en) | 1998-05-05 | 2001-05-15 | Peter Urwyler | Personal earphone assembly for mounting upon eyeglasses |
IT1302767B1 (en) | 1998-09-15 | 2000-09-29 | Dada Srl | Anti-balloon device for spinning machine or twisting frames, particularly in twisting frames for elastic yarns, and in coating and covering machines |
AUPQ439299A0 (en) | 1999-12-01 | 1999-12-23 | Silverbrook Research Pty Ltd | Interface system |
EP1194806A4 (en) | 1999-06-21 | 2008-07-23 | Microoptical Corp | Eyeglass display lens system employing off-axis optical design |
US20070116318A1 (en) | 2000-01-10 | 2007-05-24 | Tom Rickards | Hearing protection and communication assembly |
US7150526B2 (en) | 2000-06-02 | 2006-12-19 | Oakley, Inc. | Wireless interactive headset |
US6325507B1 (en) | 2000-06-02 | 2001-12-04 | Oakley, Inc. | Eyewear retention system extending across the top of a wearer's head |
JP2002027072A (en) | 2000-07-10 | 2002-01-25 | Shinichi Yamada | Portable telephone set |
US6629076B1 (en) | 2000-11-27 | 2003-09-30 | Carl Herman Haken | Method and device for aiding speech |
US6769767B2 (en) | 2001-04-30 | 2004-08-03 | Qr Spex, Inc. | Eyewear with exchangeable temples housing a transceiver forming ad hoc networks with other devices |
US7013009B2 (en) | 2001-06-21 | 2006-03-14 | Oakley, Inc. | Eyeglasses with wireless communication features |
US6850166B2 (en) | 2001-06-28 | 2005-02-01 | Nokia Mobile Phones Limited | Ancillary wireless detector |
US6729726B2 (en) | 2001-10-06 | 2004-05-04 | Stryker Corporation | Eyewear for hands-free communication |
US7313246B2 (en) * | 2001-10-06 | 2007-12-25 | Stryker Corporation | Information system using eyewear for communication |
MXPA05001079A (en) | 2002-07-26 | 2005-06-03 | Oakley Inc | Wireless interactive headset. |
US7430300B2 (en) | 2002-11-18 | 2008-09-30 | Digisenz Llc | Sound production systems and methods for providing sound inside a headgear unit |
US7512448B2 (en) | 2003-01-10 | 2009-03-31 | Phonak Ag | Electrode placement for wireless intrabody communication between components of a hearing system |
US6879443B2 (en) | 2003-04-25 | 2005-04-12 | The Microoptical Corporation | Binocular viewing system |
US7221966B2 (en) | 2003-11-04 | 2007-05-22 | Ultra Electronics Audio Pack, Inc. | Wireless communication systems for masks or helmets |
US7747003B2 (en) | 2004-07-29 | 2010-06-29 | Legacy Ip Llc | Wireless headset apparatus and methods |
US7079876B2 (en) | 2003-12-23 | 2006-07-18 | Isaac Levy | Wireless telephone headset built into eyeglasses |
US20060052146A1 (en) | 2004-09-09 | 2006-03-09 | Shu-Fong Ou | Heated mounted display device with mobile phone functions |
US20060153409A1 (en) | 2005-01-10 | 2006-07-13 | Ming-Hsiang Yeh | Structure of a pair of glasses |
US7827011B2 (en) * | 2005-05-03 | 2010-11-02 | Aware, Inc. | Method and system for real-time signal classification |
US20070049363A1 (en) * | 2005-08-23 | 2007-03-01 | Green Howard D | System and method for remotely controlling a device or system with voice commands |
JP2007174612A (en) | 2005-11-28 | 2007-07-05 | Tetsuo Suda | Eyeglasses with transmitting/receiving function and portable transceiver |
US7869768B1 (en) | 2006-08-10 | 2011-01-11 | Natan Vishlitzky | Techniques for controlling speaker volume of a portable communications device |
JP4973299B2 (en) | 2007-01-19 | 2012-07-11 | ソニー株式会社 | Optical communication apparatus and optical communication method |
US8638970B2 (en) | 2007-07-31 | 2014-01-28 | Burton Technologies, Llc | In-ear adapter for earbuds |
US8498425B2 (en) | 2008-08-13 | 2013-07-30 | Onvocal Inc | Wearable headset with self-contained vocal feedback and vocal command |
US8243973B2 (en) * | 2008-09-09 | 2012-08-14 | Rickards Thomas M | Communication eyewear assembly |
JP3149350U (en) | 2008-12-12 | 2009-03-26 | 光 宮戸 | Headset type real-time communication terminal device |
US9341843B2 (en) | 2010-02-28 | 2016-05-17 | Microsoft Technology Licensing, Llc | See-through near-eye display glasses with a small scale image source |
AU2010347741A1 (en) | 2010-03-10 | 2012-09-13 | Energy Telecom, Inc. | Communication eyewear assembly |
US9330499B2 (en) | 2011-05-20 | 2016-05-03 | Microsoft Technology Licensing, Llc | Event augmentation with real-time information |
US9075249B2 (en) * | 2012-03-07 | 2015-07-07 | Google Inc. | Eyeglass frame with input and output functionality |
-
2012
- 2012-12-13 US US13/713,789 patent/US8744113B1/en not_active Expired - Fee Related
-
2013
- 2013-11-14 EP EP13863341.7A patent/EP2932734A1/en not_active Withdrawn
- 2013-11-14 JP JP2015547952A patent/JP2016505889A/en active Pending
- 2013-11-14 WO PCT/US2013/069994 patent/WO2014092921A1/en active Application Filing
Cited By (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10330440B2 (en) * | 2014-11-26 | 2019-06-25 | Philip Lyren | Target analysis and recommendation |
US20190331458A1 (en) * | 2014-11-26 | 2019-10-31 | Philip Lyren | Target Analysis and Recommendation |
US10724830B2 (en) * | 2014-11-26 | 2020-07-28 | Philip Lyren | Target analysis and recommendation |
US11002513B2 (en) * | 2014-11-26 | 2021-05-11 | Philip Lyren | Target analysis and recommendation |
US11320242B2 (en) * | 2014-11-26 | 2022-05-03 | Philip Lyren | Target analysis and recommendation |
US20230009410A1 (en) * | 2014-11-26 | 2023-01-12 | Philip Lyren | Target Analysis and Recommendation |
US11614306B2 (en) * | 2014-11-26 | 2023-03-28 | Philip Lyren | Target analysis and recommendation |
CN107636897A (en) * | 2015-01-12 | 2018-01-26 | 华为技术有限公司 | Multiband antenna for wearable glasses |
JP2019502964A (en) * | 2016-01-22 | 2019-01-31 | エシロール・アンテルナシオナル | Head mounted device including environmental sensing module |
US20190056707A1 (en) * | 2017-08-18 | 2019-02-21 | The Boeing Company | Methods and Systems for Controlling Operation of Equipment Based on Biometric Data |
US11837874B2 (en) * | 2018-11-20 | 2023-12-05 | Lg Electronics Inc. | Wireless charging device |
USD933635S1 (en) * | 2020-04-17 | 2021-10-19 | Bose Corporation | Audio accessory |
Also Published As
Publication number | Publication date |
---|---|
WO2014092921A1 (en) | 2014-06-19 |
US8744113B1 (en) | 2014-06-03 |
EP2932734A1 (en) | 2015-10-21 |
JP2016505889A (en) | 2016-02-25 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US8744113B1 (en) | Communication eyewear assembly with zone of safety capability | |
JP6522527B2 (en) | Wearable smart device for hazard detection and warning based on image and audio data | |
US10055971B2 (en) | Safety communicator—convergence of body vitals and toxic gas parameters into smartphone app to enhance safety monitoring | |
US9547972B2 (en) | Methods and systems for emergency alerts | |
EP3304525B1 (en) | Cardiopulmonary resuscitation guidance method, computer program product and system | |
WO2016086440A1 (en) | Wearable guiding device for the blind | |
EP3436814B1 (en) | Gas detector with attachment sensor | |
US9952046B1 (en) | Cellular phone and personal protective equipment usage monitoring system | |
US11890494B2 (en) | Retrofittable mask mount system for cognitive load reducing platform | |
US20100080418A1 (en) | Portable suspicious individual detection apparatus, suspicious individual detection method, and computer-readable medium | |
CN104427960A (en) | Adaptive visual assistive device | |
US20220217308A1 (en) | Camera Glasses for Law Enforcement Accountability | |
WO2016086441A1 (en) | Indoor positioning system for totally blind population | |
CN102573542A (en) | System for providing notification of positional information | |
US11328692B2 (en) | Head-mounted situational awareness system and method of operation | |
KR101569880B1 (en) | Apparatus for generating image signal and controlling system | |
US11156856B2 (en) | Eyewear with wearing status detector | |
JP2017080382A (en) | Portable information terminal and automatic imaging device equipped with the same | |
JP6120444B2 (en) | Wearable device | |
WO2017073775A1 (en) | Automatic imaging device and wearable imaging device | |
JP5982674B1 (en) | Automatic imaging device | |
EP3815067B1 (en) | Improved arboriculture safety system | |
KR20170122415A (en) | Smart helmet based on bio-signals for augmented reality of worker centric connection with csf platform | |
CN112155276A (en) | Helmet and helmet system | |
KR101572807B1 (en) | Method, apparatus and system for transmitting image signal by wearable device |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: ENERGY TELECOM, INC., FLORIDA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:RICKARDS, TOM;REEL/FRAME:032040/0722 Effective date: 20140123 |
|
FEPP | Fee payment procedure |
Free format text: MAINTENANCE FEE REMINDER MAILED (ORIGINAL EVENT CODE: REM.) |
|
LAPS | Lapse for failure to pay maintenance fees |
Free format text: PATENT EXPIRED FOR FAILURE TO PAY MAINTENANCE FEES (ORIGINAL EVENT CODE: EXP.) |
|
STCH | Information on status: patent discontinuation |
Free format text: PATENT EXPIRED DUE TO NONPAYMENT OF MAINTENANCE FEES UNDER 37 CFR 1.362 |
|
FP | Lapsed due to failure to pay maintenance fee |
Effective date: 20180603 |