US20170048611A1 - Automatically controlled directional speaker, and lamp thereof - Google Patents
Automatically controlled directional speaker, and lamp thereof Download PDFInfo
- Publication number
- US20170048611A1 US20170048611A1 US14/885,046 US201514885046A US2017048611A1 US 20170048611 A1 US20170048611 A1 US 20170048611A1 US 201514885046 A US201514885046 A US 201514885046A US 2017048611 A1 US2017048611 A1 US 2017048611A1
- Authority
- US
- United States
- Prior art keywords
- sound
- face
- automatically controlled
- image capture
- directional speaker
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/161—Detection; Localisation; Normalisation
-
- G06K9/00228—
-
- G06K9/00268—
-
- H04N5/2256—
-
- H04N5/23238—
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R1/00—Details of transducers, loudspeakers or microphones
- H04R1/20—Arrangements for obtaining desired frequency or directional characteristics
- H04R1/32—Arrangements for obtaining desired frequency or directional characteristics for obtaining desired directional characteristic only
- H04R1/34—Arrangements for obtaining desired frequency or directional characteristics for obtaining desired directional characteristic only by using a single transducer with sound reflecting, diffracting, directing or guiding means
- H04R1/345—Arrangements for obtaining desired frequency or directional characteristics for obtaining desired directional characteristic only by using a single transducer with sound reflecting, diffracting, directing or guiding means for loudspeakers
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04S—STEREOPHONIC SYSTEMS
- H04S7/00—Indicating arrangements; Control arrangements, e.g. balance control
- H04S7/30—Control circuits for electronic adaptation of the sound field
- H04S7/302—Electronic adaptation of stereophonic sound system to listener position or orientation
- H04S7/303—Tracking of listener position or orientation
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/56—Cameras or camera modules comprising electronic image sensors; Control thereof provided with illuminating means
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/61—Control of cameras or camera modules based on recognised objects
- H04N23/611—Control of cameras or camera modules based on recognised objects where the recognised objects include parts of the human body
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/698—Control of cameras or camera modules for achieving an enlarged field of view, e.g. panoramic image capture
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R1/00—Details of transducers, loudspeakers or microphones
- H04R1/02—Casings; Cabinets ; Supports therefor; Mountings therein
- H04R1/028—Casings; Cabinets ; Supports therefor; Mountings therein associated with devices performing functions other than acoustics, e.g. electric candles
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R2201/00—Details of transducers, loudspeakers or microphones covered by H04R1/00 but not provided for in any of its subgroups
- H04R2201/02—Details casings, cabinets or mounting therein for transducers covered by H04R1/02 but not provided for in any of its subgroups
- H04R2201/025—Transducer mountings or cabinet supports enabling variable orientation of transducer of cabinet
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R2205/00—Details of stereophonic arrangements covered by H04R5/00 but not provided for in any of its subgroups
- H04R2205/024—Positioning of loudspeaker enclosures for spatial sound reproduction
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R2430/00—Signal processing covered by H04R, not provided for in its groups
- H04R2430/01—Aspects of volume control, not necessarily automatic, in sound systems
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R5/00—Stereophonic arrangements
- H04R5/02—Spatial or constructional arrangements of loudspeakers
Definitions
- the technical field relates to a broadcasting apparatus, and more particularly to An automatically controlled directional speaker and capable of automatically performing face recognition to broadcast a sound to a designated person and combining with a lamp to improve the convenience of living.
- Sound amplifying device is a widely and commonly used device capable of converting current frequency into sound to play various types of audio files. Sound amplifying devices may be combined and assembled to form a playing device such as a sound box.
- the sound field of the sound amplifying device is limited by its placing position. In other words, a user may hear a sound coming from the sound amplifying device clearly only within a certain range.
- Our sense of sound depends on an analysis result processed by auditory sense of our brain, so that we can determine and feel the elements such as the direction, distance, and volume of a sound. For example, if a user wants to listen to a stereo or a radio indoors, the sound heard by the user will vary when the user is moving, since sound amplifying device is placed at a fixed position. If the user moves away from the sound amplifying device, the sound heard by the user will gradually become lower and more unclearly due to the factors of an increased distance or an deviated angle. To sense the sound in a full range, it is necessary to install a plurality of sound amplifying devices at different positions of a room or buy an expensive sound amplifying device with a stereo surround sound effect. Either way incurs a high setup cost.
- a manual remote control may be used to control the broadcasting direction of the sound amplifying device, but this method requires users to keep controlling the sound amplifying device manually according to the moving status of a listening target, and the method of controlling the direction of the sound amplifying device just achieves the effect of hearing the sound from a listening target and cannot adjust the sound effect of the listening target. In addition, if other people are near the listening target, these people will be affected. Obviously, the aforementioned method is not a good solution.
- lamp is also a necessary electric/electronic device.
- incandescent lamps at early stage to LED lamps extensively used at present various types of lamps are installed indoor and outdoor for illumination, and the lamps are connected to mains power to obtain the startup power, and a control circuit is generally provided for adjusting the light emission of the lamps.
- both sound amplifying device and lamp are electronic devices, they are usually installed separately, and their space occupying percentage is increased accordingly, so that the using space available for users is decreased.
- an automatically controlled directional speaker that is capable of detecting and identifying whether or not there is any person in the surrounding environment and automatically guiding and broadcasting sounds towards a designated person.
- Another objective of this disclosure is to provide an LED lamp combined with the aforementioned automatically controlled directional speaker, so that the LED lamp has the effects of playing sound in a direction towards a designated person automatically and providing illumination at the same time.
- this disclosure provides an automatically controlled directional speaker, and the speaker comprises: a sound amplifying device; an image capture device, for real time detecting the status of surrounding environment to generate at least one image information; a computing device, communicatively coupled to the image capture device, for determining whether or not the image information includes at least one face information; an azimuth control motor, coupled to the sound amplifying device, for controlling the sound amplifying device to output a sound azimuth according to the face information; and an amplitude control unit, communicatively coupled to the sound amplifying device, for controlling the volume of an output sound according to the face information.
- the computing device performs an analysis by a fuzzy algorithm.
- the image capture device has an infrared focusing device or an ultrasonic focusing device
- the face information includes a relative face distance and a relative face azimuth
- the relative face distance and the relative face azimuth are determined by the infrared focusing device or the ultrasonic focusing device. Therefore, the relative distance and azimuth of a person with respect to the automatically controlled directional speaker can be determined to facilitate the sound amplifying device to broadcast more accurately.
- the speaker of this disclosure further comprises an LED assist lamp installed adjacent to a side of the image capture device to provide sufficient brightness to surrounding environment and make the image information clearer.
- the image capture device includes a wide-angle camera lens, or the image capture device is mounted onto a pivot member provided for adjusting both shooting angle and range of the image capture device in top, bottom, left and right sides.
- the speaker of this disclosure further comprises a movable waveguide structure installed at a sound output position of the sound amplifying device and having a sound input port, a sound output port, and a sound channel included between the sound input port and the sound output port; and the sound input port and the sound output port have adjustable relative aperture sizes respectively; wherein the relative diameter and size of the sound input port and the sound output port may be adjusted to further adjust the volume of the output sound of the sound amplifying device to provide a better listening effect.
- this disclosure further discloses an LED illumination lamp of the automatically controlled directional speaker, characterized in that the whole LED lamp is a bulb lamp or a down light including the computing device installed therein and provided for performing an analysis by a fuzzy algorithm.
- the speaker of is disclosure may also be combined with a lamp, so that the lamp has both illumination and broadcasting functions, and such lamp can broadcast sound to a designated person automatically.
- the speaker of this disclosure has the effects of detecting and determining whether or not there are other people in the surrounding environment, and automatically adjusting the output sound to a designated person or area, so that a user can listen to the sound very well even when the user is in a moving status.
- the speaker may be set to allow a designated person to listen to the sound, so as to prevent the sound from affecting other people.
- the speaker of this disclosure may be combined with a lamp to form an LED illumination lamp and provide both illumination and broadcasting effects. With an automatic face detection function, the LED illumination lamp can broadcast sounds towards a designated person and improve the convenience of use.
- FIG. 1 is a schematic block diagram of a first preferred embodiment of this disclosure
- FIG. 2 is a perspective view of the first preferred embodiment of this disclosure
- FIG. 3 is a first schematic view showing an application of the first preferred embodiment of this disclosure
- FIG. 4A is a second schematic view showing an application of the first preferred embodiment of this disclosure.
- FIG. 4B is a third schematic view showing an application of the first preferred embodiment of this disclosure.
- FIG. 5 is a perspective view of another implementation mode of the first preferred embodiment of this disclosure.
- FIG. 6 is a schematic view of a second preferred embodiment of the present invention.
- FIG. 7 is a perspective view of another implementation mode of the second preferred embodiment of this disclosure.
- the automatically controlled directional speaker 1 comprises a sound amplifying device 10 , an image capture device 11 , a computing device 12 , an azimuth control motor 13 and an amplitude control unit 14 .
- the image capture device 11 is provided for real time detecting the status of surrounding environment to generate at least one image information
- the computing device 12 is communicatively coupled to the image capture device 11 for determining whether or not there is at least one face information in the image information generated by the image capture device 11 .
- the image capture device 11 has an infrared focusing device 111 or an ultrasonic focusing device, wherein the principle of infrared light or ultrasound is used to detect the status of an environment and obtain the image information.
- the face information further includes a relative face distance and a relative face azimuth, and the relative face distance and the relative face azimuth are determined by the infrared focusing device 111 or the ultrasonic focusing device.
- the image capture device 11 having the infrared focusing device 111 is used for demonstration.
- the azimuth control motor 13 is coupled to the sound amplifying device 10 for controlling the azimuth of an output sound of the sound amplifying device 10 according to the relative face azimuth in the face information.
- the amplitude control unit 14 is communicatively coupled to the sound amplifying device 10 for controlling the volume of the output sound of the sound amplifying device 10 according to the relative face distance in the face information. Therefore, the automatically controlled directional speaker 1 in accordance with this disclosure detects the surrounding environment by the image capture device 11 , and then the computing device 12 analyzes the detected image information to determine whether or not the face information exists.
- the image capture device 11 includes a wide-angle camera lens for providing an accurate detection of the image information in a wide space.
- the image capture device 11 captures the image information
- the computing device 12 analyzes the image information to determine whether or not the face information is included.
- face detection is used for determining whether or not a face exists in an image by a characteristic capture method and providing information such as the face position and quantity.
- the face detection is mainly divided into two major computing methods. One of the two computing methods is based on characteristic and the other one is based learning.
- the characteristic based method is subdivided into the following methods. (1) Knowledge conditions constituting a typical face are used for coding, and face characteristics are localized. Geometric properties of a face figure are used for face region labeling. For example, an oval shape is used as a basis for determining a face figure.
- an edge detection method is used to capture an oval region in an image and the oval region is further analyzed to determine whether it is a face.
- Face characteristics such as the structural organization of a face, skin color, face symmetry and distribution of face characteristic positions are used as a basis for determining and writing an algorithm.
- a template matching method is used to compare an input image with a plurality of existing face samples and calculate the correlation between the image and the face samples to determine whether or not there is a face image in the image.
- Such template matching method is also uses face characteristics for the determination.
- Each of the characteristics above is not used individually, but the characteristics are combined and used in one or more methods to perform face detection. For instance, a skin color method used for detecting a face usually needs to combine the characteristics to improve the accuracy due to the interference by background.
- the distribution of a face is learned by neural network by capturing changes of the external appearance of the face of a large quantity of trained images to learn and modulate a plurality of models, and then comparing the whole image by a fixed range to locate the position of the face.
- Such method is relatively more difficult, since certain characteristics are not obvious, and the non-obvious characteristics may cause a failure of the detection.
- this method analyzes the color information of the skin color of a face by using color division of an input image and processes the color information by a morphologic algorithm to locate an area that may have a face, and finally uses geometric shape of the image to identify and eliminate an area that may not be a face by the oval shape together with the aspect ratio of a face profile, so as to label a face position.
- the aforementioned method combines a plurality of algorithms and logical computation to determine the face position.
- the computing logic may be adjusted as needed or other processing procedures may be supplemented.
- the technology of connected region may be used to further locate the position and region of a plurality of faces in an image.
- the traditional set theory and its derived algorithms are called a fuzzy algorithm which belongs to the area of fuzzy set theory.
- a characteristic function of the traditional set theory is selected from either “0” or “1”, such that the membership function of the fuzzy set may be selected from any value from 0 to 1.
- an element of the fuzzy set may fall between 0 and 1 with respect to the degree of membership of this set, so that fuzzy set has infinite membership functions to cope with various different conditions and factors.
- the relation between an element and a set is no longer binary, but a value of degree of membership is assigned according to the degree similar to element and set.
- the fuzzy algorithm is a quick effective computing mechanism.
- the computing device 12 of this disclosure adopts a fuzzy algorithm to determine whether or not the image information captured by the image capture device 11 has the face information.
- the computing device 12 of this disclosure analyzes the inputted image information to form a plurality of elements and applies the rule of thumb of human being to create each related fuzzy set and obtain a value of the degree of membership of each element with respect to the fuzzy sets, and then combines the values of the degree of membership to derive whether or not the face information is in the image information.
- a face is generally composed of an oval-like profile, facial features and skin color, so that after the computing device 12 receives the image information, the fuzzy algorithm analyzes the value of the degree of membership of each element of the image information with respect to the aforementioned face characteristics, so as to determine whether or not the face information is in the image information.
- this disclosure may have various models due to the difference of the setup conditions to achieve the best broadcasting effect, and the general classification is based on distance, angle, azimuth and number of people, and each model executed corresponsive to the basis of each classification is described below.
- the factor of distance if the automatically controlled directional speaker 1 comes with far, middle, and near broadcasting modes, a user may set the automatically controlled directional speaker 1 to broadcast to at least one listener 2 in a specific distance.
- the image capture device 11 automatically detects the surrounding environment to form the image information. After the computing device 12 analyzes the image information and confirms that the image information includes the face information (i.e. there is a person in the environment where the automatically controlled directional speaker 1 is situated, and then the infrared focusing device 111 or the ultrasonic focusing device determines the relative face distance and the relative face azimuth in the face information.
- the fuzzy algorithm will determine a value of the degree of membership of the relative face distance corresponsive to the far, middle, near distance of each fuzzy set and will further determine whether or not the face information belongs to the listener 2 at a closer distance. If yes, then the sound amplifying device 10 will be driven to broadcast to the listener 2 according to the selected near-distance playing mode so as to provide the automatically controlled directional speaker 1 with a more diversified and accurate broadcasting effect.
- the automatically controlled directional speaker 1 may adjust the output amplitude and volume of the sound amplifying device 10 according to the distance mode by the amplitude control unit 14 .
- the division of azimuth may be determined by the aforementioned method. If a simple and easy division is used, the azimuth control motor 13 may drives the sound amplifying device 10 to move in the up, down, left and right directions according to the face information. For a more accurate differentiation, the azimuth control motor 13 may be set to control the sound amplifying device 10 to move in the up, down, left and right directions, as well as the upper left, lower left, upper right and upper left directions. Since the definition of the aforementioned division of azimuth or angle is not clear, the azimuth control motor 13 may be unable to determine the relative face azimuth or determine the azimuth of the listener 2 by the infrared focusing device 111 or the ultrasonic focusing device.
- the fuzzy algorithm is used to obtain the content of the relative face azimuth and the value of the degree of membership with respect to each azimuth and angle of the fuzzy set, so that the azimuth control motor 13 can drive the sound amplifying device 10 to be adjusted to a better position for broadcasting.
- the automatically controlled directional speaker 1 may detect a plurality of face information in the environment or a plurality of people in the environment, the automatically controlled directional speaker 1 is designated to broadcast sounds towards a particular person according to the setup mode. For example, quantity is used as a basis for the setup, so that the automatically controlled directional speaker 1 may be set to broadcast sounds towards a place with more people or a place with less people.
- the computing device determines that there are many people in an environment and they are scattered, and the number of people at a place is greater than the number of people at another place, then the automatically controlled directional speaker 1 will automatically broadcast sounds towards the place with more people or the place with less people. In an example as shown in FIG.
- the automatically controlled directional speaker 1 broadcasts sounds towards an area with more listeners 2 .
- the computing device 12 determines that there are people in an environment and different relative face distances with respect to the automatically controlled directional speaker 1 , the aforementioned distance determination method may be used together to achieve the effect of broadcasting to one or more listeners 2 at a specific distance.
- the automatically controlled directional speaker 1 may be installed in a relatively darker environment.
- this disclosure further comprises an LED assist lamp 15 installed adjacent to a side of the image capture device 11 to provide sufficient light to the environment, so that the image information captured by the image capture device 11 has sufficient brightness and clarity.
- the image capture device 11 may be installed at a pivot member 16 , and the pivot member 16 is provided for the image capture device 11 to adjusting the angle and range of shooting to up, down, left and right directions in order to enhance the shooting range of the image capture device 11 and provide a broader scope of the content of the image information.
- this disclosure further comprises a movable waveguide structure 17 , installed at a sound output position of the sound amplifying device 10 , and having a sound input port 171 , a sound output port 172 and a sound channel 173 included between the sound input port and the sound output port.
- the relative diameter and size of the sound input port 171 and the sound output port 172 may be adjusted separately.
- the output sound amplitude of the sound amplifying device can be adjusted according to the relative face distance by the amplitude control unit 14 first, and then the relative diameter and size of the sound input port 171 and the sound output port 172 are adjusted, and finally the amplitude and volume of the sound outputted from the sound amplifying device 10 are changed by the movable waveguide structure 17 , so that the listeners can sense the sound better.
- This preferred embodiment discloses an LED illumination lamp 3 having the aforementioned automatically controlled directional speaker 1 and characterized in that the LED illumination lamp 3 is a bulb lamp, so that the LED illumination lamp 3 can provide the illumination effect and the LED illumination lamp 3 uses the image capture device 11 to automatically detect the surrounding environment of the LED illumination lamp 3 to generate the image information, and further uses the computing device 12 to analyze whether or not there is the face information, or determine whether or not there are people in the environment, and then controls the azimuth, amplitude and volume of the output sound of the sound amplifying device 10 according to the face information, so that a designated person may hear the output sound of the sound amplifying device 10 and improve the practicality of the LED illumination lamp 3 .
- the automatically controlled directional speaker 1 of this preferred embodiment are similar to those of the first preferred embodiment, and thus they will not be repeated.
- the LED illumination lamp 3 of this implementation mode is a down light installed at an indoor ceiling, so that the automatically controlled directional speaker 1 can provide indoor lighting while using a face detection method to broadcast sounds towards a designated person automatically from the sound amplifying device 10 to improve the practicality of the LED illumination lamp 3 .
- the automatically controlled directional speaker of this disclosure determines whether or not there are any people in the surrounding environment by face detection, and automatically redirects and broadcasts sounds to an area where a person is there to achieve the broadcasting with automatically controlled directivity.
- the image capture device 11 detects and captures the surrounding environment to generate the image information, and then the computing device 12 performs the face recognition analysis to obtain the face information, while combining with the infrared focusing device 111 or the ultrasonic focusing device to determine the relative face distance and the relative face azimuth in the face information to facilitate controlling the sound output direction and the output amplitude and volume of the sound amplifying device 10 .
- this disclosure eliminates the inconvenience of the traditional speaker with regard to distance and broadcasting effect, so that users may have the best listening condition all the time. Meanwhile, this disclosure comes with various sound playing modes to cope with the installation environment or user requirements, and the automatically controlled directional speaker 1 can broadcast sound more accurately as needed.
- this disclosure may be integrated with a lamp for use, so that the LED illumination lamp 3 is capable of providing illumination and automatically searching a person in order to broadcast sound to such person and achieve the light and sound effects concurrently.
Landscapes
- Engineering & Computer Science (AREA)
- Health & Medical Sciences (AREA)
- Physics & Mathematics (AREA)
- Otolaryngology (AREA)
- Acoustics & Sound (AREA)
- Signal Processing (AREA)
- Human Computer Interaction (AREA)
- Oral & Maxillofacial Surgery (AREA)
- General Health & Medical Sciences (AREA)
- General Physics & Mathematics (AREA)
- Multimedia (AREA)
- Theoretical Computer Science (AREA)
- Circuit For Audible Band Transducer (AREA)
- Studio Devices (AREA)
- Image Processing (AREA)
- Rehabilitation Tools (AREA)
- Arrangement Of Elements, Cooling, Sealing, Or The Like Of Lighting Devices (AREA)
- Obtaining Desirable Characteristics In Audible-Bandwidth Transducers (AREA)
Abstract
Description
- The technical field relates to a broadcasting apparatus, and more particularly to An automatically controlled directional speaker and capable of automatically performing face recognition to broadcast a sound to a designated person and combining with a lamp to improve the convenience of living.
- Sound amplifying device is a widely and commonly used device capable of converting current frequency into sound to play various types of audio files. Sound amplifying devices may be combined and assembled to form a playing device such as a sound box.
- The sound field of the sound amplifying device is limited by its placing position. In other words, a user may hear a sound coming from the sound amplifying device clearly only within a certain range. Our sense of sound depends on an analysis result processed by auditory sense of our brain, so that we can determine and feel the elements such as the direction, distance, and volume of a sound. For example, if a user wants to listen to a stereo or a radio indoors, the sound heard by the user will vary when the user is moving, since sound amplifying device is placed at a fixed position. If the user moves away from the sound amplifying device, the sound heard by the user will gradually become lower and more unclearly due to the factors of an increased distance or an deviated angle. To sense the sound in a full range, it is necessary to install a plurality of sound amplifying devices at different positions of a room or buy an expensive sound amplifying device with a stereo surround sound effect. Either way incurs a high setup cost.
- When many people are in the room, the sound played by the sound amplifying device may affect other people, since the sound field range of the sound amplifying device cannot be controlled or the sound cannot be listed by just the designated person only. Therefore, the conventional sound amplifying device fails to satisfy the user requirement and the effect of playing sound to a designated person situated in an environment with many people or a public place. To solve the aforementioned problem, a manual remote control may be used to control the broadcasting direction of the sound amplifying device, but this method requires users to keep controlling the sound amplifying device manually according to the moving status of a listening target, and the method of controlling the direction of the sound amplifying device just achieves the effect of hearing the sound from a listening target and cannot adjust the sound effect of the listening target. In addition, if other people are near the listening target, these people will be affected. Obviously, the aforementioned method is not a good solution.
- Besides the sound amplifying device, lamp is also a necessary electric/electronic device. From incandescent lamps at early stage to LED lamps extensively used at present, various types of lamps are installed indoor and outdoor for illumination, and the lamps are connected to mains power to obtain the startup power, and a control circuit is generally provided for adjusting the light emission of the lamps. Although both sound amplifying device and lamp are electronic devices, they are usually installed separately, and their space occupying percentage is increased accordingly, so that the using space available for users is decreased.
- As science and technology advance rapidly, users have increasingly higher demand for the quality of life, so that the discloser of this disclosure further improves the performance and practicality of the sound amplifying device and reduces the occupied space while providing users a more convenient design of an automatically controlled directional speaker and its lamp, in hope of overcoming the aforementioned drawbacks of the prior art.
- Therefore, it is an objective of this disclosure to provide an automatically controlled directional speaker that is capable of detecting and identifying whether or not there is any person in the surrounding environment and automatically guiding and broadcasting sounds towards a designated person.
- Another objective of this disclosure is to provide an LED lamp combined with the aforementioned automatically controlled directional speaker, so that the LED lamp has the effects of playing sound in a direction towards a designated person automatically and providing illumination at the same time.
- To achieve the aforementioned and other objectives, this disclosure provides an automatically controlled directional speaker, and the speaker comprises: a sound amplifying device; an image capture device, for real time detecting the status of surrounding environment to generate at least one image information; a computing device, communicatively coupled to the image capture device, for determining whether or not the image information includes at least one face information; an azimuth control motor, coupled to the sound amplifying device, for controlling the sound amplifying device to output a sound azimuth according to the face information; and an amplitude control unit, communicatively coupled to the sound amplifying device, for controlling the volume of an output sound according to the face information. Preferably, the computing device performs an analysis by a fuzzy algorithm.
- Wherein, the image capture device has an infrared focusing device or an ultrasonic focusing device, and the face information includes a relative face distance and a relative face azimuth; and the relative face distance and the relative face azimuth are determined by the infrared focusing device or the ultrasonic focusing device. Therefore, the relative distance and azimuth of a person with respect to the automatically controlled directional speaker can be determined to facilitate the sound amplifying device to broadcast more accurately.
- In addition, the speaker of this disclosure further comprises an LED assist lamp installed adjacent to a side of the image capture device to provide sufficient brightness to surrounding environment and make the image information clearer.
- To improve the image information of the image capture device in order to cover a wider visual range of an image, the image capture device includes a wide-angle camera lens, or the image capture device is mounted onto a pivot member provided for adjusting both shooting angle and range of the image capture device in top, bottom, left and right sides.
- In addition, the speaker of this disclosure further comprises a movable waveguide structure installed at a sound output position of the sound amplifying device and having a sound input port, a sound output port, and a sound channel included between the sound input port and the sound output port; and the sound input port and the sound output port have adjustable relative aperture sizes respectively; wherein the relative diameter and size of the sound input port and the sound output port may be adjusted to further adjust the volume of the output sound of the sound amplifying device to provide a better listening effect.
- In another preferred embodiment, this disclosure further discloses an LED illumination lamp of the automatically controlled directional speaker, characterized in that the whole LED lamp is a bulb lamp or a down light including the computing device installed therein and provided for performing an analysis by a fuzzy algorithm. The speaker of is disclosure may also be combined with a lamp, so that the lamp has both illumination and broadcasting functions, and such lamp can broadcast sound to a designated person automatically.
- The speaker of this disclosure has the effects of detecting and determining whether or not there are other people in the surrounding environment, and automatically adjusting the output sound to a designated person or area, so that a user can listen to the sound very well even when the user is in a moving status. The speaker may be set to allow a designated person to listen to the sound, so as to prevent the sound from affecting other people. In addition, the speaker of this disclosure may be combined with a lamp to form an LED illumination lamp and provide both illumination and broadcasting effects. With an automatic face detection function, the LED illumination lamp can broadcast sounds towards a designated person and improve the convenience of use.
-
FIG. 1 is a schematic block diagram of a first preferred embodiment of this disclosure; -
FIG. 2 is a perspective view of the first preferred embodiment of this disclosure; -
FIG. 3 is a first schematic view showing an application of the first preferred embodiment of this disclosure; -
FIG. 4A is a second schematic view showing an application of the first preferred embodiment of this disclosure; -
FIG. 4B is a third schematic view showing an application of the first preferred embodiment of this disclosure; -
FIG. 5 is a perspective view of another implementation mode of the first preferred embodiment of this disclosure; -
FIG. 6 is a schematic view of a second preferred embodiment of the present invention; and -
FIG. 7 is a perspective view of another implementation mode of the second preferred embodiment of this disclosure. - The technical content of this disclosure will become apparent with the detailed description of preferred embodiments and the illustration of related drawings as follows.
- With reference to
FIGS. 1, 2, 3, 4A and 4B for a schematic block diagram, a perspective view, and first to third schematic views of applications of an automatically controlleddirectional speaker 1 in accordance with the first preferred embodiment of this disclosure respectively, the automatically controlleddirectional speaker 1 comprises asound amplifying device 10, animage capture device 11, acomputing device 12, anazimuth control motor 13 and anamplitude control unit 14. - The
image capture device 11 is provided for real time detecting the status of surrounding environment to generate at least one image information, and thecomputing device 12 is communicatively coupled to theimage capture device 11 for determining whether or not there is at least one face information in the image information generated by theimage capture device 11. Preferably, theimage capture device 11 has an infrared focusingdevice 111 or an ultrasonic focusing device, wherein the principle of infrared light or ultrasound is used to detect the status of an environment and obtain the image information. The face information further includes a relative face distance and a relative face azimuth, and the relative face distance and the relative face azimuth are determined by the infrared focusingdevice 111 or the ultrasonic focusing device. In this embodiment, theimage capture device 11 having the infrared focusingdevice 111 is used for demonstration. - The
azimuth control motor 13 is coupled to thesound amplifying device 10 for controlling the azimuth of an output sound of thesound amplifying device 10 according to the relative face azimuth in the face information. Theamplitude control unit 14 is communicatively coupled to thesound amplifying device 10 for controlling the volume of the output sound of thesound amplifying device 10 according to the relative face distance in the face information. Therefore, the automatically controlleddirectional speaker 1 in accordance with this disclosure detects the surrounding environment by theimage capture device 11, and then thecomputing device 12 analyzes the detected image information to determine whether or not the face information exists. If there is any face information, the relative face azimuth and the relative face distance will be calculated, so that theazimuth control motor 13 and theamplitude control unit 14 can control thesound amplifying device 10 to be aligned precisely at the position of that person to play sounds and adjust the distance between that person and thesound amplifying device 10 to adjust the volume of the output sound. To broaden the range of viewed images in the detection of the image information by theimage capture device 11, theimage capture device 11 includes a wide-angle camera lens for providing an accurate detection of the image information in a wide space. - In this disclosure as described above, the
image capture device 11 captures the image information, and thecomputing device 12 analyzes the image information to determine whether or not the face information is included. In short, face detection is used for determining whether or not a face exists in an image by a characteristic capture method and providing information such as the face position and quantity. The face detection is mainly divided into two major computing methods. One of the two computing methods is based on characteristic and the other one is based learning. The characteristic based method is subdivided into the following methods. (1) Knowledge conditions constituting a typical face are used for coding, and face characteristics are localized. Geometric properties of a face figure are used for face region labeling. For example, an oval shape is used as a basis for determining a face figure. In other words, an edge detection method is used to capture an oval region in an image and the oval region is further analyzed to determine whether it is a face. (2) Face characteristics such as the structural organization of a face, skin color, face symmetry and distribution of face characteristic positions are used as a basis for determining and writing an algorithm. (3) A template matching method is used to compare an input image with a plurality of existing face samples and calculate the correlation between the image and the face samples to determine whether or not there is a face image in the image. Such template matching method is also uses face characteristics for the determination. Each of the characteristics above is not used individually, but the characteristics are combined and used in one or more methods to perform face detection. For instance, a skin color method used for detecting a face usually needs to combine the characteristics to improve the accuracy due to the interference by background. - In the learning based method, the distribution of a face is learned by neural network by capturing changes of the external appearance of the face of a large quantity of trained images to learn and modulate a plurality of models, and then comparing the whole image by a fixed range to locate the position of the face. Such method is relatively more difficult, since certain characteristics are not obvious, and the non-obvious characteristics may cause a failure of the detection.
- For example, if skin color is used as an element for the determination of an image face detection, and the characteristics of a face figure is uses as an element for assisting the determination, this method analyzes the color information of the skin color of a face by using color division of an input image and processes the color information by a morphologic algorithm to locate an area that may have a face, and finally uses geometric shape of the image to identify and eliminate an area that may not be a face by the oval shape together with the aspect ratio of a face profile, so as to label a face position. The aforementioned method combines a plurality of algorithms and logical computation to determine the face position. To eliminate the influence of external environment factors such as the color information of an image that may be affected by a light color or the color deviation of an image capture equipment to result in a deviation of the overall image information and the ideal value, the computing logic may be adjusted as needed or other processing procedures may be supplemented. To improve the accuracy of interpretation, the technology of connected region may be used to further locate the position and region of a plurality of faces in an image. In summation of the description above, several algorithms and characteristics are used to locate and determine the face position and region in an image, and set up and eliminate the changing conditions of the algorithms to cope with the difference of a face due to light or angle.
- In the setup of each of the aforementioned algorithms, the determination procedure of each algorithm provides accurate results. In short, such system is a system applying traditional set theory and algorithm. In the traditional set theory, the relation between a single element and a set is not a binary relation. As a single element to the degree of membership of a set is binary, the element is either equal to “0” or “1”. Therefore, a face recognition system built by any of the aforementioned algorithms (including the algorithm of dividing an image, and the algorithm of using face characteristics for determination) provides binary results, such as “True” and “False”, or “Yes” and “No”.
- The traditional set theory and its derived algorithms are called a fuzzy algorithm which belongs to the area of fuzzy set theory. In the fuzzy set theory, a characteristic function of the traditional set theory is selected from either “0” or “1”, such that the membership function of the fuzzy set may be selected from any value from 0 to 1. In other words, an element of the fuzzy set may fall between 0 and 1 with respect to the degree of membership of this set, so that fuzzy set has infinite membership functions to cope with various different conditions and factors. Based on the algorithm of the fuzzy set, the relation between an element and a set is no longer binary, but a value of degree of membership is assigned according to the degree similar to element and set. In the computation of a large quantity of information or the classification of each element into a range which is a fuzzy and uncertain cluster, the fuzzy algorithm is a quick effective computing mechanism. The
computing device 12 of this disclosure adopts a fuzzy algorithm to determine whether or not the image information captured by theimage capture device 11 has the face information. - The
computing device 12 of this disclosure analyzes the inputted image information to form a plurality of elements and applies the rule of thumb of human being to create each related fuzzy set and obtain a value of the degree of membership of each element with respect to the fuzzy sets, and then combines the values of the degree of membership to derive whether or not the face information is in the image information. Theoretically, a face is generally composed of an oval-like profile, facial features and skin color, so that after thecomputing device 12 receives the image information, the fuzzy algorithm analyzes the value of the degree of membership of each element of the image information with respect to the aforementioned face characteristics, so as to determine whether or not the face information is in the image information. - According to the differences of the installing space and environment of the automatically controlled
directional speaker 1 and the desired broadcasting setup conditions, this disclosure may have various models due to the difference of the setup conditions to achieve the best broadcasting effect, and the general classification is based on distance, angle, azimuth and number of people, and each model executed corresponsive to the basis of each classification is described below. As to the factor of distance, if the automatically controlleddirectional speaker 1 comes with far, middle, and near broadcasting modes, a user may set the automatically controlleddirectional speaker 1 to broadcast to at least onelistener 2 in a specific distance. Since the definition of a range of the far, middle, and near distance may not allow the automatically controlleddirectional speaker 1 to obtain the relative face distance, a quick determination of the range of the distance between a person and theimage capture device 10 is made. Similarly, the fuzzy algorithm is used to derive that the face information falls within what distance range by the rule of thumb. InFIGS. 3 and 4A , theimage capture device 11 automatically detects the surrounding environment to form the image information. After thecomputing device 12 analyzes the image information and confirms that the image information includes the face information (i.e. there is a person in the environment where the automatically controlleddirectional speaker 1 is situated, and then the infrared focusingdevice 111 or the ultrasonic focusing device determines the relative face distance and the relative face azimuth in the face information. If the selected mode of the automatically controlleddirectional speaker 1 is to broadcast to thelistener 2 at a near distance, the fuzzy algorithm will determine a value of the degree of membership of the relative face distance corresponsive to the far, middle, near distance of each fuzzy set and will further determine whether or not the face information belongs to thelistener 2 at a closer distance. If yes, then thesound amplifying device 10 will be driven to broadcast to thelistener 2 according to the selected near-distance playing mode so as to provide the automatically controlleddirectional speaker 1 with a more diversified and accurate broadcasting effect. In addition, the automatically controlleddirectional speaker 1 may adjust the output amplitude and volume of thesound amplifying device 10 according to the distance mode by theamplitude control unit 14. - Similarly, the division of azimuth may be determined by the aforementioned method. If a simple and easy division is used, the
azimuth control motor 13 may drives thesound amplifying device 10 to move in the up, down, left and right directions according to the face information. For a more accurate differentiation, theazimuth control motor 13 may be set to control thesound amplifying device 10 to move in the up, down, left and right directions, as well as the upper left, lower left, upper right and upper left directions. Since the definition of the aforementioned division of azimuth or angle is not clear, theazimuth control motor 13 may be unable to determine the relative face azimuth or determine the azimuth of thelistener 2 by the infrared focusingdevice 111 or the ultrasonic focusing device. Similar to the determination of distance, the fuzzy algorithm is used to obtain the content of the relative face azimuth and the value of the degree of membership with respect to each azimuth and angle of the fuzzy set, so that theazimuth control motor 13 can drive thesound amplifying device 10 to be adjusted to a better position for broadcasting. - Since the automatically controlled
directional speaker 1 may detect a plurality of face information in the environment or a plurality of people in the environment, the automatically controlleddirectional speaker 1 is designated to broadcast sounds towards a particular person according to the setup mode. For example, quantity is used as a basis for the setup, so that the automatically controlleddirectional speaker 1 may be set to broadcast sounds towards a place with more people or a place with less people. InFIG. 4B , if the computing device determines that there are many people in an environment and they are scattered, and the number of people at a place is greater than the number of people at another place, then the automatically controlleddirectional speaker 1 will automatically broadcast sounds towards the place with more people or the place with less people. In an example as shown inFIG. 4B , the automatically controlleddirectional speaker 1 broadcasts sounds towards an area withmore listeners 2. In another situation, thecomputing device 12 determines that there are people in an environment and different relative face distances with respect to the automatically controlleddirectional speaker 1, the aforementioned distance determination method may be used together to achieve the effect of broadcasting to one ormore listeners 2 at a specific distance. - With reference to
FIG. 5 for a perspective view of another implementation mode of this disclosure, the automatically controlleddirectional speaker 1 may be installed in a relatively darker environment. To avoid the situation of having insufficient light or low brightness to analyze and determine the contents of the image information, this disclosure further comprises an LED assistlamp 15 installed adjacent to a side of theimage capture device 11 to provide sufficient light to the environment, so that the image information captured by theimage capture device 11 has sufficient brightness and clarity. Theimage capture device 11 may be installed at apivot member 16, and thepivot member 16 is provided for theimage capture device 11 to adjusting the angle and range of shooting to up, down, left and right directions in order to enhance the shooting range of theimage capture device 11 and provide a broader scope of the content of the image information. - Wherein, this disclosure further comprises a
movable waveguide structure 17, installed at a sound output position of thesound amplifying device 10, and having asound input port 171, asound output port 172 and asound channel 173 included between the sound input port and the sound output port. The relative diameter and size of thesound input port 171 and thesound output port 172 may be adjusted separately. Therefore, the output sound amplitude of the sound amplifying device can be adjusted according to the relative face distance by theamplitude control unit 14 first, and then the relative diameter and size of thesound input port 171 and thesound output port 172 are adjusted, and finally the amplitude and volume of the sound outputted from thesound amplifying device 10 are changed by themovable waveguide structure 17, so that the listeners can sense the sound better. - With reference to
FIG. 6 for a perspective view of the second preferred embodiment of this disclosure, the elements and characteristics of this embodiment is substantially the same as those of the first preferred embodiment, and thus they will not be repeated. This preferred embodiment discloses anLED illumination lamp 3 having the aforementioned automatically controlleddirectional speaker 1 and characterized in that theLED illumination lamp 3 is a bulb lamp, so that theLED illumination lamp 3 can provide the illumination effect and theLED illumination lamp 3 uses theimage capture device 11 to automatically detect the surrounding environment of theLED illumination lamp 3 to generate the image information, and further uses thecomputing device 12 to analyze whether or not there is the face information, or determine whether or not there are people in the environment, and then controls the azimuth, amplitude and volume of the output sound of thesound amplifying device 10 according to the face information, so that a designated person may hear the output sound of thesound amplifying device 10 and improve the practicality of theLED illumination lamp 3. As to the details of the automatically controlleddirectional speaker 1 of this preferred embodiment are similar to those of the first preferred embodiment, and thus they will not be repeated. - With reference to
FIG. 7 for a perspective view of another implementation mode of the second preferred embodiment of this disclosure, theLED illumination lamp 3 of this implementation mode is a down light installed at an indoor ceiling, so that the automatically controlleddirectional speaker 1 can provide indoor lighting while using a face detection method to broadcast sounds towards a designated person automatically from thesound amplifying device 10 to improve the practicality of theLED illumination lamp 3. - In summation of the description above, the automatically controlled directional speaker of this disclosure determines whether or not there are any people in the surrounding environment by face detection, and automatically redirects and broadcasts sounds to an area where a person is there to achieve the broadcasting with automatically controlled directivity. The
image capture device 11 detects and captures the surrounding environment to generate the image information, and then thecomputing device 12 performs the face recognition analysis to obtain the face information, while combining with the infrared focusingdevice 111 or the ultrasonic focusing device to determine the relative face distance and the relative face azimuth in the face information to facilitate controlling the sound output direction and the output amplitude and volume of thesound amplifying device 10. Therefore, this disclosure eliminates the inconvenience of the traditional speaker with regard to distance and broadcasting effect, so that users may have the best listening condition all the time. Meanwhile, this disclosure comes with various sound playing modes to cope with the installation environment or user requirements, and the automatically controlleddirectional speaker 1 can broadcast sound more accurately as needed. In addition, this disclosure may be integrated with a lamp for use, so that theLED illumination lamp 3 is capable of providing illumination and automatically searching a person in order to broadcast sound to such person and achieve the light and sound effects concurrently.
Claims (11)
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
TW104126511A | 2015-08-14 | ||
TW104126511 | 2015-08-14 | ||
TW104126511A TW201707471A (en) | 2015-08-14 | 2015-08-14 | Automatically controlled directional speaker and lamp thereof enabling mobile users to stay in the best listening condition, preventing the sound from affecting others when broadcasting, and improving the convenience of use in life |
Publications (2)
Publication Number | Publication Date |
---|---|
US20170048611A1 true US20170048611A1 (en) | 2017-02-16 |
US9591399B1 US9591399B1 (en) | 2017-03-07 |
Family
ID=57908385
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/885,046 Active US9591399B1 (en) | 2015-08-14 | 2015-10-16 | Automatically controlled directional speaker, and lamp thereof |
Country Status (5)
Country | Link |
---|---|
US (1) | US9591399B1 (en) |
JP (1) | JP2017038342A (en) |
CN (1) | CN106470370A (en) |
DE (1) | DE102015117867B4 (en) |
TW (1) | TW201707471A (en) |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107543051A (en) * | 2017-10-31 | 2018-01-05 | 苏州尤伦斯照明科技有限公司 | Multifunctional hand-held stereo desk lamp |
US20180046864A1 (en) * | 2016-08-10 | 2018-02-15 | Vivint, Inc. | Sonic sensing |
US10535360B1 (en) * | 2017-05-25 | 2020-01-14 | Tp Lab, Inc. | Phone stand using a plurality of directional speakers |
US11507389B2 (en) * | 2016-09-29 | 2022-11-22 | Hewlett-Packard Development Company, L.P. | Adjusting settings on computing devices based on location |
US20220377487A1 (en) * | 2021-05-18 | 2022-11-24 | Nanning Fulian Fugui Precision Industrial Co., Ltd. | Directional sound transmission method, electronic device and readable storage medium |
Families Citing this family (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107817723A (en) * | 2017-11-02 | 2018-03-20 | 成都吱吖科技有限公司 | A kind of intelligent sound orientation control method for playing back of human body location aware |
US10873797B2 (en) * | 2017-12-29 | 2020-12-22 | Harman International Industries, Incorporated | Rotating loudspeaker |
TWI695632B (en) * | 2018-02-22 | 2020-06-01 | 圓展科技股份有限公司 | Audio control device and control method thereof |
CN108882106B (en) * | 2018-08-06 | 2020-01-21 | 湖州宇声电子厂 | Intelligent sound box with infrared sensing function |
USD924455S1 (en) * | 2018-11-28 | 2021-07-06 | Omid Minaye | Speaker lamp |
CN111314821A (en) * | 2018-12-12 | 2020-06-19 | 深圳市冠旭电子股份有限公司 | Intelligent sound box playing method and device and intelligent sound box |
US11095976B2 (en) | 2019-01-08 | 2021-08-17 | Vizio, Inc. | Sound system with automatically adjustable relative driver orientation |
US10999677B2 (en) * | 2019-05-29 | 2021-05-04 | Asahi Kasei Kabushiki Kaisha | Sound reproducing apparatus having multiple directional speakers and sound reproducing method |
JP6767664B1 (en) * | 2019-10-28 | 2020-10-14 | 富士通クライアントコンピューティング株式会社 | Information processing systems, information processing equipment and programs |
US11343602B2 (en) | 2020-02-21 | 2022-05-24 | Harman International Industries, Incorporated | Passive auto-orienting loudspeaker assembly |
KR20230122451A (en) * | 2022-02-14 | 2023-08-22 | 현대자동차주식회사 | Smart speaker and method for providing sound using the same |
CN114519914A (en) * | 2022-03-14 | 2022-05-20 | 国网山西省电力公司超高压变电分公司 | A voice prompt for transformer safety construction for electric power overhauls |
Family Cites Families (20)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP3209350B2 (en) * | 1991-09-10 | 2001-09-17 | ソニー株式会社 | Horn speaker |
JPH0631299U (en) * | 1992-09-25 | 1994-04-22 | 株式会社学習研究社 | Loudspeaker combined with sound collector |
US6160896A (en) * | 1998-05-08 | 2000-12-12 | Motorola, Inc. | Variable frequency response microphone porting system |
US6940545B1 (en) * | 2000-02-28 | 2005-09-06 | Eastman Kodak Company | Face detecting camera and method |
AU2003230067A1 (en) * | 2002-05-14 | 2003-11-11 | Koninklijke Philips Electronics N.V. | Dialog control for an electric apparatus |
JP4219719B2 (en) * | 2003-03-28 | 2009-02-04 | 三菱電機エンジニアリング株式会社 | Imaging device with speaker |
US7469060B2 (en) * | 2004-11-12 | 2008-12-23 | Honeywell International Inc. | Infrared face detection and recognition system |
JP4669340B2 (en) * | 2005-07-28 | 2011-04-13 | 富士通株式会社 | Information processing apparatus, information processing method, and information processing program |
JP2007228401A (en) * | 2006-02-24 | 2007-09-06 | Mitsubishi Electric Engineering Co Ltd | Sound luminaire |
KR20090016289A (en) * | 2007-08-10 | 2009-02-13 | 엘지전자 주식회사 | A apparatus and a method of notebook computer speaker controlling |
JP2010026551A (en) * | 2008-07-15 | 2010-02-04 | Seiko Epson Corp | Display system, and control method for display system |
US20100061586A1 (en) * | 2008-09-10 | 2010-03-11 | Sony Corporation | Tv with viewer-adapted height and angle adjustment |
JP2010206451A (en) * | 2009-03-03 | 2010-09-16 | Panasonic Corp | Speaker with camera, signal processing apparatus, and av system |
JP2011250745A (en) * | 2010-06-03 | 2011-12-15 | Aanetto:Kk | Method and system for threatening harmful animal |
WO2012091185A1 (en) * | 2010-12-27 | 2012-07-05 | Lg Electronics Inc. | Display device and method of providing feedback for gestures thereof |
JP5644593B2 (en) * | 2011-03-07 | 2014-12-24 | ソニー株式会社 | Speaker device |
CN103460718A (en) * | 2011-03-28 | 2013-12-18 | 株式会社尼康 | Electronic device and information transmission system |
JP2012205240A (en) * | 2011-03-28 | 2012-10-22 | Nikon Corp | Electronic device and information transfer system |
TWI458362B (en) * | 2012-06-22 | 2014-10-21 | Wistron Corp | Auto-adjusting audio display method and apparatus thereof |
JP5488732B1 (en) * | 2013-03-05 | 2014-05-14 | パナソニック株式会社 | Sound playback device |
-
2015
- 2015-08-14 TW TW104126511A patent/TW201707471A/en unknown
- 2015-08-20 CN CN201510514502.0A patent/CN106470370A/en active Pending
- 2015-10-16 US US14/885,046 patent/US9591399B1/en active Active
- 2015-10-20 DE DE102015117867.2A patent/DE102015117867B4/en not_active Expired - Fee Related
- 2015-11-13 JP JP2015223081A patent/JP2017038342A/en active Pending
Cited By (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20180046864A1 (en) * | 2016-08-10 | 2018-02-15 | Vivint, Inc. | Sonic sensing |
US10579879B2 (en) * | 2016-08-10 | 2020-03-03 | Vivint, Inc. | Sonic sensing |
US11354907B1 (en) | 2016-08-10 | 2022-06-07 | Vivint, Inc. | Sonic sensing |
US11507389B2 (en) * | 2016-09-29 | 2022-11-22 | Hewlett-Packard Development Company, L.P. | Adjusting settings on computing devices based on location |
US10535360B1 (en) * | 2017-05-25 | 2020-01-14 | Tp Lab, Inc. | Phone stand using a plurality of directional speakers |
US11355135B1 (en) * | 2017-05-25 | 2022-06-07 | Tp Lab, Inc. | Phone stand using a plurality of microphones |
CN107543051A (en) * | 2017-10-31 | 2018-01-05 | 苏州尤伦斯照明科技有限公司 | Multifunctional hand-held stereo desk lamp |
US20220377487A1 (en) * | 2021-05-18 | 2022-11-24 | Nanning Fulian Fugui Precision Industrial Co., Ltd. | Directional sound transmission method, electronic device and readable storage medium |
Also Published As
Publication number | Publication date |
---|---|
TW201707471A (en) | 2017-02-16 |
US9591399B1 (en) | 2017-03-07 |
CN106470370A (en) | 2017-03-01 |
DE102015117867B4 (en) | 2021-01-28 |
JP2017038342A (en) | 2017-02-16 |
DE102015117867A1 (en) | 2017-02-16 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US9591399B1 (en) | Automatically controlled directional speaker, and lamp thereof | |
US11268690B2 (en) | Luminaire system with integrated, dynamic audio visual control | |
US9451379B2 (en) | Sound field analysis system | |
US12003673B2 (en) | Acoustic echo cancellation control for distributed audio devices | |
CN114208209B (en) | Audio processing system, method and medium | |
US11546688B2 (en) | Loudspeaker device, method, apparatus and device for adjusting sound effect thereof, and medium | |
CN110677803A (en) | Dynamic optimal site calibration | |
US11652287B2 (en) | Antenna systems for wireless communication in luminaires | |
US20220270601A1 (en) | Multi-modal smart audio device system attentiveness expression | |
Scheibler et al. | Blinkies: Open source sound-to-light conversion sensors for large-scale acoustic sensing and applications | |
US9980044B2 (en) | Balance adjustment control method for sound/illumination devices | |
TWM542308U (en) | Automatically controlled directional speaker and lamp thereof | |
CN113141285A (en) | Immersive voice interaction method and system | |
CN109040903A (en) | A kind of wisdom classroom with can automatic regulating volume height speaker | |
EP3970143B1 (en) | A lighting device | |
Vryzas et al. | Investigating Multimodal Audiovisual Event Detection and Localization | |
TWI842055B (en) | Audio system with dynamic target listening spot and ambient object interference cancelation | |
TWI842056B (en) | Audio system with dynamic target listening spot and ambient object interference cancelation | |
CN114389732A (en) | Digital sound console tuning system and method | |
Lee | Introduction to Research at the APL |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: UNITY OPTO TECHNOLOGY CO., LTD., TAIWAN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:WU, CHING-HUEI;CHUANG, KAI-CHENG;REEL/FRAME:036880/0192 Effective date: 20151015 |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
AS | Assignment |
Owner name: EPISTAR CORPORATION, TAIWAN Free format text: SECURITY INTEREST;ASSIGNOR:UNITY OPTO TECHNOLOGY CO., LTD.;REEL/FRAME:052291/0387 Effective date: 20200212 |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 4 |
|
AS | Assignment |
Owner name: UNITY OPTO TECHNOLOGY CO., LTD.,, TAIWAN Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:EPISTAR CORPORATION;REEL/FRAME:060250/0674 Effective date: 20220309 Owner name: EPISTAR CORPORATION, TAIWAN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:UNITY OPTO TECHNOLOGY CO., LTD.;REEL/FRAME:059496/0205 Effective date: 20220309 |