US20150171973A1 - Proximity-based and acoustic control of media devices for media presentations - Google Patents

Proximity-based and acoustic control of media devices for media presentations Download PDF

Info

Publication number
US20150171973A1
US20150171973A1 US14/105,127 US201314105127A US2015171973A1 US 20150171973 A1 US20150171973 A1 US 20150171973A1 US 201314105127 A US201314105127 A US 201314105127A US 2015171973 A1 US2015171973 A1 US 2015171973A1
Authority
US
United States
Prior art keywords
user
media device
content
media
wireless
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US14/105,127
Inventor
Michael Edward Smith Luna
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
JB IP Acquisition LLC
Original Assignee
AliphCom LLC
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority to US14/105,127 priority Critical patent/US20150171973A1/en
Application filed by AliphCom LLC filed Critical AliphCom LLC
Assigned to ALIPHCOM reassignment ALIPHCOM ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: LUNA, MICHAEL EDWARD SMITH
Assigned to BLACKROCK ADVISORS, LLC reassignment BLACKROCK ADVISORS, LLC SECURITY INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: ALIPH, INC., ALIPHCOM, BODYMEDIA, INC., MACGYVER ACQUISITION LLC, PROJECT PARIS ACQUISITION LLC
Publication of US20150171973A1 publication Critical patent/US20150171973A1/en
Assigned to BLACKROCK ADVISORS, LLC reassignment BLACKROCK ADVISORS, LLC SECURITY INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: ALIPH, INC., ALIPHCOM, BODYMEDIA, INC., MACGYVER ACQUISITION LLC, PROJECT PARIS ACQUISITION LLC
Assigned to BLACKROCK ADVISORS, LLC reassignment BLACKROCK ADVISORS, LLC CORRECTIVE ASSIGNMENT TO CORRECT THE APPLICATION NO. 13870843 PREVIOUSLY RECORDED ON REEL 036500 FRAME 0173. ASSIGNOR(S) HEREBY CONFIRMS THE SECURITY INTEREST. Assignors: ALIPH, INC., ALIPHCOM, BODYMEDIA, INC., MACGYVER ACQUISITION, LLC, PROJECT PARIS ACQUISITION LLC
Assigned to JB IP ACQUISITION LLC reassignment JB IP ACQUISITION LLC ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: ALIPHCOM, LLC, BODYMEDIA, INC.
Assigned to J FITNESS LLC reassignment J FITNESS LLC UCC FINANCING STATEMENT Assignors: JB IP ACQUISITION, LLC
Assigned to J FITNESS LLC reassignment J FITNESS LLC SECURITY INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: JB IP ACQUISITION, LLC
Assigned to J FITNESS LLC reassignment J FITNESS LLC UCC FINANCING STATEMENT Assignors: JAWBONE HEALTH HUB, INC.
Assigned to ALIPHCOM LLC reassignment ALIPHCOM LLC RELEASE BY SECURED PARTY (SEE DOCUMENT FOR DETAILS). Assignors: BLACKROCK ADVISORS, LLC
Assigned to J FITNESS LLC reassignment J FITNESS LLC RELEASE BY SECURED PARTY (SEE DOCUMENT FOR DETAILS). Assignors: JAWBONE HEALTH HUB, INC., JB IP ACQUISITION, LLC
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F1/00Details not covered by groups G06F3/00 - G06F13/00 and G06F21/00
    • G06F1/16Constructional details or arrangements
    • G06F1/1613Constructional details or arrangements for portable computers
    • G06F1/1633Constructional details or arrangements of portable computers not specific to the type of enclosures covered by groups G06F1/1615 - G06F1/1626
    • G06F1/1684Constructional details or arrangements related to integrated I/O peripherals not covered by groups G06F1/1635 - G06F1/1675
    • G06F1/1698Constructional details or arrangements related to integrated I/O peripherals not covered by groups G06F1/1635 - G06F1/1675 the I/O peripheral being a sending/receiving arrangement to establish a cordless communication link, e.g. radio or infrared link, integrated cellular phone
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04BTRANSMISSION
    • H04B11/00Transmission systems employing sonic, ultrasonic or infrasonic waves
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/016Input arrangements with force or tactile feedback as computer generated output to the user
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/017Gesture based interaction, e.g. based on a set of recognized hand gestures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/02Input arrangements using manually operated switches, e.g. using keyboards or dials
    • G06F3/0202Constructional details or processes of manufacture of the input device
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/02Input arrangements using manually operated switches, e.g. using keyboards or dials
    • G06F3/023Arrangements for converting discrete items of information into a coded form, e.g. arrangements for interpreting keyboard generated codes as alphanumeric codes, operand codes or instruction codes
    • G06F3/0238Programmable keyboards
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/72Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
    • H04M1/724User interfaces specially adapted for cordless or mobile telephones
    • H04M1/72403User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality
    • H04M1/72409User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality by interfacing with external accessories
    • H04M1/72412User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality by interfacing with external accessories using two-way short-range wireless interfaces
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R1/00Details of transducers, loudspeakers or microphones
    • H04R1/08Mouthpieces; Microphones; Attachments therefor
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04WWIRELESS COMMUNICATION NETWORKS
    • H04W4/00Services specially adapted for wireless communication networks; Facilities therefor
    • H04W4/80Services using short range communication, e.g. near-field communication [NFC], radio-frequency identification [RFID] or low energy communication
    • H04W76/023
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04WWIRELESS COMMUNICATION NETWORKS
    • H04W76/00Connection management
    • H04W76/10Connection setup
    • H04W76/14Direct-mode setup
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04WWIRELESS COMMUNICATION NETWORKS
    • H04W8/00Network data management
    • H04W8/005Discovery of network devices, e.g. terminals
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/72Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
    • H04M1/724User interfaces specially adapted for cordless or mobile telephones
    • H04M1/72403User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality
    • H04M1/72442User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality for playing music files
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M2250/00Details of telephonic subscriber devices
    • H04M2250/02Details of telephonic subscriber devices including a Bluetooth interface
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M2250/00Details of telephonic subscriber devices
    • H04M2250/12Details of telephonic subscriber devices including a sensor for measuring a physical value, e.g. temperature or motion

Definitions

  • Embodiments relate generally to electronic and electronic hardware, computer software, wired communications, wireless communications, and wireless network communications, portable, wearable, and stationary media devices. More specifically, disclosed are media devices that detect proximity of users and/or user devices and take actions and handle content on the user devices after detecting presence of users and/or user devices.
  • Conventional paradigms for media devices require a user to take some action using a finger press to a touch screen or press a button, or the like, in order to initiate some function on the device, such as listening to music, for example.
  • Conventional media devices are not configured to recognize and act on user preferences as to how the media device serves the user's needs based on changing circumstances and changing environments the user and media device are subject to.
  • conventional media devices are typically personal devices that are mostly if not always used solely by the user and are therefore not well adapted to servicing the needs of friends, guests, or the like who may want to share content on their devices with the user.
  • FIG. 1 depicts a block diagram of one example of a media device according to an embodiment of the present application
  • FIG. 2B depicts example scenarios for another media device being configured using a configuration from a previously configured media device according to an embodiment of the present application
  • FIG. 3 depicts one example of a flow diagram of a process for installing an application on a user device and configuring a first media device using the application according to an embodiment of the present application
  • FIGS. 4A and 4B depict example flow diagrams for processes for configuring an un-configured media device according to embodiments of the present application
  • FIG. 6 depicts a block diagram of one example of a proximity detection island according to embodiments of the present application.
  • FIG. 8A is a top plan view depicting an example of proximity detection island coverage according to embodiments of the present application.
  • FIG. 8B is a front side view depicting an example of proximity detection island coverage according to embodiments of the present application.
  • FIG. 8C is a side view depicting an example of proximity detection island coverage according to embodiments of the present application.
  • FIG. 9 is a top plan view of a media device including proximity detection islands configured to detect presence according to embodiments of the present application.
  • FIG. 10 depicts one example of a flow for presence detection, notification, and media device readiness according to embodiments of the present application
  • FIG. 11 depicts another example of a flow for presence detection, notification, and media device readiness according to embodiments of the present application
  • FIG. 12 depicts yet another example of a flow for presence detection, notification, and media device readiness according to embodiments of the present application
  • FIG. 14 depicts one example of proximity detection islands associated with specific device functions according to embodiments of the present application.
  • FIG. 16 depicts another example of content handling from user devices subsequent to proximity detection according to embodiments of the present application.
  • FIG. 17 depicts one example of content handling from a data capable wristband or wristwatch subsequent to proximity detection according to embodiments of the present application
  • FIG. 18 depicts another example of content handling from a data capable wristband or wristwatch subsequent to proximity detection according to embodiments of the present application
  • FIG. 19 depicts one example of a flow for content handling on a media device post proximity detection according to embodiments of the present application.
  • FIG. 20 depicts one example of a flow for storing, recording, and queuing content post proximity detection according to embodiments of the present application
  • FIG. 22 depicts another example of a media device handling, storing, queuing, and taking action on content from a plurality of user devices according to embodiments of the present application
  • FIG. 23 depicts one example of a flow for recording user content on a media device while the media device handles current content according to embodiments of the present application
  • FIG. 24 depicts one example of queuing action for user content in a queue of a media player according to embodiments of the present application
  • FIG. 25 depicts one example of a flow for wirelessly communicating encoded content between a wireless user device and a wireless media device according to embodiments of the present application
  • FIG. 26 depicts another example of a flow for wirelessly communicating encoded content between a wireless user device and a wireless media device according to embodiments of the present application
  • FIG. 27B depicts a block diagram of another example of wirelessly communicating encoded content between a wireless user device and a wireless media device according to embodiments of the present application.
  • FIG. 28 depicts one example of a frequency spectrum for acoustic-based wireless communication between a wireless user device and a wireless media device according to embodiments of the present application.
  • FIG. 1 depicts a block diagram of one embodiment of a media device 100 having systems including but not limited to a controller 101 , a data storage (DS) system 103 , a input/output (I/O) system 105 , a radio frequency (RF) system 107 , an audio/video (AN) system 109 , a power system 111 , and a proximity sensing (PROX) system 113 .
  • a bus 110 enables electrical communication between the controller 101 , DS system 103 , I/O system 105 , RF system 107 , AV system 109 , power system 111 , and PROX system 113 .
  • Power bus 112 supplies electrical power from power system 111 to the controller 101 , DS system 103 , I/O system 105 , RF system 107 , AV system 109 , and PROX system 113 .
  • Power system 111 may include a power source internal to the media device 100 such as a battery (e.g., AAA or AA batteries) or a rechargeable battery (e.g., such as a lithium ion or nickel metal hydride type battery, etc.) denoted as BAT 135 .
  • Power system 111 may be electrically coupled with a port 114 for connecting an external power source (not shown) such as a power supply that connects with an external AC or DC power source. Examples include but are not limited to a wall wart type of power supply that converts AC power to DC power or AC power to AC power at a different voltage level.
  • port 114 may be a connector (e.g., an IEC connector) for a power cord that plugs into an AC outlet or other type of connecter, such as a universal serial bus (USB) connector.
  • Power system 111 provides DC power for the various systems of media device 100 .
  • Power system 111 may convert AC or DC power into a form usable by the various systems of media device 100 .
  • Power system 111 may provide the same or different voltages to the various systems of media device 100 .
  • the external power source may be used to power the power system 111 , recharge BAT 135 , or both.
  • power system 111 on its own or under control or controller 101 may be configured for power management to reduce power consumption of media device 100 , by for example, reducing or disconnecting power from one or more of the systems in media device 100 when those systems are not in use or are placed in a standby or idle mode.
  • Power system 111 may also be configured to monitor power usage of the various systems in media device 100 and to report that usage to other systems in media device 100 and/or to other devices (e.g., including other media devices 100 ) using one or more of the I/O system 105 , RF system 107 , and AV system 109 , for example. Operation and control of the various functions of power system 111 may be externally controlled by other devices (e.g., including other media devices 100 ).
  • Controller 101 controls operation of media device 100 and may include a non-transitory computer readable medium, such as executable program code to enable control and operation of the various systems of media device 100 .
  • DS 103 may be used to store executable code used by controller 101 in one or more data storage mediums such as ROM, RAM, SRAM, RAM, SSD, Flash, etc., for example.
  • Controller 101 may include but is not limited to one or more of a microprocessor ( ⁇ P), a microcontroller ( ⁇ P), a digital signal processor (DSP), a baseband processor, an application specific integrated circuit (ASIC), just to name a few.
  • Processors used for controller 101 may include a single core or multiple cores (e.g., dual core, quad core, etc.).
  • Port 116 may be used to electrically couple controller 101 to an external device (not shown).
  • DS system 103 may include but is not limited to non-volatile memory (e.g., Flash memory), SRAM, DRAM, ROM, SSD, just to name a few.
  • non-volatile memory e.g., Flash memory
  • SRAM static random access memory
  • DRAM dynamic random access memory
  • ROM read-only memory
  • SSD solid state drive
  • DS 103 may be electrically coupled with a port 128 for connecting an external memory source (e.g., USB Flash drive, SD, SDHC, SDXC, microSD, Memory Stick, CF, SSD, etc.).
  • an external memory source e.g., USB Flash drive, SD, SDHC, SDXC, microSD, Memory Stick, CF, SSD, etc.
  • Port 128 may be a USB or mini USB port for a Flash drive or a card slot for a Flash memory card.
  • DS 103 includes data storage for configuration data, denoted as CFG 125 , used by controller 101 to control operation of media device 100 and its various systems.
  • DS 103 may include memory designate for use by other systems in media device 100 (e.g., MAC addresses for WiFi 130 , network passwords, data for settings and parameters for A/V 109 , and other data for operation and/or control of media device 100 , etc.).
  • DS 103 may also store data used as an operating system (OS) for controller 101 . If controller 101 includes a DSP, then DS 103 may store data, algorithms, program code, an OS, etc. for use by the DSP, for example.
  • one or more systems in media device 100 may include their own data storage systems.
  • I/O system 105 may be used to control input and output operations between the various systems of media device 100 via bus 110 and between systems external to media device 100 via port 118 .
  • Port 118 may be a connector (e.g., USB, HDMI, Ethernet, fiber optic, Toslink, Firewire, IEEE 1394, or other) or a hard wired (e.g., captive) connection that facilitates coupling I/O system 105 with external systems.
  • port 118 may include one or more switches, buttons, or the like, used to control functions of the media device 100 such as a power switch, a standby power mode switch, a button for wireless pairing, an audio muting button, an audio volume control, an audio mute button, a button for connecting/disconnecting from a WiFi network, an infrared (IR) transceiver, just to name a few.
  • switches such as a power switch, a standby power mode switch, a button for wireless pairing, an audio muting button, an audio volume control, an audio mute button, a button for connecting/disconnecting from a WiFi network, an infrared (IR) transceiver, just to name a few.
  • IR infrared
  • I/O system 105 may also control indicator lights, audible signals, or the like (not shown) that give status information about the media device 100 , such as a light to indicate the media device 100 is powered up, a light to indicate the media device 100 is in wireless communication (e.g., WiFi, Bluetooth®, WiMAX, cellular, etc.), a light to indicate the media device 100 is Bluetooth® paired, in Bluetooth® pairing mode, Bluetooth® communication is enabled, a light to indicate the audio and/or microphone is muted, just to name a few.
  • Audible signals may be generated by the I/O system 105 or via the AV system 107 to indicate status, etc. of the media device 100 .
  • I/O system 105 may use optical technology to wirelessly communicate with other media devices 100 or other devices. Examples include but are not limited to infrared (IR) transmitters, receivers, transceivers, an IR LED, and an IR detector, just to name a few. I/O system 105 may include an optical transceiver OPT 185 that includes an optical transmitter 185 t (e.g., an IR LED) and an optical receiver 185 r (e.g., a photo diode).
  • IR infrared
  • OPT 185 optical transceiver OPT 185 that includes an optical transmitter 185 t (e.g., an IR LED) and an optical receiver 185 r (e.g., a photo diode).
  • OPT 185 may include the circuitry necessary to drive the optical transmitter 185 t with encoded signals and to receive and decode signals received by the optical receiver 185 r .
  • Bus 110 may be used to communicate signals to and from OPT 185 .
  • OPT 185 may be used to transmit and receive IR commands consistent with those used by infrared remote controls used to control AV equipment, televisions, computers, and other types of systems and consumer electronics devices.
  • the IR commands may be used to control and configure the media device 100 , or the media device 100 may use the IR commands to configure/re-configure and control other media devices or other user devices, for example.
  • RF system 107 includes at least one RF antenna 124 that is electrically coupled with a plurality of radios (e.g., RF transceivers) including but not limited to a Bluetooth® (BT) transceiver 120 , a WiFi transceiver 130 (e.g., for wireless communications over a wireless and/or WiMAX network), and a proprietary Ad Hoc (AH) transceiver 140 pre-configured (e.g., at the factory) to wirelessly communicate with a proprietary Ad Hoc wireless network (AH-WiFi) (not shown).
  • AH 140 and AH-WiFi are configured to allow wireless communications between similarly configured media devices (e.g., an ecosystem comprised of a plurality of similarly configured media devices) as will be explained in greater detail below.
  • RF system 107 may include more or fewer radios than depicted in FIG. 1 and the number and type of radios will be application dependent. Furthermore, radios in RF system 107 need not be transceivers, RF system 107 may include radios that transmit only or receive only, for example. Optionally, RF system 107 may include a radio 150 configured for RF communications using a proprietary format, frequency band, or other existent now or to be implemented in the future. Radio 150 may be used for cellular communications (e.g., 3G, 4G, or other), for example.
  • Radio 150 may be used for cellular communications (e.g., 3G, 4G, or other), for example.
  • Antenna 124 may be configured to be a de-tunable antenna such that it may be de-tuned 129 over a wide range of RF frequencies including but not limited to licensed bands, unlicensed bands, WiFi, WiMAX, cellular bands, Bluetooth®, from about 2.0 GHz to about 6.0 GHz range, and broadband, just to name a few.
  • PROX system 113 may use the de-tuning 129 capabilities of antenna 124 to sense proximity of the user, other people, the relative locations of other media devices 100 , just to name a few.
  • Radio 150 e.g., a transceiver or other transceiver in RF 107
  • Radio 150 may be used in conjunction with the de-tuning capabilities of antenna 124 to sense proximity, to detect and or spatially locate other RF sources such as those from other media devices 100 , devices of a user, just to name a few.
  • RF system 107 may include a port 123 configured to connect the RF system 107 with an external component or system, such as an external RF antenna, for example.
  • the transceivers depicted in FIG. 1 are non-limiting examples of the type of transceivers that may be included in RF system 107 .
  • RF system 107 may include a first transceiver configured to wirelessly communicate using a first protocol, a second transceiver configured to wirelessly communicate using a second protocol, a third transceiver configured to wirelessly communicate using a third protocol, and so on.
  • One of the transceivers in RF system 107 may be configured for short range RF communications, such as within a range from about 1 meter to about 15 meters, or less, for example.
  • Another one of the transceivers in RF system 107 may be configured for long range RF communications, such any range up to about 50 meters or more, for example.
  • Short range RF may include Bluetooth®; whereas, long range RF may include WiFi, WiMAX, cellular, and Ad Hoc wireless, for example.
  • AV system 109 includes at least one audio transducer, such as a loud speaker 160 , a microphone 170 , or both.
  • AV system 109 further includes circuitry such as amplifiers, preamplifiers, or the like as necessary to drive or process signals to/from the audio transducers.
  • AV system 109 may include a display (DISP) 180 , a video device (VID) 190 (e.g., an image capture device, video camera, still camera, or a web CAM, etc.), or both.
  • DISP 180 may be a display and/or touch screen (e.g., a LCD, OLED, or flat panel display) for displaying video media, information relating to operation of media device 100 , content available to or operated on by the media device 100 , playlists for media, date and/or time of day, alpha-numeric text and characters, caller ID, file/directory information, a GUI, just to name a few.
  • a port 122 may be used to electrically couple AV system 109 with an external device and/or external signals. Port 122 may be a USB, HDMI, Firewire/IEEE-1394, 3.5 mm audio jack, or other.
  • port 122 may be a 3.5 mm audio jack for connecting an external speaker, headphones, earphones, etc. for listening to audio content being processed by media device 100 .
  • port 122 may be a 3.5 mm audio jack for connecting an external microphone or the audio output from an external device.
  • SPK 160 may include but is not limited to one or more active or passive audio transducers such as woofers, concentric drivers, tweeters, super tweeters, midrange drivers, sub-woofers, passive radiators, just to name a few.
  • MIC 170 may include one or more microphones and the one or more microphones may have any polar pattern suitable for the intended application including but not limited to omni-directional, directional, bi-directional, uni-directional, bi-polar, uni-polar, any variety of cardioid pattern, and shotgun, for example.
  • MIC 170 may be configured for mono, stereo, or other.
  • MIC 170 may be configured to be responsive (e.g., generate an electrical signal in response to sound) to any frequency range including but not limited to ultrasonic, infrasonic, from about 20 Hz to about 20 kHz, and any range within or outside of human hearing.
  • the audio transducer of AV system 109 may serve dual roles as both a speaker and a microphone.
  • Circuitry in AV system 109 may include but is not limited to a digital-to-analog converter (DAC) and algorithms for decoding and playback of media files such as MP3, FLAG, AIFF, ALAC, WAV, MPEG, QuickTime, AVI, compressed media files, uncompressed media files, and lossless media files, just to name a few, for example.
  • a DAC may be used by AV system 109 to decode wireless data from a user device or from any of the radios in RF system 107 .
  • AV system 109 may also include an analog-to-digital converter (ADC) for converting analog signals, from MIC 170 for example, into digital signals for processing by one or more system in media device 100 .
  • ADC analog-to-digital converter
  • Media device 100 may be used for a variety of applications including but not limited to wirelessly communicating with other wireless devices, other media devices 100 , wireless networks, and the like for playback of media (e.g., streaming content), such as audio, for example.
  • media e.g., streaming content
  • the actual source for the media need not be located on a user's device (e.g., smart phone, MP3 player, iPod, iPhone, iPad, Android, laptop, PC, etc.).
  • media files to be played back on media device 100 may be located on the Internet, a web site, or in the Cloud, and media device 100 may access (e.g., over a WiFi network via WiFi 130 ) the files, process data in the files, and initiate playback of the media files.
  • Media device 100 may access or store in its memory a playlist or favorites list and playback content listed in those lists.
  • media device 100 will store content (e.g., files) to be played back on the media device 100 or on another media device 100 .
  • Media device 100 may include a housing, a chassis, an enclosure or the like, denoted in FIG. 1 as 199 .
  • the actual shape, configuration, dimensions, materials, features, design, ornamentation, aesthetics, and the like of housing 199 will be application dependent and a matter of design choice. Therefore, housing 199 need not have the rectangular form depicted in FIG. 1 or the shape, configuration etc., depicted in the Drawings of the present application. None precludes housing 199 from comprising one or more structural elements, that is, the housing 199 may be comprised of several housings that form media device 100 .
  • Housing 199 may be configured to be worn, mounted, or otherwise connected to or carried by a human being.
  • housing 199 may be configured as a wristband, an earpiece, a headband, a headphone, a headset, an earphone, a hand held device, a portable device, a desktop device, just to name a few.
  • housing 199 may be configured as speaker, a subwoofer, a conference call speaker, an intercom, a media playback device, just to name a few. If configured as a speaker, then the housing 199 may be configured as a variety of speaker types including but not limited to a left channel speaker, a right channel speaker, a center channel speaker, a left rear channel speaker, a right rear channel speaker, a subwoofer, a left channel surround speaker, a right channel surround speaker, a left channel height speaker, a right channel height speaker, any speaker in a 3.1, 5.1, 7.1, 9.1 or other surround sound format including those having two or more subwoofers or having two or more center channels, for example. In other examples, housing 199 may be configured to include a display (e.g., DISP 180 ) for viewing video, serving as a touch screen interface for a user, providing an interface for a GUI, for example.
  • a display e.g., DISP 180
  • PROX system 113 may include one or more sensors denoted as SEN 195 that are configured to sense 197 an environment 198 external to the housing 199 of media device 100 .
  • SEN 195 and/or other systems in media device 100 (e.g., antenna 124 , SPK 160 , MIC 170 , etc.)
  • PROX system 113 senses 197 an environment 198 that is external to the media device 100 (e.g., external to housing 199 ).
  • PROX system 113 may be used to sense one or more of proximity of the user or other persons to the media device 100 or other media devices 100 .
  • PROX system 113 may use a variety of sensor technologies for SEN 195 including but not limited to ultrasound, infrared (IR), passive infrared (PIR), optical, acoustic, vibration, light, ambient light sensor (ALS), IR proximity sensors, LED emitters and detectors, RGB LED's, RF, temperature, capacitive, capacitive touch, inductive, just to name a few.
  • SEN 195 may use a variety of sensor technologies for SEN 195 including but not limited to ultrasound, infrared (IR), passive infrared (PIR), optical, acoustic, vibration, light, ambient light sensor (ALS), IR proximity sensors, LED emitters and detectors, RGB LED's, RF, temperature, capacitive, capacitive touch, inductive, just to name a few.
  • PROX system 113 may be configured to sense location of users or other persons, user devices, and other media devices 100 , without limitation.
  • Output signals from PROX system 113 may be used to configure media device 100 or other media devices 100 , to re-configure and/or re-purpose media device 100 or other media devices 100 (e.g., change a role the media device 100 plays for the user, based on a user profile or configuration data), just to name a few.
  • a plurality of media devices 100 in an eco-system of media devices 100 may collectively use their respective PROX system 113 and/or other systems (e.g., RF 107 , de-tunable antenna 124 , AV 109 , etc.) to accomplish tasks including but not limited to changing configuration, re-configuring one or more media devices, implement user specified configurations and/or profiles, insertion and/or removal of one or more media devices in an eco-system, just to name a few.
  • other systems e.g., RF 107 , de-tunable antenna 124 , AV 109 , etc.
  • PROX 113 may include one or more proximity detection islands PSEN 520 as will be discussed in greater detail in FIGS. 5-6 .
  • PSEN 520 may be positioned at one or more locations on chassis 199 and configured to sense an approach of a user or other person towards the media device 100 or to sense motion or gestures of a user or other person by a portion of the body such as a hand for example.
  • PSEN 520 may be used in conjunction with or in place of one or more of SEN 195 , OPT 185 , SPK 160 , MIC 170 , RF 107 and/or de-tunable 129 antenna 124 to sense proximity and/or presence in an environment surrounding the media device 100 , for example.
  • PSEN 520 may be configured to take or cause an action to occur upon detection of an event (e.g., an approach or gesture by user 201 or other) such as emitting light (e.g., via an LED), generating a sound or announcement (e.g., via SPK 160 ), causing a vibration (e.g., via SPK 160 or a vibration motor), display information (e.g., via DISP 180 ), trigger haptic feedback, for example.
  • PSEN 520 may be included in I/O 105 instead of PROX 113 or be shared between one or more systems of media device 100 .
  • components, circuitry, and functionality of PSEN 520 may vary among a plurality of PSEN 520 sensors in media device 100 such that all PSEN 520 are not identical.
  • a scenario 200 a depicts one example of a media device (e.g., media device 100 of FIG. 1 or a similarly provisioned media device) being configured for the first time by a user 201 .
  • media device is denoted as 100 a to illustrate that it is the first time the media device 100 a is being configured.
  • the first configuration of media device 100 a may be after it is purchased, acquired, borrowed, or otherwise by user 201 , that is, the first time may be the initial out-of-the-box configuration of media device 100 a when it is new.
  • Scenario 200 a depicts a desirable user experience for user 201 to achieve the objective of making the configuring of media device 100 a as easy, straight forward, and fast as possible.
  • scenario 200 a may include media device 100 a to be configured, for example, initially by user 201 using a variety of devices 202 including but not limited to a smartphone 210 , a tablet 220 , a laptop computer 230 , a data capable wristband or the like 240 , a desktop PC or server 250 , . . . etc.
  • devices 202 including but not limited to a smartphone 210 , a tablet 220 , a laptop computer 230 , a data capable wristband or the like 240 , a desktop PC or server 250 , . . . etc.
  • tablet 220 although the description may apply to any of the other devices 202 as well.
  • controller 101 may command RF system 107 to electrically couple 224 , transceiver BT 120 with antenna 124 , and command BT 120 to begin listening 126 for a BT pairing signal from device 220 .
  • user 201 as part of the initialization process may have already used a Bluetooth® menu on tablet 220 to activate the BT radio and associated software in tablet 220 to begin searching (e.g., via RF) for a BT device to pair with. Pairing may require a code (e.g., a PIN number or code) be entered by the user 201 for the device being paired with, and the user 201 may enter a specific code or a default code such as “0000”, for example.
  • a code e.g., a PIN number or code
  • BT 120 need not be used for wireless communication between media device 100 a and the user's device (e.g., tablet 220 or other).
  • Controller 101 after a successful BT pairing, may command RF system 107 to electrically couple 228 , WiFi 130 with antenna 124 and wireless communications between tablet 220 and media device 100 a (see 260 , 226 ) may occur over a wireless network (e.g., WiFi or WiMAX) or other as denoted by wireless access point 270 .
  • a wireless network e.g., WiFi or WiMAX
  • tablet 220 requires a non-transitory computer readable medium that includes data and/or executable code to form a configuration (CFG) 125 for media device 100 a .
  • the non-transitory computer readable medium will be denoted as an application (APP) 225 .
  • APP 225 resides on or is otherwise accessible by tablet 220 or media device 100 a .
  • User 201 uses APP 225 (e.g., through a GUI, menu, drop down boxes, or the like) to make selections that comprise the data and/or executable code in the CFG 125 .
  • APP 225 may be obtained by tablet 220 in a variety of ways.
  • the media device 100 a includes instructions (e.g., on its packaging or in a user manual) for a website on the Internet 250 where the APP 225 may be downloaded.
  • Tablet 220 may use its WiFi or Cellular RF systems to communicate with wireless access point 270 (e.g., a cell tower or wireless router) to connect 271 with the website and download APP 255 which is stored on tablet 220 as APP 225 .
  • wireless access point 270 e.g., a cell tower or wireless router
  • tablet 220 may scan or otherwise image a bar code or TAG operative to connect the tablet 220 with a location (e.g., on the Internet 250 ) where the APP 225 may be found and downloaded.
  • Tablet 220 may have access to an applications store such as Google Play for Android devices, the Apple App Store for iOS devices, or the Windows 8 App Store for Windows 8 devices.
  • the APP 225 may then be downloaded from the app store.
  • media device 100 a may be preconfigured to either provide (e.g., over the BT 120 or WiFi 130 ) an address or other location that is communicated to tablet 220 and the tablet 220 uses the information to locate and download the APP 225 .
  • media device 100 a may be preloaded with one or more versions of APP 225 for use in different device operating systems (OS), such as one version for Android, another for iOS, and yet another for Windows 8, etc.
  • OS device operating systems
  • media device 100 a may use its wireless systems (e.g., BT 120 or WiFi 130 ) to determine if the preloaded versions are out of date and need to be replaced with newer versions, which the media device 100 a obtains, downloads, and subsequently makes available for download to tablet 220 .
  • wireless systems e.g., BT 120 or WiFi 130
  • the user 201 may use the APP 225 to select various options, commands, settings, etc. for CFG 125 according to the user's preferences, needs, media device ecosystem, etc., for example.
  • CFG 125 is downloaded (e.g., using BT 120 or WiFi 130 ) into DS system 103 in media device 100 a .
  • Controller 101 may use the CFG 125 and/or other executable code to control operation of media device 100 a .
  • the source for APP 225 may be obtained from a variety of locations including but not limited to: the Internet 250 ; a file or the like stored in the Cloud; a web site; a server farm; a FTP site; a drop box; an app store; a manufactures web site; or the like, just to name a few.
  • APP 225 may be installed using other processes including but not limited to: dragging and dropping the appropriate file into a directory, folder, desktop or the like on tablet 220 ; emailing the APP 225 as an attachment, a compressed or ZIP file; cutting and pasting the App 225 , just to name a few.
  • CFG 125 may include data such as the name and password for a wireless network (e.g., 270 ) so that WiFi 130 may connect with (see 226 ) and use the wireless network for future wireless communications, data for configuring subsequently purchased devices 100 , data to access media for playback, just to name a few.
  • a wireless network e.g., 270
  • WiFi 130 may connect with (see 226 ) and use the wireless network for future wireless communications
  • data for configuring subsequently purchased devices 100 data to access media for playback, just to name a few.
  • user 201 may update CFG 125 as the needs of the user 201 change over time, that is, APP 225 may be used to re-configure an existing CFG 125 .
  • APP 225 may be configured to check for updates and to query the user 201 to accept the updates such that if an update is accepted an updated version of the APP 225 may be installed on tablet 220 or on any of the other devices 202 .
  • APP 225 and CFG 125 may be installed on devices 202 and/or media device 100 a using the process described above.
  • APP 225 or some other program may be used to perform software, firmware, or data updates on device 100 a .
  • DS system 103 on device 100 a may include storage set aside for executable code (e.g., an operating system) and data used by controller 101 and/or the other systems depicted in FIG. 1 .
  • FIG. 2B where a several example scenarios of how a previously configured media device 100 a that includes CFG 125 may be used to configure another media device 100 b that is initially un-configured.
  • media device 100 a is already powered up or is turned on (e.g., by user 201 ) or is otherwise activated such that its RF system 107 is operational.
  • media device 100 a is powered up and configured to detect RF signatures from other powered up media devices using its RF system 107 .
  • RF proximity broadly means within adequate signal strength range of the BT transceivers 120 , WiFi transceivers 130 , or any other transceivers in RF system 107 , RF systems in the users devices (e.g., 202 , 220 ), and other wireless devices such as wireless routers, WiFi networks (e.g., 270 ), WiMAX networks, and cellular networks, for example.
  • Adequate signal strength range is any range that allows for reliable RF communications between wireless devices.
  • adequate signal strength range may be determined by the BT specification, but is subject to change as the BT specification and technology evolve. For example, adequate signal strength range for BT 120 may be approximately 10 meters (e.g., ⁇ 30 feet). For WiFi 130 , adequate signal strength range may vary based on parameters such as distance from and signal strength of the wireless network, and structures that interfere with the WiFi signal. However, in most typical wireless systems adequate signal strength range is usually greater than 10 meters.
  • media device 100 b is powered up and at stage 290 c its BT 120 and the BT 120 of media device 100 a recognize each other.
  • each media device ( 100 a , 100 b ) may be pre-configured (e.g., at the factory) to broadcast a unique RF signature or other wireless signature (e.g., acoustic) at power up and/or when it detects the unique signature of another device.
  • the unique RF signature may include status information including but not limited to the configuration state of a media device.
  • Each BT 120 may be configured to allow communications with and control by another media device based on the information in the unique RF signature.
  • media device 100 b transmits RF information that includes data that informs other listening BT 120 's (e.g., BT 120 in 100 a ) that media device 100 b is un-configured (e.g., has no CFG 125 ).
  • media devices 100 a and 100 b negotiate the necessary protocols and/or handshakes that allow media device 100 a to gain access to DS 103 of media device 100 b .
  • media device 100 b is ready to receive CFG 125 from media device 100 a , and at stage 290 f the CFG 125 from media device 100 a is transmitted to media device 100 b and is replicated (e.g., copied, written, etc.) in the DS 103 of media device 100 b , such that media device 100 b becomes a configured media device.
  • Data in CFG 125 may include information on wireless network 270 , including but not limited to wireless network name, wireless password, MAC addresses of other media devices, media specific configuration such as speaker type (e.g., left, right, center channel), audio mute, microphone mute, etc. Some configuration data may be subservient to other data or dominant to other data.
  • media device 100 a , media device 100 b , and user device 220 may wirelessly communicate 291 with one another over wireless network 270 using the WiFi systems of user device 220 and WiFi 130 of media devices 100 a and 100 b.
  • APP 225 may be used to input the above data into CFG 125 , for example using a GUI included with the APP 225 .
  • User 201 enters data and makes menu selections (e.g., on a touch screen display) that will become part of the data for the CFG 125 .
  • APP 225 may also be used to update and/or re-configure an existing CFG 125 on a configured media device.
  • other configured or un-configured media devices in the user's ecosystem may be updated and/or re-configured by a previously updated and/or re-configured media device as described herein, thereby relieving the user 201 from having to perform the update and/or re-configure on several media devices.
  • the APP 225 or a location provided by the APP 225 may be used to specify playlists, media sources, file locations, and the like.
  • APP 225 may be installed on more than one user device 202 and changes to APP 225 on one user device may later by replicated on the APP 225 on other user devices by a synching or update process, for example.
  • APP 225 may be stored on the internet or in the Cloud and any changes to APP 225 may be implemented in versions of the APP 225 on various user devices 202 by merely activating the APP 225 on that device and the APP 225 initiates a query process to see if any updates to the APP are available, and if so, then the APP 225 updates itself to make the version on the user device current with the latest version.
  • FIG. 2B includes an alternate scenario 200 b that may be used to configure a newly added media device, that is, an un-configured media device (e.g., 100 b ).
  • media device 100 a which is assumed to already have its WiFi 130 configured for communications with wireless network 270 , transmits over its BT 120 the necessary information for media device 100 b to join wireless network 270 .
  • media device 100 b After stage 290 d , media device 100 b , media device 100 a , and tablet 220 are connected 291 to wireless network 270 and may communicate wirelessly with one another via network 270 . Furthermore, at stage 290 d , media device 100 b is still in an un-configured state. Next, at stage 290 e , APP 225 is active on tablet 220 and wirelessly accesses the status of media devices 100 a and 100 b .
  • APP 225 determines that media device 100 b is un-configured and APP 225 acts to con figure 100 b by harvesting CFG 125 (e.g., getting a copy of) from configured media device 100 a by wirelessly 293 a obtaining CFG 125 from media device 100 a and wirelessly 293 b transmitting the harvested CFG 125 to media device 100 b .
  • Media device 100 b uses its copy of CFG 125 to configure itself thereby placing it in a configured state.
  • FIG. 2B depicts yet another example scenario where after stage 290 d , the APP 225 or any one of the media devices 100 a , 100 b , may access 295 the CFG 125 for media device 100 b from an external location, such as the Internet, the cloud, etc. as denoted by 250 where a copy of CFG 125 may be located and accessed for download into media device 100 b .
  • APP 255 , media device 100 b , or media device 100 a may access the copy of CFG 125 from 250 and wirelessly install it on media device 100 b.
  • adding a new media device to his/her ecosystem of similarly provisioned media devices does not require un-pairing with one or more already configured devices and then pairing with the new device to be added to the ecosystem. Instead, one of the already configured devices (e.g., media device 100 a having CFG 125 installed) may negotiate with the APP 225 and/or the new device to be added to handle the configuration of the new device (e.g., device 100 b ).
  • provisioned media devices broadly means devices including some, all, or more of the systems depicted in FIG. 1 and designed (e.g., by the same manufacture or to the same specifications and/or standards) to operate with one another in a seamless manner as media devices are added to or removed from an ecosystem.
  • a flow diagram 300 depicts one example of configuring a first media device using an application installed on a user device as was described above in regards to FIG. 2A .
  • a Bluetooth® (BT) discovery mode is activated on a user device such as the examples 202 of user devices depicted in FIG. 2A .
  • a GUI on the user device includes a menu for activating BT discovery mode, after which, the user device waits to pick up a BT signal of a device seeking to pair with the user's device.
  • a first media device e.g., 100 a
  • is powered up if not already powered up).
  • a BT pairing mode is activated on the first media device.
  • Examples of activating BT pairing mode include but are not limited to pushing a button or activating a switch on the first media device that places the first media device in BT pairing mode such that its BT 120 is activated to generate a RF signal that the user's device may discover while in discovery mode.
  • I/O system 105 of media device 100 may receive 118 as a signal the activation of BT pairing mode by actuation of the switch or button and that signal is processed by controller 101 to command RF system 107 to activate BT 120 in pairing mode.
  • a display e.g., DISP 180
  • the user's device and the first media device negotiate the BT pairing process, and if BT pairing is successful, then the flow continues at stage 310 . If BT pairing is not successful, then the flow repeats at the stage 206 until successful BT pairing is achieved.
  • the user device is connected to a wireless network (if not already connected) such as a WiFi, WiMAX, or cellular (e.g., 3G or 4G) network.
  • the wireless network may be used to install an application (e.g., APP 225 ) on the user's device.
  • the location of the APP may be provided with the media device or after successful BT pairing, the media device may use its BT 120 to transmit data to the user's device and that data includes a location (e.g., a URI or URL) for downloading or otherwise accessing the APP.
  • the user uses the APP to select settings for a configuration (e.g., CFG 125 ) for the first media device.
  • the user's device installs the APP on the first media device. The installation may occur in a variety of ways (see FIG.
  • a determination of whether or not the first media device is connected with a wireless network may be made at a stage 318 . If the first media device is already connected with a wireless network the “YES” branch may be taken and the flow may terminate at stage 320 .
  • the “NO” branch may be taken and the flow continues at a stage 322 where data in the CFG is used to connect WiFi 130 with a wireless network and the flow may terminate at a stage 324 .
  • the CFG may contain the information necessary for a successful connection between WiFi 130 and the wireless network, such as wireless network name and wireless network password, etc.
  • a flow diagram 400 a depicts one example of a process for configuring an un-configured media device “B” (e.g., un-configured media device 100 b at stage 290 b of FIG. 2B ) using a configured media device “A” (e.g., media device 100 a having CFG 125 of FIG. 2B ).
  • a configured media device “A” e.g., media device 100 a having CFG 125 of FIG. 2B
  • an already configured media device “A” is powered up.
  • the RF system e.g., RF system 107 of FIG. 1
  • the RF system is configured to detect RF signals from other “powered up” media devices.
  • an un-configured media device “B” (e.g., un-configured media device 100 b at stage 290 b of FIG. 2B ) is powered up.
  • the RF system of un-configured media device “B” is activated.
  • the respective RF systems of the configured “A” and un-configured “B” media devices are configured to recognize each other (e.g., via their respective BT 120 transceivers or another transceiver in the RF system).
  • a “YES” branch is taken to a stage 412 where the configured media device “A” transmits its configuration (e.g., CFG 125 ) to the un-configured media device “B” (e.g., see stages 290 e and 290 f in FIG. 2B ). If the configured “A” and un-configured “B” media devices do not recognize each other, then a “NO” branch is taken and the flow may return to an earlier stage (e.g., stage 404 to retry the recognition process.
  • media device “B” may be connected with a wireless network (e.g., via WiFi 130 ).
  • the CFG 125 that was copied to media device “B” may include information such as wireless network name and password and WiFi 130 is configured to effectuate the connection with the wireless network based on that information.
  • media device “A” may transmit the necessary information to media device “B” (e.g., using BT 120 ) at any stage of flow 400 a , such as at the stage 408 , for example.
  • the flow may terminate at a stage 420 .
  • FIG. 4B depicts another example of a process for configuring an un-configured media device “B” (e.g., un-configured media device 100 b at stage 290 b of FIG. 2B ) using a configured media device “A” (e.g., media device 100 a having CFG 125 of FIG. 2B ).
  • a configured media device “A” e.g., media device 100 a having CFG 125 of FIG. 2B
  • an already configured media device “A” is powered up.
  • the RF system of configured media device “A” is activated (e.g., RF system 107 of FIG. 1 ).
  • the RF system is configured to detect RF signals from other “powered up” media devices.
  • an un-configured media device “B” (e.g., un-configured media device 100 b at stage 290 b of FIG. 2B ) is powered up.
  • the RF system of un-configured media device “b” is activated (e.g., RF system 107 of FIG. 1 ).
  • the respective RF systems of the configured “A” and un-configured “B” media devices are configured to recognize each other (e.g., via their respective BT 120 transceivers or another transceiver in the RF system).
  • a “YES” branch is taken to a stage 432 where the configured media device “A” transmits information for a wireless network to the un-configured media device “B” (e.g., see stage 290 b in FIG. 2B ) and that information is used by the un-configured media device “B” to connect with a wireless network as was described above in regards to FIGS. 2B and 4A . If the configured “A” and un-configured “B” media devices do not recognize each other, then a “NO” branch is taken and the flow may return to an earlier stage (e.g., stage 424 to retry the recognition process.
  • the information for the wireless network is used by the un-configured media device “B” to effectuate a connection to the wireless network.
  • a user device is connected with the wireless network and an application (APP) running on the user device (e.g., APP 225 in FIG. 2B ) is activated. Stage 436 may be skipped if the user device is already connected to the wireless network.
  • APP application
  • Un-configured media device “B” may include registers, circuitry, data, program code, memory addresses, or the like that may be used to determine that the media device is un-configured.
  • the un-configured status of media device “B” may be wirelessly broadcast using any of its wireless resources or other systems, such as RF 107 and/or AV 109 .
  • the APP is aware of configured media device “A” presence on the wireless network and detects that media device “A” is presently in a configured state and therefore has a status of “configured.”
  • the APP harvests the configuration (CFG) (e.g., CFG 125 of FIG. 2B ) from configured media device “A”, and at a stage 442 copies (e.g., via a wireless transmission over the wireless network) the CFG to the un-configured media device “B.”
  • CFG configuration
  • previously un-configured media device “B” becomes a configured media device “B” by virtue of having CFG resident in its system (e.g., CFG 125 in DS system 103 in FIG. 1 ).
  • the flow may terminate at a stage 446 .
  • the APP may obtain the CFG from a location other than the configured media device “A”, such as the Internet or the Cloud as depicted in FIG. 2B . Therefore, at the stage 440 , the APP may download the CFG from a web site, from Cloud storage, or other locations on the Internet or an intranet for example.
  • additional media devices that are added by the user or are encountered by the user may be configured without the user (e.g., user 201 ) having to break a BT pairing with one media device and then establishing another BT pairing with a media device the user is adding to his/her media device ecosystem.
  • Existing media devices that are configured e.g., have CFG 125
  • configured media devices may be configured to arbitrate among themselves as to which of the configured devices will act to configured the newly added un-configured media device.
  • the existing media device that was configured last in time e.g., by a date stamp on its CFG 125
  • the existing media device that was configured first in time e.g., by a date stamp on its CFG 125
  • the existing media device that was configured first in time may be the one selected to configure the newly added un-configured media device.
  • the APP 225 on the user device 220 or other may be configured to make the configuration process as seamless as possible and may only prompt the user 201 that the APP 225 has detected an un-configured media device and query the user 201 as to whether or not the user 201 wants the APP 225 to configure the un-configured media device (e.g., media device 100 b ). If the user replies “YES”, then the APP 225 may handle the configuration process working wirelessly with the configured and un-configured media devices. If the user 201 replies “NO”, then the APP 225 may postpone the configuration for a later time when the user 201 is prepared to consummate the configuration of the un-configured media device. In other examples, the user 201 may want configuration of un-configured media devices to be automatic upon detection of the un-configured media device(s). Here the APP and/or configured media devices would automatically act to configure the un-configured media device(s).
  • APP 225 may be configured (e.g., by the user 201 ) to automatically configure any newly detected un-configured media devices that are added to the user's 201 ecosystem and the APP 225 may merely inform the user 201 that it is configuring the un-configured media devices and inform the user 201 when configuration is completed, for example.
  • subsequently added un-configured media devices may be automatically configured by an existing configured media device by each media device recognizing other media devices (e.g., via wireless systems), determining the status (e.g., configured or un-configured) of each media device, and then using the wireless systems (e.g., RF 107 , AV 109 , I/O 105 , OPT 185 , PROX 113 ) of a configured media device to configure the un-configured media device without having to resort to the APP 225 on the user's device 220 to intervene in the configuration process.
  • the wireless systems e.g., RF 107 , AV 109 , I/O 105 , OPT 185 , PROX 113
  • the configured media devices and the un-configured media devices arbitrate and effectuate the configuring of un-configured media devices without the aid of APP 225 or user device 220 .
  • the controller 101 and/or CFG 125 may include instructions for configuring media devices in an ecosystem using one or more systems in the media devices themselves.
  • the structures and/or functions of any of the above-described features may be implemented in software, hardware, firmware, circuitry, or in any combination thereof.
  • the structures and constituent elements above, as well as their functionality may be aggregated with one or more other structures or elements.
  • the elements and their functionality may be subdivided into constituent sub-elements, if any.
  • the above-described techniques may be implemented using various types of programming or formatting languages, frameworks, scripts, syntax, applications, protocols, objects, or techniques.
  • module may refer, for example, to an algorithm or a portion thereof, and/or logic implemented in either hardware circuitry or software, or a combination thereof. These may be varied and are not limited to the examples or descriptions provided.
  • Software, firmware, algorithms, executable computer readable code, program instructions for execution on a computer, or the like may be embodied in a non-transitory computer readable medium.
  • FIG. 5 a profile view depicts one example 500 of media device 100 that may include on a top surface 199 s of chassis 199 , a plurality of control elements 503 - 512 and one or more proximity detection islands (four are depicted) denoted as 520 .
  • Media device 100 may include one or more speakers 160 , one or more microphones 170 , a display 180 , a section 550 for other functions such as SEN 195 , VID 109 , or other, and antenna 124 which may be tunable 129 .
  • Each proximity detection island 520 may be configured to detect 597 proximity of one or more persons, such as user 201 as will be described in greater detail below.
  • the layout and position of the elements on chassis 199 of media device 100 are examples only and actual layout and position of any elements will be application specific and/or a matter of design choice, including ergonomic and esthetic considerations.
  • detection of presence of user 201 may occur with or without the presence of one or more user devices 202 , such as user devices 210 and 220 depicted in FIG. 5 .
  • Circuitry and/or software associated with operation of proximity detection islands 520 may work in conjunction with other systems in media device 100 to detect presence of one or more user devices 202 , such as RF system 107 detecting RF signals 563 and/or 565 (e.g., via antenna 124 ) from user devices 210 and 220 or MIC 170 detecting sound, for example.
  • Detection of presence may be signaled by media device 100 in a variety of ways including but not limited to light (e.g., from 520 and/or 503 - 512 ), sound (e.g., from SPK 160 ), vibration (e.g., from SPK 160 or other), haptic feedback, tactile feedback, display of information (e.g., DISP 180 ), RF transmission (e.g., 126 ), just to name a few.
  • SPK 160 and DISP 180 may be positioned on a front surface 199 f of chassis 199 .
  • a bottom surface 199 b of chassis 199 may be configured to rest on a surface such as a table, desk, cabinet, or the like.
  • Other elements of media device 100 may be positioned on a rear surface 199 r of chassis 199 .
  • Non-limiting examples of control elements 503 - 512 include a plurality of controls 512 (e.g., buttons, switches and/or touch surfaces) that may have functions that are fixed or change based on different scenarios as will be described below, controls 503 and 507 for volume up and volume down, control 509 for muting volume or BT paring, control 506 for initiating or pausing playback of content, control 504 for fast reversing playback or skipping backward one track, and control 508 for fast forwarding playback or skipping forward one track. Some are all of the control elements 504 - 512 may serve multiple rolls based on changing scenarios.
  • controls 512 e.g., buttons, switches and/or touch surfaces
  • controls 503 and 507 for volume up and volume down
  • control 509 for muting volume or BT paring
  • control 506 for initiating or pausing playback of content
  • control 504 for fast reversing playback or skipping backward one track
  • control 508 for fast forwarding playback or skipping
  • controls 503 and 507 may be used to increase “+” and decrease “ ⁇ ” brightness of display 180 .
  • Control 509 may be used to transfer or pick up a phone call or other content on a user device 202 , for example.
  • Proximity detection islands 520 and/or control elements 503 - 512 may be backlit (e.g., using LED's or the like) for night or low-light visibility.
  • a block diagram 600 depicts one example of a proximity detection island 520 .
  • Proximity detection island 520 may be implemented using a variety of technologies and circuit topologies and the example depicted in FIG. 6 is just one such non-limiting example and the present application is not limited to the arrangement of elements depicted in FIG. 6 .
  • One or more proximity detection islands 520 may be positioned on, connected with, carried by or otherwise mounted on media device 100 .
  • proximity detection island 520 may be mounted on a top surface 199 t of chassis 199 .
  • a structure 650 made from an optically transmissive material such as glass, plastic, a film, an optically transparent or translucent material, or the like.
  • Structure 650 may be made from a material that allows light 603 , 607 , 617 , and 630 to pass through it in both directions, that is, bi-directionally. Structure 650 may include apertures 652 defined by regions 651 (e.g., an opaque or optically reflective/absorptive material) used for providing optical access (e.g., via apertures 652 ) to an environment ENV 198 external to the media device 100 for components of the proximity detection island 520 . Structure 650 may be configured to mount flush with top surface 199 t , for example. In some examples, structure 650 may not include regions 651 .
  • regions 651 e.g., an opaque or optically reflective/absorptive material
  • Proximity detection island 520 may include at least one LED 601 (e.g., an infrared LED-IR LED) electrically coupled with driver circuitry 610 and configured to emit IR radiation 603 , at least one IR optical detector 605 (e.g., a PIN diode) electrically coupled with an analog-to-digital converter ADC 612 and configured to generate a signal in response to IR radiation 607 incident on detector 605 , and at least one indicator light 616 electrically coupled with driver circuitry 614 and configured to generate colored light 617 .
  • indicator light 616 comprises a RGB LED configured to emit light 617 in a gambit of colors indicative of status as will be described below.
  • RGB LED 616 may include four terminals, one of which coupled with circuit ground, a red “R” terminal, a green “G” terminal, and a blue “B” terminal, all of which are electrically connected with appropriate circuitry in driver 614 and with die within RGB LED 616 to effectuate generation of various colors of light in response to signals from driver 614 .
  • RGB LED 616 may include semiconductor die for LED's that generate red, green, and blue light that are electrically coupled with ground and the R, G, and B terminals, respectively.
  • element 616 may be replaced by discrete LED's (e.g., separate red, green, white, and blue LED's) or a single non-RGB LED or other light emitting device may be used for 616 .
  • the various colors may be associated with different users who approach and are detected in proximity of the media device and/or different user devices that are detected by the media device. Therefore, if there are four users/and our user devices detected, then: the color blue may be associated with user #1; yellow with user #2; green with user #3; and red with user #4.
  • Some users and or user devices may be indicated using alternating colors of light such as switching/flashing between red and green, blue and yellow, blue and green, etc.
  • other types of LED's may be combined with RGB LED 616 , such as a white LED, for example, to increase the number of color combinations possible.
  • proximity detection island 520 may include at least one light sensor for sensing ambient light conditions in the ENV 198 , such as ambient light sensor ALS 618 .
  • ALS 618 may be electrically coupled with circuitry CKT 620 configured to process signals from ALS 618 , such as optical sensor 609 (e.g., a PIN diode) in response to ambient light 630 incident on optical sensor 609 .
  • Signals from CKT 620 may be further processed by ADC 622 .
  • the various drivers, circuitry, and ADC's of proximity detection island 520 may be electrically coupled with a controller (e.g., a ⁇ C, a ⁇ P, an ASIC, or controller 101 of FIG.
  • a controller e.g., a ⁇ C, a ⁇ P, an ASIC, or controller 101 of FIG.
  • Proximity detection island 520 may include auditory system AUD 624 configured to generate sound or produce vibrations in response to presence detection or other signals.
  • AUD 624 may be mechanically coupled 641 with chassis 199 to cause chassis 199 to vibrate or make sound in response to presence detection or other signals.
  • AUD 624 may use SPK 160 to generate sound or vibration.
  • AUD 624 may use a vibration motor, such as the type used in smartphones to cause vibration when a phone call or notification is received.
  • AUD 624 may use a piezoelectric film that deforms in response to an AC or DC signal applied to the film, the deformation generating sound and/or vibration.
  • AUD 624 may be connected with or mechanically coupled with one or more of the control elements and/or one or more of the proximity detection islands 520 depicted in FIG. 5 to provide haptic and/or tactile feedback.
  • media may generate sound (e.g., from SPK 160 ) in a rich variety of tones and volume levels to convey information and/or media device status to the user.
  • a tone and volume level may be used to indicate the power status of the media device 100 , such as available charge in BAT 135 of power system 111 .
  • the volume of the tone may be louder when BAT 135 is fully charged and lower for reduced levels of charge in BAT 135 .
  • Other tones and volume levels may be used to indicate the media device 100 is ready to receive input from the user or user device, the media device 100 is in wireless communications with a WiFi router or network, cellular service, broadband service, ad hoc WiFi network, other BT enabled devices, for example.
  • Proximity detection island 520 may be configured to detect presence of a user 201 (or other person) that enters 671 an environment 198 the media device 100 is positioned in.
  • entry 671 by user 201 may include a hand 601 h or other portion of the user 201 body passing within optical detection range of proximity detection island 520 , such as hand 601 h passing over 672 the proximity detection island 520 , for example.
  • IR radiation 603 from IRLED 603 exiting through portal 652 reflects off hand 601 h and the reflected IR radiation 607 enters portal 652 and is incident on IR detector 605 causing a signal to be generated by ADC 612 , the signal being indicative of presence being detected.
  • RGB LED 616 may be used to generate one or more colors of light that indicate to user 201 that the user's presence has been detected and the media device is ready to take some action based on that detection.
  • the action taken will be application specific and may depend on actions the user 201 programmed into CFG 125 using APP 225 , for example.
  • the action taken and/or the colors emitted by RGB LED 616 may depend on the presence and/or detection of a user device 210 in conjunction with or instead of detection of presence of user 201 (e.g., RF 565 from device 210 by RF 107 ).
  • proximity detection island 520 may optionally include ambient light sensor ALS 618 configured to detect ambient light 630 present in ENV 198 such as a variety of ambient light sources including but not limited to natural light sources such as sunny ambient 631 , partially cloudy ambient 633 , inclement weather ambient 634 , cloudy ambient 635 , and night ambient 636 , and artificial light ambient 632 (e.g., electronic light sources).
  • ALS 618 may work in conjunction with IRLED 610 and/or IR detector 605 to compensate for or reduce errors in presence detection that are impacted by ambient light 630 , such as IR background noise caused by IR radiation from 632 or 631 , for example.
  • IR background noise may reduce a signal-to-noise ratio of IR detector 605 and cause false presence detection signals to be generated by ADC 612 .
  • ALS 618 may be used to detect low ambient light 630 condition such as moonlight from 636 or a darkened room (e.g., light 632 is off), and generate a signal consistent with the low ambient light 630 condition that is used to control operation of proximity detection island 520 and/or other systems in media device 100 .
  • low ambient light 630 condition such as moonlight from 636 or a darkened room (e.g., light 632 is off)
  • RGB LED 616 may emit light 617 at a reduced intensity to prevent the user 201 from being startled or blinded by the light 617 .
  • AUD 624 may be reduced in volume or vibration magnitude or may be muted.
  • audible notifications e.g., speech or music from SPK 160
  • from media device 100 may be reduced in volume or muted under low light or no light conditions (see FIG. 9 ).
  • Structure 650 may be electrically coupled 681 with capacitive touch circuitry 680 such that structure 650 is operative as a capacitive touch switch that generates a signal when a user (e.g., hand 601 h ) touches a portion of structure 650 .
  • Capacitive touch circuitry 680 may communicate 682 a signal to other systems in media device 100 (e.g., I/O 105 ) that process the signal to determine that the structure 650 has been touched and initiate an action based on the signal.
  • a user's touch of structure 650 may trigger driver 614 to activate RGB LED 616 to emit light 617 to acknowledge the touch has been received and processed by media device 100 .
  • FIG. 7 where top plan views of different examples of proximity detection island 520 configurations are depicted.
  • the various example configurations and shapes are depicted as positioned on top surface 199 t of chassis 199
  • the present application is not so limited and proximity detection islands 520 may be positioned on other surfaces/portions of media device 100 and may have shapes different than that depicted.
  • media device 100 may include more or fewer proximity detection islands 520 than depicted in FIG. 7 and the proximity detection islands 520 need not be symmetrically positioned relative to one another.
  • Actual shapes of the proximity detection islands 520 may be application specific and may be based on esthetic considerations.
  • Configuration 702 depicts five rectangular shaped proximity detection islands 520 positioned on top surface 199 t with four positioned proximate to four corners of the top surface 199 t and one proximately centered on top surface 199 t .
  • Configuration 704 depicts three circle shaped proximity detection islands 520 proximately positioned at the left, right, and center of top surface 199 t .
  • Configuration 706 depicts four hexagon shaped proximity detection islands 520 proximately positioned at the left, right, and two at the center of top surface 199 t .
  • configuration 708 depicts two triangle shaped proximity detection islands 520 proximately positioned at the left, right of top surface 199 t .
  • Proximity detection islands 520 may be configured to operate independently of one another, or in cooperation with one another.
  • Each proximity detection island 520 may be designed to have a coverage pattern configured to detect presence of user 201 when the user 201 or portion of the user body (e.g., hand 801 h ) enters the coverage pattern.
  • the coverage pattern may be semicircular 810 or circular 830 , for example.
  • Semicircular 810 coverage pattern may extend outward a distance R1 (e.g., approximately 1.5 meters) from proximity detection island 520 and may span a distance D1 about a center 871 of proximity detection island 520 .
  • Semicircular 810 coverage patterns of the four proximity detection islands 520 may not overlap one another such that there may be a coverage gap X1 and Y1 between the adjacent coverage patterns 810 .
  • Entry 825 of hand 801 h or entry 820 of user 201 may cause one or more of the proximity detection islands 520 to indicate 840 that a presence has been detected, by emitting a color of light from RGB LED 616 , for example.
  • the coverage pattern may be circular 830 and cover a 360 degree radius 870 about a center point 871 of proximity detection island 520 .
  • Circular 830 coverage pattern 830 may or may not overlap the circular 830 pattern of the other proximity detection islands 520 .
  • FIG. 8C depicts a front view 800 b of media device 100 and a coverage pattern 860 that has an angular profile Q about center point 871 .
  • Hand 801 h entering 825 into the coverage pattern 860 is detected by proximity detection island 520 and detection of hand 810 triggers light 840 being generate by RGB LED 616 of proximity detection island 520 .
  • Detection of hand 810 may also cause information “Info” to be displayed on DISP 180 and/or sound 845 to be generated by SPK 160 .
  • a side view 800 c of media device 100 is depicted with proximity detection island 520 having angular profile a about center point 871 for a coverage pattern 880 .
  • Hand 801 h entering 825 into the coverage pattern 880 is detected by proximity detection island 520 and detection of hand 810 triggers light 840 being generate by RGB LED 616 of proximity detection island 520 and AUD 624 generating vibration 847 .
  • FIG. 9 where a top plan view 900 of media device 100 depicts four proximity detection islands 520 denoted as I1, I2, I3, and I4. Furthermore, control elements 503 - 512 are depicted on top surface 199 t .
  • hand 901 h enters into proximity detection range of at least proximity detection island I1 and triggers generation of light ( 917 a-d) from one or more of the islands (I1, I2, I3, I4) such as light 617 from RGB LED 616 of FIG. 6 , for example.
  • Presence detection by proximity detection island I1 may cause a variety of response from media device 100 including but not limited to signaling that presence has been detected using light ( 917 a-d), generating sound 845 from SPK 160 , vibration 847 , displaying info 840 on DISP 180 , capturing and acting on content C from user device 220 , establishing wireless communications 126 with user device 220 or other wireless device (e.g., a wireless router), just to name a few.
  • Presence detection by proximity detection island I1 may cause media device 100 to notify user 901 that his/her presence has been detected and the media device is ready to receive input or some other action from user 901 .
  • Input and/or action from user 901 may comprise user 901 actuating one of the control elements 503 - 512 , touching or selecting an icon displayed on DISP 180 , issuing a verbal command or speech detected by MIC 170 .
  • media device 100 may emit light 917 c from proximity detection island I3. If the user device 220 is present and also detected by media device 100 (e.g., via RF signals 126 and/or 563 ), then the media device 100 may indicate that presence of the user device 220 is detected and may take one or more actions based on detecting presence of the user device 220 . If user device 220 is one that is recognized by media device 100 , then light 917 c from proximity detection island I3 may be emitted with a specific color assigned to the user device 220 , such as green for example.
  • a specific color assigned to the user device 220 such as green for example.
  • Recognition of user device 220 may occur due to the user device 220 having been previously BT paired with media device 100 , user device 220 having a wireless identifier such as a MAC address or SSID stored in or pre-registered in media device 100 or in a wireless network (e.g., a wireless router) the media device 100 and user device 220 are in wireless communications with, for example.
  • DISP 180 may display info 840 consistent with recognition of user device 220 and may display via a GUI or the like, icons or menu selections for the user 201 to choose from, such as an icon to offer the user 201 a choice to transfer content C from user device 220 to the media device 100 , to switch from BT wireless communication to WiFi wireless communication, for example.
  • CFG 125 may automatically transfer the phone conversation from user device 220 to the media device 100 such that MIC 170 and SPK 160 are enabled so that media device 100 serves as a speaker phone or conference call phone and media device 100 handles the content C of the phone call.
  • CFG 125 or other programming of media device 100 may operate to offer the user 201 the option of transferring the content C by displaying the offer on DISP 180 or via one of the control elements 503 - 512 .
  • control element 509 may blink (e.g., via backlight) to indicate to user 201 that actuating control element 509 will cause content C to be transferred from user device 220 to media device 100 .
  • control elements 503 - 512 may correspond to menu selections displayed on DISP 180 and/or a display on the user device 220 .
  • control elements 512 may correspond to six icons on DISP 180 (see 512 ′ in FIG. 8 ) and user 201 may actuate one of the control elements 512 to initiate whatever action is associated with the corresponding icon on DISP 180 , such as selecting a playlist for media to be played back on media device 100 .
  • the user 201 may select one of the icons 512 ′ on DISP 180 to effectuate the action.
  • content C comprises an alarm, task, or calendar event the user 201 has set in the user device 220 , that content C may be automatically transferred or transferred by user action using DISP 180 or control elements 503 - 512 , to media device 100 . Therefore, a wake up alarm set on user device 220 may actually be implemented on the media device 100 after the transfer, even if the user device 220 is powered down at the time the alarm is set to go off.
  • any alarm, task, or calendar event that has not been processed by the media device 100 may be transferred back to the user device 220 or updated on the user device so that still pending alarm, task, or calendar events may be processed by the user device when it is not in proximity of the media device 100 (e.g., when user 201 leaves for a business trip).
  • CFG 125 and APP 225 as described above may be used to implement and control content C handling between media device 100 and user devices.
  • control elements 503 - 512 may be implemented as capacitive touch switches. Furthermore, some or all of the control elements 503 - 512 may be backlit (e.g., using LED's, light pipes, etc.). For example, control elements 512 may be implemented as capacitive touch switches and they may optionally be backlit. In some examples, after presence is detected by one or more of the proximity detection islands (I1, I2, I3, I4), one or more of the control elements 503 - 512 may be backlit or have its back light blink or otherwise indicate to user 201 that some action is to be taken by the user 201 , such as actuating (e.g., touching) one or more of the backlit and/or blinking control elements 512 .
  • proximity detection islands may be configured to serve as capacitive touch switches or another type of switch, such that pressing, touching, or otherwise actuating one or more of the proximity detection islands (I1, I2, I3, I4) results in some action being taken by media device 100 .
  • actions taken by media device 100 subsequent to detecting presence via proximity detection islands (I1, I2, I3, I4) and/or other systems such as RF 107 , SEN 195 , MIC 170 may be determined in part on ambient light conditions as sensed by ALS 618 in proximity detection islands (I1, I2, I3, I4).
  • ambient light 630 is bright (e.g., 631 or 632 )
  • brightness of DISP 180 may be increased
  • light 917 a - d from islands may be increased
  • volume from SPK 160 may be nominal or increased because the ambient light 630 conditions are consistent with waking hours were light intensity and volume may not be a distraction to user 201 .
  • ambient light 630 is dim or dark (e.g., 636 )
  • brightness of DISP 180 may be decreased
  • light 917 a - d from islands may be decreased
  • volume from SPK 160 may be reduced or muted because the ambient light 630 conditions are consistent with non-waking hours were light intensity and volume may be a distraction to or startle user 201 .
  • Other media device 100 functions such as volume level, for example, may be determined based on ambient light 630 conditions (e.g., as detected by ALS 618 of island I4).
  • volume VH of SPK 160 may be higher (e.g., more bars); whereas, under low ambient light 630 conditions, volume VL of SPK 160 may be lower (e.g., fewer bars) or may be muted entirely VM. Conditions other than ambient light 630 may cause media device 100 to control volume as depicted in FIG. 9 .
  • FIG. 10 depicts one example of a flow 1000 for presence detection, notification, and media device readiness.
  • a query as to whether or not an approach is detected by one or more of the proximity detection islands (e.g., I1, I2, I3, I4) is made.
  • the query may be by controller CNTL 640 or controller 101 , for example. If one or more of the proximity detection islands have detected presence, then a YES branch is taken. If no presence is detected by one or more of the proximity detection islands, then a NO branch is taken and the flow 1000 may return to the stage 1002 to wait for one or more of the proximity detection islands to detect a presence.
  • the YES branch takes flow 1000 to a stage 1004 where a notification is executed by the media device 100 using light, sound, or vibration to notify a user that presence has been detected, for example, using one or more colors of light (e.g., from RGB LED's 616 ) and/or an auditory cue (e.g., from SPK 160 , vibration from 847 , or from a passive radiator used as one of the SPK 160 ).
  • the media device 100 indicates that it is ready to receive input from a user and/or user device (e.g., user 201 or a user device 220 via RF 107 ).
  • a query is made as to whether or not an input is received from a user.
  • a YES branch is taken to a stage 1010 where the media device 100 takes an appropriate action based on the type of user input received and the flow may terminate after the stage 1010 .
  • Appropriate actions taken by media device 100 will be application dependent and may be determined in whole or in part by APP 225 , CFG 125 , executable program code, hardware, etc.
  • Inputs from the user includes but is not limited to actuation of one or more of the control elements 503 - 512 , touching an icon or other area of DISP 180 , issuing a spoken command or speech detected by MIC 170 , taking an action on user device 220 that is wirelessly communicated to media device 100 , just to name a few.
  • a NO branch is taken and the flow 1000 may continue at a stage 1012 where flow 1000 may enter into a wait period of predetermined time (e.g., of approximately 15 seconds or one minute, etc.). If a user input is received before the wait period is over, then a NO branch may be taken to the stage 1010 . If the wait period is over, then a YES branch may be taken and flow 1000 may resume at the stage 1002 .
  • a wait period of predetermined time e.g., of approximately 15 seconds or one minute, etc.
  • FIG. 11 depicts another example of a flow 1100 for presence detection, notification, and media device readiness.
  • a query as to whether an approach is detected by one or more of the proximity detection islands is made. If one or more of the proximity detection islands have detected presence, then a YES branch is taken. If no presence is detected by one or more of the proximity detection islands, then a NO branch is taken and the flow 1100 may return to the stage 1102 to wait for one or more of the proximity detection islands to detect a presence.
  • the YES branch takes flow 1100 to a stage 1104 where a query is made as to whether or not ambient light (e.g., ambient light 630 as detected by ALS 618 of FIG.
  • ambient light e.g., ambient light 630 as detected by ALS 618 of FIG.
  • any notification by media device 100 in response to detecting presence at the stage 1102 is modified.
  • One or more of light, sound, or vibration may be used by media device 100 to indicate to a user that its presence has been detected. The light, sound, or vibration are altered to comport with the ambient light conditions, such as described above in regard to ambient light 630 in FIG. 9 , for example.
  • notification of presence being detected occurs using one or more of light, sound, or vibration without modification.
  • the media device 100 indicates that it is ready to receive input from a user and/or user device (e.g., user 201 or a user device 220 via RF 107 ).
  • a query is made as to whether or not an input is received from a user. If an input is received from the user and/or user device, then a YES branch is taken to a stage 1114 where the media device 100 takes an appropriate action based on the type of user input received and the flow may terminate after the stage 1114 .
  • a NO branch is taken and the flow 1110 may continue at a stage 1116 where flow 1100 may enter into a wait period of predetermined time (e.g., of approximately 15 seconds or one minute, etc.). If a user input is received before the wait period is over, then a NO branch may be taken to the stage 1114 . If the wait period is over, then a YES branch may be taken and flow 1100 may resume at the stage 1102 . Actions taken at the stage 1114 may include those described above in reference to FIG. 10 .
  • FIG. 12 depicts yet another example of a flow 1200 for presence detection, notification, and media device readiness.
  • a query as to whether an approach is detected by one or more of the proximity detection islands is made. If one or more of the proximity detection islands have detected presence, then a YES branch is taken. If no presence is detected by one or more of the proximity detection islands, then a NO branch is taken and the flow 1200 may return to the stage 1202 to wait for one or more of the proximity detection islands to detect a presence.
  • the proximity detection islands e.g., I1, I2, I3, I4
  • the YES branch takes flow 1200 to a stage 1204 where a query is made as to whether or not detection of RF (e.g., by RF 107 using antenna 124 ) is a factor to be taken into consideration in the media devices response to having detected a presence at the stage 1202 . If RF detection is not a factor, then a NO branch is taken and the flow 1200 continues to a stage 1206 . If RF detection is a factor, then a YES branch is taken and flow 1200 continues at a stage 1208 where any notification by media device 100 in response to detecting presence at the stage 1202 is modified. One or more of light, sound, or vibration may be used by media device 100 to indicate to a user that its presence has been detected.
  • the light, sound, or vibration are altered to comport with the detection of RF (e.g., from a user device 220 ), such as described above in regards to user device 220 in FIG. 9 , for example.
  • notification of presence being detected occurs using one or more of light, sound, or vibration without modification.
  • the media device 100 indicates that it is ready to receive input from a user and/or user device (e.g., user 201 or a user device 220 via RF 107 ).
  • a query is made as to whether or not an input is received from a user.
  • a YES branch is taken to a stage 1214 where the media device 100 takes an appropriate action based on the type of user input received and the flow may terminate after the stage 1214 . If no input is received from the user and/or user device, then a NO branch is taken and the flow 1200 may continue at a stage 1216 where flow 1200 may enter into a wait period of predetermined time (e.g., of approximately 15 seconds or one minute, etc.). If a user input is received before the wait period is over, then a NO branch may be taken to the stage 1214 . If the wait period is over, then a YES branch may be taken and flow 1200 may resume at the stage 1202 . Actions taken at the stage 1214 may include those described above in reference to FIGS. 9 and 10 .
  • FIG. 13 depicts one example 1300 of presence detection using proximity detection islands and/or other systems responsive to wireless detection of different users (e.g., hands 1300 a - d ) and/or different user devices (e.g., 220 a - 220 d ).
  • different users e.g., hands 1300 a - d
  • different user devices e.g., 220 a - 220 d
  • FIG. 13 depicts one example 1300 of presence detection using proximity detection islands and/or other systems responsive to wireless detection of different users (e.g., hands 1300 a - d ) and/or different user devices (e.g., 220 a - 220 d ).
  • four users denoted by hands 1300 a - d and their respective user devices 220 a - 220 b enter 925 proximity detection range of one or more of the proximity detection islands (I1, I2, I3, I4).
  • four users and four user devices are depicted, there may be more
  • Detection of user devices 220 a - 220 b may be through wireless means such as RF 107 (e.g., via antenna 124 / 129 ) and its various transceivers wirelessly communicating 126 or wirelessly detecting RF 563 from those user devices.
  • RF 107 e.g., via antenna 124 / 129
  • Detection of user devices 220 a - 220 b may be through wireless means such as RF 107 (e.g., via antenna 124 / 129 ) and its various transceivers wirelessly communicating 126 or wirelessly detecting RF 563 from those user devices.
  • RF 107 e.g., via antenna 124 / 129
  • Island I2 notifies user via light 1317 b that his/her presence has been detected.
  • User device 220 b may be carried by the user at the same time or at approximately the same time as the user's presence is detected by island I2.
  • RF 107 may detect RF 563 , may attempt to wirelessly connect 126 , or be in wireless 126 communications with user device 220 b . Accordingly, notifications and actions described above in regards to flow 1200 of FIG. 12 may occur in media device 100 in response to detecting presence 597 at or near the same time as detecting RF from a user device. Media device 100 may emit sound 1345 , vibrate 847 , display information info on DISP 180 , generate light 1317 a - 1317 d , await actuation of one or more of the control elements 503 - 512 , or other action(s), for example.
  • hands 1300 a , 1300 c , and 1300 d may be detected 597 by one or more of the proximity detection islands (I1, I2, I3, I4) along with RF 563 from user devices 220 a , 220 c , and 220 d being detected by RF 107 .
  • Media device 100 may take appropriate action(s) and make appropriate notification(s) as described herein in response to proximity detection and RF detection occurring in close time proximity to one another, simultaneously, nearly simultaneously, or in some sequence.
  • a range for RF transmissions may typically be greater than a detection range for the proximity detection islands (I1, I2, I3, I4)
  • the RF signatures or signals of user device 220 a - d may be detected by RF 107 before the proximity detection islands (I1, I2, I3, I4) detect presence of the users 1300 a - d .
  • RF 107 may detect RF 563 before the user device emitting RF 563 is approximately 10 meters or more away from media device 100 (e.g., for BT transmissions) or much more than 10 meters away for other wireless technologies (e.g., for WiFi transmissions). Therefore, in some examples, RF 107 will detect RF signals prior to proximity detection islands (I1, I2, I3, I4) detecting presence 597 .
  • Users devices 220 a - 220 d may be pre-registered or otherwise associated or known by media device 100 (e.g., via CFG 125 or other) and the actions taken and notifications given by the media device 100 may depended on and may be different for each of the user devices 220 a - 220 d .
  • media device 100 may establish or re-establish BT pairing (e.g., via BT 120 in RF 107 ) with 220 a and content C on 220 a (e.g., a phone conversation) may be transferred to media device 100 for handling via SPK 160 and MIC 170 .
  • CFG 125 and/or APP 225 on 220 a may affect how media device and user device 220 a operate post detection.
  • post detection 597 & 563 and notification for user device 220 d may result in content C (e.g., music from MP3 files) on 220 d being played back 1345 on media device 100 .
  • Control elements 503 - 512 may be activated (if not already activated) to play/pause ( 506 ), fast forward ( 508 ), fast reverse ( 504 ), increase volume ( 503 ), decrease volume ( 507 ), or mute volume ( 509 ).
  • Control elements 512 may be used to select among various play lists or other media on user device 220 d.
  • content C on user device 220 c may, post detection and notification, be displayed on DISP 180 .
  • a web page that was currently being browsed on 220 c may be transferred to media device 100 for viewing and browsing, and a data payload associated with the browsing may also be transferred to media device 100 .
  • content C comprises a video
  • the display and playback functions of the video may be transferred to media device 100 for playback and control, as well as the data payload for the video.
  • media device 100 may be transferred back in part or whole to the user devices depicted, when the user is no longer detectable via islands to proximity detection islands (I1, I2, I3, I4) or other systems of media device 100 , by user command, or by user actuating one of the control elements 503 - 512 or an icon or the like on DISP 180 , for example.
  • islands to proximity detection islands I1, I2, I3, I4
  • other systems of media device 100 by user command, or by user actuating one of the control elements 503 - 512 or an icon or the like on DISP 180 , for example.
  • FIG. 14 depicts one example 1400 of proximity detection islands associated with specific device functions.
  • functions that may be assigned to or fixed to a proximity detection island include but are not limited to “Set Up” of media device 100 , “BT Paring” between media device 100 and one or more BT equipped devices, “Shut-Off” of media device 100 (e.g., power off or placing media device 100 in a standby mode, a low power consumption mode, or a sleep mode), and “Content” being handled by media device 100 , such as the last media filed that was played on, the last buffered channel, the last playlist that was being accessed by, or the last Internet site or stream being handled by media device 100 .
  • One or more of proximity detection islands may serve as indicators for the functions associated with them or may serve to actuate those functions by pressing or touching a surface of the island (e.g., as a switch or capacitive touch switch or button, see FIG. 6 ).
  • a finger of hand 1400 h may touch structure 650 of island I2 to activate the “BT Pairing” between the media device 100 and user device 220 , the touch activating the capacitive touch function of island I2 (e.g., causing island I2 to serve as a switch).
  • Island I2 may emit light 1417 b to acknowledge the touch by hand 1400 h .
  • CFG 125 and/or APP 225 may be used to assign and re-assign functions to one or more of the proximity detection islands (I1, I2, I3, I4) and the functions assigned and the proximity islands they are assigned to may be user dependent and/or user device dependent. As another example, pressing or touching island I4 may turn power off to the media device 100 , or may place media device 100 in a low power, standby, or sleep mode.
  • one or more of the control elements 503 - 512 or an icon or the like on DISP 180 may be actuated or selected by a user in connection with one of the functions assigned to proximity detection islands (I1, I2, I3, I4).
  • control element 512 that is nearest 1427 to island I2 may be actuated by the user.
  • proximity detection islands I1, I2, I3, I4 may be associated with different users whose presence has been detected by one or more of the islands. For example, if proximity of four users (U1, U2, U3, U4) has been detected by any of the islands, then U1 may be associated with I4, U2 with I1, U3 with I2, and U4 with I3. Association with an island may be used to provide notifications to the user, such as using light from RGB LED 616 to notify the user of status (e.g., BT pairing status) or other information.
  • FIG. 15 depicts one example 1500 of content handling from a user device subsequent to proximity detection by islands 520 and/or wireless systems of media device 100 .
  • User 1500 h is detected 1540 by proximity detection island 520 which emits light 1517 , sound 1545 , vibration 847 , and display of information info on DISP 180 to indicate that media device 100 has detected presence and is ready to receive user input.
  • User device 220 may also have been detected by a transceiver RXTX 1507 in RF 107 .
  • RXTX 1507 may represent any transceiver in RF 107 such as BT 120 , WiFi 130 , AH 140 , or other 150 .
  • Media device 100 may be wirelessly connected with user device 220 using a variety of wireless paths such as a direct wireless connection 126 between media device 100 and user device 220 , and wireless connections 1565 and 1563 via wireless router 1570 , for example.
  • Content C on user device 220 may be handled or otherwise stored or routed to media device from the user device 220 or from Cloud 1550 using a variety of wireless paths.
  • Cloud 1550 may represent the Internet, an intranet, a server farm, a download site, a music store, and application store, Cloud storage, a web site, just to name a few.
  • Information including but not limited to content C, data D, a playlist PL, a stream or streaming service S, and a URL, just to name a few.
  • Cloud 1550 may also be presently on user device or wirelessly accessible to user device 220 via wireless connections 1561 , 1563 , 1567 , 126 , 1569 , and 1565 . Some of the wireless connections may be made through wireless router 1570 or media device 100 (e.g., via WiFi 130 ).
  • content C or other information resident or accessible to user device 220 may be handled by media device 100 .
  • C comprises media files such as MP3 files
  • those files may be wirelessly accessed by media device 100 by copying the files to DS 103 (e.g., in Flash memory 145 ) thereby taking the data payload and wireless bandwidth from the user device 220 to the media device 100 .
  • Media device 100 may use it wireless systems to access 1569 or 1565 and 1567 the information from Cloud 1550 and either store the information locally in DA 103 or wirelessly access the information as it is played back or otherwise consumed or used by media device 100 .
  • APP 225 and CFG 125 may include information and executable instructions that orchestrate the handling of content between media device 100 , user device 220 , and Cloud 1550 .
  • a playlist PL on user device 220 may be located in Cloud 1550 and media files associated with music/videos in the PL may be found at URL in Cloud 1550 .
  • Media device 100 may access the media files from the location specified by the URL and wirelessly stream the media files, or media device may copy a portion of those media files to DS 103 and then playback those files from its own memory (e.g., Flash 145 ).
  • user 1500 h may be one of many users who have content to be accessed and/or handled by media device 100 .
  • Post detection, songs, play lists, content, of other information on user device 220 or from Cloud 1550 may be placed in a queue with other information of similar type.
  • the queue for songs may comprise Song 1 through Song N and songs on user device 220 that were active at the time of proximity detection may be placed in some order within the queue, such as Song 4 for being fourth in line in queue for playback on media device 100 .
  • Other information such as play lists PL 1-PL N or other content such as C 1-CN may be placed in a queue for subsequent action to be taken on the information once it has moved to the top of the queue.
  • the information on user device 220 or from Cloud 1550 may be buffered in media device 100 by storing buffered data in DS 103 .
  • FIG. 16 depicts another example of content handling from user devices subsequent to proximity detection.
  • a plurality of users 1601 a - 1601 n and their associated user device 220 are detected by media device 100 are queued into DS 103 on media device 100 for handling or are buffered BUFF into DS 103 in some order.
  • Detection of each user and or user device may be indicated with one or more different colors of light 1517 , different sounds 1545 , different vibration 847 patterns, or different info on DISP 180 .
  • buffering BUFF occurs in storage 1635 provided in Cloud 1550 .
  • users 1601 a - 1601 n have information on their respective user devices 220 that may be handled by media device 100 such as Song 1-Song N, PL 1-PL N, C 1-C N.
  • the information from the plurality of users 1601 a - 1601 n is queue and/or buffered BUFF on media device 100 and/or in Cloud 1550 , that is, media device may handle all of the information internally, in Cloud 1550 , or some combination of media device 100 and Cloud 1550 .
  • a data storage capacity of the information exceeds a storage capacity of DS 103 , then some or all of the data storage may be off loaded to Cloud 1550 (e.g., using Cloud storage or a server farm).
  • Information from users 1601 a - 1601 n may be played back or otherwise handled by media device 100 in the order in which proximity of the user was detected or in some other order such as a random order or a shuffle play order.
  • DISP 180 may have an icon RDM which may be selected for random playback.
  • FIG. 17 depicts one example of content handling from a data capable wristband or wristwatch subsequent to proximity detection by a media device.
  • a hand 1700 h of a user may comprise a user device in the form of a data capable wristband or wristwatch denoted as 1740 .
  • Wristband 1740 may include information “I” that is stored in the wristband 1740 and is wirelessly accessible using a variety of wireless connections between media device 100 , wireless router 1570 , and Cloud 1750 .
  • Media device 100 may serve as a wireless hub for wristband 1740 allowing wristband 1740 to send and retrieve information from Cloud 1750 via wireless connections between media device 100 and wireless router 1570 and/or Cloud 1750 .
  • wristband 1740 may use BT to wirelessly communicate with media device 100 and media device 100 uses its WiFi 130 to wirelessly communicate with other resources such as Cloud 1750 and router 1570 .
  • Detection 1540 of hand 1700 h and/or device 1740 may trigger the emission of light 1517 , generation of sound 1545 , vibration 847 , and display of information info on DISP 180 .
  • Information “I” included in wristband 1740 may include but is not limited to alarms A, notifications N, content C, data D, and a URL. Upon detection of proximity, any of the information “I” may be wirelessly communicated from wristband 1740 to media device 100 where the information “I” may be queued (A 1-A N; D 1-D N, N1-N n; and C 1-C N) and/or buffered BUFF as described above. In some examples, post detection, wristband 1740 may wirelessly retrieve and/or store the information “I” from the media device 100 , the Cloud 1750 , or both. As one example, if wristband 1740 includes one or more alarms A, post detection those alarms A may be handled by media device 100 .
  • a plurality of users 1801 a - 1801 n and their respective wristwatches 1740 are detected by one or more proximity detection islands 520 of media device 100 and/or or other systems such as RF 107 . Detection of each user and or device 1740 may be indicated with one or more different colors of light 1517 , different sounds 1545 , different vibration 847 patterns, or different info on DISP 180 .
  • each wristwatch 1740 includes information “I” specific to its user and as each of these users and wristwatches come into proximity and are detected, information “I” may be queued, buffered BUFF, or otherwise stored or handled by media device 100 or in Cloud 1750 .
  • data D may include exercise, nutrition, dietary data, and biometric information collected from or sensed via sensors carried by the wristwatch 1740 .
  • Data D may be transferred to media device 100 or Cloud 1750 and accessed via a URL to a web page of a user.
  • the data D may be shared among other users via their web pages.
  • some or all of users 1801 a - 1801 n may be consent to sharing their information “I” through media device 100 , Cloud 1750 , or both.
  • Users 1801 a - 1801 n may view each other's information “I” on DISP 180 or go to a URL in Cloud 1750 or the like to view each other's information “I”.
  • Information “I” that is displayer on DISP 180 may be buffered BUFF, queued (A 1-A N; D 1-D N, N1-N n; and C 1-C N), or otherwise stored on media device 100 (e.g., in DS 103 ) for each user to query as desired.
  • a non-transitory computer readable medium such as CFG 125 and/or APP 225 may be used to determine actions taken by wristwatch 1740 (e.g., via APP 225 ) and media device (e.g., via CFG 125 ).
  • one example of a flow 1900 for content C handling on a media device 100 or other location, post proximity detection includes the media device 100 accessing the content C at a stage 1902 .
  • accessing may include negotiating the necessary permissions, user names and passwords, or other tasks necessary to gain access to the content C on a user device or located elsewhere (e.g., in the Cloud, on a website, or on the Internet).
  • Accessing the content C may include wirelessly connecting with the user device or other source of the content C.
  • the media device 100 makes a determination is made as to the type of the content C, such as a media file (e.g., music, video, pictures), a web page (e.g., a URL), a file, a document (e.g., a PDF file), for example.
  • the media device 100 makes a determination as to a status of the content C. Examples of status include but are not limited to static content C (e.g., a file) and dynamic content C (e.g., a stream or a file currently being accessed or played back).
  • the media device 100 handles the content C based on its type and status from stages 1904 and 1906 .
  • media device 100 queries the user devices to see if there is additional content C to be handled by the media device 100 . If additional content exists, then a YES branch may be taken and flow 1900 may return to stage 1902 . If no additional content C is to be handled, then a NO branch may be taken and at a stage 1912 a decision to terminate previously handled content C may be made.
  • a user device may have handed over content C handling to media device 100 post proximity detection, but when the user device moves out of RF and/or proximity detection range (e.g., the user leaves with his/her user device in tow), then media device 100 may release or otherwise divorce handling of the content C. If previously handled content C does not require termination, then a NO branch may be taken and flow 1900 may end. On the other hand, if previously handled content C requires termination, then a YES branch may be taken to a stage 1914 were the previously handled content C is released by the media device 100 .
  • Release by media device 100 includes but is not limited to wirelessly transferring the content C back to the user device or other location, deleting the content C from memory in the media device 100 or other location, saving, writing or redirecting the content C to a location such as /dev/null or a waste basket/trash can, halting streaming or playback of the content C, storing the content C to a temporary location, just to name a few.
  • FIG. 20 depicts one example of a flow 2000 for storing, recording, and queuing content C on a media device 100 or other location post proximity detection.
  • media device 100 may determine a size (e.g., file size) of the content C at a stage 2002 . The size determination may be made in order for the media device 100 to determine if the media device 100 has the memory resources to handle and/or store the content C. If the media device 100 cannot accommodate content C due to size, then media device 100 may select another source for the content C or access the content from the user device or other location where it is stored.
  • the media device 100 determines whether or not the content C is dynamic.
  • Examples of dynamic content C include but are not limited to content C on a user device that is currently being accessed or played back on the user device.
  • the dynamic content C may reside on the user device or may be accessed from another location (e.g., the Cloud or Internet). If the content C is not dynamic (e.g., is static such as file), then a NO branch may be taken to a stage 2010 where the media device 100 selects an appropriate location to store content C based on its size from the stage 2002 .
  • Examples of appropriate locations include but are not limited to a user device, the Cloud, the Internet, an intranet, network attached storage (NAS), a server, and DS 103 of media device 100 (e.g., in Flash memory 145 ).
  • media device 100 may include a memory card slot for a SD card, microSD card, Memory Stick, SSD, CF card, or the like, or a USB connector that will accommodate a USB thumb drive or USB hard drive, and those memory devices may comprise an appropriate location to store content C.
  • the content C is stored to the selected location. If the content C is dynamic, then a YES branch may be taken to a stage 2006 where memory device 100 selects an appropriate location to record the dynamic content C based on the size of the content C. Appropriate locations include but are not limited to those described above for the stage 2010 .
  • the media device 100 records the dynamic content to the selected location.
  • the selected location may be a buffer such as BUFF described above.
  • the media device 100 may playback other content C (e.g., an mp3 or mpeg file) while recording the content C to the selected location.
  • other content C e.g., an mp3 or mpeg file
  • media device 100 may begin to handle the content C from the various user devices as described in reference to FIGS. 19 and 20 .
  • users U1 and U3 have static content C to be handled by media device 100 and user U2 has dynamic content C.
  • media device 100 begins to record and store the dynamic content C from U2 (e.g., U2 was streaming video); however, the recording is not complete and media device 100 handles the content C from U1 next, followed by the content C of U3.
  • Content C from U1 comprises a playlist for songs stored in the Cloud and C from U3 comprises alarms A, notifications N, and data D from a data capable wristband/wristwatch.
  • Media device 100 handles and stores the content C from U3 in its internal memory (e.g., DS 103 ) and queues U3 content first for display, playback, or other on media device 100 .
  • Media device 100 accesses the songs from U1's playlist from the Cloud and queues U1 next in the queue behind U3 for playback on the SPK 160 of media device 100 .
  • the recording is complete on U2's dynamic content C and the video stream is recorded on NAS and media device 100 has accesses to the NAS via WiFi 130 .
  • U2 is queued behind U1 for playback using DISP 180 and SPK 160 of media device 100 .
  • the media device may display U3's content C on DISP 180 while playing back U1's mp3 songs over SPK 160 , even thou U1 is behind U3 in the queue.
  • U1's content is primarily played back using the media device's 100 audio systems (e.g., SPK 160 ) and U3's content C is primarily visual and is displayed using the media device's 100 video systems (e.g., DISP 180 ).
  • Servicing content C from U3 and U1 at the same time may mean temporarily bumping visual display of U1's playlist on DISP 180 to display U3's content C.
  • FIG. 21 where one example 2100 of a media device 100 handling, storing, queuing, and taking action on content from a plurality of user devices is depicted.
  • four users denoted by hands 2100 a - d move within proximity detection range of islands 520 , are detected 2140 , and the users are notified 2117 of the detection, as described above.
  • the four users 2100 a - d each have their respective user devices UD1-UD4 having content C1-C4.
  • the order in which the user devices are discovered by the media device is UD2; UD4; UD3; and UD1 and the content C on those devices are queued in the same order as the detection as denoted by C2; C4; C3; and C1 in diagram 2180 .
  • the media device 100 , the user devices UD1-UD4, wireless router 2170 , and Cloud 2150 are all able to wirelessly communicate with one another as denoted by 2167 .
  • C2 comprises a playlist and songs, is static, and each song is stored in a mp3 file in memory internal to UD2.
  • media device queues C2 first and stores C2 in a SDHC card 2121 such that the playlist and mp3 files now reside in SDHC 2121 .
  • C1 and C4 both comprise information stored in a data capable wristband/wristwatch.
  • C1 and C4 are static content.
  • Media device queues C4 behind C2, and stores C4 in Cloud 2150 .
  • C3 comprises dynamic content in the form of an audio book being played back on UD3 at the time it was detected by media device 100 .
  • C3 is queued behind C4 and is recorded on NAS 2122 for later playback on media device 100 .
  • C1 is queued behind C3 and is stored in Cloud 2150 .
  • the queuing order need not be the order in which content C is played back or otherwise acted on by media device 100 .
  • media device has ordered action to be taken on the queued content in the order of C1 and C4 first, C2 second and C3 third.
  • C3 may be third in order because it may still be recording to NAS 2122 .
  • the information comprising C1 and C4 may be quickly displayed on DISP 180 for its respective users to review.
  • the size of data represented by C1 and C4 may be much smaller than that of C2 and C3. Therefore, while C3 is recording to NAS 2122 and C2 is being copied from UD2 into SDHC 2121 , action is taken to display C1 and C4 on DISP 180 .
  • Action is then taken on C2 and a portion of the playlist from C2 is displayed on DISP 180 with the song currently being played highlighted in that list of songs.
  • the music for the song currently being played is output on SPK 160 .
  • the recording of C3 is completed and DISP 180 displays the title, author, current chapter, and publisher of the audio book.
  • Action on C3 may be put on hold pending C2 completing playback of the songs stored in SDHC 2121 .
  • media device 100 handled the various types of content C and operated on one type of content (recording C3) while other content (C1 & C4, C2) were being acted on, such as displaying C1 and C4 or playback of mp3 files from C2.
  • C3 if UD2 moves 2133 out of RF range of media device 100 , C2 may be released from the queue and action on C2 may stop and the next item of content in the queue is acted on (e.g., C3).
  • FIG. 21 is a non-limiting example and nothing precludes one of the users taking action to change the queuing order or the order in which the media device acts on queued content.
  • CFG 125 and/or APP 225 may be used to determine content queuing and an order in which queued content is acted on by media device 100 .
  • One of the users may have super user capability (e.g., via that user's APP 225 and/or CFG 125 ) that allows the super user to override or otherwise control content handling on media device 100 .
  • FIG. 22 depicts another example 2200 of a media device handling, storing, queuing, and taking action on content from a plurality of user devices.
  • a plurality of users 2200 a - 2200 n have approached media device 100 and have be detected by a proximity island 520 .
  • a plurality of user devices UDa-UDn, having content Ca-Cn, are in wireless communications 2167 as described above.
  • the content Ca-Cn from the user devices is queued in the order the user devices were detected by media device 100 .
  • Content Ca-Cn may be stored and/or accessed by media device 100 from any location that may be directly accessed or wirelessly accessed by media device 100 such as in DS 103 (directly accessed), NAS 2122 , the user devices UDa-UDn, the Cloud 2250 , etc.
  • Media device 100 may take action on the queued content in any order including but not limited to random order, the order in which it is queued, or commanded order, just to name a few.
  • Media device 100 may be configured to operate in a “party mode” where each of the users 2200 a - 2200 n in proximity of the media device 100 desires to have their content played back on the media device 100 .
  • Media device 100 may harvest all of the content and then act on it by randomly playing back content from Ca-Cn, allowing one of the users to control playback, like a DJ, or allowing a super user UDM to control playback order and content out of Ca-Cn.
  • One of the users may touch or otherwise actuate one of the control elements 503 - 512 and/or one of the proximity detector islands 520 or an icon on DISP 180 to have their content acted on by media device 100 .
  • Content in Ca-Cn may be released by media device 100 if the user device associated with that content moves out of RF range of the media device 100 .
  • a flow 2300 for recording user content on a media device while the media device handles current content is depicted.
  • entry of a user e.g., hand of a user
  • the user is notified that media device 100 has detected the user's presence (e.g., using light, sound, vibration, etc.).
  • media device 100 may use RF system 107 to detect RF signals being transmitted by a user device (e.g., 220 ) as described above.
  • the media device 100 and the user device wirelessly connect with each other (e.g., using WiFi 130 or BT 120 ).
  • content currently being handled by media device 100 is displayed on the media device 100 (e.g., DISP 180 ) or on a display of the user device, or both, for example.
  • APP 225 or other software and/or hardware may be used to display the current content being handled on media device 100 on the user device.
  • a request from the user device to the media device 100 for the media device 100 to handle user content from the user device is received.
  • the media device 100 harvests the user content from the user device (e.g., wirelessly copies, streams, or otherwise accesses the user content).
  • the user content may reside on the user device or may be located elsewhere at a location the media device 100 or user device may access, such as the Cloud, the Internet, an intranet, NAS, or other, for example.
  • the media device 100 begins recording the user content while continuing playback of the content currently being handled by the media device 100 .
  • the media device 100 based on a size of the user content (e.g., file size in MB or GB) may record the user content to memory internal to the media device 100 or to a location external to the media device 100 (e.g., NAS, the Cloud, a server, the Internet). Content that was being handled by the media device 100 continues with little or no interruption while the user content is recorded.
  • a determination may be made to queue the user content relative to the current content being handled by the media device 100 . If no queuing action is to be taken, then a NO branch may be taken and the flow 2300 may terminate. However, if the user content is to be queued, then a YES branch may be taken to a stage 2322 where a queuing action is applied to the user content.
  • Queuing action may mean any action taken by the media device 100 (e.g., via controller 101 , CFG 125 , hardware, or software) and/or user device (e.g., via APP 225 ) that affects the queuing of content on the media device 100 .
  • Queuing action may include but is not limited: to waiting for the user content to complete recording and then placing the user content in a queuing order relative to other content already queued on the media device 100 (e.g., at the back of the queue); bumping content presently at the front of the queue once the user content has completed recording and beginning playback of the recorded user content; placing the user content behind the content currently being handled by the media device 100 such that the user content will be next in line for playback; moving the user content to the front of the queue; randomly placing the user content in the queue; allowing the user of the user device to control the queuing of the user content; allowing a DJ or other user to control the queuing of the user content; and allowing each user that is detected by the proximity detection islands, have one or more items in their content harvested and pushed to the top of the queue or placed next in line in the queue; and placing the user content in a queue deck with other content, shuffling the deck and playing on of the items of content from the deck, and re-shuffling the
  • example 2400 of queuing action for user content in a queue of a media player is depicted.
  • example 2400 there are at least seven users U1-U7 and at least seven user devices UD1-UD7.
  • all seven users have approached media device 100 , have been detected 2140 and notified 2117 by proximity island 520 , and all user devices have been detected and wirelessly connected with media device 100 .
  • DISP 180 is displaying the queued order of the playlist as Song for UD1 currently being played back because it is underlined (e.g., over SPK 160 ), with Songs for UD2 and UD3 being next in the playlist.
  • User content for UD1-UD3 may reside in DS 103 or other location such as NAS 2122 or Cloud 2250 .
  • the Action for the queuing order in queue 2480 is “Play In Order”, so C1 is first, C2 is second, and C3 is third in the playback order as displayed on DISP 180 .
  • UD7 also wirelessly connected and had its user content C7 harvested by media device 100 .
  • Media device 100 begins the process of recording 2490 the content into DS 103 (e.g., into Flash 145 ).
  • other user devices may also have their user content harvested.
  • intervening user content will be placed ahead of C7 until C7 has completed 2492 recording 2492 .
  • C7 Upon completion of recording, C7 is positioned 2482 in the playlist below some already queued user content and ahead or other user content lower in the queue. In other examples, C7 may be queued in the order it was presented to the media device 100 and the media device 100 begins the recording 2490 process and allows C7 to be played back when it moves to the top of queue, but if C7 has not completed recording 2492 , then media device 100 begins the playback 2493 of C7 from a buffer BUFF 2421 where a portion of recorded C7 is stored. The playback from BUFF 2421 may continue until the recording catches up with the buffered content or is completed 2492 .
  • one of the users or user devices may have super user (e.g., UM) or other form of override authority and that user may order the queue to their liking and control the order of playback of user content.
  • Queue 2480 and/or the user content being queued need not reside in memory internal to media device 100 and may be located externally in NAS 2122 , a USB Hard Drive, Cloud 2250 , and a server, just to name a few.
  • media device 100 may delete or bump user content from queue 2480 if the wireless connection 2167 between media device 100 and the user device is broken or interrupted for a predetermined amount of time, such as two minutes, for example.
  • the “Play In Order” example depicted is a non-limiting example and one skilled in the art will appreciate that the queuing may be ordered in a variety of ways and may be determined by executable program code fixed in a non-transitory medium, such as in DS 103 , Flash 145 , CFG 125 , and APP 225 , just to name a few. Therefore, controller 101 or a controller in a user device may execute the program code that determines and controls queuing of user content on the media device 100 .
  • the wireless communications link may be a previously established link via a paring process (e.g., BT paring), a wireless network link (e.g., WiFi, AdHoc WiFi, WiMAX, Cellular, NFC) or other.
  • the wireless communications link may be established over a WiFi network that the wireless user device and the wireless media device are using and/or have previously used, for example.
  • the wireless user device and the wireless media device may have previously been paired with one another, such as through a BT link or the like.
  • the wireless communications link may be established between one or more wireless user devices and one or more wireless media devices.
  • content may be accessed by the wireless user device (e.g., 2701 of FIG. 27A ) and that content may reside locally with the wireless user device, reside external to the wireless user device or both.
  • content C in the form of media e.g., music, images, video, movies, etc.
  • memory internal to the wireless user device e.g., Flash memory or other non-volatile memory
  • content C may reside external to the wireless user device in a location such as the Cloud, the Internet, network attached storage (NAS), a web site, a web page, etc.
  • Data source 2503 broadly includes without limitation any source of data that may be accessed for the content at the stage 2504 , using one or more communications links including but not limited to wireless links and wired links.
  • data about the content C that was accessed may be accessed by the wireless user device.
  • the descriptive data DD may comprise metadata, descriptive metadata, tag, or other information/data related to, is about, that describes attributes, or is associated with the content C that was accessed.
  • the DD may already be included or otherwise associated with the content C and the stage 2506 may be skipped, for that particular datum of accessed content C, for example.
  • the DD may be accessed internal to the wireless user device, external to the wireless user device or both.
  • Data source 2505 broadly includes without limitation any source of the DD that may be accessed at the stage 2506 , using one or more communications links including but not limited to wireless links and wired links.
  • the DD may comprise information about the music such as song title, artists, sidemen, producer, cover art, lyrics, images, photographs, liner notes, playing time, time index, location or address of the file or data that comprises the music (e.g., on NAS, local memory, Cloud, the Internet, URI, URL, web page, etc.).
  • a digital audio file e.g., MP3, FLAG, AIFF, WAV, RA, AU, AAC, ATRAC, Apple Lossless, WMA, MPEG-4, etc.
  • the DD may comprise information about the music such as song title, artists, sidemen, producer, cover art, lyrics, images, photographs, liner notes, playing time, time index, location or address of the file or data that comprises the music (e.g., on NAS, local memory, Cloud, the Internet, URI, URL, web page, etc.).
  • the DD may comprise proprietary (e.g., a format and/or data structure proprietary to the wireless media device(s) 100 ), known, accepted, or standardized formats, or some combination of those, for example. Actual format, structure, content, and source for the DD may be application dependent and is not limited to the examples described herein.
  • proprietary e.g., a format and/or data structure proprietary to the wireless media device(s) 100
  • known, accepted, or standardized formats or some combination of those, for example.
  • Actual format, structure, content, and source for the DD may be application dependent and is not limited to the examples described herein.
  • the DD may be encoded into a format that may be decoded by the wireless media devices (e.g., 100 ).
  • Encoding at the stage 2508 may comprise stripping, ignoring, or otherwise parsing the DD to extract only those portions of the DD that may be encoded into the format that may be decoded by the wireless media devices.
  • Encoding at the stage 2508 may comprise encrypting or otherwise adding data security to the DD that is encoded.
  • Encoding at the stage 2508 may comprise data compression or other process for reducing a size (e.g., number of bytes) of the encoded DD.
  • Data source 2507 broadly includes without limitation any source of data that may be accessed at the stage 2508 to effectuate the encoding of the DD.
  • data source 2507 may comprise algorithms and/or data embodied in a non-transitory computer readable medium that is electronically accessed by the wireless user device or external compute engine and executed on one or more processors to encode the DD.
  • Data source 2509 broadly includes without limitation any source of data that may be accessed at the stage 2510 to effectuate the embedding of the encoded the DD to form the EC.
  • the EC may be stored or otherwise saved to one or more locations as denoted by data store 2511 .
  • Data store 2511 may be internal to the wireless user device, external to the wireless user device or both.
  • data store 2511 may be memory internal to the wireless user device (e.g., Flash memory).
  • data store 2511 may be external memory disposed in the wireless media device (e.g., 100 ).
  • data store 2511 may be external memory disposed in the Cloud, the Internet, NAS or other location.
  • a determination may be made by the wireless media device, the wireless user device or both, for handling of the EC.
  • Handling may comprise without limitation the wireless media device playing back the EC, queuing the EC for playback, storing the EC, assigning handling of the EC to one or more other wireless media devices, streaming content associated with the EC (e.g., executing a playlist encoded and/or included in the EC), wirelessly communicating the EC to another device, just to name a few. If a YES branch is taken from the stage 2514 , then the flow 2500 may transition to a stage 2516 where the EC is handled by the wireless media device.
  • Data source 2515 broadly includes without limitation any source of data that may be accessed at the stage 2516 to effectuate the handling of the EC.
  • data source 2515 may comprise CFG 225 or may comprise an operating system (OS) for the wireless media device (e.g., code in DS 103 ) that determines how EC is to be handled by the wireless media device(s).
  • OS operating system
  • flow 2055 may transition to some other stage, such as a stage 2518 .
  • the stage 2518 may select the YES branch to indicate there are one or more other items of content C to be processed (e.g., the second, third, . . . to the last of the 103 items).
  • the YES branch from the stage 2518 may transition to some other stage of flow 2500 as denoted by 2520 .
  • 2520 may comprise, consider a scenario where all of the 103 items of content C are accessed at the stage 2504 and all of the DD for all 103 items is accessed at the stage 2506 . Then 2520 may comprise a transition of flow 2500 to the stage 2508 where the DD for each item of content C is encoded, followed by the stage 2510 , etc.
  • Data source 2517 broadly includes without limitation any source of data that may be accessed at the stage 2518 to make the determination of whether or not more content C is to be processed by flow 2500 .
  • data source 2517 may be a counter that starts at the number of items of content C to be processed by flow 2500 (e.g., 103 items) and is decremented by one each time an item of content C has been processed.
  • Stage 2518 may access data source 2517 to determine if its value is zero (“0”) (e.g., all 103 items have been processed) or some other value that indicates completion or no more content C.
  • the wireless user device may access content C as was described above in reference to the stage 2504 in flow 2500 and data source 2603 may be accessed as described above for data source 2503 .
  • data store 2511 may be accessed at the stage 2604 .
  • Data store 2511 may comprise one or more items of EC (e.g., encoded content from the stage 2510 ) as was described above.
  • a determination may be made as to whether or not the content C includes encoded descriptive data (DD) as described above in reference to the stages 2506 - 2510 of flow 2500 . If a YES branch is taken, then flow 2600 may transition to a stage 2608 as will be described below. If a NO branch is taken, then flow 2600 may transition to another stage, such as the stage 2506 of flow 2500 as described above. Stages 2506 - 2510 of the flow 2500 may be executed to access, encode, and embed the DD into the content C. After stage 2510 , flow 2600 may resume at the stage 2608 .
  • DD encoded descriptive data
  • the encoded DD may be data about one or more items of unrelated content C that are encoded into the OC for subsequent decoding by the wireless media device (e.g., post stage 2614 of flow 2600 ).
  • the encoded DD may comprise content C in the form of a location for a song the wireless user device is commanding the wireless media device to handle (e.g., to playback the song on the wireless media device).
  • Wireless user device may use a speaker to acoustically transmit EC that comprises the OC startup tone with the encoded DD having the location of the song to be handled.
  • a microphone on the wireless media device may receive the acoustic transmission which is subsequently decoded by the wireless media device to extract the location of the song from the encoded DD and handle processing of the song.
  • the OC e.g., signature tone
  • the OC may include a link, pointer, or address (e.g., the encoded DD) to content C to be handled by the wireless media device.
  • the information decoded from the encoded DD may be used by the wireless media device to access and handle the content C. Therefore, in this example, the EC itself does not include the content C to be handled, but rather information for accessing the content C to be handled.
  • Block diagram 2700 a includes at least one wireless media device 100 , at least one wireless user device 2701 (e.g., a pad, tablet, or smartphone), and optionally other systems and/or resources such as wireless router 2760 , data storage 2765 (e.g., NAS), server/PC 2762 , and resource 2750 (e.g., Cloud or the Internet).
  • Wireless (RF) communications links between the various elements depicted are denoted as RF signals 2710 , 2720 , 2730 , and 2770 .
  • a user 201 may be present in an environment ENV 198 as described above and the user 201 may be the user of one or more wireless user devices, such as wireless user device 2701 , for example.
  • wireless media device 100 may include one or more systems that sense 197 the ENV 198 for presence of users (e.g., 201 ) and wireless devices (e.g., 2701 ).
  • wireless user devices such as device 2701
  • devices 2701 Actual configurations of wireless user devices, such as device 2701 , will vary and the examples provided herein are for purposes of explaining the present application and are not to be construed as limiting in any way the types and configurations of wireless user devices that may interact with wireless media devices, such as wireless media device 100 .
  • Wireless user device 2701 may include a plurality of systems including but not limited to one or more radios for transmitting, receiving, or both, RF signals 2710 for wireless communications (e.g., WiFi, IEEE 802.1, NFC, BT, BT low energy, etc.), a power supply (e.g., lithium ion rechargeable battery), one or more microphones 2702 or similar transducer for generating a signal 2702 c from incident sound (e.g., sound in ENV 198 ), one or more antennas 2714 coupled with the one or more radios, one or more speakers 2704 or similar transducers for generating sound in response to an applied signal 2704 c (e.g., from an amplifier), a display 2703 (e.g., touch screen, OLED, LCD, etc.) for displaying information such as a GUI or an application running on device 2701 , data storage (e.g., Flash memory) for storing an OS, application software, APPS, algorithms, data, media files, content C, etc., and one
  • wireless user device 2701 may include APP 225 configured for use (e.g., communication, command, control, content handling, etc.) with one or more wireless media devices 100 .
  • APP 225 may have access to content C that may reside on the device 2701 , external to the device 2701 or both.
  • APP 225 may be one or more algorithms and/or data embodied in a non-transitory computer readable medium, such as a data storage system of device 2701 , for example.
  • Access to content C by APP 225 may include managing and/or presenting the content C on display 2703 in a form the user 201 may interact with, such as the GUI or other form of displaying information.
  • the GUI of the device 2701 or a GUI generated by APP 225 may display content C in various formats.
  • the GUI displays at least a portion of the content C as one or more playlists P1-P3, with each playlist comprised of one or more songs or other media type.
  • playlist P3 includes a plurality of songs denoted as S1-S7.
  • APP 225 via the GUI, may display one or more wireless media devices that the user device 2701 is wirelessly linked with and/or has been previously wirelessly linked with, as denoted by WMD's 100 - 100 vi .
  • WMD's 100 - 100 vi There may be more or fewer wireless media devices than depicted as denoted by 2715 .
  • User 201 may have commanded the device 2701 to display the content C depicted in the GUI or the display may be controlled by APP 225 or other algorithm and/or system of device 2701 .
  • Wireless media device 100 may include a plurality of systems as described above in reference to FIG. 1 .
  • the various systems that may be included in media device 100 are depicted in block diagram form without implementation specific details such as actual connections to the systems, etc.
  • Systems in media device 100 may include but are not limited to PROX 113 (e.g., proximity detection island(s) 520 ) for sensing 197 the ENV 198 , RF 107 which may include one or more radios Tx/Rx 2777 coupled with one or more antennas 124 , some of which may be de-tunable 129 , DS 103 which may include CFG 125 and content C (e.g., content being handled by 100 ), controller 101 , A/V 109 which may include display 180 , one or more image capture devices 109 c , one or more MCI's 160 and associated amplification/processing circuitry 2771 , and one or more SPK's 170 and associated amplification/processing circuitry 2773 , and chassis 199 which may
  • Controller 101 and/or algorithms executing on controller 101 or other circuitry in media device 100 may decode ( 2772 , 2778 ) signals (e.g., encoded signals) received by the various systems or encode signals to be transmitted by the various systems.
  • signals e.g., encoded signals
  • encoded RF, acoustic, or optical signals transmitted from wireless user device 2701 may be decoded by circuitry and/or algorithms in media device 100 to generate decoded signals.
  • the media device 100 may use circuitry and/or algorithms to encode signals for wireless transmission from the media device 100 to user device 2701 , other media devices 100 , or other devices or systems such as resource 2750 , for example.
  • Wireless transmission from the media device 100 using its various systems may comprise one or more of RF ( 107 ), acoustic ( 109 ), or optical ( 180 , 185 , 190 ) signal transmission.
  • Wireless signal reception by the media device 100 using its various systems may comprise one or more of RF ( 107 ), acoustic ( 109 ), or optical ( 185 , 190 ) signal reception.
  • Wireless user device 2701 may transmit and receive wireless signals using similar RF, acoustic and optical systems as the media device 100 (e.g., 2714 , 2704 , 2702 , and 2703 ).
  • user device 2701 and media device 100 may establish a wireless communication link (e.g., at the stage 2502 ) with each other using any of their RF, acoustic, or optical systems.
  • RF 107 may detect RF signal 2710 from user device 2710 and the APP 225 and/or CFG 125 may orchestrate establishing the wireless link, such as a BT link (e.g., via paring) or via WiFi link (e.g., using wireless router 2760 ), for example.
  • the wireless communications link may be a direct link between media device 100 and user device 2701 or an indirect link, such as through a wireless network (e.g., IEEE 802.1 or wireless router 2760 ).
  • the wireless communications link may be comprise each device recognizing the other using whatever wireless protocols they have previously used in communicating with each other and initiating wireless communication of data between each other (e.g., handshakes or one or more stages of flow 2500 and/or flow 2600 ).
  • the wireless communications link may be one or more of acoustic, RF, or optical and the mode of wireless communication may reversibly switch between one or more of the RF, acoustic, or optical links during interaction between user device 2701 and media device 100 .
  • the wireless communications link may be established between one or more wireless user devices and one or more wireless media devices even thou FIGS. 27A-27B depict only one of each device.
  • the media device 100 upon detecting presence of user 201 and/or user device 2701 may transmit an acoustic signal via SPK 170 that may be detected by MIC 2702 and signal 2702 c may be decoded or otherwise processed to establish the wireless link using one of its RF systems (e.g., BT, NFC, WiFi, etc.).
  • Information about a WiFi network e.g., network name and password
  • the wireless communications may be all acoustic, all RF, or all optical (e.g., using infrared (e.g., an IrLED) communications via OPT 185 to an optoelectronic receiving device (e.g., a photo diode) on 2701 ).
  • infrared e.g., an IrLED
  • OPT 185 an optoelectronic receiving device
  • user device 2710 may access content C (e.g., at the stage 2504 ) which may be internal, external of both to the user device 2710 .
  • content C e.g., at the stage 2504
  • user 201 may have selected preferences that are stored or otherwise coded in APP 225 , that want content C to be handled by media device 100 when both devices are disposed in ENV 198 (e.g., are in close proximity of each other).
  • the content C that comprises S7 would be accessed (e.g., via a read operation) by the user device 2701 .
  • the user device 2701 may then access (e.g., at the stage 2506 ) descriptive data DD about song S7 from a data source that may be internal, external or both to user device 2701 .
  • the DD may comprise the location (e.g., address, file folder, URI, URL or other descriptor) of the song S7, playing time of S7, song title for S7, artist or group that performed S7, and a time index, if any, to begin playback of S7, for example.
  • the DD may then be encoded into a format that may be decoded by the wireless media device 100 (e.g., at the stage 2508 ).
  • Decoding may be done using hardware (e.g., 101 ), software (e.g., CFG 125 or other algorithms) or both.
  • the encoding may be in a format that is proprietary to wireless media devices 100 .
  • Decoding of the encoded DD may be accomplished using processing in the analog domain, digital domain or both. Circuitry including but not limited to a DSP, an analog-to-digital-converter (ADC) or a digital-to-analog-converter (DAC) may be used in the decoding process.
  • the DSP or other compute engine may execute one or more algorithms to effectuate the decoding of the encoded DD.
  • the encoded DD may be embedded or otherwise linked or associated with the content C that was accessed (e.g., at the stage 2510 ).
  • the encoded DD may be embedded into a file that comprises S7, included in one or more field of a data packet in which S7 will be transmitted to media device 100 , may be concatenated into a file that comprises S7 (e.g., as a header or footer), may be included in a wrapper file or object that includes S7, or may be wirelessly transmitted in some sequence that includes data for the encoded DD and data for S7, for example.
  • the encoded DD may be wirelessly transmitted first, followed by wireless transmission of the accessed content C associated with the encoded DD, such as S7.
  • the user device 2701 may access as the content a plurality of items (e.g., a plurality of songs and/or playlists, etc.) that are all described in the DD and included in the encoded DD.
  • the accessed content C may comprise one or more of the playlists P1-P3 and all of the songs in those playlists.
  • the DD may comprise information about the playlists, their respective songs, and the locations of the playlists and songs, for example.
  • the encoded DD that is embedded or otherwise associated with the accessed content C would include that information as well.
  • the content may be referred to as encoded content (EC) that may comprise a single datum or a plurality of datum that are wirelessly communicated to the wireless media device 100 either directly from the user device 2701 using one of the aforementioned wireless links, from an external system (e.g., 2750 , 2760 ) or both.
  • encoded content for song S7 in playlist P3 is denoted as S7-EC.
  • stages 2502 - 2510 of flow 2500 may be processed (e.g., by a processor in 2701 ) to create S7-EC from S7 as denoted by dashed line “a”, followed by some or all of stages 2512 - 2516 being processed to communicate S7-EC to wireless media device 100 as denoted by dashed line “b”.
  • a single datum for the EC may comprise the encoded DD embedded in the same file, data structure, or packet as the accessed content C.
  • S7-EC may be a single file or data structure that includes the content for song S7 plus the encoded DD for song S7.
  • S7-EC may comprise a packet formatted to include one of more fields for the S7's content C (e.g., a data payload) and one or more fields for the encoded DD for S7, as well as other fields used in packets for communicating data, such as error correction, check sums, headers, footers, etc.
  • the plurality of datum for the EC may comprise multiple datum's that are wirelessly transmitted to the media device 100 or a first subset of the plurality of datum that are wirelessly transmitted to the media device 100 and a second subset of datum that are accessed by the wireless media device 100 based on information contained in the first subset.
  • the first subset may include the encoded DD that upon being decoded by the media device 100 , may include information the media device 100 uses to access the datum in the second subset, which may comprise the content C to be accessed and optionally handled by the media device 100 .
  • the first subset may include data for a web page, Internet address, URI, URL or the like where the content C in the second subset can be accessed by the media device 100 .
  • the encoded DD in the first subset may include a URL 2758 that links to a web site or other address in resource 2750 (e.g., the Internet or Cloud) where the content C in the second subset may be accessed from, such as from a data storage system 2757 .
  • a URL 2758 that links to a web site or other address in resource 2750 (e.g., the Internet or Cloud) where the content C in the second subset may be accessed from, such as from a data storage system 2757 .
  • the EC may be acted on (e.g., handled) by the wireless media device 100 such as denoted by the stage 2514 - 2516 of flow 2500 .
  • Handling of the EC may broadly include the media device 100 or other media devices 100 in communication with media device 100 and/or wireless user device 2701 performing one or more actions on the EC.
  • the actions taken may be determined in whole or in part by one or more of the CFG's 125 of one or more of the media devices 100 , the APP 225 on user device 2701 , a command by user 201 (e.g., via the GUI, voice command, etc.), hardware, or algorithms, for example.
  • the user 201 may have selected song S7 for playback on media device 100 and the handling of the subsequent EC may be to initiate playback of S7 on media device 100 (e.g., via SPK's 170 ) and/or on one or more other wireless media devices 100 in wireless communication with media device 100 (e.g., via BT link, WiFi link, AdHoc WiFi link, etc.).
  • S7 is a multi-channel media file (e.g., surround sound, DTS®, Dolby®, etc.)
  • the various encoded channels of sound may be assigned (e.g., via the CFG's 125 ) to a plurality of different wireless media devices 100 in the same ecosystem and/or disposed in ENV 198 .
  • the EC comprise video, images, or audio-video content
  • handling by media device 100 may comprise playback of the image or video content on display 180 and audio content (if any) over SPK's 170 .
  • handling of EC by media device 100 may not be immediate, may be delayed, may be queued, or may be stored for later use.
  • EC may be a link, pointer, or address to other content to be accessed or handled by the media device(s) 100 and handling the EC may comprise acting on the link, pointer, or address and accessing and/or handling the content C there according to data in the EC, the CFG's 125 , etc.
  • the EC that is wirelessly communicated (e.g., at the stage 2512 ) to the media device 100 , upon decoding may reveal a location on NAS 2765 where the content C to be accessed resides and the media device 100 may wirelessly stream the content C for playback over the media device 100 or other media devices 100 .
  • wireless media device 100 and wireless network/router 2760 may be wirelessly linked (e.g., via WiFi, IEEE 802.1 or other).
  • the content C to be accessed may reside in more than one location specified by the EC, such as in NAS 2765 , server 2762 , and data storage 2754 of resource 2750 , for example.
  • NAS 2765 and server 2762 may be in communication ( 2763 , 2764 ) with router 2760 and the content C on NAS 2765 and server 2762 may be wirelessly communicated 2730 to the media device 100 for handling.
  • content C in data storage 2754 of resource 2750 may be wirelessly communicated 2720 (e.g., using cellular tower 2756 ) from resource 2750 to media device 100 for handling.
  • Content C being handled may be wirelessly transferred in total to the media device(s) 100 , be streamed, buffered, stored, or handled in chunks or portions that may be determined by factors including but not limited to wireless bandwidth, download times, a data size (e.g., in megabytes, gigabytes, etc.) of the content C, data storage capacity of media device 100 (e.g., in DS 103 ), copyright and/or digital media rights, persistence of the content C, the type of content C, just to name a few.
  • a data size e.g., in megabytes, gigabytes, etc.
  • data storage capacity of media device 100 e.g., in DS 103
  • persistence of the content C the type of content C, just to name a few.
  • the wireless communication of the EC may comprise one or more acoustic signals 2776 generated by SPK 2704 of user device 2701 that are received 2776 by MIC 160 of media device 100 .
  • Signals 160 c generated by MIC 160 may be decoded 2772 to extract data from the encoded DD that is included the EC.
  • the wireless communication of the EC (e.g., at the stage 2512 ) may comprise one or more RF signals 2710 transmitted by user device 2701 and received by Tx/Rx 2777 of media device 100 .
  • the RF signal may be decoded 2778 to extract data from the encoded DD that is included the EC.
  • the wireless communication of the EC may comprise one or more wireless signals 2710 and one or more acoustic signals 2776 that may be transmitted in some sequence or simultaneously by user device 2701 .
  • the encoded DD portion of the EC may be acoustically transmitted, followed by wireless RF transmission of the content C associated with the EC.
  • This sequence may be used when the data that comprises the content C is large in size and a higher bandwidth and/or error correction capabilities of the wireless protocol used for the RF transmission may make it more ideal than acoustic transmission of the content C.
  • ambient noise AN 2799 in ENV 198 is of sufficient amplitude (e.g., in dB's) to prevent MIC 160 from accurately receiving 2776 the acoustic transmission 2776 from SPK 2704 , then user device 2701 , media device 100 or both may act to switch from acoustic to RF wireless transmission of the EC and/or its associated content C.
  • At least a portion of the EC may be optically wirelessly transmitted (e.g., at the stage 2512 ) using optical systems of the user device 2701 and media device 100 .
  • OPT 185 in I/O 105 may be used to receive 185 r optical signals from user device 2701 or to transmit 185 t optical signals to the user device 2701 .
  • User device 2701 may include an optical transmit 185 t and receive 185 r system that may be similar to that of media device 100 and/or communicates using a wireless protocol (e.g., IrDA or other) that is compatible with or identical to those used by media device 100 .
  • a wireless protocol e.g., IrDA or other
  • infrared LED's may be used in conjunction with an opto-electronic device such as a photo diode, PIN diode, or the like for receiving the optical signals and converting them into signals to be decoded 2782 or otherwise processed in the devices 2701 and 100 .
  • RF wireless and/or acoustic wireless communications between media device 100 and user device 100 may supplement and/or replace the optical wireless communications.
  • the EC may comprise a bar code, TAG, or other image based symbol that is encoded with information that may be captured by the media device 100 (e.g., via 190 ), processed, and output as a decoded signal.
  • the stages 2506 - 2510 may result in an encoded image (e.g., a TAG or other) denoted as EC 1 being displayed on display 2703 of user device 2710 .
  • the user device 2710 and media device 100 may be positioned 2792 relative to each other such that image capture device 190 may image 190 i the encoded image EC 1 off of the display 2703 and generate a signal that is decoded 2781 to extract the encoded DD from EC 1 .
  • EC 1 may include data that describes a location where the content C associated with EC 1 may be accessed by the media device 100 , such as NAS 2765 , data storage system 2757 and/or 2754 in resource 2750 , or the wireless user device 2701 .
  • EC 1 may be generated (e.g., via flow 2500 or 2600 ) by user device 2701 , server 2762 , compute engine 2752 in resource 2750 or some combination of the foregoing, for example.
  • the captured image 190 i of EC 1 upon being decoded by media device 100 may include data instructing media device 100 to access song S7 from playlist P3 on wireless user device 2701 using one or more of the wireless RF links between those two wireless devices (e.g., WiFi via router 2760 or directly using a BT link).
  • the wireless RF links between those two wireless devices e.g., WiFi via router 2760 or directly using a BT link.
  • the user device 2701 and wireless media device may wirelessly communicate with each other using the systems as described above in reference to FIG. 27A and/or flows 2500 or 2600 .
  • the actual content to be handled by media device 100 such as song S9 in playlist P3 is referenced by another item of content that may be unrelated to the actual content of S9, denoted as other content OC.
  • content of S9 may already include DD that is encoded into the content that comprises S9.
  • Content S9 may be accessed (e.g., at the stage 2604 of flow 2600 ) and subsequently analyzed to determine whether or not S9 includes encoded DD. If not, then the DD may be accessed and embedded or associated with the OC.
  • wireless media devices 100 at power up or when activated from a standby or low power mode to an active mode, emit an signature acoustic signal encoded with information that other media devices 100 may capture (e.g., via their MIC's 160 ) and decode.
  • Each wireless media device in an ecosystem of media devices or in ENV 198 may announce, recognize, and wirelessly link with one another in response to a received signature acoustic signal.
  • the signature acoustic signal may be modulated in frequency (FM), in amplitude (AM), pulse width, pulse shape (e.g., sine wave, saw tooth wave, etc.) to encode information that may be decoded by other wireless media devices.
  • FM frequency
  • AM amplitude
  • pulse width e.g., sine wave, saw tooth wave, etc.
  • Adding encoded DD to the signature acoustic signal may comprise one form of EC.
  • the signature acoustic signal (signature tone hereinafter) may comprise one form of content C (e.g., in DS 103 ) that may be totally unrelated to any content C the user device 2701 is presenting for handling by the media device 100 .
  • the OC that may be encoded with DD about the content C to be handled such as the location of the content C on the user device 2701 , resource 2750 , wireless network 2760 , or other wireless media device 100 , for example.
  • the OC may be wirelessly transmitted to media device 100 as denoted by dashed line “a” and media device 100 may decode the OC to extract data for a location of content, such as S9 or S9 plus EC denoted as S9-EC as denoted by dashed line “b” or “C”.
  • the OC may be appended, embedded, or otherwise associated with the content for song S9 to form encoded content S9-EC as denoted by dashed lines “d” and “e”, and media device 100 may access the content S9-EC as denoted by dashed line “C”.
  • the OC may be sonically screamed out 2776 to the media device 100 using SPK 2704 .
  • MIC 160 receives the acoustic signal 2776 and signals 2772 are decoded to yield information regarding the content for song S9, such as where S9 may be accessed from by the media device 100 , how to handle the content S9 or encoded content S9-EC after it has been accessed by the media device 100 , etc.
  • the sonic screams may be wirelessly transmitted from media device 100 to user device 2701 to wirelessly communicate information between the devices. Therefore, the user device 2701 , the media device(s) 100 or both may use the sonic screams to communicate information including encoded content EC.
  • the signature acoustic signal (e.g., the sonic scream) may be in a frequency range 2800 that is within (W) a range 2801 of human hearing, that is above (A) 2803 the range of human hearing, or that is below (B) 2805 the range of human hearing.
  • the signature acoustic signal may be an infrasonic low frequency signal that is approximately 20 Hz or less, an ultrasonic frequency that is approximately 20 kHz or more, or an acoustic range that is from about 20 Hz to about 20 kHz, for example.
  • Ambient noise AN 2799 may determine which frequency range is best suited for wirelessly communicating EC or other data to/from devices 2710 and 100 .

Abstract

Embodiments relate generally to electronic and electronic hardware, computer software, wired communications, wireless communications, and wireless network communications, portable, wearable, and stationary media devices. RF transceivers and/or audio system in each media device may be used to wirelessly communicate between media devices and/or user devices and allow configuration, content, and other data to be wirelessly transmitted from one media device and/or user device to another media device. One or more user devices in proximity of one or more media devices, post detection, may wirelessly communicate (e.g., using RF, light, sound/acoustics) with the one or more media devices and the one or more media devices may orchestrate wireless handling (e.g., using RF, light, sound/acoustics) of content from the user device and/or from a wirelessly accessible location such as the Cloud, Internet, data storage device, or other location as indicated by an application (APP), content, data or both from the user device.

Description

    CROSS-REFERENCE TO RELATED APPLICATIONS
  • This application is related to the following U.S. patent applications: U.S. patent application Ser. No. 13/802,646, Filed on Mar. 13, 2013, having Attorney Docket Number ALI-230, and Titled “Proximity-Based Control Of Media Devices For Media Presentations”, all of which are herein incorporated by reference in their entirety for all purposes.
  • FIELD
  • Embodiments relate generally to electronic and electronic hardware, computer software, wired communications, wireless communications, and wireless network communications, portable, wearable, and stationary media devices. More specifically, disclosed are media devices that detect proximity of users and/or user devices and take actions and handle content on the user devices after detecting presence of users and/or user devices.
  • BACKGROUND
  • Conventional paradigms for media devices require a user to take some action using a finger press to a touch screen or press a button, or the like, in order to initiate some function on the device, such as listening to music, for example. Conventional media devices are not configured to recognize and act on user preferences as to how the media device serves the user's needs based on changing circumstances and changing environments the user and media device are subject to. Furthermore, conventional media devices are typically personal devices that are mostly if not always used solely by the user and are therefore not well adapted to servicing the needs of friends, guests, or the like who may want to share content on their devices with the user.
  • Thus, there is a need for devices, systems, methods, and software that allow a user to configure (e.g., wirelessly using one or more of RF, acoustics, optics, etc.) a media device to detect a user and/or user device, take an action based on having detected presence, and allow for content from a user device or many devices to be handled by the media device, post detection.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • Various embodiments or examples (“examples”) of the present application are disclosed in the following detailed description and the accompanying drawings. The drawings are not necessarily to scale:
  • FIG. 1 depicts a block diagram of one example of a media device according to an embodiment of the present application;
  • FIG. 2A depicts one example of a configuration scenario for a user device and a media device according to an embodiment of the present application;
  • FIG. 2B depicts example scenarios for another media device being configured using a configuration from a previously configured media device according to an embodiment of the present application;
  • FIG. 3 depicts one example of a flow diagram of a process for installing an application on a user device and configuring a first media device using the application according to an embodiment of the present application;
  • FIGS. 4A and 4B depict example flow diagrams for processes for configuring an un-configured media device according to embodiments of the present application;
  • FIG. 5 depicts a profile view of one example of a media device including control elements and proximity detection islands according to embodiments of the present application;
  • FIG. 6 depicts a block diagram of one example of a proximity detection island according to embodiments of the present application;
  • FIG. 7 depicts a top plan view of different examples of proximity detection island configurations according to embodiments of the present application;
  • FIG. 8A is a top plan view depicting an example of proximity detection island coverage according to embodiments of the present application;
  • FIG. 8B is a front side view depicting an example of proximity detection island coverage according to embodiments of the present application;
  • FIG. 8C is a side view depicting an example of proximity detection island coverage according to embodiments of the present application;
  • FIG. 9 is a top plan view of a media device including proximity detection islands configured to detect presence according to embodiments of the present application;
  • FIG. 10 depicts one example of a flow for presence detection, notification, and media device readiness according to embodiments of the present application;
  • FIG. 11 depicts another example of a flow for presence detection, notification, and media device readiness according to embodiments of the present application;
  • FIG. 12 depicts yet another example of a flow for presence detection, notification, and media device readiness according to embodiments of the present application;
  • FIG. 13 depicts one example of presence detection using proximity detection islands and/or other systems responsive to wireless detection of different users and/or different user devices according to embodiments of the present application;
  • FIG. 14 depicts one example of proximity detection islands associated with specific device functions according to embodiments of the present application;
  • FIG. 15 depicts one example of content handling from a user device subsequent to proximity detection according to embodiments of the present application;
  • FIG. 16 depicts another example of content handling from user devices subsequent to proximity detection according to embodiments of the present application;
  • FIG. 17 depicts one example of content handling from a data capable wristband or wristwatch subsequent to proximity detection according to embodiments of the present application;
  • FIG. 18 depicts another example of content handling from a data capable wristband or wristwatch subsequent to proximity detection according to embodiments of the present application;
  • FIG. 19 depicts one example of a flow for content handling on a media device post proximity detection according to embodiments of the present application;
  • FIG. 20 depicts one example of a flow for storing, recording, and queuing content post proximity detection according to embodiments of the present application;
  • FIG. 21 depicts one example of a media device handling, storing, queuing, and taking action on content from a plurality of user devices according to embodiments of the present application;
  • FIG. 22 depicts another example of a media device handling, storing, queuing, and taking action on content from a plurality of user devices according to embodiments of the present application;
  • FIG. 23 depicts one example of a flow for recording user content on a media device while the media device handles current content according to embodiments of the present application;
  • FIG. 24 depicts one example of queuing action for user content in a queue of a media player according to embodiments of the present application;
  • FIG. 25 depicts one example of a flow for wirelessly communicating encoded content between a wireless user device and a wireless media device according to embodiments of the present application;
  • FIG. 26 depicts another example of a flow for wirelessly communicating encoded content between a wireless user device and a wireless media device according to embodiments of the present application;
  • FIG. 27A depicts a block diagram of one example of wirelessly communicating encoded content between a wireless user device and a wireless media device according to embodiments of the present application;
  • FIG. 27B depicts a block diagram of another example of wirelessly communicating encoded content between a wireless user device and a wireless media device according to embodiments of the present application; and
  • FIG. 28 depicts one example of a frequency spectrum for acoustic-based wireless communication between a wireless user device and a wireless media device according to embodiments of the present application.
  • DETAILED DESCRIPTION
  • Various embodiments or examples may be implemented in numerous ways, including as a system, a process, a method, an apparatus, a user interface, or a series of program instructions on a non-transitory computer readable medium such as a computer readable storage medium or a computer network where the program instructions are sent over optical, electronic, or wireless communication links. In general, operations of disclosed processes may be performed in an arbitrary order, unless otherwise provided in the claims.
  • A detailed description of one or more examples is provided below along with accompanying figures. The detailed description is provided in connection with such examples, but is not limited to any particular example. The scope is limited only by the claims and numerous alternatives, modifications, and equivalents are encompassed. Numerous specific details are set forth in the following description in order to provide a thorough understanding. These details are provided for the purpose of example and the described techniques may be practiced according to the claims without some or all of these specific details. For clarity, technical material that is known in the technical fields related to the examples has not been described in detail to avoid unnecessarily obscuring the description.
  • FIG. 1 depicts a block diagram of one embodiment of a media device 100 having systems including but not limited to a controller 101, a data storage (DS) system 103, a input/output (I/O) system 105, a radio frequency (RF) system 107, an audio/video (AN) system 109, a power system 111, and a proximity sensing (PROX) system 113. A bus 110 enables electrical communication between the controller 101, DS system 103, I/O system 105, RF system 107, AV system 109, power system 111, and PROX system 113. Power bus 112 supplies electrical power from power system 111 to the controller 101, DS system 103, I/O system 105, RF system 107, AV system 109, and PROX system 113.
  • Power system 111 may include a power source internal to the media device 100 such as a battery (e.g., AAA or AA batteries) or a rechargeable battery (e.g., such as a lithium ion or nickel metal hydride type battery, etc.) denoted as BAT 135. Power system 111 may be electrically coupled with a port 114 for connecting an external power source (not shown) such as a power supply that connects with an external AC or DC power source. Examples include but are not limited to a wall wart type of power supply that converts AC power to DC power or AC power to AC power at a different voltage level. In other examples, port 114 may be a connector (e.g., an IEC connector) for a power cord that plugs into an AC outlet or other type of connecter, such as a universal serial bus (USB) connector. Power system 111 provides DC power for the various systems of media device 100. Power system 111 may convert AC or DC power into a form usable by the various systems of media device 100. Power system 111 may provide the same or different voltages to the various systems of media device 100. In applications where a rechargeable battery is used for BAT 135, the external power source may be used to power the power system 111, recharge BAT 135, or both. Further, power system 111 on its own or under control or controller 101 may be configured for power management to reduce power consumption of media device 100, by for example, reducing or disconnecting power from one or more of the systems in media device 100 when those systems are not in use or are placed in a standby or idle mode. Power system 111 may also be configured to monitor power usage of the various systems in media device 100 and to report that usage to other systems in media device 100 and/or to other devices (e.g., including other media devices 100) using one or more of the I/O system 105, RF system 107, and AV system 109, for example. Operation and control of the various functions of power system 111 may be externally controlled by other devices (e.g., including other media devices 100).
  • Controller 101 controls operation of media device 100 and may include a non-transitory computer readable medium, such as executable program code to enable control and operation of the various systems of media device 100. DS 103 may be used to store executable code used by controller 101 in one or more data storage mediums such as ROM, RAM, SRAM, RAM, SSD, Flash, etc., for example. Controller 101 may include but is not limited to one or more of a microprocessor (μP), a microcontroller (μP), a digital signal processor (DSP), a baseband processor, an application specific integrated circuit (ASIC), just to name a few. Processors used for controller 101 may include a single core or multiple cores (e.g., dual core, quad core, etc.). Port 116 may be used to electrically couple controller 101 to an external device (not shown).
  • DS system 103 may include but is not limited to non-volatile memory (e.g., Flash memory), SRAM, DRAM, ROM, SSD, just to name a few. In that the media device 100 in some applications is designed to be compact, portable, or to have a small size footprint, memory in DS 103 will typically be solid state memory (e.g., no moving or rotating components); however, in some application a hard disk drive (HDD) or hybrid HDD may be used for all or some of the memory in DS 103. In some examples, DS 103 may be electrically coupled with a port 128 for connecting an external memory source (e.g., USB Flash drive, SD, SDHC, SDXC, microSD, Memory Stick, CF, SSD, etc.). Port 128 may be a USB or mini USB port for a Flash drive or a card slot for a Flash memory card. In some examples as will be explained in greater detail below, DS 103 includes data storage for configuration data, denoted as CFG 125, used by controller 101 to control operation of media device 100 and its various systems. DS 103 may include memory designate for use by other systems in media device 100 (e.g., MAC addresses for WiFi 130, network passwords, data for settings and parameters for A/V 109, and other data for operation and/or control of media device 100, etc.). DS 103 may also store data used as an operating system (OS) for controller 101. If controller 101 includes a DSP, then DS 103 may store data, algorithms, program code, an OS, etc. for use by the DSP, for example. In some examples, one or more systems in media device 100 may include their own data storage systems.
  • I/O system 105 may be used to control input and output operations between the various systems of media device 100 via bus 110 and between systems external to media device 100 via port 118. Port 118 may be a connector (e.g., USB, HDMI, Ethernet, fiber optic, Toslink, Firewire, IEEE 1394, or other) or a hard wired (e.g., captive) connection that facilitates coupling I/O system 105 with external systems. In some examples port 118 may include one or more switches, buttons, or the like, used to control functions of the media device 100 such as a power switch, a standby power mode switch, a button for wireless pairing, an audio muting button, an audio volume control, an audio mute button, a button for connecting/disconnecting from a WiFi network, an infrared (IR) transceiver, just to name a few. I/O system 105 may also control indicator lights, audible signals, or the like (not shown) that give status information about the media device 100, such as a light to indicate the media device 100 is powered up, a light to indicate the media device 100 is in wireless communication (e.g., WiFi, Bluetooth®, WiMAX, cellular, etc.), a light to indicate the media device 100 is Bluetooth® paired, in Bluetooth® pairing mode, Bluetooth® communication is enabled, a light to indicate the audio and/or microphone is muted, just to name a few. Audible signals may be generated by the I/O system 105 or via the AV system 107 to indicate status, etc. of the media device 100. Audible signals may be used to announce Bluetooth® status, powering up or down the media device 100, muting the audio or microphone, an incoming phone call, a new message such as a text, email, or SMS, just to name a few. In some examples, I/O system 105 may use optical technology to wirelessly communicate with other media devices 100 or other devices. Examples include but are not limited to infrared (IR) transmitters, receivers, transceivers, an IR LED, and an IR detector, just to name a few. I/O system 105 may include an optical transceiver OPT 185 that includes an optical transmitter 185 t (e.g., an IR LED) and an optical receiver 185 r (e.g., a photo diode). OPT 185 may include the circuitry necessary to drive the optical transmitter 185 t with encoded signals and to receive and decode signals received by the optical receiver 185 r. Bus 110 may be used to communicate signals to and from OPT 185. OPT 185 may be used to transmit and receive IR commands consistent with those used by infrared remote controls used to control AV equipment, televisions, computers, and other types of systems and consumer electronics devices. The IR commands may be used to control and configure the media device 100, or the media device 100 may use the IR commands to configure/re-configure and control other media devices or other user devices, for example.
  • RF system 107 includes at least one RF antenna 124 that is electrically coupled with a plurality of radios (e.g., RF transceivers) including but not limited to a Bluetooth® (BT) transceiver 120, a WiFi transceiver 130 (e.g., for wireless communications over a wireless and/or WiMAX network), and a proprietary Ad Hoc (AH) transceiver 140 pre-configured (e.g., at the factory) to wirelessly communicate with a proprietary Ad Hoc wireless network (AH-WiFi) (not shown). AH 140 and AH-WiFi are configured to allow wireless communications between similarly configured media devices (e.g., an ecosystem comprised of a plurality of similarly configured media devices) as will be explained in greater detail below. RF system 107 may include more or fewer radios than depicted in FIG. 1 and the number and type of radios will be application dependent. Furthermore, radios in RF system 107 need not be transceivers, RF system 107 may include radios that transmit only or receive only, for example. Optionally, RF system 107 may include a radio 150 configured for RF communications using a proprietary format, frequency band, or other existent now or to be implemented in the future. Radio 150 may be used for cellular communications (e.g., 3G, 4G, or other), for example. Antenna 124 may be configured to be a de-tunable antenna such that it may be de-tuned 129 over a wide range of RF frequencies including but not limited to licensed bands, unlicensed bands, WiFi, WiMAX, cellular bands, Bluetooth®, from about 2.0 GHz to about 6.0 GHz range, and broadband, just to name a few. As will be discussed below, PROX system 113 may use the de-tuning 129 capabilities of antenna 124 to sense proximity of the user, other people, the relative locations of other media devices 100, just to name a few. Radio 150 (e.g., a transceiver) or other transceiver in RF 107, may be used in conjunction with the de-tuning capabilities of antenna 124 to sense proximity, to detect and or spatially locate other RF sources such as those from other media devices 100, devices of a user, just to name a few. RF system 107 may include a port 123 configured to connect the RF system 107 with an external component or system, such as an external RF antenna, for example. The transceivers depicted in FIG. 1 are non-limiting examples of the type of transceivers that may be included in RF system 107. RF system 107 may include a first transceiver configured to wirelessly communicate using a first protocol, a second transceiver configured to wirelessly communicate using a second protocol, a third transceiver configured to wirelessly communicate using a third protocol, and so on. One of the transceivers in RF system 107 may be configured for short range RF communications, such as within a range from about 1 meter to about 15 meters, or less, for example. Another one of the transceivers in RF system 107 may be configured for long range RF communications, such any range up to about 50 meters or more, for example. Short range RF may include Bluetooth®; whereas, long range RF may include WiFi, WiMAX, cellular, and Ad Hoc wireless, for example.
  • AV system 109 includes at least one audio transducer, such as a loud speaker 160, a microphone 170, or both. AV system 109 further includes circuitry such as amplifiers, preamplifiers, or the like as necessary to drive or process signals to/from the audio transducers. Optionally, AV system 109 may include a display (DISP) 180, a video device (VID) 190 (e.g., an image capture device, video camera, still camera, or a web CAM, etc.), or both. DISP 180 may be a display and/or touch screen (e.g., a LCD, OLED, or flat panel display) for displaying video media, information relating to operation of media device 100, content available to or operated on by the media device 100, playlists for media, date and/or time of day, alpha-numeric text and characters, caller ID, file/directory information, a GUI, just to name a few. A port 122 may be used to electrically couple AV system 109 with an external device and/or external signals. Port 122 may be a USB, HDMI, Firewire/IEEE-1394, 3.5 mm audio jack, or other. For example, port 122 may be a 3.5 mm audio jack for connecting an external speaker, headphones, earphones, etc. for listening to audio content being processed by media device 100. As another example, port 122 may be a 3.5 mm audio jack for connecting an external microphone or the audio output from an external device. In some examples, SPK 160 may include but is not limited to one or more active or passive audio transducers such as woofers, concentric drivers, tweeters, super tweeters, midrange drivers, sub-woofers, passive radiators, just to name a few. MIC 170 may include one or more microphones and the one or more microphones may have any polar pattern suitable for the intended application including but not limited to omni-directional, directional, bi-directional, uni-directional, bi-polar, uni-polar, any variety of cardioid pattern, and shotgun, for example. MIC 170 may be configured for mono, stereo, or other. MIC 170 may be configured to be responsive (e.g., generate an electrical signal in response to sound) to any frequency range including but not limited to ultrasonic, infrasonic, from about 20 Hz to about 20 kHz, and any range within or outside of human hearing. In some applications, the audio transducer of AV system 109 may serve dual roles as both a speaker and a microphone.
  • Circuitry in AV system 109 may include but is not limited to a digital-to-analog converter (DAC) and algorithms for decoding and playback of media files such as MP3, FLAG, AIFF, ALAC, WAV, MPEG, QuickTime, AVI, compressed media files, uncompressed media files, and lossless media files, just to name a few, for example. A DAC may be used by AV system 109 to decode wireless data from a user device or from any of the radios in RF system 107. AV system 109 may also include an analog-to-digital converter (ADC) for converting analog signals, from MIC 170 for example, into digital signals for processing by one or more system in media device 100.
  • Media device 100 may be used for a variety of applications including but not limited to wirelessly communicating with other wireless devices, other media devices 100, wireless networks, and the like for playback of media (e.g., streaming content), such as audio, for example. The actual source for the media need not be located on a user's device (e.g., smart phone, MP3 player, iPod, iPhone, iPad, Android, laptop, PC, etc.). For example, media files to be played back on media device 100 may be located on the Internet, a web site, or in the Cloud, and media device 100 may access (e.g., over a WiFi network via WiFi 130) the files, process data in the files, and initiate playback of the media files. Media device 100 may access or store in its memory a playlist or favorites list and playback content listed in those lists. In some applications, media device 100 will store content (e.g., files) to be played back on the media device 100 or on another media device 100.
  • Media device 100 may include a housing, a chassis, an enclosure or the like, denoted in FIG. 1 as 199. The actual shape, configuration, dimensions, materials, features, design, ornamentation, aesthetics, and the like of housing 199 will be application dependent and a matter of design choice. Therefore, housing 199 need not have the rectangular form depicted in FIG. 1 or the shape, configuration etc., depicted in the Drawings of the present application. Nothing precludes housing 199 from comprising one or more structural elements, that is, the housing 199 may be comprised of several housings that form media device 100. Housing 199 may be configured to be worn, mounted, or otherwise connected to or carried by a human being. For example, housing 199 may be configured as a wristband, an earpiece, a headband, a headphone, a headset, an earphone, a hand held device, a portable device, a desktop device, just to name a few.
  • In other examples, housing 199 may be configured as speaker, a subwoofer, a conference call speaker, an intercom, a media playback device, just to name a few. If configured as a speaker, then the housing 199 may be configured as a variety of speaker types including but not limited to a left channel speaker, a right channel speaker, a center channel speaker, a left rear channel speaker, a right rear channel speaker, a subwoofer, a left channel surround speaker, a right channel surround speaker, a left channel height speaker, a right channel height speaker, any speaker in a 3.1, 5.1, 7.1, 9.1 or other surround sound format including those having two or more subwoofers or having two or more center channels, for example. In other examples, housing 199 may be configured to include a display (e.g., DISP 180) for viewing video, serving as a touch screen interface for a user, providing an interface for a GUI, for example.
  • PROX system 113 may include one or more sensors denoted as SEN 195 that are configured to sense 197 an environment 198 external to the housing 199 of media device 100. Using SEN 195 and/or other systems in media device 100 (e.g., antenna 124, SPK 160, MIC 170, etc.), PROX system 113 senses 197 an environment 198 that is external to the media device 100 (e.g., external to housing 199). PROX system 113 may be used to sense one or more of proximity of the user or other persons to the media device 100 or other media devices 100. PROX system 113 may use a variety of sensor technologies for SEN 195 including but not limited to ultrasound, infrared (IR), passive infrared (PIR), optical, acoustic, vibration, light, ambient light sensor (ALS), IR proximity sensors, LED emitters and detectors, RGB LED's, RF, temperature, capacitive, capacitive touch, inductive, just to name a few. PROX system 113 may be configured to sense location of users or other persons, user devices, and other media devices 100, without limitation. Output signals from PROX system 113 may be used to configure media device 100 or other media devices 100, to re-configure and/or re-purpose media device 100 or other media devices 100 (e.g., change a role the media device 100 plays for the user, based on a user profile or configuration data), just to name a few. A plurality of media devices 100 in an eco-system of media devices 100 may collectively use their respective PROX system 113 and/or other systems (e.g., RF 107, de-tunable antenna 124, AV 109, etc.) to accomplish tasks including but not limited to changing configuration, re-configuring one or more media devices, implement user specified configurations and/or profiles, insertion and/or removal of one or more media devices in an eco-system, just to name a few.
  • In other examples, PROX 113 may include one or more proximity detection islands PSEN 520 as will be discussed in greater detail in FIGS. 5-6. PSEN 520 may be positioned at one or more locations on chassis 199 and configured to sense an approach of a user or other person towards the media device 100 or to sense motion or gestures of a user or other person by a portion of the body such as a hand for example. PSEN 520 may be used in conjunction with or in place of one or more of SEN 195, OPT 185, SPK 160, MIC 170, RF 107 and/or de-tunable 129 antenna 124 to sense proximity and/or presence in an environment surrounding the media device 100, for example. PSEN 520 may be configured to take or cause an action to occur upon detection of an event (e.g., an approach or gesture by user 201 or other) such as emitting light (e.g., via an LED), generating a sound or announcement (e.g., via SPK 160), causing a vibration (e.g., via SPK 160 or a vibration motor), display information (e.g., via DISP 180), trigger haptic feedback, for example. In some examples, PSEN 520 may be included in I/O 105 instead of PROX 113 or be shared between one or more systems of media device 100. In other examples, components, circuitry, and functionality of PSEN 520 may vary among a plurality of PSEN 520 sensors in media device 100 such that all PSEN 520 are not identical.
  • Simple Out-of-the-Box User Experience
  • Attention is now directed to FIG. 2A, where a scenario 200 a depicts one example of a media device (e.g., media device 100 of FIG. 1 or a similarly provisioned media device) being configured for the first time by a user 201. For purposes of explanation, in FIG. 2A media device is denoted as 100 a to illustrate that it is the first time the media device 100 a is being configured. For example, the first configuration of media device 100 a may be after it is purchased, acquired, borrowed, or otherwise by user 201, that is, the first time may be the initial out-of-the-box configuration of media device 100 a when it is new. Scenario 200 a depicts a desirable user experience for user 201 to achieve the objective of making the configuring of media device 100 a as easy, straight forward, and fast as possible.
  • To that end, in FIG. 2A, scenario 200 a may include media device 100 a to be configured, for example, initially by user 201 using a variety of devices 202 including but not limited to a smartphone 210, a tablet 220, a laptop computer 230, a data capable wristband or the like 240, a desktop PC or server 250, . . . etc. For purposes of simplifying explanation, the following description will focus on tablet 220, although the description may apply to any of the other devices 202 as well. Upon initial power up of media device 100 a, controller 101 may command RF system 107 to electrically couple 224, transceiver BT 120 with antenna 124, and command BT 120 to begin listening 126 for a BT pairing signal from device 220. Here, user 201 as part of the initialization process may have already used a Bluetooth® menu on tablet 220 to activate the BT radio and associated software in tablet 220 to begin searching (e.g., via RF) for a BT device to pair with. Pairing may require a code (e.g., a PIN number or code) be entered by the user 201 for the device being paired with, and the user 201 may enter a specific code or a default code such as “0000”, for example.
  • Subsequently, after tablet 220 and media device 100 a have successfully BT paired with one another, the process of configuring media device 100 a to service the specific needs of user 201 may begin. In some examples, after successful BT pairing, BT 120 need not be used for wireless communication between media device 100 a and the user's device (e.g., tablet 220 or other). Controller 101, after a successful BT pairing, may command RF system 107 to electrically couple 228, WiFi 130 with antenna 124 and wireless communications between tablet 220 and media device 100 a (see 260, 226) may occur over a wireless network (e.g., WiFi or WiMAX) or other as denoted by wireless access point 270. Post-pairing, tablet 220 requires a non-transitory computer readable medium that includes data and/or executable code to form a configuration (CFG) 125 for media device 100 a. For purposes of explanation, the non-transitory computer readable medium will be denoted as an application (APP) 225. APP 225 resides on or is otherwise accessible by tablet 220 or media device 100 a. User 201 uses APP 225 (e.g., through a GUI, menu, drop down boxes, or the like) to make selections that comprise the data and/or executable code in the CFG 125.
  • APP 225 may be obtained by tablet 220 in a variety of ways. In one example, the media device 100 a includes instructions (e.g., on its packaging or in a user manual) for a website on the Internet 250 where the APP 225 may be downloaded. Tablet 220 may use its WiFi or Cellular RF systems to communicate with wireless access point 270 (e.g., a cell tower or wireless router) to connect 271 with the website and download APP 255 which is stored on tablet 220 as APP 225. In another example, tablet 220 may scan or otherwise image a bar code or TAG operative to connect the tablet 220 with a location (e.g., on the Internet 250) where the APP 225 may be found and downloaded. Tablet 220 may have access to an applications store such as Google Play for Android devices, the Apple App Store for iOS devices, or the Windows 8 App Store for Windows 8 devices. The APP 225 may then be downloaded from the app store. In yet another example, after pairing, media device 100 a may be preconfigured to either provide (e.g., over the BT 120 or WiFi 130) an address or other location that is communicated to tablet 220 and the tablet 220 uses the information to locate and download the APP 225. In another example, media device 100 a may be preloaded with one or more versions of APP 225 for use in different device operating systems (OS), such as one version for Android, another for iOS, and yet another for Windows 8, etc. In that OS versions and/or APP 225 are periodically updated, media device 100 a may use its wireless systems (e.g., BT 120 or WiFi 130) to determine if the preloaded versions are out of date and need to be replaced with newer versions, which the media device 100 a obtains, downloads, and subsequently makes available for download to tablet 220.
  • Regardless of how the APP 225 is obtained, once the APP 225 is installed on any of the devices 202, the user 201 may use the APP 225 to select various options, commands, settings, etc. for CFG 125 according to the user's preferences, needs, media device ecosystem, etc., for example. After the user 201 finalizes the configuration process, CFG 125 is downloaded (e.g., using BT 120 or WiFi 130) into DS system 103 in media device 100 a. Controller 101 may use the CFG 125 and/or other executable code to control operation of media device 100 a. In FIG. 2A, the source for APP 225 may be obtained from a variety of locations including but not limited to: the Internet 250; a file or the like stored in the Cloud; a web site; a server farm; a FTP site; a drop box; an app store; a manufactures web site; or the like, just to name a few. APP 225 may be installed using other processes including but not limited to: dragging and dropping the appropriate file into a directory, folder, desktop or the like on tablet 220; emailing the APP 225 as an attachment, a compressed or ZIP file; cutting and pasting the App 225, just to name a few.
  • CFG 125 may include data such as the name and password for a wireless network (e.g., 270) so that WiFi 130 may connect with (see 226) and use the wireless network for future wireless communications, data for configuring subsequently purchased devices 100, data to access media for playback, just to name a few. By using the APP 225, user 201 may update CFG 125 as the needs of the user 201 change over time, that is, APP 225 may be used to re-configure an existing CFG 125. Furthermore, APP 225 may be configured to check for updates and to query the user 201 to accept the updates such that if an update is accepted an updated version of the APP 225 may be installed on tablet 220 or on any of the other devices 202. Although the previous discussion has focused on installing the APP 225 and CFG 125, one skilled in the art will appreciate that other data may be installed on devices 202 and/or media device 100 a using the process described above. As one example, APP 225 or some other program may be used to perform software, firmware, or data updates on device 100 a. DS system 103 on device 100 a may include storage set aside for executable code (e.g., an operating system) and data used by controller 101 and/or the other systems depicted in FIG. 1.
  • Moving on to FIG. 2B, where a several example scenarios of how a previously configured media device 100 a that includes CFG 125 may be used to configure another media device 100 b that is initially un-configured. In scenario 200 b, media device 100 a is already powered up or is turned on (e.g., by user 201) or is otherwise activated such that its RF system 107 is operational. Accordingly, at stage 290 a, media device 100 a is powered up and configured to detect RF signatures from other powered up media devices using its RF system 107. At stage 290 b another media device denoted as 100 b is introduced into RF proximity of media device 100 a and is powered up so that its RF system 107 is operational and configured to detect RF signatures from other powered up media devices (e.g., signature of media device 100 a). Here RF proximity broadly means within adequate signal strength range of the BT transceivers 120, WiFi transceivers 130, or any other transceivers in RF system 107, RF systems in the users devices (e.g., 202, 220), and other wireless devices such as wireless routers, WiFi networks (e.g., 270), WiMAX networks, and cellular networks, for example. Adequate signal strength range is any range that allows for reliable RF communications between wireless devices. For BT enabled devices, adequate signal strength range may be determined by the BT specification, but is subject to change as the BT specification and technology evolve. For example, adequate signal strength range for BT 120 may be approximately 10 meters (e.g., ˜30 feet). For WiFi 130, adequate signal strength range may vary based on parameters such as distance from and signal strength of the wireless network, and structures that interfere with the WiFi signal. However, in most typical wireless systems adequate signal strength range is usually greater than 10 meters.
  • At stage 290 b, media device 100 b is powered up and at stage 290 c its BT 120 and the BT 120 of media device 100 a recognize each other. For example, each media device (100 a, 100 b) may be pre-configured (e.g., at the factory) to broadcast a unique RF signature or other wireless signature (e.g., acoustic) at power up and/or when it detects the unique signature of another device. The unique RF signature may include status information including but not limited to the configuration state of a media device. Each BT 120 may be configured to allow communications with and control by another media device based on the information in the unique RF signature. Accordingly, at the stage 290 c, media device 100 b transmits RF information that includes data that informs other listening BT 120's (e.g., BT 120 in 100 a) that media device 100 b is un-configured (e.g., has no CFG 125).
  • At stage 290 d, media devices 100 a and 100 b negotiate the necessary protocols and/or handshakes that allow media device 100 a to gain access to DS 103 of media device 100 b. At stage 290 e, media device 100 b is ready to receive CFG 125 from media device 100 a, and at stage 290 f the CFG 125 from media device 100 a is transmitted to media device 100 b and is replicated (e.g., copied, written, etc.) in the DS 103 of media device 100 b, such that media device 100 b becomes a configured media device.
  • Data in CFG 125 may include information on wireless network 270, including but not limited to wireless network name, wireless password, MAC addresses of other media devices, media specific configuration such as speaker type (e.g., left, right, center channel), audio mute, microphone mute, etc. Some configuration data may be subservient to other data or dominant to other data. After the stage 290 f, media device 100 a, media device 100 b, and user device 220 may wirelessly communicate 291 with one another over wireless network 270 using the WiFi systems of user device 220 and WiFi 130 of media devices 100 a and 100 b.
  • APP 225 may be used to input the above data into CFG 125, for example using a GUI included with the APP 225. User 201 enters data and makes menu selections (e.g., on a touch screen display) that will become part of the data for the CFG 125. APP 225 may also be used to update and/or re-configure an existing CFG 125 on a configured media device. Subsequent to the update and/or re-configuring, other configured or un-configured media devices in the user's ecosystem may be updated and/or re-configured by a previously updated and/or re-configured media device as described herein, thereby relieving the user 201 from having to perform the update and/or re-configure on several media devices. The APP 225 or a location provided by the APP 225 may be used to specify playlists, media sources, file locations, and the like. APP 225 may be installed on more than one user device 202 and changes to APP 225 on one user device may later by replicated on the APP 225 on other user devices by a synching or update process, for example. APP 225 may be stored on the internet or in the Cloud and any changes to APP 225 may be implemented in versions of the APP 225 on various user devices 202 by merely activating the APP 225 on that device and the APP 225 initiates a query process to see if any updates to the APP are available, and if so, then the APP 225 updates itself to make the version on the user device current with the latest version.
  • Media devices 100 a and 100 b having their respective WiFi 130 enabled to communicate with wireless network 270, tablet 220, or other wireless devices of user 201. FIG. 2B includes an alternate scenario 200 b that may be used to configure a newly added media device, that is, an un-configured media device (e.g., 100 b). For example, at stage 290 d, media device 100 a, which is assumed to already have its WiFi 130 configured for communications with wireless network 270, transmits over its BT 120 the necessary information for media device 100 b to join wireless network 270. After stage 290 d, media device 100 b, media device 100 a, and tablet 220 are connected 291 to wireless network 270 and may communicate wirelessly with one another via network 270. Furthermore, at stage 290 d, media device 100 b is still in an un-configured state. Next, at stage 290 e, APP 225 is active on tablet 220 and wirelessly accesses the status of media devices 100 a and 100 b. APP 225 determines that media device 100 b is un-configured and APP 225 acts to configure 100 b by harvesting CFG 125 (e.g., getting a copy of) from configured media device 100 a by wirelessly 293 a obtaining CFG 125 from media device 100 a and wirelessly 293 b transmitting the harvested CFG 125 to media device 100 b. Media device 100 b uses its copy of CFG 125 to configure itself thereby placing it in a configured state.
  • After all the devices 220, 100 a, 100 b, are enabled for wireless communications with one another, FIG. 2B depicts yet another example scenario where after stage 290 d, the APP 225 or any one of the media devices 100 a, 100 b, may access 295 the CFG 125 for media device 100 b from an external location, such as the Internet, the cloud, etc. as denoted by 250 where a copy of CFG 125 may be located and accessed for download into media device 100 b. APP 255, media device 100 b, or media device 100 a, may access the copy of CFG 125 from 250 and wirelessly install it on media device 100 b.
  • In the example scenarios depicted in FIG. 2B, it should be noted that after the pairing of media device 100 a and tablet 220 in FIG. 2A, the configuration of media device 100 b in FIG. 2B did not require tablet 220 to use its BT features to pair with media device 100 b to effectuate the configuration of media device 100 b. Moreover, there was no need for the BT pairing between tablet 220 and media device 100 a to be broken in order to effectuate the configuration of media device 100 b. Furthermore, there is no need for table 220 and media devices 100 a and/or 100 b to be BT paired at all with tablet 220 in order to configure media device 100 b. Accordingly, from the standpoint of user 201, adding a new media device to his/her ecosystem of similarly provisioned media devices does not require un-pairing with one or more already configured devices and then pairing with the new device to be added to the ecosystem. Instead, one of the already configured devices (e.g., media device 100 a having CFG 125 installed) may negotiate with the APP 225 and/or the new device to be added to handle the configuration of the new device (e.g., device 100 b). Similarly provisioned media devices broadly means devices including some, all, or more of the systems depicted in FIG. 1 and designed (e.g., by the same manufacture or to the same specifications and/or standards) to operate with one another in a seamless manner as media devices are added to or removed from an ecosystem.
  • Reference is now made to FIG. 3 where a flow diagram 300 depicts one example of configuring a first media device using an application installed on a user device as was described above in regards to FIG. 2A. At a stage 302 a Bluetooth® (BT) discovery mode is activated on a user device such as the examples 202 of user devices depicted in FIG. 2A. Typically, a GUI on the user device includes a menu for activating BT discovery mode, after which, the user device waits to pick up a BT signal of a device seeking to pair with the user's device. At a stage 304 a first media device (e.g., 100 a) is powered up (if not already powered up). At stage 306 a BT pairing mode is activated on the first media device. Examples of activating BT pairing mode include but are not limited to pushing a button or activating a switch on the first media device that places the first media device in BT pairing mode such that its BT 120 is activated to generate a RF signal that the user's device may discover while in discovery mode. I/O system 105 of media device 100 may receive 118 as a signal the activation of BT pairing mode by actuation of the switch or button and that signal is processed by controller 101 to command RF system 107 to activate BT 120 in pairing mode. In other examples, after powering up the first media device, a display (e.g., DISP 180) may include a touch screen interface and/or GUI that guides a user to activate the BT pairing mode on the first media device.
  • At a stage 308 the user's device and the first media device negotiate the BT pairing process, and if BT pairing is successful, then the flow continues at stage 310. If BT pairing is not successful, then the flow repeats at the stage 206 until successful BT pairing is achieved. At stage 310 the user device is connected to a wireless network (if not already connected) such as a WiFi, WiMAX, or cellular (e.g., 3G or 4G) network. At a stage 312, the wireless network may be used to install an application (e.g., APP 225) on the user's device. The location of the APP (e.g., on the Internet or in the Cloud) may be provided with the media device or after successful BT pairing, the media device may use its BT 120 to transmit data to the user's device and that data includes a location (e.g., a URI or URL) for downloading or otherwise accessing the APP. At a stage 314, the user uses the APP to select settings for a configuration (e.g., CFG 125) for the first media device. After the user completes the configuration, at a stage 316 the user's device installs the APP on the first media device. The installation may occur in a variety of ways (see FIG. 2A) including but not limited to: using the BT capabilities of each device (e.g., 220 and 100 a) to install the CFG; using the WiFi capabilities of each device to install the CFG; and having the first media device (e.g., 100 a) fetch the CFG from an external source such as the Internet or Cloud using its WiFi 130; just to name a few. Optionally, at stages 318-324 a determination of whether or not the first media device is connected with a wireless network may be made at a stage 318. If the first media device is already connected with a wireless network the “YES” branch may be taken and the flow may terminate at stage 320. On the other hand, if the first media device is not connected with a wireless network the “NO” branch may be taken and the flow continues at a stage 322 where data in the CFG is used to connect WiFi 130 with a wireless network and the flow may terminate at a stage 324. The CFG may contain the information necessary for a successful connection between WiFi 130 and the wireless network, such as wireless network name and wireless network password, etc.
  • Now reference is made to FIG. 4A, where a flow diagram 400 a depicts one example of a process for configuring an un-configured media device “B” (e.g., un-configured media device 100 b at stage 290 b of FIG. 2B) using a configured media device “A” (e.g., media device 100 a having CFG 125 of FIG. 2B). At a stage 402 an already configured media device “A” is powered up. At a stage 404 the RF system (e.g., RF system 107 of FIG. 1) of configured media device “A” is activated. The RF system is configured to detect RF signals from other “powered up” media devices. At a stage 406, an un-configured media device “B” (e.g., un-configured media device 100 b at stage 290 b of FIG. 2B) is powered up. At a stage 408 the RF system of un-configured media device “B” is activated. At stage 408, the respective RF systems of the configured “A” and un-configured “B” media devices are configured to recognize each other (e.g., via their respective BT 120 transceivers or another transceiver in the RF system). At a stage 410, if the configured “A” and un-configured “B” media devices recognize each other, then a “YES” branch is taken to a stage 412 where the configured media device “A” transmits its configuration (e.g., CFG 125) to the un-configured media device “B” (e.g., see stages 290 e and 290 f in FIG. 2B). If the configured “A” and un-configured “B” media devices do not recognize each other, then a “NO” branch is taken and the flow may return to an earlier stage (e.g., stage 404 to retry the recognition process. Optionally, after being configured, media device “B” may be connected with a wireless network (e.g., via WiFi 130). At a stage 414 a determination is made as to whether or not media device “B” is connected to a wireless network. If already connected, then a “YES” branch is taken and the process may terminate at a stage 416. However, if not connected with a wireless network, then a “NO” branch is taken and media device “B” is connected to the wireless network at a stage 418. For example, the CFG 125 that was copied to media device “B” may include information such as wireless network name and password and WiFi 130 is configured to effectuate the connection with the wireless network based on that information. Alternatively, media device “A” may transmit the necessary information to media device “B” (e.g., using BT 120) at any stage of flow 400 a, such as at the stage 408, for example. After the wireless network connection is made, the flow may terminate at a stage 420.
  • Attention is now directed to FIG. 4B, where a flow diagram 400 b depicts another example of a process for configuring an un-configured media device “B” (e.g., un-configured media device 100 b at stage 290 b of FIG. 2B) using a configured media device “A” (e.g., media device 100 a having CFG 125 of FIG. 2B). At a stage 422 an already configured media device “A” is powered up. At a stage 424 the RF system of configured media device “A” is activated (e.g., RF system 107 of FIG. 1). The RF system is configured to detect RF signals from other “powered up” media devices. At a stage 426, an un-configured media device “B” (e.g., un-configured media device 100 b at stage 290 b of FIG. 2B) is powered up. At a stage 428 the RF system of un-configured media device “b” is activated (e.g., RF system 107 of FIG. 1). At the stage 428, the respective RF systems of the configured “A” and un-configured “B” media devices are configured to recognize each other (e.g., via their respective BT 120 transceivers or another transceiver in the RF system). At a stage 430, if the configured “A” and un-configured “B” media devices recognize each other, then a “YES” branch is taken to a stage 432 where the configured media device “A” transmits information for a wireless network to the un-configured media device “B” (e.g., see stage 290 b in FIG. 2B) and that information is used by the un-configured media device “B” to connect with a wireless network as was described above in regards to FIGS. 2B and 4A. If the configured “A” and un-configured “B” media devices do not recognize each other, then a “NO” branch is taken and the flow may return to an earlier stage (e.g., stage 424 to retry the recognition process. At a stage 434, the information for the wireless network is used by the un-configured media device “B” to effectuate a connection to the wireless network. At a stage 436, a user device is connected with the wireless network and an application (APP) running on the user device (e.g., APP 225 in FIG. 2B) is activated. Stage 436 may be skipped if the user device is already connected to the wireless network. The APP is aware of un-configured media device “B” presence on the wireless network and at a stage 438 detects that media device “B” is presently in an un-configured state and therefore has a status of “un-configured.” Un-configured media device “B” may include registers, circuitry, data, program code, memory addresses, or the like that may be used to determine that the media device is un-configured. The un-configured status of media device “B” may be wirelessly broadcast using any of its wireless resources or other systems, such as RF 107 and/or AV 109. At a stage 440, the APP is aware of configured media device “A” presence on the wireless network and detects that media device “A” is presently in a configured state and therefore has a status of “configured.” The APP harvests the configuration (CFG) (e.g., CFG 125 of FIG. 2B) from configured media device “A”, and at a stage 442 copies (e.g., via a wireless transmission over the wireless network) the CFG to the un-configured media device “B.” At a stage 444, previously un-configured media device “B” becomes a configured media device “B” by virtue of having CFG resident in its system (e.g., CFG 125 in DS system 103 in FIG. 1). After media device “B” has been configured, the flow may terminate at a stage 446. In other examples, the APP may obtain the CFG from a location other than the configured media device “A”, such as the Internet or the Cloud as depicted in FIG. 2B. Therefore, at the stage 440, the APP may download the CFG from a web site, from Cloud storage, or other locations on the Internet or an intranet for example.
  • In the examples depicted in FIGS. 2A-4B, after one of the media devices is configured, additional media devices that are added by the user or are encountered by the user may be configured without the user (e.g., user 201) having to break a BT pairing with one media device and then establishing another BT pairing with a media device the user is adding to his/her media device ecosystem. Existing media devices that are configured (e.g., have CFG 125) may be used to configure a new media device using the wireless systems (e.g., acoustic, optical, RF) of the media devices in the ecosystem. If multiple configured media devices are present in the ecosystem when the user adds a new un-configured media device, configured media devices may be configured to arbitrate among themselves as to which of the configured devices will act to configured the newly added un-configured media device. For example, the existing media device that was configured last in time (e.g., by a date stamp on its CFG 125) may be the one selected to configure the newly added un-configured media device. Alternatively, the existing media device that was configured first in time (e.g., by a date stamp on its CFG 125) may be the one selected to configure the newly added un-configured media device. The APP 225 on the user device 220 or other, may be configured to make the configuration process as seamless as possible and may only prompt the user 201 that the APP 225 has detected an un-configured media device and query the user 201 as to whether or not the user 201 wants the APP 225 to configure the un-configured media device (e.g., media device 100 b). If the user replies “YES”, then the APP 225 may handle the configuration process working wirelessly with the configured and un-configured media devices. If the user 201 replies “NO”, then the APP 225 may postpone the configuration for a later time when the user 201 is prepared to consummate the configuration of the un-configured media device. In other examples, the user 201 may want configuration of un-configured media devices to be automatic upon detection of the un-configured media device(s). Here the APP and/or configured media devices would automatically act to configure the un-configured media device(s).
  • APP 225 may be configured (e.g., by the user 201) to automatically configure any newly detected un-configured media devices that are added to the user's 201 ecosystem and the APP 225 may merely inform the user 201 that it is configuring the un-configured media devices and inform the user 201 when configuration is completed, for example. Moreover, in other examples, once a user 201 configures a media device using the APP 225, subsequently added un-configured media devices may be automatically configured by an existing configured media device by each media device recognizing other media devices (e.g., via wireless systems), determining the status (e.g., configured or un-configured) of each media device, and then using the wireless systems (e.g., RF 107, AV 109, I/O 105, OPT 185, PROX 113) of a configured media device to configure the un-configured media device without having to resort to the APP 225 on the user's device 220 to intervene in the configuration process. That is, the configured media devices and the un-configured media devices arbitrate and effectuate the configuring of un-configured media devices without the aid of APP 225 or user device 220. In this scenario, the controller 101 and/or CFG 125 may include instructions for configuring media devices in an ecosystem using one or more systems in the media devices themselves.
  • In at least some examples, the structures and/or functions of any of the above-described features may be implemented in software, hardware, firmware, circuitry, or in any combination thereof. Note that the structures and constituent elements above, as well as their functionality, may be aggregated with one or more other structures or elements. Alternatively, the elements and their functionality may be subdivided into constituent sub-elements, if any. As software, the above-described techniques may be implemented using various types of programming or formatting languages, frameworks, scripts, syntax, applications, protocols, objects, or techniques. As hardware and/or firmware, the above-described techniques may be implemented using various types of programming or integrated circuit design languages, including hardware description languages, such as any register transfer language (“RTL”) configured to design field-programmable gate arrays (“FPGAs”), application-specific integrated circuits (“ASICs”), or any other type of integrated circuit. According to some embodiments, the term “module” may refer, for example, to an algorithm or a portion thereof, and/or logic implemented in either hardware circuitry or software, or a combination thereof. These may be varied and are not limited to the examples or descriptions provided. Software, firmware, algorithms, executable computer readable code, program instructions for execution on a computer, or the like may be embodied in a non-transitory computer readable medium.
  • Media Device with Proximity Detection
  • Attention is now directed to FIG. 5 where a profile view depicts one example 500 of media device 100 that may include on a top surface 199 s of chassis 199, a plurality of control elements 503-512 and one or more proximity detection islands (four are depicted) denoted as 520. Media device 100 may include one or more speakers 160, one or more microphones 170, a display 180, a section 550 for other functions such as SEN 195, VID 109, or other, and antenna 124 which may be tunable 129. Each proximity detection island 520 may be configured to detect 597 proximity of one or more persons, such as user 201 as will be described in greater detail below. The layout and position of the elements on chassis 199 of media device 100 are examples only and actual layout and position of any elements will be application specific and/or a matter of design choice, including ergonomic and esthetic considerations. As will be described in greater detail below, detection of presence of user 201 may occur with or without the presence of one or more user devices 202, such as user devices 210 and 220 depicted in FIG. 5. Circuitry and/or software associated with operation of proximity detection islands 520 may work in conjunction with other systems in media device 100 to detect presence of one or more user devices 202, such as RF system 107 detecting RF signals 563 and/or 565 (e.g., via antenna 124) from user devices 210 and 220 or MIC 170 detecting sound, for example. Detection of presence may be signaled by media device 100 in a variety of ways including but not limited to light (e.g., from 520 and/or 503-512), sound (e.g., from SPK 160), vibration (e.g., from SPK 160 or other), haptic feedback, tactile feedback, display of information (e.g., DISP 180), RF transmission (e.g., 126), just to name a few. SPK 160 and DISP 180 may be positioned on a front surface 199 f of chassis 199. A bottom surface 199 b of chassis 199 may be configured to rest on a surface such as a table, desk, cabinet, or the like. Other elements of media device 100 may be positioned on a rear surface 199 r of chassis 199.
  • Non-limiting examples of control elements 503-512 include a plurality of controls 512 (e.g., buttons, switches and/or touch surfaces) that may have functions that are fixed or change based on different scenarios as will be described below, controls 503 and 507 for volume up and volume down, control 509 for muting volume or BT paring, control 506 for initiating or pausing playback of content, control 504 for fast reversing playback or skipping backward one track, and control 508 for fast forwarding playback or skipping forward one track. Some are all of the control elements 504-512 may serve multiple rolls based on changing scenarios. For example, for playback of video content or for information displayed on display 180 (e.g., a touch screen), controls 503 and 507 may be used to increase “+” and decrease “−” brightness of display 180. Control 509 may be used to transfer or pick up a phone call or other content on a user device 202, for example. Proximity detection islands 520 and/or control elements 503-512 may be backlit (e.g., using LED's or the like) for night or low-light visibility.
  • Moving on to FIG. 6, a block diagram 600 depicts one example of a proximity detection island 520. Proximity detection island 520 may be implemented using a variety of technologies and circuit topologies and the example depicted in FIG. 6 is just one such non-limiting example and the present application is not limited to the arrangement of elements depicted in FIG. 6. One or more proximity detection islands 520 may be positioned on, connected with, carried by or otherwise mounted on media device 100. For example, proximity detection island 520 may be mounted on a top surface 199 t of chassis 199. A structure 650 made from an optically transmissive material such as glass, plastic, a film, an optically transparent or translucent material, or the like. Structure 650 may be made from a material that allows light 603, 607, 617, and 630 to pass through it in both directions, that is, bi-directionally. Structure 650 may include apertures 652 defined by regions 651 (e.g., an opaque or optically reflective/absorptive material) used for providing optical access (e.g., via apertures 652) to an environment ENV 198 external to the media device 100 for components of the proximity detection island 520. Structure 650 may be configured to mount flush with top surface 199 t, for example. In some examples, structure 650 may not include regions 651.
  • Proximity detection island 520 may include at least one LED 601 (e.g., an infrared LED-IR LED) electrically coupled with driver circuitry 610 and configured to emit IR radiation 603, at least one IR optical detector 605 (e.g., a PIN diode) electrically coupled with an analog-to-digital converter ADC 612 and configured to generate a signal in response to IR radiation 607 incident on detector 605, and at least one indicator light 616 electrically coupled with driver circuitry 614 and configured to generate colored light 617. As depicted, indicator light 616 comprises a RGB LED configured to emit light 617 in a gambit of colors indicative of status as will be described below. Here, RGB LED 616 may include four terminals, one of which coupled with circuit ground, a red “R” terminal, a green “G” terminal, and a blue “B” terminal, all of which are electrically connected with appropriate circuitry in driver 614 and with die within RGB LED 616 to effectuate generation of various colors of light in response to signals from driver 614. For example, RGB LED 616 may include semiconductor die for LED's that generate red, green, and blue light that are electrically coupled with ground and the R, G, and B terminals, respectively. One skilled in the art will appreciate that element 616 may be replaced by discrete LED's (e.g., separate red, green, white, and blue LED's) or a single non-RGB LED or other light emitting device may be used for 616. The various colors may be associated with different users who approach and are detected in proximity of the media device and/or different user devices that are detected by the media device. Therefore, if there are four users/and our user devices detected, then: the color blue may be associated with user #1; yellow with user #2; green with user #3; and red with user #4. Some users and or user devices may be indicated using alternating colors of light such as switching/flashing between red and green, blue and yellow, blue and green, etc. In other examples other types of LED's may be combined with RGB LED 616, such as a white LED, for example, to increase the number of color combinations possible.
  • Optionally, proximity detection island 520 may include at least one light sensor for sensing ambient light conditions in the ENV 198, such as ambient light sensor ALS 618. ALS 618 may be electrically coupled with circuitry CKT 620 configured to process signals from ALS 618, such as optical sensor 609 (e.g., a PIN diode) in response to ambient light 630 incident on optical sensor 609. Signals from CKT 620 may be further processed by ADC 622. The various drivers, circuitry, and ADC's of proximity detection island 520 may be electrically coupled with a controller (e.g., a μC, a μP, an ASIC, or controller 101 of FIG. 1) that is electrically coupled with a bus 645 (e.g., bus 110 of FIG. 1) that communicates signals between proximity detection island 520 and other systems of media device 100. Proximity detection island 520 may include auditory system AUD 624 configured to generate sound or produce vibrations in response to presence detection or other signals. AUD 624 may be mechanically coupled 641 with chassis 199 to cause chassis 199 to vibrate or make sound in response to presence detection or other signals. In some examples AUD 624 may use SPK 160 to generate sound or vibration. In other examples AUD 624 may use a vibration motor, such as the type used in smartphones to cause vibration when a phone call or notification is received. In yet another example, AUD 624 may use a piezoelectric film that deforms in response to an AC or DC signal applied to the film, the deformation generating sound and/or vibration. In yet other examples, AUD 624 may be connected with or mechanically coupled with one or more of the control elements and/or one or more of the proximity detection islands 520 depicted in FIG. 5 to provide haptic and/or tactile feedback. Upon detecting and acknowledging an approach by a user and/or user device, media may generate sound (e.g., from SPK 160) in a rich variety of tones and volume levels to convey information and/or media device status to the user. For example, a tone and volume level may be used to indicate the power status of the media device 100, such as available charge in BAT 135 of power system 111. The volume of the tone may be louder when BAT 135 is fully charged and lower for reduced levels of charge in BAT 135. Other tones and volume levels may be used to indicate the media device 100 is ready to receive input from the user or user device, the media device 100 is in wireless communications with a WiFi router or network, cellular service, broadband service, ad hoc WiFi network, other BT enabled devices, for example.
  • Proximity detection island 520 may be configured to detect presence of a user 201 (or other person) that enters 671 an environment 198 the media device 100 is positioned in. Here, entry 671 by user 201 may include a hand 601 h or other portion of the user 201 body passing within optical detection range of proximity detection island 520, such as hand 601 h passing over 672 the proximity detection island 520, for example. IR radiation 603 from IRLED 603 exiting through portal 652 reflects off hand 601 h and the reflected IR radiation 607 enters portal 652 and is incident on IR detector 605 causing a signal to be generated by ADC 612, the signal being indicative of presence being detected. RGB LED 616 may be used to generate one or more colors of light that indicate to user 201 that the user's presence has been detected and the media device is ready to take some action based on that detection. The action taken will be application specific and may depend on actions the user 201 programmed into CFG 125 using APP 225, for example. The action taken and/or the colors emitted by RGB LED 616 may depend on the presence and/or detection of a user device 210 in conjunction with or instead of detection of presence of user 201 (e.g., RF 565 from device 210 by RF 107).
  • As described above, proximity detection island 520 may optionally include ambient light sensor ALS 618 configured to detect ambient light 630 present in ENV 198 such as a variety of ambient light sources including but not limited to natural light sources such as sunny ambient 631, partially cloudy ambient 633, inclement weather ambient 634, cloudy ambient 635, and night ambient 636, and artificial light ambient 632 (e.g., electronic light sources). ALS 618 may work in conjunction with IRLED 610 and/or IR detector 605 to compensate for or reduce errors in presence detection that are impacted by ambient light 630, such as IR background noise caused by IR radiation from 632 or 631, for example. IR background noise may reduce a signal-to-noise ratio of IR detector 605 and cause false presence detection signals to be generated by ADC 612.
  • ALS 618 may be used to detect low ambient light 630 condition such as moonlight from 636 or a darkened room (e.g., light 632 is off), and generate a signal consistent with the low ambient light 630 condition that is used to control operation of proximity detection island 520 and/or other systems in media device 100. As one example, if user approaches 671 proximity detection island 520 in low light or no light conditions as signaled by ALS 618, RGB LED 616 may emit light 617 at a reduced intensity to prevent the user 201 from being startled or blinded by the light 617. Further, under low light or no light conditions AUD 624 may be reduced in volume or vibration magnitude or may be muted. Additionally, audible notifications (e.g., speech or music from SPK 160) from media device 100 may be reduced in volume or muted under low light or no light conditions (see FIG. 9).
  • Structure 650 may be electrically coupled 681 with capacitive touch circuitry 680 such that structure 650 is operative as a capacitive touch switch that generates a signal when a user (e.g., hand 601 h) touches a portion of structure 650. Capacitive touch circuitry 680 may communicate 682 a signal to other systems in media device 100 (e.g., I/O 105) that process the signal to determine that the structure 650 has been touched and initiate an action based on the signal. A user's touch of structure 650 may trigger driver 614 to activate RGB LED 616 to emit light 617 to acknowledge the touch has been received and processed by media device 100.
  • Reference is now made to FIG. 7, where top plan views of different examples of proximity detection island 520 configurations are depicted. Although the various example configurations and shapes are depicted as positioned on top surface 199 t of chassis 199, the present application is not so limited and proximity detection islands 520 may be positioned on other surfaces/portions of media device 100 and may have shapes different than that depicted. Furthermore, media device 100 may include more or fewer proximity detection islands 520 than depicted in FIG. 7 and the proximity detection islands 520 need not be symmetrically positioned relative to one another. Actual shapes of the proximity detection islands 520 may be application specific and may be based on esthetic considerations. Configuration 702 depicts five rectangular shaped proximity detection islands 520 positioned on top surface 199 t with four positioned proximate to four corners of the top surface 199 t and one proximately centered on top surface 199 t. Configuration 704 depicts three circle shaped proximity detection islands 520 proximately positioned at the left, right, and center of top surface 199 t. Configuration 706 depicts four hexagon shaped proximity detection islands 520 proximately positioned at the left, right, and two at the center of top surface 199 t. Finally, configuration 708 depicts two triangle shaped proximity detection islands 520 proximately positioned at the left, right of top surface 199 t. In some examples there may be a single proximity detection island 520. Proximity detection islands 520 may be configured to operate independently of one another, or in cooperation with one another.
  • Moving to FIG. 8A, a top plan view of proximity detection island 520 coverage is depicted. Each proximity detection island 520 may be designed to have a coverage pattern configured to detect presence of user 201 when the user 201 or portion of the user body (e.g., hand 801 h) enters the coverage pattern. Here, the coverage pattern may be semicircular 810 or circular 830, for example. Semicircular 810 coverage pattern may extend outward a distance R1 (e.g., approximately 1.5 meters) from proximity detection island 520 and may span a distance D1 about a center 871 of proximity detection island 520. Semicircular 810 coverage patterns of the four proximity detection islands 520 may not overlap one another such that there may be a coverage gap X1 and Y1 between the adjacent coverage patterns 810. Entry 825 of hand 801 h or entry 820 of user 201 may cause one or more of the proximity detection islands 520 to indicate 840 that a presence has been detected, by emitting a color of light from RGB LED 616, for example. In other examples, the coverage pattern may be circular 830 and cover a 360 degree radius 870 about a center point 871 of proximity detection island 520. Circular 830 coverage pattern 830 may or may not overlap the circular 830 pattern of the other proximity detection islands 520.
  • FIG. 8C depicts a front view 800 b of media device 100 and a coverage pattern 860 that has an angular profile Q about center point 871. Hand 801 h entering 825 into the coverage pattern 860 is detected by proximity detection island 520 and detection of hand 810 triggers light 840 being generate by RGB LED 616 of proximity detection island 520. Detection of hand 810 may also cause information “Info” to be displayed on DISP 180 and/or sound 845 to be generated by SPK 160. In FIG. 8C, a side view 800 c of media device 100 is depicted with proximity detection island 520 having angular profile a about center point 871 for a coverage pattern 880. Hand 801 h entering 825 into the coverage pattern 880 is detected by proximity detection island 520 and detection of hand 810 triggers light 840 being generate by RGB LED 616 of proximity detection island 520 and AUD 624 generating vibration 847.
  • Attention is now directed to FIG. 9, where a top plan view 900 of media device 100 depicts four proximity detection islands 520 denoted as I1, I2, I3, and I4. Furthermore, control elements 503-512 are depicted on top surface 199 t. In the example depicted, hand 901 h enters into proximity detection range of at least proximity detection island I1 and triggers generation of light (917 a-d) from one or more of the islands (I1, I2, I3, I4) such as light 617 from RGB LED 616 of FIG. 6, for example. Presence detection by proximity detection island I1 may cause a variety of response from media device 100 including but not limited to signaling that presence has been detected using light (917 a-d), generating sound 845 from SPK 160, vibration 847, displaying info 840 on DISP 180, capturing and acting on content C from user device 220, establishing wireless communications 126 with user device 220 or other wireless device (e.g., a wireless router), just to name a few. Presence detection by proximity detection island I1 may cause media device 100 to notify user 901 that his/her presence has been detected and the media device is ready to receive input or some other action from user 901. Input and/or action from user 901 may comprise user 901 actuating one of the control elements 503-512, touching or selecting an icon displayed on DISP 180, issuing a verbal command or speech detected by MIC 170.
  • As one example, upon detecting presence of user 901, media device 100 may emit light 917 c from proximity detection island I3. If the user device 220 is present and also detected by media device 100 (e.g., via RF signals 126 and/or 563), then the media device 100 may indicate that presence of the user device 220 is detected and may take one or more actions based on detecting presence of the user device 220. If user device 220 is one that is recognized by media device 100, then light 917 c from proximity detection island I3 may be emitted with a specific color assigned to the user device 220, such as green for example. Recognition of user device 220 may occur due to the user device 220 having been previously BT paired with media device 100, user device 220 having a wireless identifier such as a MAC address or SSID stored in or pre-registered in media device 100 or in a wireless network (e.g., a wireless router) the media device 100 and user device 220 are in wireless communications with, for example. DISP 180 may display info 840 consistent with recognition of user device 220 and may display via a GUI or the like, icons or menu selections for the user 201 to choose from, such as an icon to offer the user 201 a choice to transfer content C from user device 220 to the media device 100, to switch from BT wireless communication to WiFi wireless communication, for example. As one example, if content C comprises a telephone conversation, the media device 100 through instructions or the like in CFG 125 may automatically transfer the phone conversation from user device 220 to the media device 100 such that MIC 170 and SPK 160 are enabled so that media device 100 serves as a speaker phone or conference call phone and media device 100 handles the content C of the phone call. If the transfer of content C is not automatic, CFG 125 or other programming of media device 100 may operate to offer the user 201 the option of transferring the content C by displaying the offer on DISP 180 or via one of the control elements 503-512. For example, control element 509 may blink (e.g., via backlight) to indicate to user 201 that actuating control element 509 will cause content C to be transferred from user device 220 to media device 100.
  • In some examples, control elements 503-512 may correspond to menu selections displayed on DISP 180 and/or a display on the user device 220. For example, control elements 512 may correspond to six icons on DISP 180 (see 512′ in FIG. 8) and user 201 may actuate one of the control elements 512 to initiate whatever action is associated with the corresponding icon on DISP 180, such as selecting a playlist for media to be played back on media device 100. Or the user 201 may select one of the icons 512′ on DISP 180 to effectuate the action.
  • As one example, if content C comprises an alarm, task, or calendar event the user 201 has set in the user device 220, that content C may be automatically transferred or transferred by user action using DISP 180 or control elements 503-512, to media device 100. Therefore, a wake up alarm set on user device 220 may actually be implemented on the media device 100 after the transfer, even if the user device 220 is powered down at the time the alarm is set to go off. When the user device is powered up, any alarm, task, or calendar event that has not been processed by the media device 100 may be transferred back to the user device 220 or updated on the user device so that still pending alarm, task, or calendar events may be processed by the user device when it is not in proximity of the media device 100 (e.g., when user 201 leaves for a business trip). CFG 125 and APP 225 as described above may be used to implement and control content C handling between media device 100 and user devices.
  • Some or all of the control elements 503-512 may be implemented as capacitive touch switches. Furthermore, some or all of the control elements 503-512 may be backlit (e.g., using LED's, light pipes, etc.). For example, control elements 512 may be implemented as capacitive touch switches and they may optionally be backlit. In some examples, after presence is detected by one or more of the proximity detection islands (I1, I2, I3, I4), one or more of the control elements 503-512 may be backlit or have its back light blink or otherwise indicate to user 201 that some action is to be taken by the user 201, such as actuating (e.g., touching) one or more of the backlit and/or blinking control elements 512. In some examples, proximity detection islands (I1, I2, I3, I4) may be configured to serve as capacitive touch switches or another type of switch, such that pressing, touching, or otherwise actuating one or more of the proximity detection islands (I1, I2, I3, I4) results in some action being taken by media device 100.
  • In FIG. 9, actions taken by media device 100 subsequent to detecting presence via proximity detection islands (I1, I2, I3, I4) and/or other systems such as RF 107, SEN 195, MIC 170, may be determined in part on ambient light conditions as sensed by ALS 618 in proximity detection islands (I1, I2, I3, I4). As one example, if ambient light 630 is bright (e.g., 631 or 632), then brightness of DISP 180 may be increased, light 917 a-d from islands may be increased, and volume from SPK 160 may be nominal or increased because the ambient light 630 conditions are consistent with waking hours were light intensity and volume may not be a distraction to user 201. On the other hand, if ambient light 630 is dim or dark (e.g., 636), then brightness of DISP 180 may be decreased, light 917 a-d from islands may be decreased, and volume from SPK 160 may be reduced or muted because the ambient light 630 conditions are consistent with non-waking hours were light intensity and volume may be a distraction to or startle user 201. Other media device 100 functions such as volume level, for example, may be determined based on ambient light 630 conditions (e.g., as detected by ALS 618 of island I4). As one example, under bright ambient light 630 conditions, volume VH of SPK 160 may be higher (e.g., more bars); whereas, under low ambient light 630 conditions, volume VL of SPK 160 may be lower (e.g., fewer bars) or may be muted entirely VM. Conditions other than ambient light 630 may cause media device 100 to control volume as depicted in FIG. 9.
  • FIG. 10 depicts one example of a flow 1000 for presence detection, notification, and media device readiness. At a stage 1002 a query as to whether or not an approach is detected by one or more of the proximity detection islands (e.g., I1, I2, I3, I4) is made. Here, the query may be by controller CNTL 640 or controller 101, for example. If one or more of the proximity detection islands have detected presence, then a YES branch is taken. If no presence is detected by one or more of the proximity detection islands, then a NO branch is taken and the flow 1000 may return to the stage 1002 to wait for one or more of the proximity detection islands to detect a presence. The YES branch takes flow 1000 to a stage 1004 where a notification is executed by the media device 100 using light, sound, or vibration to notify a user that presence has been detected, for example, using one or more colors of light (e.g., from RGB LED's 616) and/or an auditory cue (e.g., from SPK 160, vibration from 847, or from a passive radiator used as one of the SPK 160). At as stage 1006, the media device 100 indicates that it is ready to receive input from a user and/or user device (e.g., user 201 or a user device 220 via RF 107). At a stage 1008 a query is made as to whether or not an input is received from a user. If an input is received from the user and/or user device, then a YES branch is taken to a stage 1010 where the media device 100 takes an appropriate action based on the type of user input received and the flow may terminate after the stage 1010. Appropriate actions taken by media device 100 will be application dependent and may be determined in whole or in part by APP 225, CFG 125, executable program code, hardware, etc. Inputs from the user includes but is not limited to actuation of one or more of the control elements 503-512, touching an icon or other area of DISP 180, issuing a spoken command or speech detected by MIC 170, taking an action on user device 220 that is wirelessly communicated to media device 100, just to name a few. If no input is received from the user and/or user device, then a NO branch is taken and the flow 1000 may continue at a stage 1012 where flow 1000 may enter into a wait period of predetermined time (e.g., of approximately 15 seconds or one minute, etc.). If a user input is received before the wait period is over, then a NO branch may be taken to the stage 1010. If the wait period is over, then a YES branch may be taken and flow 1000 may resume at the stage 1002.
  • FIG. 11 depicts another example of a flow 1100 for presence detection, notification, and media device readiness. At a stage 1102 a query as to whether an approach is detected by one or more of the proximity detection islands (e.g., I1, I2, I3, I4) is made. If one or more of the proximity detection islands have detected presence, then a YES branch is taken. If no presence is detected by one or more of the proximity detection islands, then a NO branch is taken and the flow 1100 may return to the stage 1102 to wait for one or more of the proximity detection islands to detect a presence. The YES branch takes flow 1100 to a stage 1104 where a query is made as to whether or not ambient light (e.g., ambient light 630 as detected by ALS 618 of FIG. 6) is a factor to be taken into consideration in the media devices response to having detected a presence at the stage 1102. If ambient light is not a factor, then a NO branch is taken and the flow 1100 continues to a stage 1106. If ambient light is a factor, then a YES branch is taken and flow 1100 continues at a stage 1108 where any notification by media device 100 in response to detecting presence at the stage 1102 is modified. One or more of light, sound, or vibration may be used by media device 100 to indicate to a user that its presence has been detected. The light, sound, or vibration are altered to comport with the ambient light conditions, such as described above in regard to ambient light 630 in FIG. 9, for example. At the stage 1106, notification of presence being detected occurs using one or more of light, sound, or vibration without modification. At a stage 1110, the media device 100 indicates that it is ready to receive input from a user and/or user device (e.g., user 201 or a user device 220 via RF 107). At a stage 1112 a query is made as to whether or not an input is received from a user. If an input is received from the user and/or user device, then a YES branch is taken to a stage 1114 where the media device 100 takes an appropriate action based on the type of user input received and the flow may terminate after the stage 1114. If no input is received from the user and/or user device, then a NO branch is taken and the flow 1110 may continue at a stage 1116 where flow 1100 may enter into a wait period of predetermined time (e.g., of approximately 15 seconds or one minute, etc.). If a user input is received before the wait period is over, then a NO branch may be taken to the stage 1114. If the wait period is over, then a YES branch may be taken and flow 1100 may resume at the stage 1102. Actions taken at the stage 1114 may include those described above in reference to FIG. 10.
  • FIG. 12 depicts yet another example of a flow 1200 for presence detection, notification, and media device readiness. At a stage 1202 a query as to whether an approach is detected by one or more of the proximity detection islands (e.g., I1, I2, I3, I4) is made. If one or more of the proximity detection islands have detected presence, then a YES branch is taken. If no presence is detected by one or more of the proximity detection islands, then a NO branch is taken and the flow 1200 may return to the stage 1202 to wait for one or more of the proximity detection islands to detect a presence. The YES branch takes flow 1200 to a stage 1204 where a query is made as to whether or not detection of RF (e.g., by RF 107 using antenna 124) is a factor to be taken into consideration in the media devices response to having detected a presence at the stage 1202. If RF detection is not a factor, then a NO branch is taken and the flow 1200 continues to a stage 1206. If RF detection is a factor, then a YES branch is taken and flow 1200 continues at a stage 1208 where any notification by media device 100 in response to detecting presence at the stage 1202 is modified. One or more of light, sound, or vibration may be used by media device 100 to indicate to a user that its presence has been detected. The light, sound, or vibration are altered to comport with the detection of RF (e.g., from a user device 220), such as described above in regards to user device 220 in FIG. 9, for example. At the stage 1206, notification of presence being detected occurs using one or more of light, sound, or vibration without modification. At a stage 1210, the media device 100 indicates that it is ready to receive input from a user and/or user device (e.g., user 201 or a user device 220 via RF 107). At a stage 1212 a query is made as to whether or not an input is received from a user. If an input is received from the user and/or user device, then a YES branch is taken to a stage 1214 where the media device 100 takes an appropriate action based on the type of user input received and the flow may terminate after the stage 1214. If no input is received from the user and/or user device, then a NO branch is taken and the flow 1200 may continue at a stage 1216 where flow 1200 may enter into a wait period of predetermined time (e.g., of approximately 15 seconds or one minute, etc.). If a user input is received before the wait period is over, then a NO branch may be taken to the stage 1214. If the wait period is over, then a YES branch may be taken and flow 1200 may resume at the stage 1202. Actions taken at the stage 1214 may include those described above in reference to FIGS. 9 and 10.
  • FIG. 13 depicts one example 1300 of presence detection using proximity detection islands and/or other systems responsive to wireless detection of different users (e.g., hands 1300 a-d) and/or different user devices (e.g., 220 a-220 d). In FIG. 13 four users denoted by hands 1300 a-d and their respective user devices 220 a-220 b enter 925 proximity detection range of one or more of the proximity detection islands (I1, I2, I3, I4). Although four users and four user devices are depicted, there may be more or fewer than depicted in FIG. 13. Detection of user devices 220 a-220 b may be through wireless means such as RF 107 (e.g., via antenna 124/129) and its various transceivers wirelessly communicating 126 or wirelessly detecting RF 563 from those user devices. For example, considering just one of the users and one of the user devices, hand 1300 b enters 925 detection range of proximity detection island I2 and is detected 597 by island I2. Island I2 notifies user via light 1317 b that his/her presence has been detected. User device 220 b may be carried by the user at the same time or at approximately the same time as the user's presence is detected by island I2. Therefore, RF 107 may detect RF 563, may attempt to wirelessly connect 126, or be in wireless 126 communications with user device 220 b. Accordingly, notifications and actions described above in regards to flow 1200 of FIG. 12 may occur in media device 100 in response to detecting presence 597 at or near the same time as detecting RF from a user device. Media device 100 may emit sound 1345, vibrate 847, display information info on DISP 180, generate light 1317 a-1317 d, await actuation of one or more of the control elements 503-512, or other action(s), for example. At the same time or at different times, other users denoted by hands 1300 a, 1300 c, and 1300 d may be detected 597 by one or more of the proximity detection islands (I1, I2, I3, I4) along with RF 563 from user devices 220 a, 220 c, and 220 d being detected by RF 107. Media device 100 may take appropriate action(s) and make appropriate notification(s) as described herein in response to proximity detection and RF detection occurring in close time proximity to one another, simultaneously, nearly simultaneously, or in some sequence. In that a range for RF transmissions may typically be greater than a detection range for the proximity detection islands (I1, I2, I3, I4), in some examples the RF signatures or signals of user device 220 a-d may be detected by RF 107 before the proximity detection islands (I1, I2, I3, I4) detect presence of the users 1300 a-d. For example, RF 107 may detect RF 563 before the user device emitting RF 563 is approximately 10 meters or more away from media device 100 (e.g., for BT transmissions) or much more than 10 meters away for other wireless technologies (e.g., for WiFi transmissions). Therefore, in some examples, RF 107 will detect RF signals prior to proximity detection islands (I1, I2, I3, I4) detecting presence 597.
  • Users devices 220 a-220 d may be pre-registered or otherwise associated or known by media device 100 (e.g., via CFG 125 or other) and the actions taken and notifications given by the media device 100 may depended on and may be different for each of the user devices 220 a-220 d. For example, after detection and notification based on detecting proximity 597 and RF 563 for user device 220 a, media device 100 may establish or re-establish BT pairing (e.g., via BT 120 in RF 107) with 220 a and content C on 220 a (e.g., a phone conversation) may be transferred to media device 100 for handling via SPK 160 and MIC 170. CFG 125 and/or APP 225 on 220 a may affect how media device and user device 220 a operate post detection.
  • As another example, post detection 597 & 563 and notification for user device 220 d may result in content C (e.g., music from MP3 files) on 220 d being played back 1345 on media device 100. Control elements 503-512 may be activated (if not already activated) to play/pause (506), fast forward (508), fast reverse (504), increase volume (503), decrease volume (507), or mute volume (509). Control elements 512 may be used to select among various play lists or other media on user device 220 d.
  • In another example, content C on user device 220 c may, post detection and notification, be displayed on DISP 180. For example, a web page that was currently being browsed on 220 c may be transferred to media device 100 for viewing and browsing, and a data payload associated with the browsing may also be transferred to media device 100. If content C comprises a video, the display and playback functions of the video may be transferred to media device 100 for playback and control, as well as the data payload for the video.
  • Content C this is transferred to media device 100 may be transferred back in part or whole to the user devices depicted, when the user is no longer detectable via islands to proximity detection islands (I1, I2, I3, I4) or other systems of media device 100, by user command, or by user actuating one of the control elements 503-512 or an icon or the like on DISP 180, for example.
  • FIG. 14 depicts one example 1400 of proximity detection islands associated with specific device functions. Examples of functions that may be assigned to or fixed to a proximity detection island (I1, I2, I3, I4) include but are not limited to “Set Up” of media device 100, “BT Paring” between media device 100 and one or more BT equipped devices, “Shut-Off” of media device 100 (e.g., power off or placing media device 100 in a standby mode, a low power consumption mode, or a sleep mode), and “Content” being handled by media device 100, such as the last media filed that was played on, the last buffered channel, the last playlist that was being accessed by, or the last Internet site or stream being handled by media device 100. One or more of proximity detection islands (I1, I2, I3, I4) may serve as indicators for the functions associated with them or may serve to actuate those functions by pressing or touching a surface of the island (e.g., as a switch or capacitive touch switch or button, see FIG. 6). For example, a finger of hand 1400 h may touch structure 650 of island I2 to activate the “BT Pairing” between the media device 100 and user device 220, the touch activating the capacitive touch function of island I2 (e.g., causing island I2 to serve as a switch). Island I2 may emit light 1417 b to acknowledge the touch by hand 1400 h. CFG 125 and/or APP 225 may be used to assign and re-assign functions to one or more of the proximity detection islands (I1, I2, I3, I4) and the functions assigned and the proximity islands they are assigned to may be user dependent and/or user device dependent. As another example, pressing or touching island I4 may turn power off to the media device 100, or may place media device 100 in a low power, standby, or sleep mode.
  • In other examples, one or more of the control elements 503-512 or an icon or the like on DISP 180 may be actuated or selected by a user in connection with one of the functions assigned to proximity detection islands (I1, I2, I3, I4). For example, to activate the “BT Pairing” function of island I2, control element 512 that is nearest 1427 to island I2 may be actuated by the user. In another example, proximity detection islands (I1, I2, I3, I4) may be associated with different users whose presence has been detected by one or more of the islands. For example, if proximity of four users (U1, U2, U3, U4) has been detected by any of the islands, then U1 may be associated with I4, U2 with I1, U3 with I2, and U4 with I3. Association with an island may be used to provide notifications to the user, such as using light from RGB LED 616 to notify the user of status (e.g., BT pairing status) or other information.
  • FIG. 15 depicts one example 1500 of content handling from a user device subsequent to proximity detection by islands 520 and/or wireless systems of media device 100. User 1500 h is detected 1540 by proximity detection island 520 which emits light 1517, sound 1545, vibration 847, and display of information info on DISP 180 to indicate that media device 100 has detected presence and is ready to receive user input. User device 220 may also have been detected by a transceiver RXTX 1507 in RF 107. RXTX 1507 may represent any transceiver in RF 107 such as BT 120, WiFi 130, AH 140, or other 150. Media device 100, post detection, may be wirelessly connected with user device 220 using a variety of wireless paths such as a direct wireless connection 126 between media device 100 and user device 220, and wireless connections 1565 and 1563 via wireless router 1570, for example. Content C on user device 220 may be handled or otherwise stored or routed to media device from the user device 220 or from Cloud 1550 using a variety of wireless paths. Cloud 1550 may represent the Internet, an intranet, a server farm, a download site, a music store, and application store, Cloud storage, a web site, just to name a few. Information including but not limited to content C, data D, a playlist PL, a stream or streaming service S, and a URL, just to name a few. Although content C is depicted as being presently on user device 220, one or more of the information in Cloud 1550 may also be presently on user device or wirelessly accessible to user device 220 via wireless connections 1561, 1563, 1567, 126, 1569, and 1565. Some of the wireless connections may be made through wireless router 1570 or media device 100 (e.g., via WiFi 130).
  • In some examples, content C or other information resident or accessible to user device 220 may be handled by media device 100. For example, if C comprises media files such as MP3 files, those files may be wirelessly accessed by media device 100 by copying the files to DS 103 (e.g., in Flash memory 145) thereby taking the data payload and wireless bandwidth from the user device 220 to the media device 100. Media device 100 may use it wireless systems to access 1569 or 1565 and 1567 the information from Cloud 1550 and either store the information locally in DA 103 or wirelessly access the information as it is played back or otherwise consumed or used by media device 100. APP 225 and CFG 125 may include information and executable instructions that orchestrate the handling of content between media device 100, user device 220, and Cloud 1550. For example, a playlist PL on user device 220 may be located in Cloud 1550 and media files associated with music/videos in the PL may be found at URL in Cloud 1550. Media device 100 may access the media files from the location specified by the URL and wirelessly stream the media files, or media device may copy a portion of those media files to DS 103 and then playback those files from its own memory (e.g., Flash 145).
  • In other examples, user 1500 h may be one of many users who have content to be accessed and/or handled by media device 100. Post detection, songs, play lists, content, of other information on user device 220 or from Cloud 1550 may be placed in a queue with other information of similar type. The queue for songs may comprise Song 1 through Song N and songs on user device 220 that were active at the time of proximity detection may be placed in some order within the queue, such as Song 4 for being fourth in line in queue for playback on media device 100. Other information such as play lists PL 1-PL N or other content such as C 1-CN may be placed in a queue for subsequent action to be taken on the information once it has moved to the top of the queue. In some examples, the information on user device 220 or from Cloud 1550 may be buffered in media device 100 by storing buffered data in DS 103. FIG. 16 depicts another example of content handling from user devices subsequent to proximity detection. In FIG. 16, a plurality of users 1601 a-1601 n and their associated user device 220 are detected by media device 100 are queued into DS 103 on media device 100 for handling or are buffered BUFF into DS 103 in some order. Detection of each user and or user device may be indicated with one or more different colors of light 1517, different sounds 1545, different vibration 847 patterns, or different info on DISP 180. In some examples, buffering BUFF occurs in storage 1635 provided in Cloud 1550. In FIG. 16, users 1601 a-1601 n have information on their respective user devices 220 that may be handled by media device 100 such as Song 1-Song N, PL 1-PL N, C 1-C N. The information from the plurality of users 1601 a-1601 n is queue and/or buffered BUFF on media device 100 and/or in Cloud 1550, that is, media device may handle all of the information internally, in Cloud 1550, or some combination of media device 100 and Cloud 1550. For example, if a data storage capacity of the information exceeds a storage capacity of DS 103, then some or all of the data storage may be off loaded to Cloud 1550 (e.g., using Cloud storage or a server farm). Information from users 1601 a-1601 n may be played back or otherwise handled by media device 100 in the order in which proximity of the user was detected or in some other order such as a random order or a shuffle play order. For example, DISP 180 may have an icon RDM which may be selected for random playback.
  • FIG. 17 depicts one example of content handling from a data capable wristband or wristwatch subsequent to proximity detection by a media device. A hand 1700 h of a user may comprise a user device in the form of a data capable wristband or wristwatch denoted as 1740. Wristband 1740 may include information “I” that is stored in the wristband 1740 and is wirelessly accessible using a variety of wireless connections between media device 100, wireless router 1570, and Cloud 1750. Media device 100 may serve as a wireless hub for wristband 1740 allowing wristband 1740 to send and retrieve information from Cloud 1750 via wireless connections between media device 100 and wireless router 1570 and/or Cloud 1750. For example, wristband 1740 may use BT to wirelessly communicate with media device 100 and media device 100 uses its WiFi 130 to wirelessly communicate with other resources such as Cloud 1750 and router 1570. Detection 1540 of hand 1700 h and/or device 1740 may trigger the emission of light 1517, generation of sound 1545, vibration 847, and display of information info on DISP 180.
  • Information “I” included in wristband 1740 may include but is not limited to alarms A, notifications N, content C, data D, and a URL. Upon detection of proximity, any of the information “I” may be wirelessly communicated from wristband 1740 to media device 100 where the information “I” may be queued (A 1-A N; D 1-D N, N1-N n; and C 1-C N) and/or buffered BUFF as described above. In some examples, post detection, wristband 1740 may wirelessly retrieve and/or store the information “I” from the media device 100, the Cloud 1750, or both. As one example, if wristband 1740 includes one or more alarms A, post detection those alarms A may be handled by media device 100. Therefore, if one of the alarms A is set to go off at 6:00 pm and detection occurs at 5:50 pm, then that alarm may be handled by media device 100 using one or more of DISP 180, SPK 160, and vibration 847, for example. If another alarm is set for 5:30 am and the wristband 1740 and media device 100 are still in proximity of each another at 5:30 am, then the media device 100 may handle the 5:30 am alarm as well. The 6:00 pm and 5:30 am alarms may be queued in the alarms list as one of A 1-AN. When wristband 1740 and media device 100 are no longer in proximity of each other, any alarms not processed by media device 100 may be processed by wristband 1740.
  • In FIG. 18, a plurality of users 1801 a-1801 n and their respective wristwatches 1740 are detected by one or more proximity detection islands 520 of media device 100 and/or or other systems such as RF 107. Detection of each user and or device 1740 may be indicated with one or more different colors of light 1517, different sounds 1545, different vibration 847 patterns, or different info on DISP 180. Here, each wristwatch 1740 includes information “I” specific to its user and as each of these users and wristwatches come into proximity and are detected, information “I” may be queued, buffered BUFF, or otherwise stored or handled by media device 100 or in Cloud 1750. For example, data D may include exercise, nutrition, dietary data, and biometric information collected from or sensed via sensors carried by the wristwatch 1740. Data D may be transferred to media device 100 or Cloud 1750 and accessed via a URL to a web page of a user. The data D may be shared among other users via their web pages. For example, some or all of users 1801 a-1801 n may be consent to sharing their information “I” through media device 100, Cloud 1750, or both. Users 1801 a-1801 n may view each other's information “I” on DISP 180 or go to a URL in Cloud 1750 or the like to view each other's information “I”. Information “I” that is displayer on DISP 180 may be buffered BUFF, queued (A 1-A N; D 1-D N, N1-N n; and C 1-C N), or otherwise stored on media device 100 (e.g., in DS 103) for each user to query as desired. A non-transitory computer readable medium such as CFG 125 and/or APP 225 may be used to determine actions taken by wristwatch 1740 (e.g., via APP 225) and media device (e.g., via CFG 125).
  • In FIG. 19, one example of a flow 1900 for content C handling on a media device 100 or other location, post proximity detection includes the media device 100 accessing the content C at a stage 1902. Here, accessing may include negotiating the necessary permissions, user names and passwords, or other tasks necessary to gain access to the content C on a user device or located elsewhere (e.g., in the Cloud, on a website, or on the Internet). Accessing the content C may include wirelessly connecting with the user device or other source of the content C. At a stage 1904 the media device 100 makes a determination is made as to the type of the content C, such as a media file (e.g., music, video, pictures), a web page (e.g., a URL), a file, a document (e.g., a PDF file), for example. At a stage 1906 the media device 100 makes a determination as to a status of the content C. Examples of status include but are not limited to static content C (e.g., a file) and dynamic content C (e.g., a stream or a file currently being accessed or played back). At a stage 1908 the media device 100 handles the content C based on its type and status from stages 1904 and 1906.
  • In that there may be many user devices to service post proximity detection or more than one item of content C to be handled from one or more user devices, at a stage 1910 media device 100 queries the user devices to see if there is additional content C to be handled by the media device 100. If additional content exists, then a YES branch may be taken and flow 1900 may return to stage 1902. If no additional content C is to be handled, then a NO branch may be taken and at a stage 1912 a decision to terminate previously handled content C may be made. Here, a user device may have handed over content C handling to media device 100 post proximity detection, but when the user device moves out of RF and/or proximity detection range (e.g., the user leaves with his/her user device in tow), then media device 100 may release or otherwise divorce handling of the content C. If previously handled content C does not require termination, then a NO branch may be taken and flow 1900 may end. On the other hand, if previously handled content C requires termination, then a YES branch may be taken to a stage 1914 were the previously handled content C is released by the media device 100. Release by media device 100 includes but is not limited to wirelessly transferring the content C back to the user device or other location, deleting the content C from memory in the media device 100 or other location, saving, writing or redirecting the content C to a location such as /dev/null or a waste basket/trash can, halting streaming or playback of the content C, storing the content C to a temporary location, just to name a few.
  • FIG. 20 depicts one example of a flow 2000 for storing, recording, and queuing content C on a media device 100 or other location post proximity detection. After content C has been handled by media device 100 (e.g., stage 1908 of FIG. 19), media device 100 may determine a size (e.g., file size) of the content C at a stage 2002. The size determination may be made in order for the media device 100 to determine if the media device 100 has the memory resources to handle and/or store the content C. If the media device 100 cannot accommodate content C due to size, then media device 100 may select another source for the content C or access the content from the user device or other location where it is stored. At a stage 2004 the media device 100 determines whether or not the content C is dynamic. Examples of dynamic content C include but are not limited to content C on a user device that is currently being accessed or played back on the user device. The dynamic content C may reside on the user device or may be accessed from another location (e.g., the Cloud or Internet). If the content C is not dynamic (e.g., is static such as file), then a NO branch may be taken to a stage 2010 where the media device 100 selects an appropriate location to store content C based on its size from the stage 2002. Examples of appropriate locations include but are not limited to a user device, the Cloud, the Internet, an intranet, network attached storage (NAS), a server, and DS 103 of media device 100 (e.g., in Flash memory 145). In some examples, media device 100 may include a memory card slot for a SD card, microSD card, Memory Stick, SSD, CF card, or the like, or a USB connector that will accommodate a USB thumb drive or USB hard drive, and those memory devices may comprise an appropriate location to store content C. At a stage 2012 the content C is stored to the selected location. If the content C is dynamic, then a YES branch may be taken to a stage 2006 where memory device 100 selects an appropriate location to record the dynamic content C based on the size of the content C. Appropriate locations include but are not limited to those described above for the stage 2010. At a stage 2008 the media device 100 records the dynamic content to the selected location. The selected location may be a buffer such as BUFF described above. At a stage 2014 a determination may be made as to whether or not the recording is complete. If the recording is not complete, then a NO branch may be taken and flow 2000 may return to the stage 2008. If the recording is complete, then a YES branch may be taken to a stage 2016 where a decision to queue the content C is made. If the content C is not to be queued, then a NO branch may be taken and the flow 2000 may end. If the content C is to be queued then a YES branch may be taken and at a stage 2018 the recorded content C or stored content C (e.g., from stage 2012) is queued. Queuing may occur as described above in reference to FIGS. 15-18. Media device 100 may maintain the queue in memory, but the actual content C need not be stored internally in memory device 100 and may be located at some other location such as the Cloud or a user device, for example.
  • At the stage 2008, the media device 100 may playback other content C (e.g., an mp3 or mpeg file) while recording the content C to the selected location. For example, if three users (U1-U3) approach media device 100 with their respective user devices, are detected by one or more of the proximity detection islands (e.g., I1, I2, I3, I4) and/or by RF 107, then post detection, media device 100 may begin to handle the content C from the various user devices as described in reference to FIGS. 19 and 20. However, assume for purposes of explanation, that users U1 and U3 have static content C to be handled by media device 100 and user U2 has dynamic content C. Furthermore, assume that queuing of the content C may not be in the order in which media device 100 detected the user devices, and that order is U2, U3, U1. Now, per flows 1900 and 2000, media device 100 begins to record and store the dynamic content C from U2 (e.g., U2 was streaming video); however, the recording is not complete and media device 100 handles the content C from U1 next, followed by the content C of U3. Content C from U1 comprises a playlist for songs stored in the Cloud and C from U3 comprises alarms A, notifications N, and data D from a data capable wristband/wristwatch. Media device 100 handles and stores the content C from U3 in its internal memory (e.g., DS 103) and queues U3 content first for display, playback, or other on media device 100. Media device 100 accesses the songs from U1's playlist from the Cloud and queues U1 next in the queue behind U3 for playback on the SPK 160 of media device 100. Finally, the recording is complete on U2's dynamic content C and the video stream is recorded on NAS and media device 100 has accesses to the NAS via WiFi 130. U2 is queued behind U1 for playback using DISP 180 and SPK 160 of media device 100. In some examples, where there are not conflicts in handling content C, the media device may display U3's content C on DISP 180 while playing back U1's mp3 songs over SPK 160, even thou U1 is behind U3 in the queue. Here, there is no or minimal conflict in handling content C because U1's content is primarily played back using the media device's 100 audio systems (e.g., SPK 160) and U3's content C is primarily visual and is displayed using the media device's 100 video systems (e.g., DISP 180). Servicing content C from U3 and U1 at the same time may mean temporarily bumping visual display of U1's playlist on DISP 180 to display U3's content C.
  • Moving now to FIG. 21 where one example 2100 of a media device 100 handling, storing, queuing, and taking action on content from a plurality of user devices is depicted. In FIG. 21, four users denoted by hands 2100 a-d move within proximity detection range of islands 520, are detected 2140, and the users are notified 2117 of the detection, as described above. The four users 2100 a-d each have their respective user devices UD1-UD4 having content C1-C4. For purposes of explanation, assume the order in which the user devices are discovered by the media device (e.g., via RF 107) is UD2; UD4; UD3; and UD1 and the content C on those devices are queued in the same order as the detection as denoted by C2; C4; C3; and C1 in diagram 2180. The media device 100, the user devices UD1-UD4, wireless router 2170, and Cloud 2150 are all able to wirelessly communicate with one another as denoted by 2167.
  • C2 comprises a playlist and songs, is static, and each song is stored in a mp3 file in memory internal to UD2. As per the flows 1900 and 2000, media device queues C2 first and stores C2 in a SDHC card 2121 such that the playlist and mp3 files now reside in SDHC 2121. C1 and C4 both comprise information stored in a data capable wristband/wristwatch. C1 and C4 are static content. Media device queues C4 behind C2, and stores C4 in Cloud 2150. C3 comprises dynamic content in the form of an audio book being played back on UD3 at the time it was detected by media device 100. C3 is queued behind C4 and is recorded on NAS 2122 for later playback on media device 100. C1 is queued behind C3 and is stored in Cloud 2150.
  • However, the queuing order need not be the order in which content C is played back or otherwise acted on by media device 100. In diagram 2180, media device has ordered action to be taken on the queued content in the order of C1 and C4 first, C2 second and C3 third. C3 may be third in order because it may still be recording to NAS 2122. The information comprising C1 and C4 may be quickly displayed on DISP 180 for its respective users to review. Furthermore, the size of data represented by C1 and C4 may be much smaller than that of C2 and C3. Therefore, while C3 is recording to NAS 2122 and C2 is being copied from UD2 into SDHC 2121, action is taken to display C1 and C4 on DISP 180. Action is then taken on C2 and a portion of the playlist from C2 is displayed on DISP 180 with the song currently being played highlighted in that list of songs. The music for the song currently being played is output on SPK 160. Finally, the recording of C3 is completed and DISP 180 displays the title, author, current chapter, and publisher of the audio book. Action on C3 may be put on hold pending C2 completing playback of the songs stored in SDHC 2121.
  • Here, media device 100 handled the various types of content C and operated on one type of content (recording C3) while other content (C1 & C4, C2) were being acted on, such as displaying C1 and C4 or playback of mp3 files from C2. In FIG. 21, if UD2 moves 2133 out of RF range of media device 100, C2 may be released from the queue and action on C2 may stop and the next item of content in the queue is acted on (e.g., C3). FIG. 21 is a non-limiting example and nothing precludes one of the users taking action to change the queuing order or the order in which the media device acts on queued content. Moreover, CFG 125 and/or APP 225 may be used to determine content queuing and an order in which queued content is acted on by media device 100. One of the users may have super user capability (e.g., via that user's APP 225 and/or CFG 125) that allows the super user to override or otherwise control content handling on media device 100.
  • FIG. 22 depicts another example 2200 of a media device handling, storing, queuing, and taking action on content from a plurality of user devices. Here, a plurality of users 2200 a-2200 n have approached media device 100 and have be detected by a proximity island 520. A plurality of user devices UDa-UDn, having content Ca-Cn, are in wireless communications 2167 as described above. In diagram 2280, the content Ca-Cn from the user devices is queued in the order the user devices were detected by media device 100. Content Ca-Cn may be stored and/or accessed by media device 100 from any location that may be directly accessed or wirelessly accessed by media device 100 such as in DS 103 (directly accessed), NAS 2122, the user devices UDa-UDn, the Cloud 2250, etc.
  • Media device 100 may take action on the queued content in any order including but not limited to random order, the order in which it is queued, or commanded order, just to name a few. Media device 100 may be configured to operate in a “party mode” where each of the users 2200 a-2200 n in proximity of the media device 100 desires to have their content played back on the media device 100. Media device 100 may harvest all of the content and then act on it by randomly playing back content from Ca-Cn, allowing one of the users to control playback, like a DJ, or allowing a super user UDM to control playback order and content out of Ca-Cn. One of the users may touch or otherwise actuate one of the control elements 503-512 and/or one of the proximity detector islands 520 or an icon on DISP 180 to have their content acted on by media device 100. Content in Ca-Cn may be released by media device 100 if the user device associated with that content moves out of RF range of the media device 100.
  • In FIG. 23, a flow 2300 for recording user content on a media device while the media device handles current content is depicted. At a stage 2302 entry of a user (e.g., hand of a user) into detection range of a proximity detection island 520 of media device 100 is detected. At a stage 2304 the user is notified that media device 100 has detected the user's presence (e.g., using light, sound, vibration, etc.). At a stage 2306, media device 100 may use RF system 107 to detect RF signals being transmitted by a user device (e.g., 220) as described above. At a stage 2308, the media device 100 and the user device wirelessly connect with each other (e.g., using WiFi 130 or BT 120). At a stage 2310 content currently being handled by media device 100 (e.g., being played back or queued for playback) is displayed on the media device 100 (e.g., DISP 180) or on a display of the user device, or both, for example. APP 225 or other software and/or hardware may be used to display the current content being handled on media device 100 on the user device. At as stage 2312, a request from the user device to the media device 100 for the media device 100 to handle user content from the user device is received. At a stage 2314, the media device 100 harvests the user content from the user device (e.g., wirelessly copies, streams, or otherwise accesses the user content). The user content may reside on the user device or may be located elsewhere at a location the media device 100 or user device may access, such as the Cloud, the Internet, an intranet, NAS, or other, for example. At a stage 2316 the media device 100 begins recording the user content while continuing playback of the content currently being handled by the media device 100. As was described above in reference to FIG. 22, the media device 100, based on a size of the user content (e.g., file size in MB or GB) may record the user content to memory internal to the media device 100 or to a location external to the media device 100 (e.g., NAS, the Cloud, a server, the Internet). Content that was being handled by the media device 100 continues with little or no interruption while the user content is recorded. At a stage 2318 the user content is stored as described above and flow 2300 may terminate at the stage 2318. Optionally, at a stage 2320, a determination may be made to queue the user content relative to the current content being handled by the media device 100. If no queuing action is to be taken, then a NO branch may be taken and the flow 2300 may terminate. However, if the user content is to be queued, then a YES branch may be taken to a stage 2322 where a queuing action is applied to the user content. Queuing action may mean any action taken by the media device 100 (e.g., via controller 101, CFG 125, hardware, or software) and/or user device (e.g., via APP 225) that affects the queuing of content on the media device 100.
  • Queuing action may include but is not limited: to waiting for the user content to complete recording and then placing the user content in a queuing order relative to other content already queued on the media device 100 (e.g., at the back of the queue); bumping content presently at the front of the queue once the user content has completed recording and beginning playback of the recorded user content; placing the user content behind the content currently being handled by the media device 100 such that the user content will be next in line for playback; moving the user content to the front of the queue; randomly placing the user content in the queue; allowing the user of the user device to control the queuing of the user content; allowing a DJ or other user to control the queuing of the user content; and allowing each user that is detected by the proximity detection islands, have one or more items in their content harvested and pushed to the top of the queue or placed next in line in the queue; and placing the user content in a queue deck with other content, shuffling the deck and playing on of the items of content from the deck, and re-shuffling the deck after playback of item; just to name a few.
  • Content, including the user content that was recorded may be queued in a party mode where each user who wants their content played back on the media device 100, approaches the media device 100, is detected by the proximity detection islands, receives notification of detection, has at least one selected item of user content harvested by the media device 100, and has the item of user content played back either immediately or after the current content being played back finishes. In some examples, the queue for content playback on media device 100 is only two items of content deep and comprises the current piece of content being played back and the user content of the user who approached the media device 100 and had their content harvested as described above.
  • Now referencing FIG. 24, one example 2400 of queuing action for user content in a queue of a media player is depicted. In example 2400 there are at least seven users U1-U7 and at least seven user devices UD1-UD7. For purposes of simplifying the description, assume that all seven users have approached media device 100, have been detected 2140 and notified 2117 by proximity island 520, and all user devices have been detected and wirelessly connected with media device 100. Here user content C1, C2, and C3 has been queued in queue 2480 and DISP 180 is displaying the queued order of the playlist as Song for UD1 currently being played back because it is underlined (e.g., over SPK 160), with Songs for UD2 and UD3 being next in the playlist. User content for UD1-UD3 may reside in DS 103 or other location such as NAS 2122 or Cloud 2250. User devices UD1-UD3, in that order, were the first three devices to wirelessly connect and have their user content C1-C3 harvested by media device 100. The Action for the queuing order in queue 2480 is “Play In Order”, so C1 is first, C2 is second, and C3 is third in the playback order as displayed on DISP 180. At some point in time, UD7 also wirelessly connected and had its user content C7 harvested by media device 100. Media device 100 begins the process of recording 2490 the content into DS 103 (e.g., into Flash 145). In the meantime, other user devices (not shown) may also have their user content harvested. In that the recording 2490 of C7 is still in progress, intervening user content will be placed ahead of C7 until C7 has completed 2492 recording 2492. Upon completion of recording, C7 is positioned 2482 in the playlist below some already queued user content and ahead or other user content lower in the queue. In other examples, C7 may be queued in the order it was presented to the media device 100 and the media device 100 begins the recording 2490 process and allows C7 to be played back when it moves to the top of queue, but if C7 has not completed recording 2492, then media device 100 begins the playback 2493 of C7 from a buffer BUFF 2421 where a portion of recorded C7 is stored. The playback from BUFF 2421 may continue until the recording catches up with the buffered content or is completed 2492.
  • As described above, one of the users or user devices may have super user (e.g., UM) or other form of override authority and that user may order the queue to their liking and control the order of playback of user content. Queue 2480 and/or the user content being queued need not reside in memory internal to media device 100 and may be located externally in NAS 2122, a USB Hard Drive, Cloud 2250, and a server, just to name a few. In some examples, media device 100 may delete or bump user content from queue 2480 if the wireless connection 2167 between media device 100 and the user device is broken or interrupted for a predetermined amount of time, such as two minutes, for example. The “Play In Order” example depicted is a non-limiting example and one skilled in the art will appreciate that the queuing may be ordered in a variety of ways and may be determined by executable program code fixed in a non-transitory medium, such as in DS 103, Flash 145, CFG 125, and APP 225, just to name a few. Therefore, controller 101 or a controller in a user device may execute the program code that determines and controls queuing of user content on the media device 100.
  • Attention is now directed to FIG. 25 where one example of a flow 2500 for wirelessly communicating encoded content between a wireless user device and a wireless media device is depicted. At a stage 2502 a wireless communication link may be established between a wireless user device (e.g., 202, 2701, etc.) and a wireless media device (e.g., 100). The wireless communications link may be via any of the systems of either device that are configured for wireless communications including but not limited to RF (e.g., using one or more radios), acoustic, optical, and imaging, just to name a few. The wireless communications link may be a previously established link via a paring process (e.g., BT paring), a wireless network link (e.g., WiFi, AdHoc WiFi, WiMAX, Cellular, NFC) or other. The wireless communications link may be established over a WiFi network that the wireless user device and the wireless media device are using and/or have previously used, for example. The wireless user device and the wireless media device may have previously been paired with one another, such as through a BT link or the like. At the stage 2502, the wireless communications link may be established between one or more wireless user devices and one or more wireless media devices. Data source 2501 broadly includes without limitation any source of data that may be accessed for the establishment of the wireless communications link between one or more wireless user devices and one or more wireless media devices. Data source 2501 may be accessed by the wireless user device, the wireless media device or both. For example, data source 2501 may comprise the APP 225 for the wireless user device. As another example, data source 2501 may comprise the CFG 125 for the wireless media device.
  • At a stage 2504 content (e.g., C) may be accessed by the wireless user device (e.g., 2701 of FIG. 27A) and that content may reside locally with the wireless user device, reside external to the wireless user device or both. As one example, content C in the form of media (e.g., music, images, video, movies, etc.) may reside in memory internal to the wireless user device (e.g., Flash memory or other non-volatile memory) and be accessed (e.g., a read operation) by a controller, processor or the like. As another example, content C may reside external to the wireless user device in a location such as the Cloud, the Internet, network attached storage (NAS), a web site, a web page, etc. Data source 2503 broadly includes without limitation any source of data that may be accessed for the content at the stage 2504, using one or more communications links including but not limited to wireless links and wired links.
  • At a stage 2506 data about the content C that was accessed (e.g., descriptive data—DD) may be accessed by the wireless user device. The descriptive data DD may comprise metadata, descriptive metadata, tag, or other information/data related to, is about, that describes attributes, or is associated with the content C that was accessed. In some examples, the DD may already be included or otherwise associated with the content C and the stage 2506 may be skipped, for that particular datum of accessed content C, for example. In other examples, the DD may be accessed internal to the wireless user device, external to the wireless user device or both. Data source 2505 broadly includes without limitation any source of the DD that may be accessed at the stage 2506, using one or more communications links including but not limited to wireless links and wired links. As one example, if content C is music in the form of a digital audio file (e.g., MP3, FLAG, AIFF, WAV, RA, AU, AAC, ATRAC, Apple Lossless, WMA, MPEG-4, etc.), then the DD may comprise information about the music such as song title, artists, sidemen, producer, cover art, lyrics, images, photographs, liner notes, playing time, time index, location or address of the file or data that comprises the music (e.g., on NAS, local memory, Cloud, the Internet, URI, URL, web page, etc.). The DD may comprise proprietary (e.g., a format and/or data structure proprietary to the wireless media device(s) 100), known, accepted, or standardized formats, or some combination of those, for example. Actual format, structure, content, and source for the DD may be application dependent and is not limited to the examples described herein.
  • At a stage 2508 the DD may be encoded into a format that may be decoded by the wireless media devices (e.g., 100). Encoding at the stage 2508 may comprise stripping, ignoring, or otherwise parsing the DD to extract only those portions of the DD that may be encoded into the format that may be decoded by the wireless media devices. Encoding at the stage 2508 may comprise encrypting or otherwise adding data security to the DD that is encoded. Encoding at the stage 2508 may comprise data compression or other process for reducing a size (e.g., number of bytes) of the encoded DD. Data source 2507 broadly includes without limitation any source of data that may be accessed at the stage 2508 to effectuate the encoding of the DD. For example, data source 2507 may comprise algorithms and/or data embodied in a non-transitory computer readable medium that is electronically accessed by the wireless user device or external compute engine and executed on one or more processors to encode the DD.
  • At a stage 2510, the encoded DD from the stage 2508 may be embedded into the content C to form encoded content (EC). Embedding at the stage 2510 may comprise appending the encoded DD to a datum that comprises the content C that was accessed at the stage 2504, for example. Embedding at the stage 2510 may comprise adding the encoded DD to data structure or packet that comprises a plurality of fields such as headers, data payloads (e.g., the content C), or other, where the encoded DD may be one or more of the fields in the data structure or packet, for example. Data source 2509 broadly includes without limitation any source of data that may be accessed at the stage 2510 to effectuate the embedding of the encoded the DD to form the EC. Optionally, the EC may be stored or otherwise saved to one or more locations as denoted by data store 2511. Data store 2511 may be internal to the wireless user device, external to the wireless user device or both. For example, data store 2511 may be memory internal to the wireless user device (e.g., Flash memory). As another example, data store 2511 may be external memory disposed in the wireless media device (e.g., 100). As yet another example, data store 2511 may be external memory disposed in the Cloud, the Internet, NAS or other location. The EC may be saved for future use or re-use by the wireless user device. For example, some of the stages of flow 2500 may not need to be executed (e.g., stages 2504-2510) because the EC may be accessed from data store 2511. After the EC is accessed from data store 2511, other stages of the flow 2500 may be executed on the EC.
  • At a stage 2512 the EC may be wirelessly communicated to the wireless media device (e.g., using the wireless link from the stage 2502). Data source 2513 broadly includes without limitation any source of data that may be accessed at the stage 2512 to effectuate the wireless communication of the EC to the wireless media device 100. Data source 2513 may comprise APP 225, for example. The APP 225 may include data as to which wireless media device(s) the wireless user device is linked to, any handshakes or other protocols that may be needed to wirelessly communicate the EC to the wireless media device(s), etc. As one example, if the wireless user device and wireless media device are wirelessly linked via BT paring, via an acoustic link, and a WiFi link (e.g., they may communicate with one another over a wireless network they are connected with), then data source 2513 (e.g., APP 225) may be used to decide which of the available wireless links to use to communicate the EC (e.g., BT, acoustic, or WiFi). In this example, data source 2513 may be configured to use the fastest and most secure link available, such as WiFi over BT or acoustic. A microphone on either of the wireless devices (e.g., 100 and/or 2701) may detect high levels of ambient noise (e.g., AN 2799 of FIG. 27A-27B) which may make the acoustic link less reliable than under ideal acoustic conditions (e.g., little or no ambient noise); therefore, data source 2513 may be configured to select the WiFi link to communicate the EC to the wireless media device. Here, data source 2513 (e.g., APP 225) may be executing on a processor of the wireless user device. The wireless communication of the EC to the wireless media device at the stage 2512 may be by way of an external wireless device (e.g., the Cloud, the Internet, a wireless network, etc.) and need not be by the wireless link between the wireless user device and the wireless media device. As one example, APP 225 or other algorithm may be configured to execute some or all of the stages of flow 2500 on an external resource (e.g., 2750 of FIG. 27A-27B). At the stage 2512, that external resource may wirelessly communicate the EC to the wireless media device. Data source 2501 at the stage 2502 may include information (e.g., wireless network name and password) that enables the wireless media device to establish a communications link (e.g., wired or wireless) with the external resource to facilitate the wireless communication of the EC at the stage 2512.
  • At a stage 2514 a determination may be made by the wireless media device, the wireless user device or both, for handling of the EC. Handling may comprise without limitation the wireless media device playing back the EC, queuing the EC for playback, storing the EC, assigning handling of the EC to one or more other wireless media devices, streaming content associated with the EC (e.g., executing a playlist encoded and/or included in the EC), wirelessly communicating the EC to another device, just to name a few. If a YES branch is taken from the stage 2514, then the flow 2500 may transition to a stage 2516 where the EC is handled by the wireless media device. Data source 2515 broadly includes without limitation any source of data that may be accessed at the stage 2516 to effectuate the handling of the EC. For example, data source 2515 may comprise CFG 225 or may comprise an operating system (OS) for the wireless media device (e.g., code in DS 103) that determines how EC is to be handled by the wireless media device(s). On the other hand, if a NO branch is taken at the stage 2514, then flow 2055 may transition to some other stage, such as a stage 2518.
  • At the stage 2518 a determination may be made as to whether or not more content C may be available for processing by flow 2500. For example, at the stage 2504 there may have been one-hundred-three (103) items of content C on the wireless user device, comprised of three (3) playlist with a first playlist having fifty (50) songs, a second playlist having twenty (20) songs, and a third playlist having thirty (30) songs. Assuming for purposes of explanation that a user of the wireless user device wants the one-hundred-three (103) items of content C to be encoded and handled by the wireless media device it links with at the stage 2502, then flow 2500 may process each item of the 103 items of content C. After the first item of content C is processed, the stage 2518 may select the YES branch to indicate there are one or more other items of content C to be processed (e.g., the second, third, . . . to the last of the 103 items). The YES branch from the stage 2518 may transition to some other stage of flow 2500 as denoted by 2520. As one example of what 2520 may comprise, consider a scenario where all of the 103 items of content C are accessed at the stage 2504 and all of the DD for all 103 items is accessed at the stage 2506. Then 2520 may comprise a transition of flow 2500 to the stage 2508 where the DD for each item of content C is encoded, followed by the stage 2510, etc. On the other hand, if there is no more content C to be processed, then a NO branch may be taken and flow 2500 may terminate. Data source 2517 broadly includes without limitation any source of data that may be accessed at the stage 2518 to make the determination of whether or not more content C is to be processed by flow 2500. For example, data source 2517 may be a counter that starts at the number of items of content C to be processed by flow 2500 (e.g., 103 items) and is decremented by one each time an item of content C has been processed. Stage 2518 may access data source 2517 to determine if its value is zero (“0”) (e.g., all 103 items have been processed) or some other value that indicates completion or no more content C.
  • Reference is now made to FIG. 26 where another example of a flow 2600 for wirelessly communicating encoded content between a wireless user device and a wireless media device is depicted. At a stage 2602 a wireless link may be established between a wireless user device and a wireless media device as was described above in reference to the stage 2502 in flow 2500 and data source 2601 may be accessed to establish the wireless link as described above for data source 2501.
  • At a stage 2604 the wireless user device may access content C as was described above in reference to the stage 2504 in flow 2500 and data source 2603 may be accessed as described above for data source 2503. Optionally, data store 2511 may be accessed at the stage 2604. Data store 2511 may comprise one or more items of EC (e.g., encoded content from the stage 2510) as was described above.
  • At a stage 2606 a determination may be made as to whether or not the content C includes encoded descriptive data (DD) as described above in reference to the stages 2506-2510 of flow 2500. If a YES branch is taken, then flow 2600 may transition to a stage 2608 as will be described below. If a NO branch is taken, then flow 2600 may transition to another stage, such as the stage 2506 of flow 2500 as described above. Stages 2506-2510 of the flow 2500 may be executed to access, encode, and embed the DD into the content C. After stage 2510, flow 2600 may resume at the stage 2608.
  • At the stage 2608 a determination may be made as to whether or not to embed the encoded DD into other content (OC). One example of OC may be a signature tone or startup tone used by one or more of the wireless media devices (e.g., 100) to acoustically wirelessly communicate with one another for data communication (e.g., a power up), to configure one or more wireless media device (e.g., using CFG 125), etc. APP 225 may include or have access to the signature tone as OC and may embed encoded DD into the OC for purpose of communicating information in the encoded DD to the wireless media device. The encoded DD may comprise data that is unrelated to the data in the OC. For example, the encoded DD may be data about one or more items of unrelated content C that are encoded into the OC for subsequent decoding by the wireless media device (e.g., post stage 2614 of flow 2600). For example, if the OC is the signature tone, the encoded DD may comprise content C in the form of a location for a song the wireless user device is commanding the wireless media device to handle (e.g., to playback the song on the wireless media device). Wireless user device may use a speaker to acoustically transmit EC that comprises the OC startup tone with the encoded DD having the location of the song to be handled. A microphone on the wireless media device may receive the acoustic transmission which is subsequently decoded by the wireless media device to extract the location of the song from the encoded DD and handle processing of the song. Essentially, the OC (e.g., signature tone) may include a link, pointer, or address (e.g., the encoded DD) to content C to be handled by the wireless media device. Upon receiving and decoding the EC that includes OC plus the encoded DD, the information decoded from the encoded DD may be used by the wireless media device to access and handle the content C. Therefore, in this example, the EC itself does not include the content C to be handled, but rather information for accessing the content C to be handled.
  • Accordingly, if a YES branch is taken from the stage 2608, then the OC is accessed at a stage 2610. Data source 2605 may be used at the stage 2610 to access the OC in a manner similar to that described above for data sources 2603 and/or 2503. At a stage 2612 the encoded DD is embedded into the OC to form EC as was described above in reference to flow 2500. At a stage 2614 the EC may be wirelessly communicated to the wireless media device as was described above in reference to the stage 2512 of flow 2500. Data source 2613 may be accessed at the stage 2614 as was described above in reference to stage 2512 and data source 2513. However, if a NO branch is taken at the stage 2608, then flow 2600 may transition to the stage 2614 (e.g., the OC will not be embedded with encoded DD). After the stage 2614, flow 2600 may transition to some other stage, such as the stage 2514 of flow 2500 as described above.
  • Referring now to FIG. 27A where a block diagram 2700 a of one example of wirelessly communicating encoded content between a wireless user device and a wireless media device is depicted. Block diagram 2700 a includes at least one wireless media device 100, at least one wireless user device 2701 (e.g., a pad, tablet, or smartphone), and optionally other systems and/or resources such as wireless router 2760, data storage 2765 (e.g., NAS), server/PC 2762, and resource 2750 (e.g., Cloud or the Internet). Wireless (RF) communications links between the various elements depicted are denoted as RF signals 2710, 2720, 2730, and 2770. Acoustic wireless communications between wireless user device 2710 and one or more wireless media devices 100 are denoted as 2776, 2775, and 2779 and will be described in greater detail below. A user 201 may be present in an environment ENV 198 as described above and the user 201 may be the user of one or more wireless user devices, such as wireless user device 2701, for example. As described above, wireless media device 100 may include one or more systems that sense 197 the ENV 198 for presence of users (e.g., 201) and wireless devices (e.g., 2701).
  • Actual configurations of wireless user devices, such as device 2701, will vary and the examples provided herein are for purposes of explaining the present application and are not to be construed as limiting in any way the types and configurations of wireless user devices that may interact with wireless media devices, such as wireless media device 100. Wireless user device 2701 may include a plurality of systems including but not limited to one or more radios for transmitting, receiving, or both, RF signals 2710 for wireless communications (e.g., WiFi, IEEE 802.1, NFC, BT, BT low energy, etc.), a power supply (e.g., lithium ion rechargeable battery), one or more microphones 2702 or similar transducer for generating a signal 2702 c from incident sound (e.g., sound in ENV 198), one or more antennas 2714 coupled with the one or more radios, one or more speakers 2704 or similar transducers for generating sound in response to an applied signal 2704 c (e.g., from an amplifier), a display 2703 (e.g., touch screen, OLED, LCD, etc.) for displaying information such as a GUI or an application running on device 2701, data storage (e.g., Flash memory) for storing an OS, application software, APPS, algorithms, data, media files, content C, etc., and one or more controllers or processors (e.g., SoC, Baseband processor, DSP, μP, μC, ASIC, etc.), just to name a few, for example.
  • As described above, wireless user device 2701 may include APP 225 configured for use (e.g., communication, command, control, content handling, etc.) with one or more wireless media devices 100. APP 225 may have access to content C that may reside on the device 2701, external to the device 2701 or both. APP 225 may be one or more algorithms and/or data embodied in a non-transitory computer readable medium, such as a data storage system of device 2701, for example. Access to content C by APP 225 may include managing and/or presenting the content C on display 2703 in a form the user 201 may interact with, such as the GUI or other form of displaying information. The GUI of the device 2701 or a GUI generated by APP 225 may display content C in various formats. Here, the GUI displays at least a portion of the content C as one or more playlists P1-P3, with each playlist comprised of one or more songs or other media type. For example, playlist P3 includes a plurality of songs denoted as S1-S7. There may be more or fewer playlists and/or songs as denoted by 2711 and 2713. APP 225, via the GUI, may display one or more wireless media devices that the user device 2701 is wirelessly linked with and/or has been previously wirelessly linked with, as denoted by WMD's 100-100 vi. There may be more or fewer wireless media devices than depicted as denoted by 2715. User 201 may have commanded the device 2701 to display the content C depicted in the GUI or the display may be controlled by APP 225 or other algorithm and/or system of device 2701.
  • Wireless media device 100 may include a plurality of systems as described above in reference to FIG. 1. For purposes of explanation, the various systems that may be included in media device 100 are depicted in block diagram form without implementation specific details such as actual connections to the systems, etc. Systems in media device 100 may include but are not limited to PROX 113 (e.g., proximity detection island(s) 520) for sensing 197 the ENV 198, RF 107 which may include one or more radios Tx/Rx 2777 coupled with one or more antennas 124, some of which may be de-tunable 129, DS 103 which may include CFG 125 and content C (e.g., content being handled by 100), controller 101, A/V 109 which may include display 180, one or more image capture devices 109 c, one or more MCI's 160 and associated amplification/processing circuitry 2771, and one or more SPK's 170 and associated amplification/processing circuitry 2773, and chassis 199 which may have any shape or configuration. Controller 101 and/or algorithms executing on controller 101 or other circuitry in media device 100 may decode (2772, 2778) signals (e.g., encoded signals) received by the various systems or encode signals to be transmitted by the various systems. For example, encoded RF, acoustic, or optical signals transmitted from wireless user device 2701 may be decoded by circuitry and/or algorithms in media device 100 to generate decoded signals. Similarly, the media device 100 may use circuitry and/or algorithms to encode signals for wireless transmission from the media device 100 to user device 2701, other media devices 100, or other devices or systems such as resource 2750, for example. Wireless transmission from the media device 100 using its various systems may comprise one or more of RF (107), acoustic (109), or optical (180, 185, 190) signal transmission. Wireless signal reception by the media device 100 using its various systems may comprise one or more of RF (107), acoustic (109), or optical (185, 190) signal reception. Wireless user device 2701 may transmit and receive wireless signals using similar RF, acoustic and optical systems as the media device 100 (e.g., 2714, 2704, 2702, and 2703).
  • Now referring back to flow 2500 of FIG. 25, user device 2701 and media device 100 may establish a wireless communication link (e.g., at the stage 2502) with each other using any of their RF, acoustic, or optical systems. In FIG. 27A, RF 107 may detect RF signal 2710 from user device 2710 and the APP 225 and/or CFG 125 may orchestrate establishing the wireless link, such as a BT link (e.g., via paring) or via WiFi link (e.g., using wireless router 2760), for example. The wireless communications link may be a direct link between media device 100 and user device 2701 or an indirect link, such as through a wireless network (e.g., IEEE 802.1 or wireless router 2760). If media device 100 and user device 2701 have previously encountered and/or linked with one another (e.g., via BT or WiFi), then the wireless communications link may be comprise each device recognizing the other using whatever wireless protocols they have previously used in communicating with each other and initiating wireless communication of data between each other (e.g., handshakes or one or more stages of flow 2500 and/or flow 2600).
  • The wireless communications link may be one or more of acoustic, RF, or optical and the mode of wireless communication may reversibly switch between one or more of the RF, acoustic, or optical links during interaction between user device 2701 and media device 100. The wireless communications link may be established between one or more wireless user devices and one or more wireless media devices even thou FIGS. 27A-27B depict only one of each device. As one example, the media device 100 upon detecting presence of user 201 and/or user device 2701 may transmit an acoustic signal via SPK 170 that may be detected by MIC 2702 and signal 2702 c may be decoded or otherwise processed to establish the wireless link using one of its RF systems (e.g., BT, NFC, WiFi, etc.). Information about a WiFi network (e.g., network name and password) may be encoded in the acoustic signal or may comprise data in CFG 125 and/or APP 225 that is accessed to effectuate the wireless link. In some applications, the wireless communications may be all acoustic, all RF, or all optical (e.g., using infrared (e.g., an IrLED) communications via OPT 185 to an optoelectronic receiving device (e.g., a photo diode) on 2701).
  • After establishing the wireless link, user device 2710 may access content C (e.g., at the stage 2504) which may be internal, external of both to the user device 2710. For example, user 201 may have selected preferences that are stored or otherwise coded in APP 225, that want content C to be handled by media device 100 when both devices are disposed in ENV 198 (e.g., are in close proximity of each other). As one example, if the user 201 was currently playing song S7 in playlist P3 when the wireless link was established (e.g., at the stage 2502), then the content C that comprises S7 would be accessed (e.g., via a read operation) by the user device 2701. The user device 2701 may then access (e.g., at the stage 2506) descriptive data DD about song S7 from a data source that may be internal, external or both to user device 2701. The DD may comprise the location (e.g., address, file folder, URI, URL or other descriptor) of the song S7, playing time of S7, song title for S7, artist or group that performed S7, and a time index, if any, to begin playback of S7, for example. The DD may then be encoded into a format that may be decoded by the wireless media device 100 (e.g., at the stage 2508). Decoding may be done using hardware (e.g., 101), software (e.g., CFG 125 or other algorithms) or both. The encoding may be in a format that is proprietary to wireless media devices 100. Decoding of the encoded DD may be accomplished using processing in the analog domain, digital domain or both. Circuitry including but not limited to a DSP, an analog-to-digital-converter (ADC) or a digital-to-analog-converter (DAC) may be used in the decoding process. The DSP or other compute engine may execute one or more algorithms to effectuate the decoding of the encoded DD.
  • The encoded DD may be embedded or otherwise linked or associated with the content C that was accessed (e.g., at the stage 2510). For example, the encoded DD may be embedded into a file that comprises S7, included in one or more field of a data packet in which S7 will be transmitted to media device 100, may be concatenated into a file that comprises S7 (e.g., as a header or footer), may be included in a wrapper file or object that includes S7, or may be wirelessly transmitted in some sequence that includes data for the encoded DD and data for S7, for example. As one example, the encoded DD may be wirelessly transmitted first, followed by wireless transmission of the accessed content C associated with the encoded DD, such as S7. Although a single item of content C (S7) is described, the user device 2701 may access as the content a plurality of items (e.g., a plurality of songs and/or playlists, etc.) that are all described in the DD and included in the encoded DD. As one example, the accessed content C may comprise one or more of the playlists P1-P3 and all of the songs in those playlists. The DD may comprise information about the playlists, their respective songs, and the locations of the playlists and songs, for example. The encoded DD that is embedded or otherwise associated with the accessed content C would include that information as well. After embedding or otherwise associating the encoded DD with the content C, the content may be referred to as encoded content (EC) that may comprise a single datum or a plurality of datum that are wirelessly communicated to the wireless media device 100 either directly from the user device 2701 using one of the aforementioned wireless links, from an external system (e.g., 2750, 2760) or both. In FIG. 27A, the encoded content for song S7 in playlist P3 is denoted as S7-EC. For example, some or all of the stages 2502-2510 of flow 2500 may be processed (e.g., by a processor in 2701) to create S7-EC from S7 as denoted by dashed line “a”, followed by some or all of stages 2512-2516 being processed to communicate S7-EC to wireless media device 100 as denoted by dashed line “b”.
  • A single datum for the EC may comprise the encoded DD embedded in the same file, data structure, or packet as the accessed content C. For example, S7-EC may be a single file or data structure that includes the content for song S7 plus the encoded DD for song S7. As another example, S7-EC may comprise a packet formatted to include one of more fields for the S7's content C (e.g., a data payload) and one or more fields for the encoded DD for S7, as well as other fields used in packets for communicating data, such as error correction, check sums, headers, footers, etc.
  • The plurality of datum for the EC may comprise multiple datum's that are wirelessly transmitted to the media device 100 or a first subset of the plurality of datum that are wirelessly transmitted to the media device 100 and a second subset of datum that are accessed by the wireless media device 100 based on information contained in the first subset. For example, the first subset may include the encoded DD that upon being decoded by the media device 100, may include information the media device 100 uses to access the datum in the second subset, which may comprise the content C to be accessed and optionally handled by the media device 100. The first subset may include data for a web page, Internet address, URI, URL or the like where the content C in the second subset can be accessed by the media device 100. For example, the encoded DD in the first subset may include a URL 2758 that links to a web site or other address in resource 2750 (e.g., the Internet or Cloud) where the content C in the second subset may be accessed from, such as from a data storage system 2757.
  • In some examples, the EC may be acted on (e.g., handled) by the wireless media device 100 such as denoted by the stage 2514-2516 of flow 2500. Handling of the EC may broadly include the media device 100 or other media devices 100 in communication with media device 100 and/or wireless user device 2701 performing one or more actions on the EC. The actions taken may be determined in whole or in part by one or more of the CFG's 125 of one or more of the media devices 100, the APP 225 on user device 2701, a command by user 201 (e.g., via the GUI, voice command, etc.), hardware, or algorithms, for example. As one example, the user 201 may have selected song S7 for playback on media device 100 and the handling of the subsequent EC may be to initiate playback of S7 on media device 100 (e.g., via SPK's 170) and/or on one or more other wireless media devices 100 in wireless communication with media device 100 (e.g., via BT link, WiFi link, AdHoc WiFi link, etc.). As another example, if S7 is a multi-channel media file (e.g., surround sound, DTS®, Dolby®, etc.), then the various encoded channels of sound (left-front, right-front, center, left-rear, right-rear, subwoofer, etc.) may be assigned (e.g., via the CFG's 125) to a plurality of different wireless media devices 100 in the same ecosystem and/or disposed in ENV 198. As yet another example, if the EC comprise video, images, or audio-video content, handling by media device 100 may comprise playback of the image or video content on display 180 and audio content (if any) over SPK's 170. In some examples, handling of EC by media device 100 may not be immediate, may be delayed, may be queued, or may be stored for later use. In other examples, EC may be a link, pointer, or address to other content to be accessed or handled by the media device(s) 100 and handling the EC may comprise acting on the link, pointer, or address and accessing and/or handling the content C there according to data in the EC, the CFG's 125, etc. For example, the EC that is wirelessly communicated (e.g., at the stage 2512) to the media device 100, upon decoding may reveal a location on NAS 2765 where the content C to be accessed resides and the media device 100 may wirelessly stream the content C for playback over the media device 100 or other media devices 100. Here, wireless media device 100 and wireless network/router 2760 may be wirelessly linked (e.g., via WiFi, IEEE 802.1 or other). The content C to be accessed may reside in more than one location specified by the EC, such as in NAS 2765, server 2762, and data storage 2754 of resource 2750, for example. NAS 2765 and server 2762 may be in communication (2763, 2764) with router 2760 and the content C on NAS 2765 and server 2762 may be wirelessly communicated 2730 to the media device 100 for handling. Similarly, content C in data storage 2754 of resource 2750 may be wirelessly communicated 2720 (e.g., using cellular tower 2756) from resource 2750 to media device 100 for handling. Content C being handled may be wirelessly transferred in total to the media device(s) 100, be streamed, buffered, stored, or handled in chunks or portions that may be determined by factors including but not limited to wireless bandwidth, download times, a data size (e.g., in megabytes, gigabytes, etc.) of the content C, data storage capacity of media device 100 (e.g., in DS 103), copyright and/or digital media rights, persistence of the content C, the type of content C, just to name a few.
  • The wireless communication of the EC (e.g., at the stage 2512) may comprise one or more acoustic signals 2776 generated by SPK 2704 of user device 2701 that are received 2776 by MIC 160 of media device 100. Signals 160 c generated by MIC 160 may be decoded 2772 to extract data from the encoded DD that is included the EC. The wireless communication of the EC (e.g., at the stage 2512) may comprise one or more RF signals 2710 transmitted by user device 2701 and received by Tx/Rx 2777 of media device 100. The RF signal may be decoded 2778 to extract data from the encoded DD that is included the EC. The wireless communication of the EC (e.g., at the stage 2512) may comprise one or more wireless signals 2710 and one or more acoustic signals 2776 that may be transmitted in some sequence or simultaneously by user device 2701. For example, the encoded DD portion of the EC may be acoustically transmitted, followed by wireless RF transmission of the content C associated with the EC. This sequence may be used when the data that comprises the content C is large in size and a higher bandwidth and/or error correction capabilities of the wireless protocol used for the RF transmission may make it more ideal than acoustic transmission of the content C. Moreover, if ambient noise AN 2799 in ENV 198 is of sufficient amplitude (e.g., in dB's) to prevent MIC 160 from accurately receiving 2776 the acoustic transmission 2776 from SPK 2704, then user device 2701, media device 100 or both may act to switch from acoustic to RF wireless transmission of the EC and/or its associated content C.
  • In other examples, at least a portion of the EC may be optically wirelessly transmitted (e.g., at the stage 2512) using optical systems of the user device 2701 and media device 100. For example, OPT 185 in I/O 105 may be used to receive 185 r optical signals from user device 2701 or to transmit 185 t optical signals to the user device 2701. User device 2701 may include an optical transmit 185 t and receive 185 r system that may be similar to that of media device 100 and/or communicates using a wireless protocol (e.g., IrDA or other) that is compatible with or identical to those used by media device 100. For example, infrared LED's may be used in conjunction with an opto-electronic device such as a photo diode, PIN diode, or the like for receiving the optical signals and converting them into signals to be decoded 2782 or otherwise processed in the devices 2701 and 100. RF wireless and/or acoustic wireless communications between media device 100 and user device 100 may supplement and/or replace the optical wireless communications.
  • In another example, the EC may comprise a bar code, TAG, or other image based symbol that is encoded with information that may be captured by the media device 100 (e.g., via 190), processed, and output as a decoded signal. As one example, the stages 2506-2510 may result in an encoded image (e.g., a TAG or other) denoted as EC1 being displayed on display 2703 of user device 2710. The user device 2710 and media device 100 may be positioned 2792 relative to each other such that image capture device 190 may image 190 i the encoded image EC1 off of the display 2703 and generate a signal that is decoded 2781 to extract the encoded DD from EC1. For example, EC1 may include data that describes a location where the content C associated with EC1 may be accessed by the media device 100, such as NAS 2765, data storage system 2757 and/or 2754 in resource 2750, or the wireless user device 2701. EC1 may be generated (e.g., via flow 2500 or 2600) by user device 2701, server 2762, compute engine 2752 in resource 2750 or some combination of the foregoing, for example. As another example, the captured image 190 i of EC1 upon being decoded by media device 100 may include data instructing media device 100 to access song S7 from playlist P3 on wireless user device 2701 using one or more of the wireless RF links between those two wireless devices (e.g., WiFi via router 2760 or directly using a BT link).
  • Moving on to FIG. 27B, where the user device 2701 and wireless media device may wirelessly communicate with each other using the systems as described above in reference to FIG. 27A and/or flows 2500 or 2600. In FIG. 27B the actual content to be handled by media device 100, such as song S9 in playlist P3 is referenced by another item of content that may be unrelated to the actual content of S9, denoted as other content OC. Here, content of S9 may already include DD that is encoded into the content that comprises S9. Content S9 may be accessed (e.g., at the stage 2604 of flow 2600) and subsequently analyzed to determine whether or not S9 includes encoded DD. If not, then the DD may be accessed and embedded or associated with the OC. As one example, consider a scenario where wireless media devices 100, at power up or when activated from a standby or low power mode to an active mode, emit an signature acoustic signal encoded with information that other media devices 100 may capture (e.g., via their MIC's 160) and decode. Each wireless media device in an ecosystem of media devices or in ENV 198 may announce, recognize, and wirelessly link with one another in response to a received signature acoustic signal. The signature acoustic signal may be modulated in frequency (FM), in amplitude (AM), pulse width, pulse shape (e.g., sine wave, saw tooth wave, etc.) to encode information that may be decoded by other wireless media devices. Adding encoded DD to the signature acoustic signal (e.g., signature tone) may comprise one form of EC. The signature acoustic signal (signature tone hereinafter) may comprise one form of content C (e.g., in DS 103) that may be totally unrelated to any content C the user device 2701 is presenting for handling by the media device 100. However, the OC that may be encoded with DD about the content C to be handled, such as the location of the content C on the user device 2701, resource 2750, wireless network 2760, or other wireless media device 100, for example.
  • In FIG. 27B, the OC may be wirelessly transmitted to media device 100 as denoted by dashed line “a” and media device 100 may decode the OC to extract data for a location of content, such as S9 or S9 plus EC denoted as S9-EC as denoted by dashed line “b” or “C”. On the other hand, the OC may be appended, embedded, or otherwise associated with the content for song S9 to form encoded content S9-EC as denoted by dashed lines “d” and “e”, and media device 100 may access the content S9-EC as denoted by dashed line “C”. For example, at the stage 2614, the OC may be sonically screamed out 2776 to the media device 100 using SPK 2704. MIC 160 receives the acoustic signal 2776 and signals 2772 are decoded to yield information regarding the content for song S9, such as where S9 may be accessed from by the media device 100, how to handle the content S9 or encoded content S9-EC after it has been accessed by the media device 100, etc. The sonic screams may be wirelessly transmitted from media device 100 to user device 2701 to wirelessly communicate information between the devices. Therefore, the user device 2701, the media device(s) 100 or both may use the sonic screams to communicate information including encoded content EC.
  • Turning now to FIG. 28, the signature acoustic signal (e.g., the sonic scream) may be in a frequency range 2800 that is within (W) a range 2801 of human hearing, that is above (A) 2803 the range of human hearing, or that is below (B) 2805 the range of human hearing. The signature acoustic signal may be an infrasonic low frequency signal that is approximately 20 Hz or less, an ultrasonic frequency that is approximately 20 kHz or more, or an acoustic range that is from about 20 Hz to about 20 kHz, for example. Ambient noise AN 2799 may determine which frequency range is best suited for wirelessly communicating EC or other data to/from devices 2710 and 100.
  • Although the foregoing examples have been described in some detail for purposes of clarity of understanding, the above-described conceptual techniques are not limited to the details provided. There are many alternative ways of implementing the above-described conceptual techniques. The disclosed examples are illustrative and not restrictive.

Claims (23)

1. A method for proximity-based control of content on a device, comprising:
acoustically detecting a user device using a proximity system, an audio/video (A/V) system or both of a media device;
establishing an acoustic communications link between the user device and the media device;
receiving on the media device, using the link, an acoustic signal from the user device;
harvesting user content, data or both from the acoustic signal; and
transferring content handling of the user content, the data or both from the user device to the media device.
2. The method of claim 1 and further comprising: queuing the user content relative to other content currently being handled by the media device.
3. The method of claim 2, wherein the queuing comprises a preferential queuing of the user content relative to the other content.
4. The method of claim 1 and further comprising: recording, using the media device, the user content, the data or both.
5. The method of claim 4, wherein the recording includes storing the user content, the data or both in memory internal to the media device.
6. The method of claim 4, wherein the recording includes storing the user content, the data or both in memory external to the media device.
7. The method of claim 1 and further comprising: modifying a playlist on the media device based on the user content, the data or both.
8. (canceled)
9. The method of claim 1, wherein the transferring comprises initiating playback of the user content on the media device.
10. The method of claim 1, wherein the harvesting, the transferring or both include wirelessly establishing a radio frequency (RF) communication link between the user device and a RF system of the media device.
11. The method of claim 1, wherein the harvesting, the transferring or both are wirelessly accomplished by using the acoustic communications link between the user device and the media device.
12. The method of claim 1, wherein the proximity system includes at least one proximity detection island configured to detect the user device.
13. The method of claim 1, wherein the receiving comprises using a microphone included in the A/V system to receive the acoustic signal from the user device.
14.-15. (canceled)
16. A wireless device including wireless detection of users and user devices, comprising:
a controller in electrical communication with
a data storage system,
a radio frequency (RF) system including at least one RF antenna electrically coupled with a plurality of RF transceivers,
an audio/video (AN) system including a speaker for producing sound, a microphone for capturing sound,
a proximity detection system configured to wirelessly detect presence of a user, a user device or both,
wherein the RF system is configured to detect a RF signal from the user device, to establish a wireless communications link with the user device, to wirelessly access user content, data or both from the user device or a location specified by the user device using the wireless communications link,
wherein the A/V system is configured to detect an acoustic signal from the user device using the microphone, to establish an acoustic communications link with the user device using the speaker and the microphone, to wirelessly access the user content, the data or both using the acoustic communications link, and
wherein the controller is configured to harvest the user content, the data or both that was accessed using the RF system, the A/V system or both, and configured to take over handling of the user content, the data or both from the user device.
17. The device of claim 16, wherein the handling comprises playback or queuing of the user content, the data or both.
18. The device of claim 16, wherein the handling comprises recording or buffering of the user content, the data or both.
19. The device of claim 16, wherein the data includes information used by the RF system to establish the wireless communications link with the user device, and the user content, the data or both are wirelessly accessed using the wireless communications link.
20. The device of claim 16, wherein the proximity detection system includes a plurality of proximity detection islands.
21. The device of claim 16, wherein the A/V system is configured to process the acoustic signal in a frequency range that is outside a range for human hearing.
22. (canceled)
23. The device of claim 16, wherein the microphone is configured to capture the acoustic signal in a frequency range that is outside a range for human hearing.
24. The device of claim 16, wherein the speaker is configured to produce the acoustic signal in a frequency range that is outside a range for human hearing.
US14/105,127 2013-03-13 2013-12-12 Proximity-based and acoustic control of media devices for media presentations Abandoned US20150171973A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US14/105,127 US20150171973A1 (en) 2013-03-13 2013-12-12 Proximity-based and acoustic control of media devices for media presentations

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US13/802,646 US9319149B2 (en) 2013-03-13 2013-03-13 Proximity-based control of media devices for media presentations
US14/105,127 US20150171973A1 (en) 2013-03-13 2013-12-12 Proximity-based and acoustic control of media devices for media presentations

Publications (1)

Publication Number Publication Date
US20150171973A1 true US20150171973A1 (en) 2015-06-18

Family

ID=51525235

Family Applications (3)

Application Number Title Priority Date Filing Date
US13/802,646 Active US9319149B2 (en) 2013-03-13 2013-03-13 Proximity-based control of media devices for media presentations
US14/105,127 Abandoned US20150171973A1 (en) 2013-03-13 2013-12-12 Proximity-based and acoustic control of media devices for media presentations
US15/071,161 Abandoned US20160270138A1 (en) 2013-03-13 2016-03-15 Proximity-based control of media devices for media presentations

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US13/802,646 Active US9319149B2 (en) 2013-03-13 2013-03-13 Proximity-based control of media devices for media presentations

Family Applications After (1)

Application Number Title Priority Date Filing Date
US15/071,161 Abandoned US20160270138A1 (en) 2013-03-13 2016-03-15 Proximity-based control of media devices for media presentations

Country Status (6)

Country Link
US (3) US9319149B2 (en)
EP (1) EP2974064A1 (en)
AU (1) AU2014243706A1 (en)
CA (1) CA2906582A1 (en)
RU (1) RU2015143304A (en)
WO (1) WO2014160501A1 (en)

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20180076649A1 (en) * 2016-09-15 2018-03-15 Hand E Holder Products, Inc. Wireless charging apparatus
US20180184349A1 (en) * 2015-06-05 2018-06-28 Wizedsp Ltd. A hybrid wireless communication system and method
US20180310176A1 (en) * 2017-04-24 2018-10-25 Osram Sylvania Inc. Methods and Systems For Authenticating a Device to a Wireless Network
US20180352014A1 (en) * 2017-06-02 2018-12-06 Apple Inc. Alarms for a system of smart media playback devices
WO2019077594A1 (en) * 2017-10-16 2019-04-25 Sonular Ltd. Communication management and communicating between a communication device and another device
US10382929B2 (en) 2016-04-17 2019-08-13 Sonular Ltd. Communication management and communicating between a mobile communication device and another device

Families Citing this family (54)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
AU5459299A (en) * 1998-08-20 2000-03-14 Cognis Corporation A surfactant blend based on alpha-olefin sulfonates and alkyl polyglycosides
US8074581B2 (en) 2007-10-12 2011-12-13 Steelcase Inc. Conference table assembly
JP4894826B2 (en) * 2008-07-14 2012-03-14 ソニー株式会社 COMMUNICATION DEVICE, COMMUNICATION SYSTEM, NOTIFICATION METHOD, AND PROGRAM
US10631632B2 (en) 2008-10-13 2020-04-28 Steelcase Inc. Egalitarian control apparatus and method for sharing information in a collaborative workspace
US20140361954A1 (en) * 2013-06-07 2014-12-11 Lewis Epstein Personal control apparatus and method for sharing information in a collaboration workspace
US8572513B2 (en) 2009-03-16 2013-10-29 Apple Inc. Device, method, and graphical user interface for moving a current position in content at a variable scrubbing rate
US10884607B1 (en) 2009-05-29 2021-01-05 Steelcase Inc. Personal control apparatus and method for sharing information in a collaborative workspace
US9294869B2 (en) 2013-03-13 2016-03-22 Aliphcom Methods, systems and apparatus to affect RF transmission from a non-linked wireless client
US10706096B2 (en) 2011-08-18 2020-07-07 Apple Inc. Management of local and remote media items
US9002322B2 (en) 2011-09-29 2015-04-07 Apple Inc. Authentication with secondary approver
US20140354441A1 (en) * 2013-03-13 2014-12-04 Michael Edward Smith Luna System and constituent media device components and media device-based ecosystem
US9319149B2 (en) 2013-03-13 2016-04-19 Aliphcom Proximity-based control of media devices for media presentations
US11044451B2 (en) 2013-03-14 2021-06-22 Jawb Acquisition Llc Proximity-based control of media devices for media presentations
WO2014143776A2 (en) 2013-03-15 2014-09-18 Bodhi Technology Ventures Llc Providing remote interactions with host device using a wireless device
TWI526960B (en) * 2013-12-23 2016-03-21 財團法人工業技術研究院 Method and system for brokering between devices and network services
US9990129B2 (en) 2014-05-30 2018-06-05 Apple Inc. Continuity of application across devices
US10339293B2 (en) 2014-08-15 2019-07-02 Apple Inc. Authenticated device used to unlock another device
WO2016036541A2 (en) 2014-09-02 2016-03-10 Apple Inc. Phone user interface
CN113824998A (en) 2014-09-02 2021-12-21 苹果公司 Music user interface
US10275138B2 (en) 2014-09-02 2019-04-30 Sonos, Inc. Zone recognition
TWI505244B (en) * 2014-11-28 2015-10-21 Nico Technology Ltd Media player system
CN105763594B (en) * 2014-12-19 2021-05-25 阿里巴巴集团控股有限公司 Information pushing method and device
WO2016104988A1 (en) * 2014-12-23 2016-06-30 엘지전자 주식회사 Mobile terminal, audio output device and audio output system comprising same
US9474132B1 (en) * 2015-06-19 2016-10-18 Uniform Industrial Corp. Light control device and system, and method of light control
FR3049732B1 (en) * 2016-03-29 2023-04-21 Duc Long Banh SYSTEM FOR BRINGING PEOPLE INTO CONTACT EACH HAVING A TOUCH TELEPHONE TERMINAL
DK179186B1 (en) 2016-05-19 2018-01-15 Apple Inc REMOTE AUTHORIZATION TO CONTINUE WITH AN ACTION
DK201670622A1 (en) 2016-06-12 2018-02-12 Apple Inc User interfaces for transactions
CN105933539B (en) * 2016-06-22 2019-12-06 北京小米移动软件有限公司 audio playing control method and device and terminal
US10264213B1 (en) 2016-12-15 2019-04-16 Steelcase Inc. Content amplification system and method
US9730089B1 (en) * 2016-12-23 2017-08-08 Quantenna Communications, Inc. Remote controlled WiFi transceiver for wireless home networks
US10327272B2 (en) * 2017-01-05 2019-06-18 International Business Machines Corporation Using peripheral IoT devices to identify primary device replacements
US11070647B1 (en) * 2017-03-14 2021-07-20 Parallels International Gmbh Seamless cross-platform synchronization of user activities and application data between mobile and desktop devices
US11431836B2 (en) 2017-05-02 2022-08-30 Apple Inc. Methods and interfaces for initiating media playback
US10992795B2 (en) 2017-05-16 2021-04-27 Apple Inc. Methods and interfaces for home media control
US10928980B2 (en) 2017-05-12 2021-02-23 Apple Inc. User interfaces for playing and managing audio items
US20220279063A1 (en) 2017-05-16 2022-09-01 Apple Inc. Methods and interfaces for home media control
CN111343060B (en) 2017-05-16 2022-02-11 苹果公司 Method and interface for home media control
CN108513715A (en) * 2017-07-26 2018-09-07 深圳市大疆灵眸科技有限公司 Wirelessly communicate matching method, radio receiver and wireless transmitting device, system
US10949508B2 (en) * 2017-08-11 2021-03-16 Productionpal, Llc System and method to protect original music from unauthorized reproduction and use
US10599259B2 (en) * 2017-11-20 2020-03-24 Google Llc Virtual reality / augmented reality handheld controller sensing
JP6863321B2 (en) * 2018-03-22 2021-04-21 京セラドキュメントソリューションズ株式会社 Image forming device
US10616425B2 (en) * 2018-03-26 2020-04-07 Kabushiki Kaisha Toshiba Image forming apparatus and image forming method
US11010121B2 (en) 2019-05-31 2021-05-18 Apple Inc. User interfaces for audio media control
KR20220027295A (en) 2019-05-31 2022-03-07 애플 인크. User interfaces for audio media control
US10867608B1 (en) 2019-05-31 2020-12-15 Apple Inc. Multi-user configuration
DK201970533A1 (en) 2019-05-31 2021-02-15 Apple Inc Methods and user interfaces for sharing audio
US10904029B2 (en) 2019-05-31 2021-01-26 Apple Inc. User interfaces for managing controllable external devices
US11513667B2 (en) 2020-05-11 2022-11-29 Apple Inc. User interface for audio message
US11395232B2 (en) * 2020-05-13 2022-07-19 Roku, Inc. Providing safety and environmental features using human presence detection
US11202121B2 (en) 2020-05-13 2021-12-14 Roku, Inc. Providing customized entertainment experience using human presence detection
US11736767B2 (en) * 2020-05-13 2023-08-22 Roku, Inc. Providing energy-efficient features using human presence detection
US11392291B2 (en) 2020-09-25 2022-07-19 Apple Inc. Methods and interfaces for media control with dynamic feedback
US11847378B2 (en) 2021-06-06 2023-12-19 Apple Inc. User interfaces for audio routing
US11909835B2 (en) * 2021-09-17 2024-02-20 Qualcomm Incorporated Content transfer using radio frequency sensing

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20090085724A1 (en) * 2004-02-25 2009-04-02 Accenture Global Services Gmbh Rfid enabled media system and method that provides dynamic downloadable media content
US20100281261A1 (en) * 2007-11-21 2010-11-04 Nxp B.V. Device and method for near field communications using audio transducers
US20140154993A1 (en) * 2012-12-04 2014-06-05 Paul Andrew Bates Mobile Source Media Content Access

Family Cites Families (45)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
TW456112B (en) 1999-12-10 2001-09-21 Sun Wave Technology Corp Multi-function remote control with touch screen display
US7890661B2 (en) * 2001-05-16 2011-02-15 Aol Inc. Proximity synchronizing audio gateway device
US7132973B2 (en) 2003-06-20 2006-11-07 Lucent Technologies Inc. Universal soft remote control
US7562379B2 (en) 2003-12-22 2009-07-14 Sony Corporation Method and system for wireless digital multimedia presentation
US7653883B2 (en) * 2004-07-30 2010-01-26 Apple Inc. Proximity detector in handheld device
US20060161621A1 (en) * 2005-01-15 2006-07-20 Outland Research, Llc System, method and computer program product for collaboration and synchronization of media content on a plurality of media players
US8244179B2 (en) * 2005-05-12 2012-08-14 Robin Dua Wireless inter-device data processing configured through inter-device transmitted data
US20090238384A1 (en) 2006-01-05 2009-09-24 Todd Beauchamp Method and support structure for integrating audio and video components
US20080005418A1 (en) 2006-05-09 2008-01-03 Jorge Julian Interactive interface for electronic devices
US7860038B2 (en) * 2006-08-04 2010-12-28 Microsoft Corporation Wireless support for portable media player devices
US8154418B2 (en) 2008-03-31 2012-04-10 Magna Mirrors Of America, Inc. Interior rearview mirror system
US8842076B2 (en) 2008-07-07 2014-09-23 Rockstar Consortium Us Lp Multi-touch touchscreen incorporating pen tracking
US8024007B2 (en) 2008-07-28 2011-09-20 Embarq Holdings Company, Llc System and method for a projection enabled VoIP phone
DE102008046493B4 (en) 2008-09-09 2010-07-29 Fm Marketing Gmbh Multimedia arrangement with a programmable universal remote control
WO2010057057A1 (en) * 2008-11-14 2010-05-20 Wms Gaming, Inc. Storing and using casino content
US20100173585A1 (en) 2009-01-08 2010-07-08 Microsoft Corporation Seamless data communication experience
US8700097B2 (en) 2009-02-05 2014-04-15 Samsung Electronics Co., Ltd. Method and system for controlling dual-processing of screen data in mobile terminal having projector function
US8579442B2 (en) 2009-05-27 2013-11-12 Transpacific Image, Llc Advertisement content selection and presentation
US8521217B2 (en) 2009-06-10 2013-08-27 Digimarc Corporation Content sharing methods and systems
JP5409263B2 (en) 2009-10-28 2014-02-05 京セラ株式会社 Mobile electronic device and mobile phone
CN102053463B (en) 2009-11-10 2012-06-27 中强光电股份有限公司 Projector and power supply control method thereof
US20110126119A1 (en) 2009-11-20 2011-05-26 Young Daniel J Contextual presentation of information
US20110298581A1 (en) 2010-06-08 2011-12-08 Wei Hsu Universal remote controller
US9134799B2 (en) 2010-07-16 2015-09-15 Qualcomm Incorporated Interacting with a projected user interface using orientation sensors
US8449118B2 (en) 2010-08-13 2013-05-28 T-Mobile Usa, Inc. Device-adjacent ambiently displayed image
US8943541B2 (en) 2010-10-11 2015-01-27 Eldon Technology Limited Holographic 3D display
US9207782B2 (en) 2010-12-16 2015-12-08 Lg Electronics Inc. Remote controller, remote controlling method and display system having the same
KR101807286B1 (en) 2011-02-11 2017-12-08 삼성전자주식회사 Method and apparatus for performing function in mobile terminal using short range communication
US9055162B2 (en) 2011-02-15 2015-06-09 Lg Electronics Inc. Method of transmitting and receiving data, display device and mobile terminal using the same
US9179182B2 (en) 2011-04-12 2015-11-03 Kenneth J. Huebner Interactive multi-display control systems
US8929817B2 (en) 2011-05-13 2015-01-06 Nokia Corporation Sensor-based touch inquiry control
US9438996B2 (en) 2012-02-21 2016-09-06 Intertrust Technologies Corporation Systems and methods for calibrating speakers
US20130330084A1 (en) 2012-04-12 2013-12-12 O2Micro Inc. Systems and Methods for Remotely Controlling Electronic Devices
US9131266B2 (en) 2012-08-10 2015-09-08 Qualcomm Incorporated Ad-hoc media presentation based upon dynamic discovery of media output devices that are proximate to one or more users
US9071900B2 (en) 2012-08-20 2015-06-30 Nokia Technologies Oy Multi-channel recording
US9654762B2 (en) 2012-10-01 2017-05-16 Samsung Electronics Co., Ltd. Apparatus and method for stereoscopic video with motion sensors
US20140331135A1 (en) 2013-01-04 2014-11-06 SookBox LLC Digital content connectivity and control via a plurality of controllers that are treated as a single controller
US9042829B2 (en) 2013-01-04 2015-05-26 Nokia Corporation Method, apparatus, and computer program product for wireless short-range communication
US9042817B2 (en) 2013-03-07 2015-05-26 Kin-Man TSE Method and system to automatically establish NFC link between NFC enabled electronic devices based on proximate distance
US9319149B2 (en) 2013-03-13 2016-04-19 Aliphcom Proximity-based control of media devices for media presentations
US20140347565A1 (en) 2013-05-21 2014-11-27 Aliphcom Media devices configured to interface with information appliances
US20140354441A1 (en) 2013-03-13 2014-12-04 Michael Edward Smith Luna System and constituent media device components and media device-based ecosystem
US20140342660A1 (en) 2013-05-20 2014-11-20 Scott Fullam Media devices for audio and video projection of media presentations
CN105308673B (en) 2013-06-13 2018-06-22 谷歌有限责任公司 For managing the method for the output in HDMI sources, system and medium
US9647621B2 (en) 2013-06-13 2017-05-09 Google Inc. Methods, systems, and media for controlling audio of an HDMI audio system

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20090085724A1 (en) * 2004-02-25 2009-04-02 Accenture Global Services Gmbh Rfid enabled media system and method that provides dynamic downloadable media content
US20100281261A1 (en) * 2007-11-21 2010-11-04 Nxp B.V. Device and method for near field communications using audio transducers
US20140154993A1 (en) * 2012-12-04 2014-06-05 Paul Andrew Bates Mobile Source Media Content Access

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20180184349A1 (en) * 2015-06-05 2018-06-28 Wizedsp Ltd. A hybrid wireless communication system and method
US10382929B2 (en) 2016-04-17 2019-08-13 Sonular Ltd. Communication management and communicating between a mobile communication device and another device
US10735933B2 (en) 2016-04-17 2020-08-04 Sonular Ltd. Communication management and communicating between a mobile communication device and another device
US20180076649A1 (en) * 2016-09-15 2018-03-15 Hand E Holder Products, Inc. Wireless charging apparatus
US20180310176A1 (en) * 2017-04-24 2018-10-25 Osram Sylvania Inc. Methods and Systems For Authenticating a Device to a Wireless Network
US20180352014A1 (en) * 2017-06-02 2018-12-06 Apple Inc. Alarms for a system of smart media playback devices
US10805370B2 (en) * 2017-06-02 2020-10-13 Apple Inc. Alarms for a system of smart media playback devices
US11303686B2 (en) * 2017-06-02 2022-04-12 Apple Inc. Alarms for a system of smart media playback devices
US11949725B2 (en) 2017-06-02 2024-04-02 Apple Inc. Alarms for a system of smart media playback devices
WO2019077594A1 (en) * 2017-10-16 2019-04-25 Sonular Ltd. Communication management and communicating between a communication device and another device

Also Published As

Publication number Publication date
RU2015143304A (en) 2017-04-19
US20160270138A1 (en) 2016-09-15
AU2014243706A1 (en) 2015-11-05
WO2014160501A1 (en) 2014-10-02
US9319149B2 (en) 2016-04-19
US20140267002A1 (en) 2014-09-18
EP2974064A1 (en) 2016-01-20
CA2906582A1 (en) 2014-10-02

Similar Documents

Publication Publication Date Title
US11490061B2 (en) Proximity-based control of media devices for media presentations
US9319149B2 (en) Proximity-based control of media devices for media presentations
US9282423B2 (en) Proximity and interface controls of media devices for media presentations
US20140267148A1 (en) Proximity and interface controls of media devices for media presentations
US9294869B2 (en) Methods, systems and apparatus to affect RF transmission from a non-linked wireless client
US20140342660A1 (en) Media devices for audio and video projection of media presentations
US20150172878A1 (en) Acoustic environments and awareness user interfaces for media devices
US10219100B2 (en) Determining proximity for devices interacting with media devices
US20140347565A1 (en) Media devices configured to interface with information appliances
US10210739B2 (en) Proximity-based control of media devices
US9380613B2 (en) Media device configuration and ecosystem setup
CA2929455A1 (en) Proximity-based control of media devices for media presentations

Legal Events

Date Code Title Description
AS Assignment

Owner name: ALIPHCOM, CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:LUNA, MICHAEL EDWARD SMITH;REEL/FRAME:035440/0139

Effective date: 20150413

AS Assignment

Owner name: BLACKROCK ADVISORS, LLC, NEW JERSEY

Free format text: SECURITY INTEREST;ASSIGNORS:ALIPHCOM;MACGYVER ACQUISITION LLC;ALIPH, INC.;AND OTHERS;REEL/FRAME:035531/0312

Effective date: 20150428

AS Assignment

Owner name: BLACKROCK ADVISORS, LLC, NEW JERSEY

Free format text: SECURITY INTEREST;ASSIGNORS:ALIPHCOM;MACGYVER ACQUISITION LLC;ALIPH, INC.;AND OTHERS;REEL/FRAME:036500/0173

Effective date: 20150826

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION

AS Assignment

Owner name: BLACKROCK ADVISORS, LLC, NEW JERSEY

Free format text: CORRECTIVE ASSIGNMENT TO CORRECT THE APPLICATION NO. 13870843 PREVIOUSLY RECORDED ON REEL 036500 FRAME 0173. ASSIGNOR(S) HEREBY CONFIRMS THE SECURITY INTEREST;ASSIGNORS:ALIPHCOM;MACGYVER ACQUISITION, LLC;ALIPH, INC.;AND OTHERS;REEL/FRAME:041793/0347

Effective date: 20150826

AS Assignment

Owner name: JB IP ACQUISITION LLC, NEW YORK

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:ALIPHCOM, LLC;BODYMEDIA, INC.;REEL/FRAME:049805/0582

Effective date: 20180205

AS Assignment

Owner name: J FITNESS LLC, NEW YORK

Free format text: UCC FINANCING STATEMENT;ASSIGNOR:JB IP ACQUISITION, LLC;REEL/FRAME:049825/0718

Effective date: 20180205

Owner name: J FITNESS LLC, NEW YORK

Free format text: UCC FINANCING STATEMENT;ASSIGNOR:JAWBONE HEALTH HUB, INC.;REEL/FRAME:049825/0659

Effective date: 20180205

Owner name: J FITNESS LLC, NEW YORK

Free format text: SECURITY INTEREST;ASSIGNOR:JB IP ACQUISITION, LLC;REEL/FRAME:049825/0907

Effective date: 20180205

AS Assignment

Owner name: ALIPHCOM LLC, NEW YORK

Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:BLACKROCK ADVISORS, LLC;REEL/FRAME:050005/0095

Effective date: 20190529

AS Assignment

Owner name: J FITNESS LLC, NEW YORK

Free format text: RELEASE BY SECURED PARTY;ASSIGNORS:JAWBONE HEALTH HUB, INC.;JB IP ACQUISITION, LLC;REEL/FRAME:050067/0286

Effective date: 20190808