US20150105960A1 - Autonomous vehicle media control - Google Patents
Autonomous vehicle media control Download PDFInfo
- Publication number
- US20150105960A1 US20150105960A1 US14/050,719 US201314050719A US2015105960A1 US 20150105960 A1 US20150105960 A1 US 20150105960A1 US 201314050719 A US201314050719 A US 201314050719A US 2015105960 A1 US2015105960 A1 US 2015105960A1
- Authority
- US
- United States
- Prior art keywords
- media content
- vehicle
- computer
- media
- event
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 claims description 28
- 230000007257 malfunction Effects 0.000 claims description 4
- 230000008569 process Effects 0.000 description 21
- 238000004891 communication Methods 0.000 description 13
- 230000007246 mechanism Effects 0.000 description 7
- 238000010586 diagram Methods 0.000 description 6
- 230000009471 action Effects 0.000 description 4
- 230000008859 change Effects 0.000 description 4
- 238000012544 monitoring process Methods 0.000 description 4
- 230000001133 acceleration Effects 0.000 description 3
- 238000005516 engineering process Methods 0.000 description 3
- 230000005540 biological transmission Effects 0.000 description 2
- 230000001413 cellular effect Effects 0.000 description 2
- 230000006870 function Effects 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000006855 networking Effects 0.000 description 2
- 230000003287 optical effect Effects 0.000 description 2
- 230000004044 response Effects 0.000 description 2
- 230000001960 triggered effect Effects 0.000 description 2
- 230000003466 anti-cipated effect Effects 0.000 description 1
- 230000010267 cellular communication Effects 0.000 description 1
- 238000004590 computer program Methods 0.000 description 1
- 238000010276 construction Methods 0.000 description 1
- 238000013480 data collection Methods 0.000 description 1
- 230000001419 dependent effect Effects 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000018109 developmental process Effects 0.000 description 1
- 230000007613 environmental effect Effects 0.000 description 1
- 239000000835 fiber Substances 0.000 description 1
- 230000002452 interceptive effect Effects 0.000 description 1
- 230000008520 organization Effects 0.000 description 1
- 230000002085 persistent effect Effects 0.000 description 1
- 230000007704 transition Effects 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05D—SYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
- G05D1/00—Control of position, course or altitude of land, water, air, or space vehicles, e.g. automatic pilot
- G05D1/0055—Control of position, course or altitude of land, water, air, or space vehicles, e.g. automatic pilot with safety arrangements
- G05D1/0061—Control of position, course or altitude of land, water, air, or space vehicles, e.g. automatic pilot with safety arrangements for transition from automatic pilot to manual pilot and vice versa
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W50/00—Details of control systems for road vehicle drive control not related to the control of a particular sub-unit, e.g. process diagnostic or vehicle driver interfaces
- B60W50/08—Interaction between the driver and the control system
- B60W50/14—Means for informing the driver, warning the driver or prompting a driver intervention
Definitions
- FIG. 1 is a block diagram of an exemplary autonomous vehicle system including monitoring and control of media playback in a vehicle.
- FIG. 2 is a block diagram of media content with respect to an event.
- FIG. 3 is a diagram of an exemplary process for monitoring and controlling media playback in an autonomous vehicle.
- FIG. 1 is a block diagram of an exemplary autonomous vehicle system 100 including media playback and media playback control mechanisms.
- a computer 105 may be configured for communicating with one or more remote sites such as a server 125 via a network 120 , such remote site possibly including a data store 130 .
- a vehicle 101 includes a vehicle computer 105 that is configured to receive information, e.g., collected data 111 , from one or more data collectors 110 related to various components or conditions of the vehicle 101 , e.g., components such as a steering system, a braking system, a powertrain, etc.
- the computer 105 generally includes an autonomous driving module 106 that comprises instructions for autonomously, i.e., without operator input, operating the vehicle 101 , using the collected data 111 , in response to instructions received from a server 125 , etc.
- the computer 105 e.g., in the module 106 , generally includes instructions for receiving data, e.g., data 111 from one or more data collectors 110 and/or a human machine interface (HMI), such as an interactive voice response (IVR) system, a graphical user interface (GUI) including a touchscreen or the like, etc.
- HMI human machine interface
- the computer 105 may further receive information from the server 125 , e.g., relating to road conditions, traffic conditions, weather conditions, a vehicle 101 route, etc.
- the computer 105 may further include, or be communicatively coupled to a device that includes, a media module 115 for playing items of media content 116 in the vehicle 101 .
- a user device 150 e.g., a tablet computer, smart phone, or the like, could be used to play media content 116 in the vehicle 101 .
- the computing device 105 may provide instructions to a media module 115 relating to playback of media content 116 in a display within the vehicle 101 , e.g., a display of the computing device 105 , the user device 150 , etc.
- the computing device 105 may instruct the media module 115 to pause or modify playback of media content 116 .
- the computing device 105 may instruct the media module 115 to resume playback. Additionally or alternatively, playback of media content 116 may be paused or cease when autonomous driving operations are paused or cease.
- a vehicle 101 generally includes a vehicle computer 105 that includes a processor and a memory, the memory including one or more forms of computer-readable media, and storing instructions executable by the processor for performing various operations, including as disclosed herein.
- the computer 105 may include more than one computing device, e.g., controllers or the like included in the vehicle 101 for monitoring and/or controlling various vehicle components, e.g., an engine control unit (ECU), transmission control unit (TCU), etc.
- the computer 105 is generally configured for communications on a controller area network (CAN) bus or the like.
- the computer 105 may also have a connection to an onboard diagnostics connector (OBD-II).
- OBD-II onboard diagnostics connector
- the computer 105 may transmit messages to various devices in a vehicle and/or receive messages from the various devices, e.g., controllers, actuators, sensors, etc., including data collectors 110 .
- the computer 105 actually comprises multiple devices, e.g., where a media player including the media module 115 is a separate device in the vehicle 101
- the CAN bus or the like may be used for communications between devices represented as the computer 105 in this disclosure.
- the computer 105 may be configured for communicating with devices within and without the vehicle 101 via the network 120 , which, as described below, may include various wired and/or wireless networking technologies, e.g., cellular, Bluetooth, wired and/or wireless packet networks, etc.
- the network 120 may include various wired and/or wireless networking technologies, e.g., cellular, Bluetooth, wired and/or wireless packet networks, etc.
- an autonomous driving module 106 Generally included in instructions stored in and executed by the computer 105 is an autonomous driving module 106 .
- the module 106 may control various vehicle 101 components and/or operations without a driver to operate the vehicle 101 .
- the module 106 may be used to regulate vehicle 101 speed, acceleration, deceleration, steering, braking, including regular and emergency brakes, transmission control, operation of components such as lights, windshield wipers, etc.
- the module 106 may include instructions for determining an occurrence of an event according to which playback of media content 116 should be modified or paused.
- Data collectors 110 may include a variety of devices. For example, various controllers in a vehicle 101 may operate as data collectors 110 to provide data 111 via the CAN bus, e.g., data 111 relating to vehicle speed, acceleration, etc. Further, sensors or the like, cameras, global positioning system (GPS) equipment, etc., could be included in a vehicle and configured as data collectors 110 to provide data directly to the computer 105 , e.g., via a wired or wireless connection. Sensor data collectors 110 could include mechanisms such as RADAR, LADAR (also sometimes known as LIDAR), ultrasonic, etc. sensors that could be deployed to measure a distance between the vehicle 101 and other vehicles or objects.
- GPS global positioning system
- a memory of the computer 105 generally stores collected data 111 .
- Collected data 111 may include a variety of data collected in a vehicle 101 . Examples of collected data 111 are provided above, and moreover, data 111 is generally collected using one or more data collectors 110 as described above, and may additionally include data calculated therefrom in the computer 105 , and/or at the server 125 .
- collected data 111 may include any data that may be gathered by a collection device 110 and/or computed from such data. Accordingly, collected data 111 could include a variety of data related to vehicle 101 operations and/or performance, as well as data related to environmental conditions, road conditions, etc. relating to the vehicle 101 .
- the network 120 represents one or more mechanisms by which a vehicle computer 105 may communicate with a remote server 125 .
- the network 120 may be one or more of various wired or wireless communication mechanisms, including any desired combination of wired (e.g., cable and fiber) and/or wireless (e.g., cellular, wireless, satellite, microwave, and radio frequency) communication mechanisms and any desired network topology (or topologies when multiple communication mechanisms are utilized).
- Exemplary communication networks include wireless communication networks (e.g., using Bluetooth, IEEE 802.11, etc.), local area networks (LAN) and/or wide area networks (WAN), including the Internet, providing data communication services.
- the server 125 may be one or more computer servers, each generally including at least one processor and at least one memory, the memory storing instructions executable by the processor, including instructions for carrying out various steps and processes described herein.
- the server 125 may include or be communicatively coupled to a data store 130 for storing collected data 111 and/or parameters 116 .
- collected data 111 relating to road conditions, weather conditions, etc. could be stored in the data store 130 .
- Such collected data 111 from a vehicle 101 could be aggregated with collected data 111 from one or more other vehicles 101 by the server 125 , and used for the server 125 to provide instructions to one or more vehicles 101 concerning autonomous operations and/or to provide information such as traffic data or the like to the computer 105 .
- the media module 115 generally includes instructions stored in a computer memory and executable by a computer processor to play media content 116 on a display in the vehicle 101 , e.g., a display of the computing device 105 , user device 150 , etc.
- the media module 115 may be included in the computer 105 or in a separate device communicatively coupled to the computer 105 . In any case, the media module 115 may receive instructions from the computer 105 to pause, modify, resume, or otherwise control playback of media content 116 .
- “playback” can mean playing recorded content 116 , streaming live content 116 and/or content 116 downloaded via the network 120 on a real-time or near real-time basis, presenting content 116 as part of a video game, virtual reality, or the like, etc.
- media content 116 may be provided in a format such as the MPEG-1, MPEG-2 or the H.264/MPEG-4 Advanced Video Coding standards (AVC) (H.264 and MPEG-4 at present being consistent), or according to some other standard or standards.
- AVC H.264/MPEG-4 Advanced Video Coding standards
- media content 116 could be audio data formatted according to standards such as MPEG-2 Audio Layer III (MP3), Advanced Audio Coding (AAC), etc.
- MP3 MPEG-2 Audio Layer III
- AAC Advanced Audio Coding
- Media content 116 metadata may be used to identify indices or locations such as timestamps, frame numbers, etc. in the media content 116 . Such indices may be used to access one or more portions of media data 116 , e.g., such as may be specified according to pointers or the like indicated by instructions in the computer 105 and/or stored in a memory of the computer 105 .
- FIG. 2 is a diagram of media content 116 with respect to an event 205 that may be determined by the computer 105 . That is, the computer 105 may determine the occurrence of an event, and may note a location, e.g., according to an index, in the media content 116 .
- the event index 205 may substantially coincide with a pause or modify index 210 . That is, playback of media content 116 may be paused or modified at a location in the media content 116 contemporaneous with, or substantially contemporaneous with, a location of the media content 116 being played back at a time when the event arose.
- indices may also be noted by the computer 105 with respect to the media content 116 .
- the computer 105 may include instructions to identify a contextual rewind index 215 , that is, a location in the media content 116 prior to the location where playback was paused or modified from which playback of media content 116 should be re-started after having been paused for an event.
- a contextual rewind index 215 could be determined according to a number of seconds, a number of frames, etc. prior to a pause or modify location 210 , and included in instructions of the computer 105 .
- media content 116 metadata could include information for determining a contextual rewind index 215 for a particular item of media content 116 . The computer 105 could then determine, e.g., using an offset provided in the media content 116 metadata, an appropriate contextual rewind index 215 .
- a resume normal playback index 220 may specify a location where sound may be restored to the video. For example, sound could be muted to provide an informational alert in the case of an event 205 that a vehicle 101 occupant may wish to be advised of, but that does not necessitate complete attention or possible control of the vehicle 101 by an occupant. Accordingly, sound in a video item of media content 116 could be muted between a modify index 210 and a resume normal playback index 220 .
- a user device 150 may be any one of a variety of computing devices including a processor and a memory, as well as communication capabilities.
- the user device 150 may be a portable computer, tablet computer, a smart phone, etc. that includes capabilities for wireless communications using IEEE 802.11, Bluetooth, and/or cellular communications protocols.
- the user device 150 may use such communication capabilities to communicate via the network 120 and also directly with a vehicle computer 105 , e.g., using Bluetooth.
- a user device 150 may be used to carry out certain operations herein ascribed to a data collector 110 , e.g., voice recognition functions, cameras, global positioning system (GPS) functions, etc., in a user device 150 could be used to provide data 111 to the computer 105 . Further, a user device 150 could be used to provide a human machine interface (HMI) to the computer 105 . Moreover, a user device 150 could carry out operations of a media module 115 , including playing media content 116 .
- a data collector 110 e.g., voice recognition functions, cameras, global positioning system (GPS) functions, etc.
- FIG. 3 is a diagram of an exemplary process for monitoring and controlling media playback in an autonomous vehicle.
- the process 300 begins in a block 305 , in which the vehicle 101 conducts autonomous driving operations, i.e., operation of the vehicle 101 is performed in a manner partially or completely controlled by the autonomous driving module 106 .
- autonomous driving operations i.e., operation of the vehicle 101 is performed in a manner partially or completely controlled by the autonomous driving module 106 .
- all vehicle 101 operations e.g., steering, braking, speed, etc.
- the vehicle 101 may be operated in a partially autonomous (i.e., partially manual fashion, where some operations, e.g., braking, could be manually controlled by a driver, while other operations, e.g., steering, could be controlled by the computer 105 .
- the media module 115 e.g., according to instructions in the computer 105 , user device 150 , separate media player in the vehicle 101 , etc., conducts playback of media content 116 .
- the computer 105 could include or be connected to a media player, e.g., an audio player, video player, etc., in the vehicle 101 .
- a user device 150 could establish communication with the computer 105 , e.g., via a wireless networking technology such as Bluetooth, etc., whereupon the user device 150 could perform playback of media content 116 .
- the computer 105 is generally configured to provide instructions to control playback of media content 116 in the vehicle 101 as described herein.
- the computer 105 determines whether an event has occurred warranting pause or modification of playback of media content 116 .
- An event could be triggered in a variety of ways. For example, a navigation module included in, or in communication with, the computer 105 could determine that the vehicle 101 was nearing a point in a route demanding vehicle 101 occupant attention and/or for which autonomous vehicle operations may not be possible, etc., e.g., approaching intersection, entering severe traffic, entering an exit ramp of a freeway, approaching severe weather, entering a driveway, etc.
- an event could be triggered by information and/or instructions from the server 125 . For example, where the server 125 is responsible for providing instructions to the vehicle 101 for autonomous operations, the server 125 could provide an instruction for manual control of the vehicle 101 to be resumed, which would be an event warranting pause of playback of media content 116 .
- a “pause event” is an event requiring vehicle 101 occupant attention and possibly transition to manual control of vehicle 101 .
- a “modify event” is an event concerning which vehicle 101 occupant attention and notification is desirable, warranting modifying playback of media content 116 to make the playback less intrusive than it would otherwise be but not ceasing the playback altogether.
- the nature of an event could be determined from a type of media content 116 being played back.
- media content 116 could include sound and/or images related to a videogame.
- content 116 such as a video game where modifying the content 116 , e.g., to suppress some or all of the content 116 , could interfere with the user's ability to properly and completely enjoy the content, e.g., play the videogame, all events for the content 116 could be pause events.
- a “pause event” may be dependent on a location in the vehicle 101 at which media content 116 is being played. For example, where media content 116 is being played in a back seat, passenger seat, etc., the media content 116 may not be subject to a pause event. On the other hand, where the media content 116 is being played at an operator's position, is audio content played through speakers in the vehicle 101 cabin, etc., then the media content 116 may be subject to a pause event.
- the process 300 proceeds to a block 320 . Otherwise, a block 340 is executed next.
- the computer 105 identifies an index or location 205 in the media content 116 related to the event.
- event index 205 is generally substantially the same as a pause or modify index 210 , i.e., media content 116 is generally paused as soon as the computer 105 recognizes an event.
- the computer 105 generally provides a message or alert to a user via a human machine interface (HMI) included in the vehicle 101 and/or via a user device 150 .
- the message or alert may indicate the nature of the event detected as described above in the block 315 , e.g., heavy traffic, route changing, freeway exit approaching, etc.
- the message or alert may request user input.
- a user may be able to provide inputs concerning whether to assume manual control of the vehicle 101 , resume playback of the media content 116 , specify a location in the media content, e.g., a time index or the like, at which playback should be resumed, etc.
- the options concerning which the user can provide input may be varied depending on the type of media. For example, if the user is listening to or watching streaming media 116 such as a live sporting event, a vehicle 101 HMI could provide the option to resume the media feed where it was paused, e.g., at the index 210 , at some point prior to where it was paused, e.g., at an index 215 , or to resume streaming the media content 116 at the current point in a live feed, e.g., at an index 220 . Not all options make sense for all types of streaming media 116 ; for example, a user viewing a movie would not be given an option to resume a current live feed. Likewise, in the case of a video game or the like, pausing and resuming at a pause index 210 may be the only option.
- the computer 105 determines whether playback of the media content 116 should be resumed. For example, a user may provide input to resume playback, a time for which media content 116 playback should be modified, e.g., a time when a video should be muted, as specified in instructions of the computer 105 in conjunction with a particular event, may have elapsed, etc. If playback of the media content 116 is not to be resumed, e.g., based on user input, because autonomous vehicle 101 operations have ceased, etc., then the process 300 proceeds to a block 350 . Otherwise, the process 300 proceeds to a block 335 .
- the computer 105 determines a location at which to resume playback of media content 116 .
- resuming playback of media content 116 includes restoring a portion of media content 116 that may have been suppressed in conjunction with an event, e.g., restoring sound, color to video, etc.).
- an event e.g., restoring sound, color to video, etc.
- such location may be determined according to a type of media content 116 and/or a type of event according to which instructions in the computer 105 may specify a contextual rewind index 215 , a resume a normal playback index 220 , or simply resuming playback 116 from a pause or modify index 210 .
- a block 340 the computer 105 determines whether playback of the media content 116 is completed. If not, the process 300 returns to the block 310 . However, if playback of the media content 116 is completed, then a block 345 follows the block 340 .
- the computer 105 determines whether additional media content 116 , e.g., another item of media content 116 such as a song, movie, etc., is to be played. If so, the process 300 returns to the block 310 for playback of such media content 116 . Otherwise, the process 300 ends.
- additional media content 116 e.g., another item of media content 116 such as a song, movie, etc.
- the block 350 may follow the block 330 .
- the computer 105 determines whether playback of media content 116 has ceased because the vehicle 101 has ceased autonomous driving operations. If not, i.e., if autonomous driving operations are still being conducted, then the process 300 proceeds to the block 345 . Otherwise, the process 300 ends.
- Computing devices such as those discussed herein generally each include instructions executable by one or more computing devices such as those identified above, and for carrying out blocks or steps of processes described above.
- process blocks discussed above may be embodied as computer-executable instructions.
- a computer-readable medium includes any medium that participates in providing data (e.g., instructions), which may be read by a computer. Such a medium may take many forms, including, but not limited to, non-volatile media, volatile media, etc.
- Non-volatile media include, for example, optical or magnetic disks and other persistent memory.
- Volatile media include dynamic random access memory (DRAM), which typically constitutes a main memory.
- DRAM dynamic random access memory
- Computer-readable media include, for example, a floppy disk, a flexible disk, hard disk, magnetic tape, any other magnetic medium, a CD-ROM, DVD, any other optical medium, punch cards, paper tape, any other physical medium with patterns of holes, a RAM, a PROM, an EPROM, a FLASH-EEPROM, any other memory chip or cartridge, or any other medium from which a computer can read.
Abstract
Description
- A vehicle such as an automobile may be configured for autonomous driving operations. For example, the vehicle may include a central control unit or the like, i.e., a computing device having a processor and a memory, that receives data from various vehicle data collection devices such as sensors and generally also external data sources such as navigation information. The central control unit may then provide instructions to various vehicle components, e.g., actuators and the like that control steering, braking, acceleration, etc., to control vehicle operations without action, or with reduced action, by a human operator.
- An autonomous vehicle may carry one or more occupants. The autonomous vehicle is operated partially or completely without action or control by a human operator. Therefore, the autonomous vehicle may not require attention of a driver or operator. Further, vehicle occupants including a driver or operator may play various types of media in the autonomous vehicle, e.g., audio, video, video games, etc.
-
FIG. 1 is a block diagram of an exemplary autonomous vehicle system including monitoring and control of media playback in a vehicle. -
FIG. 2 is a block diagram of media content with respect to an event. -
FIG. 3 is a diagram of an exemplary process for monitoring and controlling media playback in an autonomous vehicle. -
FIG. 1 is a block diagram of an exemplaryautonomous vehicle system 100 including media playback and media playback control mechanisms. Acomputer 105 may be configured for communicating with one or more remote sites such as aserver 125 via anetwork 120, such remote site possibly including adata store 130. Avehicle 101 includes avehicle computer 105 that is configured to receive information, e.g., collecteddata 111, from one ormore data collectors 110 related to various components or conditions of thevehicle 101, e.g., components such as a steering system, a braking system, a powertrain, etc. - The
computer 105 generally includes an autonomous driving module 106 that comprises instructions for autonomously, i.e., without operator input, operating thevehicle 101, using the collecteddata 111, in response to instructions received from aserver 125, etc. Accordingly, thecomputer 105, e.g., in the module 106, generally includes instructions for receiving data, e.g.,data 111 from one ormore data collectors 110 and/or a human machine interface (HMI), such as an interactive voice response (IVR) system, a graphical user interface (GUI) including a touchscreen or the like, etc. Thecomputer 105 may further receive information from theserver 125, e.g., relating to road conditions, traffic conditions, weather conditions, avehicle 101 route, etc. - The
computer 105 may further include, or be communicatively coupled to a device that includes, amedia module 115 for playing items ofmedia content 116 in thevehicle 101. For example, a user device 150, e.g., a tablet computer, smart phone, or the like, could be used to playmedia content 116 in thevehicle 101. In any case, thecomputing device 105 may provide instructions to amedia module 115 relating to playback ofmedia content 116 in a display within thevehicle 101, e.g., a display of thecomputing device 105, the user device 150, etc. Further, when thecomputing device 105 detects one or more predetermined events, e.g., a change invehicle 101 route, a change in, or unexpected, road or traffic conditions, a change in weather conditions, an unexpected change in operation of the vehicle 101 (e.g., an equipment or component failure or malfunction, a degraded signal from asensor data collector 110, an accident, etc.) or some other event requiring attention of avehicle 101 occupant, thecomputing device 101 may instruct themedia module 115 to pause or modify playback ofmedia content 116. Once avehicle 101 occupant has been notified of an event and/or taken appropriate action and/or indicated that playback ofmedia content 116 should resume, thecomputing device 105 may instruct themedia module 115 to resume playback. Additionally or alternatively, playback ofmedia content 116 may be paused or cease when autonomous driving operations are paused or cease. - A
vehicle 101 generally includes avehicle computer 105 that includes a processor and a memory, the memory including one or more forms of computer-readable media, and storing instructions executable by the processor for performing various operations, including as disclosed herein. Further, thecomputer 105 may include more than one computing device, e.g., controllers or the like included in thevehicle 101 for monitoring and/or controlling various vehicle components, e.g., an engine control unit (ECU), transmission control unit (TCU), etc. Thecomputer 105 is generally configured for communications on a controller area network (CAN) bus or the like. Thecomputer 105 may also have a connection to an onboard diagnostics connector (OBD-II). Via the CAN bus, OBD-II, and/or other wired or wireless mechanisms, thecomputer 105 may transmit messages to various devices in a vehicle and/or receive messages from the various devices, e.g., controllers, actuators, sensors, etc., includingdata collectors 110. Alternatively or additionally, in cases where thecomputer 105 actually comprises multiple devices, e.g., where a media player including themedia module 115 is a separate device in thevehicle 101, the CAN bus or the like may be used for communications between devices represented as thecomputer 105 in this disclosure. In addition, thecomputer 105 may be configured for communicating with devices within and without thevehicle 101 via thenetwork 120, which, as described below, may include various wired and/or wireless networking technologies, e.g., cellular, Bluetooth, wired and/or wireless packet networks, etc. - Generally included in instructions stored in and executed by the
computer 105 is an autonomous driving module 106. Using data received in thecomputer 105, e.g., fromdata collectors 110, theserver 125, etc., the module 106 may controlvarious vehicle 101 components and/or operations without a driver to operate thevehicle 101. For example, the module 106 may be used to regulatevehicle 101 speed, acceleration, deceleration, steering, braking, including regular and emergency brakes, transmission control, operation of components such as lights, windshield wipers, etc. Further, the module 106 may include instructions for determining an occurrence of an event according to which playback ofmedia content 116 should be modified or paused. In conjunction with determining such an event, thecomputer 105 may further include instructions to provide a message, e.g., an alert, via an HMI of thevehicle 101, and may yet further include instructions to receivevehicle 101 occupant input, e.g., to continue playback ofmedia content 116, to halt playback ofmedia content 116, etc. -
Data collectors 110 may include a variety of devices. For example, various controllers in avehicle 101 may operate asdata collectors 110 to providedata 111 via the CAN bus, e.g.,data 111 relating to vehicle speed, acceleration, etc. Further, sensors or the like, cameras, global positioning system (GPS) equipment, etc., could be included in a vehicle and configured asdata collectors 110 to provide data directly to thecomputer 105, e.g., via a wired or wireless connection.Sensor data collectors 110 could include mechanisms such as RADAR, LADAR (also sometimes known as LIDAR), ultrasonic, etc. sensors that could be deployed to measure a distance between thevehicle 101 and other vehicles or objects. - A memory of the
computer 105 generally stores collecteddata 111. Collecteddata 111 may include a variety of data collected in avehicle 101. Examples of collecteddata 111 are provided above, and moreover,data 111 is generally collected using one ormore data collectors 110 as described above, and may additionally include data calculated therefrom in thecomputer 105, and/or at theserver 125. In general, collecteddata 111 may include any data that may be gathered by acollection device 110 and/or computed from such data. Accordingly, collecteddata 111 could include a variety of data related tovehicle 101 operations and/or performance, as well as data related to environmental conditions, road conditions, etc. relating to thevehicle 101. - Continuing with
FIG. 1 , thenetwork 120 represents one or more mechanisms by which avehicle computer 105 may communicate with aremote server 125. Accordingly, thenetwork 120 may be one or more of various wired or wireless communication mechanisms, including any desired combination of wired (e.g., cable and fiber) and/or wireless (e.g., cellular, wireless, satellite, microwave, and radio frequency) communication mechanisms and any desired network topology (or topologies when multiple communication mechanisms are utilized). Exemplary communication networks include wireless communication networks (e.g., using Bluetooth, IEEE 802.11, etc.), local area networks (LAN) and/or wide area networks (WAN), including the Internet, providing data communication services. - The
server 125 may be one or more computer servers, each generally including at least one processor and at least one memory, the memory storing instructions executable by the processor, including instructions for carrying out various steps and processes described herein. Theserver 125 may include or be communicatively coupled to adata store 130 for storing collecteddata 111 and/orparameters 116. For example, collecteddata 111 relating to road conditions, weather conditions, etc. could be stored in thedata store 130. Such collecteddata 111 from avehicle 101 could be aggregated with collecteddata 111 from one or moreother vehicles 101 by theserver 125, and used for theserver 125 to provide instructions to one ormore vehicles 101 concerning autonomous operations and/or to provide information such as traffic data or the like to thecomputer 105. - The
media module 115 generally includes instructions stored in a computer memory and executable by a computer processor to playmedia content 116 on a display in thevehicle 101, e.g., a display of thecomputing device 105, user device 150, etc. Themedia module 115 may be included in thecomputer 105 or in a separate device communicatively coupled to thecomputer 105. In any case, themedia module 115 may receive instructions from thecomputer 105 to pause, modify, resume, or otherwise control playback ofmedia content 116. Note that as used herein, “playback” can mean playing recordedcontent 116, streaminglive content 116 and/orcontent 116 downloaded via thenetwork 120 on a real-time or near real-time basis, presentingcontent 116 as part of a video game, virtual reality, or the like, etc. -
Media content 116 is generally in a digital format, e.g., compressed audio and/or video data. Themedia content 116 generally includes, according to such digital format, metadata or the like in addition to data, e.g., MPEG frames, used to render a media presentation. For example, MPEG refers to a set of standards generally promulgated by the International Standards Organization/International Electrical Commission Moving Picture Experts Group (MPEG). H.264 refers to a standard promulgated by the International Telecommunications Union (ITU). Accordingly, by way of example and not limitation,media content 116 may be provided in a format such as the MPEG-1, MPEG-2 or the H.264/MPEG-4 Advanced Video Coding standards (AVC) (H.264 and MPEG-4 at present being consistent), or according to some other standard or standards. For example,media content 116 could be audio data formatted according to standards such as MPEG-2 Audio Layer III (MP3), Advanced Audio Coding (AAC), etc. - Further, the foregoing standards generally provide for including metadata in a file of
media content 116.Media content 116 metadata may be used to identify indices or locations such as timestamps, frame numbers, etc. in themedia content 116. Such indices may be used to access one or more portions ofmedia data 116, e.g., such as may be specified according to pointers or the like indicated by instructions in thecomputer 105 and/or stored in a memory of thecomputer 105. - For example,
FIG. 2 is a diagram ofmedia content 116 with respect to anevent 205 that may be determined by thecomputer 105. That is, thecomputer 105 may determine the occurrence of an event, and may note a location, e.g., according to an index, in themedia content 116. Theevent index 205 may substantially coincide with a pause or modifyindex 210. That is, playback ofmedia content 116 may be paused or modified at a location in themedia content 116 contemporaneous with, or substantially contemporaneous with, a location of themedia content 116 being played back at a time when the event arose. - Other indices may also be noted by the
computer 105 with respect to themedia content 116. For example, thecomputer 105 may include instructions to identify acontextual rewind index 215, that is, a location in themedia content 116 prior to the location where playback was paused or modified from which playback ofmedia content 116 should be re-started after having been paused for an event. Acontextual rewind index 215 could be determined according to a number of seconds, a number of frames, etc. prior to a pause or modifylocation 210, and included in instructions of thecomputer 105. Alternatively or additionally,media content 116 metadata could include information for determining acontextual rewind index 215 for a particular item ofmedia content 116. Thecomputer 105 could then determine, e.g., using an offset provided in themedia content 116 metadata, an appropriatecontextual rewind index 215. - Further, in a case where
media content 116 playback was modified rather than paused, e.g., where sound invideo media content 116 was muted, a resumenormal playback index 220 may specify a location where sound may be restored to the video. For example, sound could be muted to provide an informational alert in the case of anevent 205 that avehicle 101 occupant may wish to be advised of, but that does not necessitate complete attention or possible control of thevehicle 101 by an occupant. Accordingly, sound in a video item ofmedia content 116 could be muted between a modifyindex 210 and a resumenormal playback index 220. - Continuing with
FIG. 1 , a user device 150 may be any one of a variety of computing devices including a processor and a memory, as well as communication capabilities. For example, the user device 150 may be a portable computer, tablet computer, a smart phone, etc. that includes capabilities for wireless communications using IEEE 802.11, Bluetooth, and/or cellular communications protocols. Further, the user device 150 may use such communication capabilities to communicate via thenetwork 120 and also directly with avehicle computer 105, e.g., using Bluetooth. Accordingly, a user device 150 may be used to carry out certain operations herein ascribed to adata collector 110, e.g., voice recognition functions, cameras, global positioning system (GPS) functions, etc., in a user device 150 could be used to providedata 111 to thecomputer 105. Further, a user device 150 could be used to provide a human machine interface (HMI) to thecomputer 105. Moreover, a user device 150 could carry out operations of amedia module 115, including playingmedia content 116. -
FIG. 3 is a diagram of an exemplary process for monitoring and controlling media playback in an autonomous vehicle. - The
process 300 begins in a block 305, in which thevehicle 101 conducts autonomous driving operations, i.e., operation of thevehicle 101 is performed in a manner partially or completely controlled by the autonomous driving module 106. For example, allvehicle 101 operations, e.g., steering, braking, speed, etc., could be controlled by the module 106 in thecomputer 105. However, it is also possible that, in the block 305, thevehicle 101 may be operated in a partially autonomous (i.e., partially manual fashion, where some operations, e.g., braking, could be manually controlled by a driver, while other operations, e.g., steering, could be controlled by thecomputer 105. - Following the block 305, in a
block 310, themedia module 115, e.g., according to instructions in thecomputer 105, user device 150, separate media player in thevehicle 101, etc., conducts playback ofmedia content 116. For example, thecomputer 105 could include or be connected to a media player, e.g., an audio player, video player, etc., in thevehicle 101. Alternatively or additionally, a user device 150 could establish communication with thecomputer 105, e.g., via a wireless networking technology such as Bluetooth, etc., whereupon the user device 150 could perform playback ofmedia content 116. In any event, thecomputer 105 is generally configured to provide instructions to control playback ofmedia content 116 in thevehicle 101 as described herein. - Following the
block 310, in a block 315, thecomputer 105 determines whether an event has occurred warranting pause or modification of playback ofmedia content 116. An event could be triggered in a variety of ways. For example, a navigation module included in, or in communication with, thecomputer 105 could determine that thevehicle 101 was nearing a point in aroute demanding vehicle 101 occupant attention and/or for which autonomous vehicle operations may not be possible, etc., e.g., approaching intersection, entering severe traffic, entering an exit ramp of a freeway, approaching severe weather, entering a driveway, etc. Alternatively or additionally, an event could be triggered by information and/or instructions from theserver 125. For example, where theserver 125 is responsible for providing instructions to thevehicle 101 for autonomous operations, theserver 125 could provide an instruction for manual control of thevehicle 101 to be resumed, which would be an event warranting pause of playback ofmedia content 116. - In general, a “pause event” is an
event requiring vehicle 101 occupant attention and possibly transition to manual control ofvehicle 101. A “modify event” is an event concerning whichvehicle 101 occupant attention and notification is desirable, warranting modifying playback ofmedia content 116 to make the playback less intrusive than it would otherwise be but not ceasing the playback altogether. Further, the nature of an event could be determined from a type ofmedia content 116 being played back. For example,media content 116 could include sound and/or images related to a videogame. In the case ofcontent 116 such as a video game where modifying thecontent 116, e.g., to suppress some or all of thecontent 116, could interfere with the user's ability to properly and completely enjoy the content, e.g., play the videogame, all events for thecontent 116 could be pause events. - Further, in some implementations, a “pause event” may be dependent on a location in the
vehicle 101 at whichmedia content 116 is being played. For example, wheremedia content 116 is being played in a back seat, passenger seat, etc., themedia content 116 may not be subject to a pause event. On the other hand, where themedia content 116 is being played at an operator's position, is audio content played through speakers in thevehicle 101 cabin, etc., then themedia content 116 may be subject to a pause event. - If an event is identified in the block 315, then the
process 300 proceeds to a block 320. Otherwise, ablock 340 is executed next. In the block 320, thecomputer 105 identifies an index orlocation 205 in themedia content 116 related to the event. As noted above,such event index 205 is generally substantially the same as a pause or modifyindex 210, i.e.,media content 116 is generally paused as soon as thecomputer 105 recognizes an event. - Following the block 320, in a
block 325 thecomputer 105 generally provides a message or alert to a user via a human machine interface (HMI) included in thevehicle 101 and/or via a user device 150. For example, the message or alert may indicate the nature of the event detected as described above in the block 315, e.g., heavy traffic, route changing, freeway exit approaching, etc. Further, the message or alert may request user input. For example, a user may be able to provide inputs concerning whether to assume manual control of thevehicle 101, resume playback of themedia content 116, specify a location in the media content, e.g., a time index or the like, at which playback should be resumed, etc. The options concerning which the user can provide input may be varied depending on the type of media. For example, if the user is listening to or watchingstreaming media 116 such as a live sporting event, avehicle 101 HMI could provide the option to resume the media feed where it was paused, e.g., at theindex 210, at some point prior to where it was paused, e.g., at anindex 215, or to resume streaming themedia content 116 at the current point in a live feed, e.g., at anindex 220. Not all options make sense for all types of streamingmedia 116; for example, a user viewing a movie would not be given an option to resume a current live feed. Likewise, in the case of a video game or the like, pausing and resuming at apause index 210 may be the only option. - Following the
block 325, in ablock 330, thecomputer 105 determines whether playback of themedia content 116 should be resumed. For example, a user may provide input to resume playback, a time for whichmedia content 116 playback should be modified, e.g., a time when a video should be muted, as specified in instructions of thecomputer 105 in conjunction with a particular event, may have elapsed, etc. If playback of themedia content 116 is not to be resumed, e.g., based on user input, becauseautonomous vehicle 101 operations have ceased, etc., then theprocess 300 proceeds to ablock 350. Otherwise, theprocess 300 proceeds to ablock 335. Depending on a type ofmedia content 116, playback could also be resumed upon the user taking over control of driving thevehicle 101. For example, playback ofaudio content 116 could be resumed when the driver takes over control. Alternatively, upon playback of a sporting event, movie, or the like, an HMI in thevehicle 101 could provide an option to resume an audio-only portion of thecontent 116. Alternatively or additionally, where a second person, i.e., a passenger in addition to avehicle 101 operator, is present, the HMI could provide the option for the second person to take control of the media playback device to resume video as well as audio portions ofcontent 116 if the driver takes back control of thevehicle 101. - In the
block 335, thecomputer 105 determines a location at which to resume playback ofmedia content 116. (In this context, resuming playback ofmedia content 116 includes restoring a portion ofmedia content 116 that may have been suppressed in conjunction with an event, e.g., restoring sound, color to video, etc.). Generally, as explained above, such location may be determined according to a type ofmedia content 116 and/or a type of event according to which instructions in thecomputer 105 may specify acontextual rewind index 215, a resume anormal playback index 220, or simply resumingplayback 116 from a pause or modifyindex 210. In any case, playback ofmedia content 116 generally begins at the determined location in theblock 335. It will be apparent that, accordingly, many implementations include creating a buffer tostore media content 116 while it is paused so that playback can be resumed at a selected index. - Following the
block 335, in ablock 340, thecomputer 105 determines whether playback of themedia content 116 is completed. If not, theprocess 300 returns to theblock 310. However, if playback of themedia content 116 is completed, then ablock 345 follows theblock 340. - In the
block 345, thecomputer 105 determines whetheradditional media content 116, e.g., another item ofmedia content 116 such as a song, movie, etc., is to be played. If so, theprocess 300 returns to theblock 310 for playback ofsuch media content 116. Otherwise, theprocess 300 ends. - The
block 350 may follow theblock 330. In theblock 350, thecomputer 105 determines whether playback ofmedia content 116 has ceased because thevehicle 101 has ceased autonomous driving operations. If not, i.e., if autonomous driving operations are still being conducted, then theprocess 300 proceeds to theblock 345. Otherwise, theprocess 300 ends. - Computing devices such as those discussed herein generally each include instructions executable by one or more computing devices such as those identified above, and for carrying out blocks or steps of processes described above. For example, process blocks discussed above may be embodied as computer-executable instructions.
- Computer-executable instructions may be compiled or interpreted from computer programs created using a variety of programming languages and/or technologies, including, without limitation, and either alone or in combination, Java™, C, C++, C#, Visual Basic, Java Script, Perl, HTML, etc. In general, a processor (e.g., a microprocessor) receives instructions, e.g., from a memory, a computer-readable medium, etc., and executes these instructions, thereby performing one or more processes, including one or more of the processes described herein. Such instructions and other data may be stored and transmitted using a variety of computer-readable media. A file in a computing device is generally a collection of data stored on a computer readable medium, such as a storage medium, a random access memory, etc.
- A computer-readable medium includes any medium that participates in providing data (e.g., instructions), which may be read by a computer. Such a medium may take many forms, including, but not limited to, non-volatile media, volatile media, etc. Non-volatile media include, for example, optical or magnetic disks and other persistent memory. Volatile media include dynamic random access memory (DRAM), which typically constitutes a main memory. Common forms of computer-readable media include, for example, a floppy disk, a flexible disk, hard disk, magnetic tape, any other magnetic medium, a CD-ROM, DVD, any other optical medium, punch cards, paper tape, any other physical medium with patterns of holes, a RAM, a PROM, an EPROM, a FLASH-EEPROM, any other memory chip or cartridge, or any other medium from which a computer can read.
- In the drawings, the same reference numbers indicate the same elements. Further, some or all of these elements could be changed. With regard to the media, processes, systems, methods, etc. described herein, it should be understood that, although the steps of such processes, etc. have been described as occurring according to a certain ordered sequence, such processes could be practiced with the described steps performed in an order other than the order described herein. It further should be understood that certain steps could be performed simultaneously, that other steps could be added, or that certain steps described herein could be omitted. In other words, the descriptions of processes herein are provided for the purpose of illustrating certain embodiments, and should in no way be construed so as to limit the claimed invention.
- Accordingly, it is to be understood that the above description is intended to be illustrative and not restrictive. Many embodiments and applications other than the examples provided would be apparent to those of skill in the art upon reading the above description. The scope of the invention should be determined, not with reference to the above description, but should instead be determined with reference to the appended claims, along with the full scope of equivalents to which such claims are entitled. It is anticipated and intended that future developments will occur in the arts discussed herein, and that the disclosed systems and methods will be incorporated into such future embodiments. In sum, it should be understood that the invention is capable of modification and variation and is limited only by the following claims.
- All terms used in the claims are intended to be given their broadest reasonable constructions and their ordinary meanings as understood by those skilled in the art unless an explicit indication to the contrary in made herein. In particular, use of the singular articles such as “a,” “the,” “said,” etc. should be read to recite one or more of the indicated elements unless a claim recites an explicit limitation to the contrary.
Claims (20)
Priority Applications (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/050,719 US9335765B2 (en) | 2013-10-10 | 2013-10-10 | Autonomous vehicle media control |
DE201410220001 DE102014220001A1 (en) | 2013-10-10 | 2014-10-02 | Autonomous vehicle media control |
CN201410532583.2A CN104581355A (en) | 2013-10-10 | 2014-10-10 | Autonomous vehicle media control |
RU2014140949A RU2014140949A (en) | 2013-10-10 | 2014-10-10 | SYSTEM AND METHOD FOR MANAGING A MEDIACONTENT OF A VEHICLE IN AN AUTONOMOUS MOTION MODE |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/050,719 US9335765B2 (en) | 2013-10-10 | 2013-10-10 | Autonomous vehicle media control |
Publications (2)
Publication Number | Publication Date |
---|---|
US20150105960A1 true US20150105960A1 (en) | 2015-04-16 |
US9335765B2 US9335765B2 (en) | 2016-05-10 |
Family
ID=52738225
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/050,719 Active 2034-01-24 US9335765B2 (en) | 2013-10-10 | 2013-10-10 | Autonomous vehicle media control |
Country Status (4)
Country | Link |
---|---|
US (1) | US9335765B2 (en) |
CN (1) | CN104581355A (en) |
DE (1) | DE102014220001A1 (en) |
RU (1) | RU2014140949A (en) |
Cited By (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9588517B2 (en) * | 2015-07-14 | 2017-03-07 | Delphi Technologies, Inc. | Automated vehicle control take-over alert timing based on infotainment activation |
US20170349184A1 (en) * | 2016-06-06 | 2017-12-07 | GM Global Technology Operations LLC | Speech-based group interactions in autonomous vehicles |
US9886034B2 (en) | 2015-11-11 | 2018-02-06 | Ford Global Technologies, Llc | Vehicle control based on connectivity of a portable device |
US9904286B2 (en) | 2015-10-13 | 2018-02-27 | Nokia Technologies Oy | Method and apparatus for providing adaptive transitioning between operational modes of an autonomous vehicle |
JP2018107662A (en) * | 2016-12-27 | 2018-07-05 | パイオニア株式会社 | Information processing device, information processing method, and program |
US20190130800A1 (en) * | 2017-11-02 | 2019-05-02 | Toyota Jidosha Kabushiki Kaisha | Movable body and advertisement providing method |
WO2019121429A1 (en) * | 2017-12-18 | 2019-06-27 | Bayerische Motoren Werke Aktiengesellschaft | Method and system for the autonomy-level-dependent setting of a mobile communication terminal |
US10362068B2 (en) * | 2015-01-12 | 2019-07-23 | Volvo Car Corporation | System and method for scheduling a video conference in an autonomous vehicle |
US10696160B2 (en) | 2018-11-28 | 2020-06-30 | International Business Machines Corporation | Automatic control of in-vehicle media |
FR3102323A1 (en) * | 2019-10-17 | 2021-04-23 | Psa Automobiles Sa | VEHICLE WITH INFO-ENTERTAINMENT DEVICE WITH CONTROLLED OPERATION DEPENDING ON THE DRIVING PHASE |
US11155260B1 (en) * | 2016-12-09 | 2021-10-26 | United Services Automobile Association (Usaa) | Autonomous vehicle entity vector-based situational awareness scoring matrix |
US20230087084A1 (en) * | 2014-08-28 | 2023-03-23 | Nagravision S.A. | Intelligent content queuing from a secondary device |
Families Citing this family (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9841965B2 (en) * | 2015-06-15 | 2017-12-12 | Lear Corporation | Centralized system for software updating vehicle components |
US9796388B2 (en) * | 2015-12-17 | 2017-10-24 | Ford Global Technologies, Llc | Vehicle mode determination |
DE102016217821A1 (en) | 2016-09-17 | 2018-03-22 | Volkswagen Aktiengesellschaft | Method for interacting with display and / or operating devices in a vehicle |
DE102016225710A1 (en) * | 2016-12-21 | 2018-06-21 | Robert Bosch Gmbh | Method for operating an autonomous vehicle |
DE102017201804A1 (en) * | 2017-02-06 | 2018-08-09 | Robert Bosch Gmbh | Method for collecting data, method for updating a scenario catalog, device, computer program and machine-readable storage medium |
US10166996B2 (en) * | 2017-02-09 | 2019-01-01 | Toyota Motor Engineering & Manufacturing North America, Inc. | Systems and methods for adaptively communicating notices in a vehicle |
DE102018201523A1 (en) | 2018-02-01 | 2019-08-01 | Audi Ag | Method for operating at least one electronic data goggles in a motor vehicle and display device for a motor vehicle |
DE102018216383B4 (en) | 2018-09-25 | 2023-12-07 | Audi Ag | Method and control device for operating a head-mounted display device in a motor vehicle |
DE102020214969A1 (en) | 2020-11-27 | 2022-06-02 | Volkswagen Aktiengesellschaft | Method and device for operating a vehicle with an automatic driving function |
DE102021118627A1 (en) | 2021-07-19 | 2023-01-19 | Bayerische Motoren Werke Aktiengesellschaft | Procedure for information displays in vehicles with autonomous driving function |
Family Cites Families (13)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6707421B1 (en) * | 1997-08-19 | 2004-03-16 | Siemens Vdo Automotive Corporation | Driver information system |
US6680694B1 (en) * | 1997-08-19 | 2004-01-20 | Siemens Vdo Automotive Corporation | Vehicle information system |
US7818116B1 (en) * | 2002-12-30 | 2010-10-19 | Mapquest, Inc. | Presenting a travel route in a ground-based vehicle |
CN2845152Y (en) * | 2004-02-20 | 2006-12-06 | 彭小毛 | Automotive audio player volume controller |
CN100429101C (en) * | 2005-09-09 | 2008-10-29 | 中国科学院自动化研究所 | Safety monitoring system for running car and monitoring method |
CN1897682A (en) * | 2006-03-10 | 2007-01-17 | 中山大学 | Video-media play and user automatic interactive system and method |
US8275307B2 (en) | 2006-07-24 | 2012-09-25 | Qualcomm Incorporated | Vehicle audio integrator |
US9976865B2 (en) * | 2006-07-28 | 2018-05-22 | Ridetones, Inc. | Vehicle communication system with navigation |
US9870130B2 (en) * | 2008-05-13 | 2018-01-16 | Apple Inc. | Pushing a user interface to a remote device |
US20100118147A1 (en) * | 2008-11-11 | 2010-05-13 | Honeywell International Inc. | Methods and apparatus for adaptively streaming video data based on a triggering event |
US8317329B2 (en) * | 2009-04-02 | 2012-11-27 | GM Global Technology Operations LLC | Infotainment display on full-windshield head-up display |
US20120290150A1 (en) * | 2011-05-13 | 2012-11-15 | John Doughty | Apparatus, system, and method for providing and using location information |
CN102984562B (en) * | 2012-11-26 | 2016-06-29 | 宁波大胜日用制品有限公司 | Car multimedia play device |
-
2013
- 2013-10-10 US US14/050,719 patent/US9335765B2/en active Active
-
2014
- 2014-10-02 DE DE201410220001 patent/DE102014220001A1/en active Pending
- 2014-10-10 RU RU2014140949A patent/RU2014140949A/en not_active Application Discontinuation
- 2014-10-10 CN CN201410532583.2A patent/CN104581355A/en active Pending
Cited By (14)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20230087084A1 (en) * | 2014-08-28 | 2023-03-23 | Nagravision S.A. | Intelligent content queuing from a secondary device |
US10362068B2 (en) * | 2015-01-12 | 2019-07-23 | Volvo Car Corporation | System and method for scheduling a video conference in an autonomous vehicle |
US9588517B2 (en) * | 2015-07-14 | 2017-03-07 | Delphi Technologies, Inc. | Automated vehicle control take-over alert timing based on infotainment activation |
US9904286B2 (en) | 2015-10-13 | 2018-02-27 | Nokia Technologies Oy | Method and apparatus for providing adaptive transitioning between operational modes of an autonomous vehicle |
US9886034B2 (en) | 2015-11-11 | 2018-02-06 | Ford Global Technologies, Llc | Vehicle control based on connectivity of a portable device |
US20170349184A1 (en) * | 2016-06-06 | 2017-12-07 | GM Global Technology Operations LLC | Speech-based group interactions in autonomous vehicles |
US11155260B1 (en) * | 2016-12-09 | 2021-10-26 | United Services Automobile Association (Usaa) | Autonomous vehicle entity vector-based situational awareness scoring matrix |
JP2018107662A (en) * | 2016-12-27 | 2018-07-05 | パイオニア株式会社 | Information processing device, information processing method, and program |
JP2021185705A (en) * | 2016-12-27 | 2021-12-09 | パイオニア株式会社 | Information processing device, information processing method, and program |
US10777105B2 (en) * | 2017-11-02 | 2020-09-15 | Toyota Jidosha Kabushiki Kaisha | Movable body and advertisement providing method |
US20190130800A1 (en) * | 2017-11-02 | 2019-05-02 | Toyota Jidosha Kabushiki Kaisha | Movable body and advertisement providing method |
WO2019121429A1 (en) * | 2017-12-18 | 2019-06-27 | Bayerische Motoren Werke Aktiengesellschaft | Method and system for the autonomy-level-dependent setting of a mobile communication terminal |
US10696160B2 (en) | 2018-11-28 | 2020-06-30 | International Business Machines Corporation | Automatic control of in-vehicle media |
FR3102323A1 (en) * | 2019-10-17 | 2021-04-23 | Psa Automobiles Sa | VEHICLE WITH INFO-ENTERTAINMENT DEVICE WITH CONTROLLED OPERATION DEPENDING ON THE DRIVING PHASE |
Also Published As
Publication number | Publication date |
---|---|
CN104581355A (en) | 2015-04-29 |
DE102014220001A1 (en) | 2015-04-16 |
US9335765B2 (en) | 2016-05-10 |
RU2014140949A (en) | 2016-04-27 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US9335765B2 (en) | Autonomous vehicle media control | |
US10834221B2 (en) | Method and system for providing predictions via artificial intelligence (AI) models using a distributed system | |
JP6779835B2 (en) | Monitoring control system, monitoring control device and monitoring control method | |
US10552695B1 (en) | Driver monitoring system and method of operating the same | |
JP6102117B2 (en) | MOBILE BODY, SYSTEM, PROGRAM, AND INFORMATION PROCESSING DEVICE | |
KR20180020886A (en) | System and method for processing traffic sound data to provide driver assistance | |
CN106056697B (en) | A kind of event-monitoring methods, devices and systems | |
JP2018010406A (en) | Monitoring system | |
US20180022290A1 (en) | Systems, Methods, And Devices For Rendering In-Vehicle Media Content Based On Vehicle Sensor Data | |
US10448164B2 (en) | Acoustic playback of a digital audio medium in a motor vehicle | |
CN104574570A (en) | Rearview mirror automobile data recorder | |
US11895432B2 (en) | Recording control apparatus, recording control method, and recording control program | |
CN108924461B (en) | Video image processing method and device | |
JP6345572B2 (en) | Traveling video recording system, drive recorder used therefor, and method for uploading recorded traveling video | |
CN115297461B (en) | Data interaction method and device, vehicle, readable storage medium and chip | |
JP2011208984A (en) | Information display device for vehicle | |
KR20100129009A (en) | Black box system for vehicles | |
US11039279B1 (en) | Automated tracking of vehicle operation and synchronized media delivery | |
KR20130140438A (en) | Blackbox for a vehicle for providing a signal detection function of the brake or the accelerator | |
JP2010257483A (en) | Driving support device and driving support method | |
CN116366615A (en) | Media stream playing method and device | |
US11750771B2 (en) | Recording control apparatus, recording control method, and recording control program | |
JP7462547B2 (en) | Information processing device, mobile object, program, and information processing method | |
JP2019121314A (en) | Determination device, information recording device, determination method, and program for determination | |
JP7348724B2 (en) | In-vehicle device and display method |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: FORD GLOBAL TECHNOLOGIES, LLC, MICHIGAN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:PILUTTI, THOMAS EDWARD;RUPP, MATTHEW Y.;TROMBLEY, ROGER ARNOLD;AND OTHERS;SIGNING DATES FROM 20131009 TO 20131010;REEL/FRAME:031381/0618 |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 4 |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1552); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 8 |