US20220046319A1 - Adapting interactions with a television user - Google Patents

Adapting interactions with a television user Download PDF

Info

Publication number
US20220046319A1
US20220046319A1 US17/480,499 US202117480499A US2022046319A1 US 20220046319 A1 US20220046319 A1 US 20220046319A1 US 202117480499 A US202117480499 A US 202117480499A US 2022046319 A1 US2022046319 A1 US 2022046319A1
Authority
US
United States
Prior art keywords
user
behavior
responses
implementations
predetermined
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
US17/480,499
Inventor
Hyehoon Yi
David Young
Lindsay Miller
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Saturn Licensing LLC
Original Assignee
Sony Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Sony Corp filed Critical Sony Corp
Priority to US17/480,499 priority Critical patent/US20220046319A1/en
Assigned to Sony Group Corporation reassignment Sony Group Corporation ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: SONY CORPORATION
Assigned to SATURN LICENSING LLC reassignment SATURN LICENSING LLC ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: Sony Group Corporation
Publication of US20220046319A1 publication Critical patent/US20220046319A1/en
Pending legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/442Monitoring of processes or resources, e.g. detecting the failure of a recording device, monitoring the downstream bandwidth, the number of times a movie has been viewed, the storage space available from the internal hard disk
    • H04N21/44213Monitoring of end-user related data
    • H04N21/44218Detecting physical presence or behaviour of the user, e.g. using sensors to detect if the user is leaving the room or changes his face expression during a TV program
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/16Sound input; Sound output
    • G06F3/167Audio in a user interface, e.g. using voice commands for navigating, audio feedback
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream, rendering scenes according to MPEG-4 scene graphs
    • H04N21/44008Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream, rendering scenes according to MPEG-4 scene graphs involving operations for analysing video streams, e.g. detecting features or characteristics in the video stream
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream, rendering scenes according to MPEG-4 scene graphs
    • H04N21/44016Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream, rendering scenes according to MPEG-4 scene graphs involving splicing one content stream with another content stream, e.g. for substituting a video clip
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/442Monitoring of processes or resources, e.g. detecting the failure of a recording device, monitoring the downstream bandwidth, the number of times a movie has been viewed, the storage space available from the internal hard disk
    • H04N21/44213Monitoring of end-user related data
    • H04N21/44222Analytics of user selections, e.g. selection of programs or purchase activity
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/472End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content
    • H04N21/47214End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content for content reservation or setting reminders; for requesting event notification, e.g. of sport results or stock market

Definitions

  • voice agents are devoid of personality, and can be perceived to be emotionless or robotic to the user (e.g., lacking any emotion or feeling of empathy). While this does not affect the usefulness, it can make the experience awkward. Attempting to create an agent with personality has shortcomings because each person has a different preference and tolerance for particular personality traits. Conventional voice agents often create an awkward and undesirable user experience.
  • Implementations generally relate to adapting interactions with a television user.
  • a system includes one or more processors, and includes logic encoded in one or more non-transitory computer-readable storage media for execution by the one or more processors.
  • the logic is operable to perform operations including interacting with a user based on a first behavior pattern, wherein the first behavior pattern includes one or more predetermined behaviors; observing one or more responses from the user, wherein the one or more responses are responsive to the one or more predetermined behaviors; modifying at least one of the predetermined behaviors, wherein the modifying of the results in at least one modified behavior; observing at least one change in one or more of the responses, wherein the at least one change is responsive to the at least one modified behavior; and interacting with the user based on a second behavior pattern, wherein the second behavior pattern includes one or more of the predetermined behaviors and the modified behavior.
  • At least one of the predetermined behaviors includes communicating with the user by voice.
  • the at least one modified behavior includes conveying affects to the user.
  • the logic when executed is further operable to perform operations comprising modifying one or more voice characteristics.
  • the at least one change in one or more of the responses comprises a positive reaction or a negative reaction.
  • the logic when executed is further operable to perform operations comprising: detecting a notification to be delivered to the user; detecting when a commercial is being shown on a television; and delivering the notification to the user during the commercial.
  • the logic when executed is further operable to perform operations comprising: determining an identity of the user; and presenting to the user a predetermined user interface that is associated with the user.
  • a non-transitory computer-readable storage medium with program instructions thereon When executed by one or more processors, the instructions are operable to perform operations including interacting with a user based on a first behavior pattern, wherein the first behavior pattern includes one or more predetermined behaviors; observing one or more responses from the user, wherein the one or more responses are responsive to the one or more predetermined behaviors; modifying at least one of the predetermined behaviors, wherein the modifying of the results in at least one modified behavior; observing at least one change in one or more of the responses, wherein the at least one change is responsive to the at least one modified behavior; and interacting with the user based on a second behavior pattern, wherein the second behavior pattern includes one or more of the predetermined behaviors and the modified behavior.
  • At least one of the predetermined behaviors includes communicating with the user by voice.
  • the at least one modified behavior includes conveying affects to the user.
  • the instructions when executed are further operable to perform operations comprising modifying one or more voice characteristics.
  • the at least one change in one or more of the responses comprises a positive reaction or a negative reaction.
  • the instructions when executed are further operable to perform operations comprising: detecting a notification to be delivered to the user; detecting when a commercial is being shown on a television; and delivering the notification to the user during the commercial.
  • the instructions when executed are further operable to perform operations comprising: determining an identity of the user; and presenting to the user a predetermined user interface that is associated with the user.
  • a method includes interacting with a user based on a first behavior pattern, wherein the first behavior pattern includes one or more predetermined behaviors; observing one or more responses from the user, wherein the one or more responses are responsive to the one or more predetermined behaviors; modifying at least one of the predetermined behaviors, wherein the modifying of the results in at least one modified behavior; observing at least one change in one or more of the responses, wherein the at least one change is responsive to the at least one modified behavior; and interacting with the user based on a second behavior pattern, wherein the second behavior pattern includes one or more of the predetermined behaviors and the modified behavior.
  • At least one of the predetermined behaviors includes communicating with the user by voice.
  • the at least one modified behavior includes conveying affects to the user.
  • the method further comprises modifying one or more voice characteristics.
  • the at least one change in one or more of the responses comprises a positive reaction or a negative reaction.
  • the method further comprises: detecting a notification to be delivered to the user; detecting when a commercial is being shown on a television; and delivering the notification to the user during the commercial.
  • FIG. 1 illustrates a block diagram of an example television environment, which may be used for some implementations described herein.
  • FIG. 2 illustrates an example flow diagram for adapting interactions with a television user, according to some implementations.
  • FIG. 3 illustrates an example flow diagram for adapting interactions with a television user, according to some implementations.
  • FIG. 4 illustrates an example flow diagram for adapting interactions with a television user, according to some implementations.
  • FIG. 5 illustrates a block diagram of an example television environment, which may be used for some implementations described herein.
  • FIG. 6 illustrates a block diagram of an example television environment, which may be used for some implementations described herein.
  • FIG. 7 illustrates a block diagram of an example computing system 700 , which may be used for some implementations described herein.
  • Implementations described herein adapt interactions with a television user.
  • a system interacts with a user based on a first behavior pattern, which includes one or more predetermined behaviors.
  • the system observes responses from the user, where the responses are responsive to the predetermined behaviors.
  • the system modifies one or more of the predetermined behaviors, which results in modified behaviors.
  • the system observes if there are any changes in the user's responses, wherein the changes are responsive to the modified behaviors.
  • the system continues interacting with the user based on a second behavior pattern, which includes one or more of the predetermined behaviors and the modified behavior.
  • FIG. 1 illustrates a block diagram of an example television environment 100 , which may be used for some implementations described herein.
  • television environment 100 includes a television 102 , which interacts with a user 104 .
  • Television 102 interacts with user 104 based on behavior patterns.
  • television 102 may interact with user 104 based on a first behavior pattern (labeled Behavior Pattern 1 ), which may elicit a response from user 104 .
  • Television 102 may modify one or more behaviors, which in turn may elicit a response from user 104 .
  • television 102 may resume interacting with user 104 with a second behavior pattern (labeled Behavior Pattern 2 ), which includes at least some of the behaviors of the first behavior pattern and the modified behavior pattern.
  • Example implementations are described in more detail herein.
  • television 102 includes a voice agent for interacting with the user.
  • Television 102 also incudes a built-in microphone and a built-in camera.
  • television 102 may observe the user's facial expressions and see how the user responds.
  • television 102 may use machine learning techniques in order to determine if particular communication styles are successful or not.
  • FIG. 1 shows one block for television 102 .
  • Block 102 may represent multiple systems and/or devices.
  • television 102 may not have all of the components shown and/or may have other elements including other types of elements instead of, or in addition to, those shown herein.
  • Television 102 may perform the implementations described herein alone or in combination with other devices.
  • the system may stored unique user profiles for each user of multiple users.
  • FIG. 2 illustrates an example flow diagram for adapting interactions with a television user, according to some implementations.
  • a method is initiated at block 202 , where the system such as television 102 interacts with a user based on a first behavior pattern, wherein the first behavior pattern includes one or more predetermined behaviors.
  • the system interacts by sounds such as by voice or with music or other sounds.
  • a predetermined behavior may include communicating with the user by voice.
  • a predetermined behavior may include providing particular information to the user (e.g., movie recommendations, etc.).
  • the predetermined behaviors may vary depending on the implementation.
  • the system may use varying voice tones (e.g., high, low, cheery or calm, etc.), speed, level of engagement (e.g., proactive or passive; long answers or short answers, etc.) and other factors to build engagement.
  • voice tones e.g., high, low, cheery or calm, etc.
  • speed, level of engagement e.g., proactive or passive; long answers or short answers, etc.
  • other factors to build engagement e.g., the system when communicating by voice may speak with higher volume or lower volume. The system may speak more quickly or more slowly.
  • the system may use a smooth voice or a less smooth voice, etc.
  • the system may use a male voice, female voice, etc.
  • the system observes one or more responses from the user, wherein the one or more responses are responsive to the one or more predetermined behaviors.
  • the system may observe facial expressions.
  • the system may determine using facial and/or pattern recognition if the user is happy or unhappy, or confused.
  • the system modifies at least one of the predetermined behaviors, wherein the modifying results in at least one modified behavior.
  • the system modifies at least one of the predetermined behaviors by modifying one or more voice characteristics. For example, the system may communicate with the user with more volume and more slowly. Alternatively, the system may communicate in a more curt, direct, to the point manner. Such variations may range from subtle to overt.
  • a modified behavior includes conveying affects to the user.
  • the system may play background music (e.g., mood music, etc.) or may make particular sounds (e.g., beeps, etc.).
  • the system may convey affects including music, sound, lights, or pictures. Such content may be retrieved from a database of appropriate effects.
  • the system may provide selected music and/or sounds and lights conforming to the general mood of the query (e.g., happy, sad, etc.).
  • the agent confirms the diagnosis of constructive user responses and takes the next appropriate interaction with the user.
  • the system may convey visual cues along with audio cues.
  • Such cues may include, for example, colors, photos, videos, or even an avatar of the system with the capacity to show emotion to the user. These may be informed by user television viewing preferences.
  • the system observes at least one change in one or more of the responses, wherein the at least one change is responsive to the at least one modified behavior.
  • the system assess the success of different modifications by observing resulting responses from the user.
  • at least one change in one or more of the responses comprises a positive reaction or a negative reaction. For example, if the user shows a positive change such as smiling more, or becoming more interactive, the system would log the positive change in the user's responses. In contrast, if the user shows a negative changes such as smiling less, or becoming less interactive, the system would log the negative change in the user's responses.
  • the responses or feedback from the user may be implicit or explicit.
  • the system may look for facial expressions.
  • the system may receive explicit feedback from user indicating that the user likes or does not like particular modified behavior from the system.
  • the system may continuously alter its interaction mode as needed or to optimize interaction with the user.
  • the system may introduce other predetermined behavior.
  • the system may proactively inform the user of upcoming sports events or a movie schedule.
  • the user might explicitly expresses positive or negative feedback (e.g., “Keep notifying me of sports events.” or “I'm not interested.” etc.).
  • the user may behave in a way that implies positive or negative reaction.
  • the user might actively use the information to schedule a TV recording, or user might sound unenthusiastic or annoyed when responding, etc.
  • the system logs these responses, and, over time, forms an appropriate level of “proactiveness” that is appropriate for the particular user.
  • the system interacts with the user based on a second behavior pattern, wherein the second behavior pattern includes one or more of the predetermined behaviors and the modified behavior.
  • the personality that the system conveys evolves and adapts to each user.
  • Such emotional or empathetic interaction to a user's query conveys a deeper sense of understanding by the system and greatly enhances the user experience. Implementations may provide non-verbal communication when appropriate, which may result in a greater trust and bonding with the system.
  • FIG. 3 illustrates an example flow diagram for adapting interactions with a television user, according to some implementations.
  • a method is initiated at block 302 , where the system detects a notification to be delivered to the user.
  • the system detects when a commercial is being shown on a television.
  • the system may detect commercials based on volume changes, black frames, set times, etc.
  • the system delivers the notification to the user during the commercial.
  • the system avoids unwanted interruptions while the user is enjoying television content. This allows the user to enjoy television with minimal interruptions (e.g., pop-up notifications) while the user is watching their primary content. This enhances the viewing experience.
  • the system may prioritize notifications.
  • the system may deliver some predetermined notifications during the primary content. For example, the system may deliver a notification from a smart doorbell (e.g., pizza delivery, etc.) or from a particular person (e.g., family member, etc.).
  • a smart doorbell e.g., pizza delivery, etc.
  • a particular person e.g., family member, etc.
  • FIG. 4 illustrates an example flow diagram for adapting interactions with a television user, according to some implementations. As described in more detail herein, implementations optimize interaction for the specific user.
  • a method is initiated at block 402 , where the system determines the identity of the user.
  • the system may use a built-in camera and facial recognition, and a built-in mic and voice recognition to detect who is using the television.
  • the system may also determine a usage signature of the user (e.g., method of operation such as remote control, voice, phone control, etc.).
  • the system determines a predetermined user interface that is associated with the user.
  • the system switches to a profile of the particular user and sub-profile (e.g., current control method), which carries some associated information about which user interface (UI) patterns and settings are preferred for that user.
  • UI user interface
  • the system may access a library of different UI patterns to be used for various situations. Each pattern may be given a priority. In some implementations the system collects all the available factors (e.g., person, ability level, control method, etc.) and weighs those factors. The system then selects the best UI pattern for the situation. The user may also manually enter settings and select some priority levels or patterns that the user prefers.
  • factors e.g., person, ability level, control method, etc.
  • the system may store and access a user's viewing preferences in order to help the system determine or suggest particular UI settings. For example, a user who has a preference for watching sporting might want notifications of upcoming sporting events. A user who has a preference for music programs might want background music to be played with some user interfaces.
  • the system presents to the user the predetermined user interface that is associated with the user.
  • a system seamlessly and automatically switches between different UI patterns and settings based on who is operating the television. For example, the system adapts to users with specific preferences (e.g., easy controls for younger users, etc.) and/or impairments (e.g., vision, hearing, mobility, etc.).
  • Implementations allow users to operate the television in the best way for them without actively hunting through settings for particular controls. Implementations also encourage smoother operation of voice by displaying voice UI on the first attempt to use it. In multi-person households, the system automatically switches between optimal methods, giving each user a highly personalized experience.
  • the system modifies the controls of to the access capability/limitations of the user by identifying the user.
  • the system accesses the user's access/capability/limitation profile, and identifies the appropriate UI for the user. For example, a user with blindness may need the television to have a screen reader function) or have voice-over enabled.
  • the system may automatically turn such features on for that user, and may turn those features off for other users.
  • the system may speak in a particular language associated with the user.
  • FIG. 5 illustrates a block diagram of an example television environment 500 , which may be used for some implementations described herein. Shown is a television 502 and a user 504 interacting with television 502 . In this example scenario, when the system recognizes user 504 with a camera 506 , the system displays a UI 508 that is appropriate for user 504 .
  • UI 508 is a menu-style UI pattern that is preferred by user 504 .
  • the system accesses a library of different UIs, which may be stored by the system or accessed in the cloud.
  • the library may include traditional menu UIs, graphical UIs, etc.
  • FIG. 6 illustrates a block diagram of an example television environment 600 , which may be used for some implementations described herein. Shown is the same television 502 as in FIG. 5 , and a different user 604 interacting with television 502 .
  • the system when the system recognizes user 604 with a camera 506 , the system displays a UI 608 that is appropriate for user 504 .
  • UI 508 is a graphical-style UI pattern that is preferred by user 604 .
  • FIG. 7 illustrates a block diagram of an example computing system 700 , which may be used for some implementations described herein.
  • computing system 700 may be used to implement television 102 of FIG. 1 , as well as to perform implementations described herein.
  • computing system 700 may include a processor 702 , an operating system 704 , a memory 706 , and an input/output (I/O) interface 708 .
  • processor 702 may be used to implement various functions and features described herein, as well as to perform the method implementations described herein. While processor 702 is described as performing implementations described herein, any suitable component or combination of components of computing system 700 or any suitable processor or processors associated with computing system 700 or any suitable system may perform the steps described. Implementations described herein may be carried out on a user device, on a server, or a combination of both.
  • Computing system 700 also includes a software application 710 , which may be stored on memory 706 or on any other suitable storage location or computer-readable medium.
  • Software application 710 provides instructions that enable processor 702 to perform the implementations described herein and other functions.
  • Software application may also include an engine such as a network engine for performing various functions associated with one or more networks and network communications.
  • the components of computing system 700 may be implemented by one or more processors or any combination of hardware devices, as well as any combination of hardware, software, firmware, etc.
  • FIG. 7 shows one block for each of processor 702 , operating system 704 , memory 706 , I/O interface 708 , and software application 710 .
  • These blocks 702 , 704 , 706 , 708 , and 710 may represent multiple processors, operating systems, memories, I/O interfaces, and software applications.
  • computing system 700 may not have all of the components shown and/or may have other elements including other types of components instead of, or in addition to, those shown herein.
  • software is encoded in one or more non-transitory computer-readable media for execution by one or more processors.
  • the software when executed by one or more processors is operable to perform the implementations described herein and other functions.
  • routines of particular embodiments including C, C++, Java, assembly language, etc.
  • Different programming techniques can be employed such as procedural or object oriented.
  • the routines can execute on a single processing device or multiple processors. Although the steps, operations, or computations may be presented in a specific order, this order may be changed in different particular embodiments. In some particular embodiments, multiple steps shown as sequential in this specification can be performed at the same time.
  • Particular embodiments may be implemented in a non-transitory computer-readable storage medium (also referred to as a machine-readable storage medium) for use by or in connection with the instruction execution system, apparatus, or device.
  • a non-transitory computer-readable storage medium also referred to as a machine-readable storage medium
  • Particular embodiments can be implemented in the form of control logic in software or hardware or a combination of both.
  • the control logic when executed by one or more processors is operable to perform the implementations described herein and other functions.
  • a tangible medium such as a hardware storage device can be used to store the control logic, which can include executable instructions.
  • Particular embodiments may be implemented by using a programmable general purpose digital computer, and/or by using application specific integrated circuits, programmable logic devices, field programmable gate arrays, optical, chemical, biological, quantum or nanoengineered systems, components and mechanisms.
  • the functions of particular embodiments can be achieved by any means as is known in the art.
  • Distributed, networked systems, components, and/or circuits can be used. Communication, or transfer, of data may be wired, wireless, or by any other means.
  • a “processor” may include any suitable hardware and/or software system, mechanism, or component that processes data, signals or other information.
  • a processor may include a system with a general-purpose central processing unit, multiple processing units, dedicated circuitry for achieving functionality, or other systems. Processing need not be limited to a geographic location, or have temporal limitations. For example, a processor may perform its functions in “real-time,” “offline,” in a “batch mode,” etc. Portions of processing may be performed at different times and at different locations, by different (or the same) processing systems.
  • a computer may be any processor in communication with a memory.
  • the memory may be any suitable data storage, memory and/or non-transitory computer-readable storage medium, including electronic storage devices such as random-access memory (RAM), read-only memory (ROM), magnetic storage device (hard disk drive or the like), flash, optical storage device (CD, DVD or the like), magnetic or optical disk, or other tangible media suitable for storing instructions (e.g., program or software instructions) for execution by the processor.
  • a tangible medium such as a hardware storage device can be used to store the control logic, which can include executable instructions.
  • the instructions can also be contained in, and provided as, an electronic signal, for example in the form of software as a service (SaaS) delivered from a server (e.g., a distributed system and/or a cloud computing system).
  • SaaS software as a service

Abstract

Implementations generally relate to adapting interactions with a television user. In some implementations, a method includes interacting with a user based on a first behavior pattern, wherein the first behavior pattern includes one or more predetermined behaviors. The method further includes observing one or more responses from the user, wherein the one or more responses are responsive to the one or more predetermined behaviors. The method further includes modifying at least one of the predetermined behaviors, wherein the modifying of the results in at least one modified behavior. The method further includes observing at least one change in one or more of the responses, wherein the at least one change is responsive to the at least one modified behavior. The method further includes interacting with the user based on a second behavior pattern, wherein the second behavior pattern includes one or more of the predetermined behaviors and the modified behavior.

Description

    CROSS REFERENCES TO RELATED APPLICATIONS
  • This application is a continuation of the following application, U.S. patent application Ser. No. 16/055,795, entitled ADAPTING INTERACTIONS WITH A TELEVISION USER, filed on Aug. 6, 2018 (020699-112600US/SYP328331US02), which is hereby incorporated by reference as if set forth in full in this application for all purposes.
  • BACKGROUND
  • Devices are increasingly using voice agents to interact with users. The current generation of voice agents are devoid of personality, and can be perceived to be emotionless or robotic to the user (e.g., lacking any emotion or feeling of empathy). While this does not affect the usefulness, it can make the experience awkward. Attempting to create an agent with personality has shortcomings because each person has a different preference and tolerance for particular personality traits. Conventional voice agents often create an awkward and undesirable user experience.
  • SUMMARY
  • Implementations generally relate to adapting interactions with a television user. In some implementations, a system includes one or more processors, and includes logic encoded in one or more non-transitory computer-readable storage media for execution by the one or more processors. When executed, the logic is operable to perform operations including interacting with a user based on a first behavior pattern, wherein the first behavior pattern includes one or more predetermined behaviors; observing one or more responses from the user, wherein the one or more responses are responsive to the one or more predetermined behaviors; modifying at least one of the predetermined behaviors, wherein the modifying of the results in at least one modified behavior; observing at least one change in one or more of the responses, wherein the at least one change is responsive to the at least one modified behavior; and interacting with the user based on a second behavior pattern, wherein the second behavior pattern includes one or more of the predetermined behaviors and the modified behavior.
  • With further regard to the system, in some implementations, at least one of the predetermined behaviors includes communicating with the user by voice. In some implementations, the at least one modified behavior includes conveying affects to the user. In some implementations, to modify at least one of the predetermined behaviors, the logic when executed is further operable to perform operations comprising modifying one or more voice characteristics. In some implementations, the at least one change in one or more of the responses comprises a positive reaction or a negative reaction. In some implementations, the logic when executed is further operable to perform operations comprising: detecting a notification to be delivered to the user; detecting when a commercial is being shown on a television; and delivering the notification to the user during the commercial. In some implementations, the logic when executed is further operable to perform operations comprising: determining an identity of the user; and presenting to the user a predetermined user interface that is associated with the user.
  • In some embodiments, a non-transitory computer-readable storage medium with program instructions thereon is provided. When executed by one or more processors, the instructions are operable to perform operations including interacting with a user based on a first behavior pattern, wherein the first behavior pattern includes one or more predetermined behaviors; observing one or more responses from the user, wherein the one or more responses are responsive to the one or more predetermined behaviors; modifying at least one of the predetermined behaviors, wherein the modifying of the results in at least one modified behavior; observing at least one change in one or more of the responses, wherein the at least one change is responsive to the at least one modified behavior; and interacting with the user based on a second behavior pattern, wherein the second behavior pattern includes one or more of the predetermined behaviors and the modified behavior.
  • With further regard to the computer-readable storage medium, in some implementations, at least one of the predetermined behaviors includes communicating with the user by voice. In some implementations, the at least one modified behavior includes conveying affects to the user. In some implementations, to modify at least one of the predetermined behaviors, the instructions when executed are further operable to perform operations comprising modifying one or more voice characteristics. In some implementations, the at least one change in one or more of the responses comprises a positive reaction or a negative reaction. In some implementations, the instructions when executed are further operable to perform operations comprising: detecting a notification to be delivered to the user; detecting when a commercial is being shown on a television; and delivering the notification to the user during the commercial. In some implementations, the instructions when executed are further operable to perform operations comprising: determining an identity of the user; and presenting to the user a predetermined user interface that is associated with the user.
  • In some implementations, a method includes interacting with a user based on a first behavior pattern, wherein the first behavior pattern includes one or more predetermined behaviors; observing one or more responses from the user, wherein the one or more responses are responsive to the one or more predetermined behaviors; modifying at least one of the predetermined behaviors, wherein the modifying of the results in at least one modified behavior; observing at least one change in one or more of the responses, wherein the at least one change is responsive to the at least one modified behavior; and interacting with the user based on a second behavior pattern, wherein the second behavior pattern includes one or more of the predetermined behaviors and the modified behavior.
  • With further regard to the method, in some implementations, at least one of the predetermined behaviors includes communicating with the user by voice. In some implementations, the at least one modified behavior includes conveying affects to the user. In some implementations, to modify at least one of the predetermined behaviors, the method further comprises modifying one or more voice characteristics. In some implementations, the at least one change in one or more of the responses comprises a positive reaction or a negative reaction. In some implementations, the method further comprises: detecting a notification to be delivered to the user; detecting when a commercial is being shown on a television; and delivering the notification to the user during the commercial.
  • A further understanding of the nature and the advantages of particular implementations disclosed herein may be realized by reference of the remaining portions of the specification and the attached drawings.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • FIG. 1 illustrates a block diagram of an example television environment, which may be used for some implementations described herein.
  • FIG. 2 illustrates an example flow diagram for adapting interactions with a television user, according to some implementations.
  • FIG. 3 illustrates an example flow diagram for adapting interactions with a television user, according to some implementations.
  • FIG. 4 illustrates an example flow diagram for adapting interactions with a television user, according to some implementations.
  • FIG. 5 illustrates a block diagram of an example television environment, which may be used for some implementations described herein.
  • FIG. 6 illustrates a block diagram of an example television environment, which may be used for some implementations described herein.
  • FIG. 7 illustrates a block diagram of an example computing system 700, which may be used for some implementations described herein.
  • DETAILED DESCRIPTION
  • Implementations described herein adapt interactions with a television user. In various implementations, a system interacts with a user based on a first behavior pattern, which includes one or more predetermined behaviors. The system observes responses from the user, where the responses are responsive to the predetermined behaviors. As described in more detail herein, the system modifies one or more of the predetermined behaviors, which results in modified behaviors. The system observes if there are any changes in the user's responses, wherein the changes are responsive to the modified behaviors. The system continues interacting with the user based on a second behavior pattern, which includes one or more of the predetermined behaviors and the modified behavior.
  • FIG. 1 illustrates a block diagram of an example television environment 100, which may be used for some implementations described herein. In some implementations, television environment 100 includes a television 102, which interacts with a user 104. Television 102 interacts with user 104 based on behavior patterns. For example, television 102 may interact with user 104 based on a first behavior pattern (labeled Behavior Pattern 1), which may elicit a response from user 104. Television 102 may modify one or more behaviors, which in turn may elicit a response from user 104. In various implementations, television 102 may resume interacting with user 104 with a second behavior pattern (labeled Behavior Pattern 2), which includes at least some of the behaviors of the first behavior pattern and the modified behavior pattern. Example implementations are described in more detail herein.
  • In various implementations, television 102 includes a voice agent for interacting with the user. Television 102 also incudes a built-in microphone and a built-in camera. As such, when the user makes different kinds of requests, television 102 may observe the user's facial expressions and see how the user responds. In various implementations, television 102 may use machine learning techniques in order to determine if particular communication styles are successful or not.
  • For ease of illustration, FIG. 1 shows one block for television 102. Block 102 may represent multiple systems and/or devices. In other implementations, television 102 may not have all of the components shown and/or may have other elements including other types of elements instead of, or in addition to, those shown herein. Television 102 may perform the implementations described herein alone or in combination with other devices. In various implementations, the system may stored unique user profiles for each user of multiple users.
  • FIG. 2 illustrates an example flow diagram for adapting interactions with a television user, according to some implementations. Referring to both FIGS. 1 and 2, a method is initiated at block 202, where the system such as television 102 interacts with a user based on a first behavior pattern, wherein the first behavior pattern includes one or more predetermined behaviors. In various implementations described herein the system interacts by sounds such as by voice or with music or other sounds. In various implementations, a predetermined behavior may include communicating with the user by voice. In some implementations, a predetermined behavior may include providing particular information to the user (e.g., movie recommendations, etc.).
  • The predetermined behaviors may vary depending on the implementation. The system may use varying voice tones (e.g., high, low, cheery or calm, etc.), speed, level of engagement (e.g., proactive or passive; long answers or short answers, etc.) and other factors to build engagement. For example, the system when communicating by voice may speak with higher volume or lower volume. The system may speak more quickly or more slowly. The system may use a smooth voice or a less smooth voice, etc. The system may use a male voice, female voice, etc.
  • At block 204, the system observes one or more responses from the user, wherein the one or more responses are responsive to the one or more predetermined behaviors. For example, the system may observe facial expressions. The system may determine using facial and/or pattern recognition if the user is happy or unhappy, or confused.
  • At block 206, the system modifies at least one of the predetermined behaviors, wherein the modifying results in at least one modified behavior. In various implementations, the system modifies at least one of the predetermined behaviors by modifying one or more voice characteristics. For example, the system may communicate with the user with more volume and more slowly. Alternatively, the system may communicate in a more curt, direct, to the point manner. Such variations may range from subtle to overt.
  • In various implementations, a modified behavior includes conveying affects to the user. For example, the system may play background music (e.g., mood music, etc.) or may make particular sounds (e.g., beeps, etc.). In another example, the system may convey affects including music, sound, lights, or pictures. Such content may be retrieved from a database of appropriate effects.
  • In some implementations, when the user queries the system, the system may provide selected music and/or sounds and lights conforming to the general mood of the query (e.g., happy, sad, etc.). With a receptive response from the user the agent confirms the diagnosis of constructive user responses and takes the next appropriate interaction with the user.
  • In some implementations, the system may convey visual cues along with audio cues. Such cues may include, for example, colors, photos, videos, or even an avatar of the system with the capacity to show emotion to the user. These may be informed by user television viewing preferences.
  • At block 208, the system observes at least one change in one or more of the responses, wherein the at least one change is responsive to the at least one modified behavior. The system assess the success of different modifications by observing resulting responses from the user. In various implementations, at least one change in one or more of the responses comprises a positive reaction or a negative reaction. For example, if the user shows a positive change such as smiling more, or becoming more interactive, the system would log the positive change in the user's responses. In contrast, if the user shows a negative changes such as smiling less, or becoming less interactive, the system would log the negative change in the user's responses.
  • In various implementations, the responses or feedback from the user may be implicit or explicit. For example, in some implementations, the system may look for facial expressions. In some implementations, the system may receive explicit feedback from user indicating that the user likes or does not like particular modified behavior from the system. The system may continuously alter its interaction mode as needed or to optimize interaction with the user.
  • In some implementations, the system may introduce other predetermined behavior. For an example, the system may proactively inform the user of upcoming sports events or a movie schedule. In response, the user might explicitly expresses positive or negative feedback (e.g., “Keep notifying me of sports events.” or “I'm not interested.” etc.). Or, the user may behave in a way that implies positive or negative reaction. For example, the user might actively use the information to schedule a TV recording, or user might sound unenthusiastic or annoyed when responding, etc. The system logs these responses, and, over time, forms an appropriate level of “proactiveness” that is appropriate for the particular user.
  • At block 210, the system interacts with the user based on a second behavior pattern, wherein the second behavior pattern includes one or more of the predetermined behaviors and the modified behavior. Over time, the personality that the system conveys evolves and adapts to each user.
  • Such emotional or empathetic interaction to a user's query conveys a deeper sense of understanding by the system and greatly enhances the user experience. Implementations may provide non-verbal communication when appropriate, which may result in a greater trust and bonding with the system.
  • Although the steps, operations, or computations may be presented in a specific order, the order may be changed in particular implementations. Other orderings of the steps are possible, depending on the particular implementation. In some particular implementations, multiple steps shown as sequential in this specification may be performed at the same time. Also, some implementations may not have all of the steps shown and/or may have other steps instead of, or in addition to, those shown herein.
  • FIG. 3 illustrates an example flow diagram for adapting interactions with a television user, according to some implementations. Referring to both FIGS. 1 and 3, a method is initiated at block 302, where the system detects a notification to be delivered to the user.
  • At block 304, the system detects when a commercial is being shown on a television. The system may detect commercials based on volume changes, black frames, set times, etc.
  • At block 306, the system delivers the notification to the user during the commercial. As a result, the system avoids unwanted interruptions while the user is enjoying television content. This allows the user to enjoy television with minimal interruptions (e.g., pop-up notifications) while the user is watching their primary content. This enhances the viewing experience.
  • In some implementations, the system may prioritize notifications. The system may deliver some predetermined notifications during the primary content. For example, the system may deliver a notification from a smart doorbell (e.g., pizza delivery, etc.) or from a particular person (e.g., family member, etc.).
  • Although the steps, operations, or computations may be presented in a specific order, the order may be changed in particular implementations. Other orderings of the steps are possible, depending on the particular implementation. In some particular implementations, multiple steps shown as sequential in this specification may be performed at the same time. Also, some implementations may not have all of the steps shown and/or may have other steps instead of, or in addition to, those shown herein.
  • FIG. 4 illustrates an example flow diagram for adapting interactions with a television user, according to some implementations. As described in more detail herein, implementations optimize interaction for the specific user. Referring to both FIGS. 1 and 4, a method is initiated at block 402, where the system determines the identity of the user. The system may use a built-in camera and facial recognition, and a built-in mic and voice recognition to detect who is using the television. The system may also determine a usage signature of the user (e.g., method of operation such as remote control, voice, phone control, etc.).
  • At block 404, the system determines a predetermined user interface that is associated with the user. The system switches to a profile of the particular user and sub-profile (e.g., current control method), which carries some associated information about which user interface (UI) patterns and settings are preferred for that user.
  • The system may access a library of different UI patterns to be used for various situations. Each pattern may be given a priority. In some implementations the system collects all the available factors (e.g., person, ability level, control method, etc.) and weighs those factors. The system then selects the best UI pattern for the situation. The user may also manually enter settings and select some priority levels or patterns that the user prefers.
  • In some implementations, the system may store and access a user's viewing preferences in order to help the system determine or suggest particular UI settings. For example, a user who has a preference for watching sporting might want notifications of upcoming sporting events. A user who has a preference for music programs might want background music to be played with some user interfaces.
  • At block 406, the system presents to the user the predetermined user interface that is associated with the user. A system seamlessly and automatically switches between different UI patterns and settings based on who is operating the television. For example, the system adapts to users with specific preferences (e.g., easy controls for younger users, etc.) and/or impairments (e.g., vision, hearing, mobility, etc.).
  • Implementations allow users to operate the television in the best way for them without actively hunting through settings for particular controls. Implementations also encourage smoother operation of voice by displaying voice UI on the first attempt to use it. In multi-person households, the system automatically switches between optimal methods, giving each user a highly personalized experience.
  • In some implementations, the system modifies the controls of to the access capability/limitations of the user by identifying the user. The system accesses the user's access/capability/limitation profile, and identifies the appropriate UI for the user. For example, a user with blindness may need the television to have a screen reader function) or have voice-over enabled. The system may automatically turn such features on for that user, and may turn those features off for other users. In another example, the system may speak in a particular language associated with the user.
  • Although the steps, operations, or computations may be presented in a specific order, the order may be changed in particular implementations. Other orderings of the steps are possible, depending on the particular implementation. In some particular implementations, multiple steps shown as sequential in this specification may be performed at the same time. Also, some implementations may not have all of the steps shown and/or may have other steps instead of, or in addition to, those shown herein.
  • FIG. 5 illustrates a block diagram of an example television environment 500, which may be used for some implementations described herein. Shown is a television 502 and a user 504 interacting with television 502. In this example scenario, when the system recognizes user 504 with a camera 506, the system displays a UI 508 that is appropriate for user 504. UI 508 is a menu-style UI pattern that is preferred by user 504.
  • In some implementations, the system accesses a library of different UIs, which may be stored by the system or accessed in the cloud. For example, the library may include traditional menu UIs, graphical UIs, etc.
  • FIG. 6 illustrates a block diagram of an example television environment 600, which may be used for some implementations described herein. Shown is the same television 502 as in FIG. 5, and a different user 604 interacting with television 502. In this example scenario, when the system recognizes user 604 with a camera 506, the system displays a UI 608 that is appropriate for user 504. UI 508 is a graphical-style UI pattern that is preferred by user 604.
  • FIG. 7 illustrates a block diagram of an example computing system 700, which may be used for some implementations described herein. For example, computing system 700 may be used to implement television 102 of FIG. 1, as well as to perform implementations described herein. In some implementations, computing system 700 may include a processor 702, an operating system 704, a memory 706, and an input/output (I/O) interface 708. In various implementations, processor 702 may be used to implement various functions and features described herein, as well as to perform the method implementations described herein. While processor 702 is described as performing implementations described herein, any suitable component or combination of components of computing system 700 or any suitable processor or processors associated with computing system 700 or any suitable system may perform the steps described. Implementations described herein may be carried out on a user device, on a server, or a combination of both.
  • Computing system 700 also includes a software application 710, which may be stored on memory 706 or on any other suitable storage location or computer-readable medium. Software application 710 provides instructions that enable processor 702 to perform the implementations described herein and other functions. Software application may also include an engine such as a network engine for performing various functions associated with one or more networks and network communications. The components of computing system 700 may be implemented by one or more processors or any combination of hardware devices, as well as any combination of hardware, software, firmware, etc.
  • For ease of illustration, FIG. 7 shows one block for each of processor 702, operating system 704, memory 706, I/O interface 708, and software application 710. These blocks 702, 704, 706, 708, and 710 may represent multiple processors, operating systems, memories, I/O interfaces, and software applications. In various implementations, computing system 700 may not have all of the components shown and/or may have other elements including other types of components instead of, or in addition to, those shown herein.
  • Although the description has been described with respect to particular embodiments thereof, these particular embodiments are merely illustrative, and not restrictive. Concepts illustrated in the examples may be applied to other examples and implementations.
  • In various implementations, software is encoded in one or more non-transitory computer-readable media for execution by one or more processors. The software when executed by one or more processors is operable to perform the implementations described herein and other functions.
  • Any suitable programming language can be used to implement the routines of particular embodiments including C, C++, Java, assembly language, etc. Different programming techniques can be employed such as procedural or object oriented. The routines can execute on a single processing device or multiple processors. Although the steps, operations, or computations may be presented in a specific order, this order may be changed in different particular embodiments. In some particular embodiments, multiple steps shown as sequential in this specification can be performed at the same time.
  • Particular embodiments may be implemented in a non-transitory computer-readable storage medium (also referred to as a machine-readable storage medium) for use by or in connection with the instruction execution system, apparatus, or device. Particular embodiments can be implemented in the form of control logic in software or hardware or a combination of both. The control logic when executed by one or more processors is operable to perform the implementations described herein and other functions. For example, a tangible medium such as a hardware storage device can be used to store the control logic, which can include executable instructions.
  • Particular embodiments may be implemented by using a programmable general purpose digital computer, and/or by using application specific integrated circuits, programmable logic devices, field programmable gate arrays, optical, chemical, biological, quantum or nanoengineered systems, components and mechanisms. In general, the functions of particular embodiments can be achieved by any means as is known in the art. Distributed, networked systems, components, and/or circuits can be used. Communication, or transfer, of data may be wired, wireless, or by any other means.
  • A “processor” may include any suitable hardware and/or software system, mechanism, or component that processes data, signals or other information. A processor may include a system with a general-purpose central processing unit, multiple processing units, dedicated circuitry for achieving functionality, or other systems. Processing need not be limited to a geographic location, or have temporal limitations. For example, a processor may perform its functions in “real-time,” “offline,” in a “batch mode,” etc. Portions of processing may be performed at different times and at different locations, by different (or the same) processing systems. A computer may be any processor in communication with a memory. The memory may be any suitable data storage, memory and/or non-transitory computer-readable storage medium, including electronic storage devices such as random-access memory (RAM), read-only memory (ROM), magnetic storage device (hard disk drive or the like), flash, optical storage device (CD, DVD or the like), magnetic or optical disk, or other tangible media suitable for storing instructions (e.g., program or software instructions) for execution by the processor. For example, a tangible medium such as a hardware storage device can be used to store the control logic, which can include executable instructions. The instructions can also be contained in, and provided as, an electronic signal, for example in the form of software as a service (SaaS) delivered from a server (e.g., a distributed system and/or a cloud computing system).
  • It will also be appreciated that one or more of the elements depicted in the drawings/figures can also be implemented in a more separated or integrated manner, or even removed or rendered as inoperable in certain cases, as is useful in accordance with a particular application. It is also within the spirit and scope to implement a program or code that can be stored in a machine-readable medium to permit a computer to perform any of the methods described above.
  • As used in the description herein and throughout the claims that follow, “a”, “an”, and “the” includes plural references unless the context clearly dictates otherwise. Also, as used in the description herein and throughout the claims that follow, the meaning of “in” includes “in” and “on” unless the context clearly dictates otherwise.
  • Thus, while particular embodiments have been described herein, latitudes of modification, various changes, and substitutions are intended in the foregoing disclosures, and it will be appreciated that in some instances some features of particular embodiments will be employed without a corresponding use of other features without departing from the scope and spirit as set forth. Therefore, many modifications may be made to adapt a particular situation or material to the essential scope and spirit.

Claims (21)

1.-20. (canceled)
21. A device comprising:
circuitry configured to
determine a user profile;
select a user interface (UI) pattern of one or more UI patterns based on a situation, wherein the situation includes the user profile;
perform an operation based on a first behavior pattern, wherein the first behavior pattern includes one or more predetermined behaviors;
receive one or more responses, wherein the one or more responses are responsive to the one or more predetermined behaviors;
modify at least one of the predetermined behaviors, wherein modifying the at least one predetermined behavior results in a modified behavior; and
perform an operation based on a second behavior pattern, wherein the second behavior pattern includes the modified behavior.
22. The device of claim 21, wherein the modified behavior includes conveying affects to a user.
23. The device of claim 21, wherein the at least one change in one or more of the responses comprises a positive reaction or a negative reaction.
24. The device of claim 21, wherein the circuitry is configured to perform the operation including:
detecting a notification to be presented; and
presenting the notification.
25. The device of claim 24, wherein the circuitry is configured to perform the operation including:
detecting when a commercial is being shown; and
presenting the notification during the commercial.
26. The device of claim 24, wherein the circuitry is further configured to perform the operation including presenting the notification based on a priority of the notification.
27. The device of claim 24, wherein the circuitry is further configured to perform the operation including detecting the notification to be presented based on a user preference.
28. The device of claim 21, wherein the circuitry is configured to determine the user profile using voice recognition.
29. The device of claim 21, wherein the circuitry is configured to determine the user profile using facial recognition.
30. The device of claim 21, wherein the circuitry is configured to access a library of different UI patterns.
31. The device of claim 21, wherein the one or more predetermined behaviors include communicating by voice.
32. The device of claim 31, wherein the circuitry is configured to modify one or more voice characteristics of the voice, and the one or more voice characteristics are modified based on one or more of tone, speed, and level of engagement, volume, and gender.
33. The device of claim 21, further comprising a display.
34. The device of claim 21, further comprising a microphone.
35. The device of claim 21, further comprising a camera.
36. A non-transitory computer-readable storage medium with program instructions stored thereon, the program instructions when executed by one or more processors are operable to perform operations comprising:
determining a user profile:
selecting a user interface (UI) pattern of the one or more UI patterns based on a situation, wherein the situation includes the user profile;
performing an operation based on a first behavior pattern, wherein the first behavior pattern includes one or more predetermined behaviors;
receiving one or more responses, wherein the one or more responses are responsive to the one or more predetermined behaviors;
modifying at least one of the predetermined behaviors, wherein modifying the at least one predetermined behavior results in a modified behavior; and
performing an operation based on a second behavior pattern, wherein the second behavior pattern includes the modified behavior.
37. A method comprising:
determining a user profile;
selecting a user interface (UI) pattern of the one or more UI patterns based on a situation, wherein the situation includes the user profile;
performing an operation based on a first behavior pattern, wherein the first behavior pattern includes one ore more predetermined behaviors;
receiving one or more responses, wherein the one or more responses are responsive to the one or more predetermined behaviors;
modifying at least one of the predetermined behaviors, wherein modifying the at least one predetermined behavior results in a modified behavior; and
performing an operation based on a second behavior pattern, wherein the second behavior pattern includes the modified behavior.
38. The method of claim 37, wherein the modified behavior includes conveying affects to a user.
39. The method of claim 37, wherein the at least one change in one or more of the responses comprises a positive reaction or a negative reaction.
40. The method of claim 37, wherein the performing of the operation including:
detecting a notification to be presented; and
presenting the notification.
US17/480,499 2018-08-06 2021-09-21 Adapting interactions with a television user Pending US20220046319A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US17/480,499 US20220046319A1 (en) 2018-08-06 2021-09-21 Adapting interactions with a television user

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US16/055,795 US11134308B2 (en) 2018-08-06 2018-08-06 Adapting interactions with a television user
US17/480,499 US20220046319A1 (en) 2018-08-06 2021-09-21 Adapting interactions with a television user

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US16/055,795 Continuation US11134308B2 (en) 2018-08-06 2018-08-06 Adapting interactions with a television user

Publications (1)

Publication Number Publication Date
US20220046319A1 true US20220046319A1 (en) 2022-02-10

Family

ID=68104677

Family Applications (2)

Application Number Title Priority Date Filing Date
US16/055,795 Active US11134308B2 (en) 2018-08-06 2018-08-06 Adapting interactions with a television user
US17/480,499 Pending US20220046319A1 (en) 2018-08-06 2021-09-21 Adapting interactions with a television user

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US16/055,795 Active US11134308B2 (en) 2018-08-06 2018-08-06 Adapting interactions with a television user

Country Status (3)

Country Link
US (2) US11134308B2 (en)
JP (1) JP7300093B2 (en)
WO (1) WO2020031044A1 (en)

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20080276186A1 (en) * 2007-03-31 2008-11-06 Sony Deutschland Gmbh Method and system for adapting a user interface of a device
US20090138805A1 (en) * 2007-11-21 2009-05-28 Gesturetek, Inc. Media preferences
US20100058255A1 (en) * 2008-08-26 2010-03-04 At&T Intellectual Property I, L.P. Method and apparatus for creating webpages in a media communication system
US20170206095A1 (en) * 2016-01-14 2017-07-20 Samsung Electronics Co., Ltd. Virtual agent

Family Cites Families (30)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5115501A (en) 1988-11-04 1992-05-19 International Business Machines Corporation Procedure for automatically customizing the user interface of application programs
US6359661B1 (en) 1996-11-05 2002-03-19 Gateway, Inc. Multiple user profile remote control
EP1176483A2 (en) 2000-07-27 2002-01-30 Yamaha Hatsudoki Kabushiki Kaisha Automatic electronic device and method for controlling such an electronic device
US20030093280A1 (en) 2001-07-13 2003-05-15 Pierre-Yves Oudeyer Method and apparatus for synthesising an emotion conveyed on a sound
WO2003058518A2 (en) 2002-01-07 2003-07-17 Stephen James Crampton Method and apparatus for an avatar user interface system
US20040034874A1 (en) 2002-08-19 2004-02-19 Hord Phillip M. Pop-up PVR advertising
US7774815B1 (en) 2002-09-30 2010-08-10 Arris Group, Inc. Context-sensitive interactive television ticker
US20070271518A1 (en) * 2006-05-16 2007-11-22 Bellsouth Intellectual Property Corporation Methods, Apparatus and Computer Program Products for Audience-Adaptive Control of Content Presentation Based on Sensed Audience Attentiveness
US20070300185A1 (en) 2006-06-27 2007-12-27 Microsoft Corporation Activity-centric adaptive user interface
US20080130958A1 (en) 2006-11-30 2008-06-05 Motorola, Inc. Method and system for vision-based parameter adjustment
US8813121B2 (en) * 2008-12-02 2014-08-19 At&T Intellectual Property I, L.P. Delaying emergency alert system messages
US20110032423A1 (en) * 2009-08-06 2011-02-10 Sony Corporation Adaptive user profiling for tv-centric home automation system
US20140168277A1 (en) * 2011-05-10 2014-06-19 Cisco Technology Inc. Adaptive Presentation of Content
US9256396B2 (en) 2011-10-10 2016-02-09 Microsoft Technology Licensing, Llc Speech recognition for context switching
KR20140052155A (en) 2012-10-19 2014-05-07 삼성전자주식회사 Display apparatus, method for controlling the display apparatus and processor for controlling the display apparatus
JP2014123872A (en) 2012-12-21 2014-07-03 Mitsubishi Electric Corp Display device and display method
US9675895B2 (en) 2013-03-13 2017-06-13 Hasbro, Inc. Three way multidirectional interactive toy
WO2015030814A1 (en) 2013-08-30 2015-03-05 Intel Corporation Mechanism for facilitating dynamic adjustments to computing device characteristics in response to changes in user viewing patterns
US9595146B2 (en) 2013-11-18 2017-03-14 Microsoft Technology Licensing, Llc Persistent user identification
JPWO2016068262A1 (en) 2014-10-29 2017-08-10 京セラ株式会社 Communication robot
US10764424B2 (en) * 2014-12-05 2020-09-01 Microsoft Technology Licensing, Llc Intelligent digital assistant alarm system for application collaboration with notification presentation
US20170220570A1 (en) * 2016-01-28 2017-08-03 Echostar Technologies L.L.C. Adjusting media content based on collected viewer data
US10091355B2 (en) 2016-02-19 2018-10-02 International Business Machines Corporation Virtual voice response agent individually configured for a user
US20170289766A1 (en) 2016-03-29 2017-10-05 Microsoft Technology Licensing, Llc Digital Assistant Experience based on Presence Detection
SG11201809397TA (en) 2016-04-26 2018-11-29 Taechyon Robotics Corp Multiple interactive personalities robot
US20180068012A1 (en) * 2016-09-07 2018-03-08 International Business Machines Corporation Chat flow tree structure adjustment based on sentiment and flow history
US10546586B2 (en) * 2016-09-07 2020-01-28 International Business Machines Corporation Conversation path rerouting in a dialog system based on user sentiment
US10361879B2 (en) 2016-11-12 2019-07-23 Fujitsu Limited Persona-based service delivery
US10586079B2 (en) * 2016-12-23 2020-03-10 Soundhound, Inc. Parametric adaptation of voice synthesis
WO2019195799A1 (en) * 2018-04-05 2019-10-10 Synaptics Incorporated Context-aware control for smart devices

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20080276186A1 (en) * 2007-03-31 2008-11-06 Sony Deutschland Gmbh Method and system for adapting a user interface of a device
US20090138805A1 (en) * 2007-11-21 2009-05-28 Gesturetek, Inc. Media preferences
US20100058255A1 (en) * 2008-08-26 2010-03-04 At&T Intellectual Property I, L.P. Method and apparatus for creating webpages in a media communication system
US20170206095A1 (en) * 2016-01-14 2017-07-20 Samsung Electronics Co., Ltd. Virtual agent

Also Published As

Publication number Publication date
WO2020031044A1 (en) 2020-02-13
JP2021533649A (en) 2021-12-02
US20200045370A1 (en) 2020-02-06
US11134308B2 (en) 2021-09-28
JP7300093B2 (en) 2023-06-29

Similar Documents

Publication Publication Date Title
US11086493B2 (en) Identifying and controlling smart devices
CN111970533B (en) Interaction method and device for live broadcast room and electronic equipment
US10104341B2 (en) Animated user identifiers
US10721527B2 (en) Device setting adjustment based on content recognition
KR101977915B1 (en) Methods, systems, and media for presenting recommended media content items
US10586131B2 (en) Multimedia conferencing system for determining participant engagement
CN106941624B (en) Processing method and device for network video trial viewing
WO2021169432A1 (en) Data processing method and apparatus of live broadcast application, electronic device and storage medium
CN111246225B (en) Information interaction method and device, electronic equipment and computer readable storage medium
US20240080520A1 (en) Systems and methods for providing media recommendations
CN115136570A (en) Integration of internet of things devices
US20230262291A1 (en) Systems and methods for providing media recommendations
CN111556352A (en) Multimedia resource sharing method and device, electronic equipment and storage medium
US20230120441A1 (en) Systems and methods for sequenced, multimodal communication
US20220046319A1 (en) Adapting interactions with a television user
CN112685599A (en) Video recommendation method and device
US11375283B2 (en) Configuring settings of a television
CN114554231A (en) Information display method and device, electronic equipment and storage medium
CN113573121A (en) Video playing method, device and storage medium
US20200112767A1 (en) Artificial intelligence optimized playback experience
CN114375567A (en) Processing interactive requests with user-specific data on a shared device
WO2020054361A1 (en) Information processing system, information processing method, and recording medium
CN113050805A (en) Intelligent interaction method and device based on input tool
WO2022103414A1 (en) Systems and methods for providing media recommendations
KR20220023453A (en) Display apparatus and control method thereof

Legal Events

Date Code Title Description
AS Assignment

Owner name: SONY GROUP CORPORATION, JAPAN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:SONY CORPORATION;REEL/FRAME:057871/0646

Effective date: 20211021

AS Assignment

Owner name: SATURN LICENSING LLC, NEW YORK

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:SONY GROUP CORPORATION;REEL/FRAME:057939/0387

Effective date: 20211027

STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: FINAL REJECTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: FINAL REJECTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED