US20190012710A1 - Sensors and sentiment analysis for rating systems - Google Patents
Sensors and sentiment analysis for rating systems Download PDFInfo
- Publication number
- US20190012710A1 US20190012710A1 US15/810,328 US201715810328A US2019012710A1 US 20190012710 A1 US20190012710 A1 US 20190012710A1 US 201715810328 A US201715810328 A US 201715810328A US 2019012710 A1 US2019012710 A1 US 2019012710A1
- Authority
- US
- United States
- Prior art keywords
- sentiment
- entertainment
- processor
- users
- user
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
- 238000004458 analytical method Methods 0.000 title description 36
- 238000000034 method Methods 0.000 claims abstract description 44
- 230000006397 emotional response Effects 0.000 claims abstract description 7
- 230000006870 function Effects 0.000 claims description 11
- 230000008569 process Effects 0.000 description 24
- 238000010586 diagram Methods 0.000 description 10
- 238000004891 communication Methods 0.000 description 8
- 238000004590 computer program Methods 0.000 description 7
- 238000005516 engineering process Methods 0.000 description 6
- 230000008901 benefit Effects 0.000 description 4
- 230000005540 biological transmission Effects 0.000 description 4
- 235000002017 Zea mays subsp mays Nutrition 0.000 description 2
- 241000482268 Zea mays subsp. mays Species 0.000 description 2
- 238000003491 array Methods 0.000 description 2
- 230000008859 change Effects 0.000 description 2
- 230000003287 optical effect Effects 0.000 description 2
- 230000002093 peripheral effect Effects 0.000 description 2
- 230000001902 propagating effect Effects 0.000 description 2
- RYGMFSIKBFXOCR-UHFFFAOYSA-N Copper Chemical compound [Cu] RYGMFSIKBFXOCR-UHFFFAOYSA-N 0.000 description 1
- 230000036772 blood pressure Effects 0.000 description 1
- 230000001413 cellular effect Effects 0.000 description 1
- 229910052802 copper Inorganic materials 0.000 description 1
- 239000010949 copper Substances 0.000 description 1
- 230000001419 dependent effect Effects 0.000 description 1
- 238000001514 detection method Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 239000000835 fiber Substances 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
- 230000010354 integration Effects 0.000 description 1
- 238000004519 manufacturing process Methods 0.000 description 1
- 230000007246 mechanism Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000003058 natural language processing Methods 0.000 description 1
- 230000029058 respiratory gaseous exchange Effects 0.000 description 1
- 230000004044 response Effects 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 230000003068 static effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06Q—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
- G06Q30/00—Commerce
- G06Q30/02—Marketing; Price estimation or determination; Fundraising
- G06Q30/0282—Rating or review of business operators or products
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/0002—Remote monitoring of patients using telemetry, e.g. transmission of vital signals via a communication network
- A61B5/0015—Remote monitoring of patients using telemetry, e.g. transmission of vital signals via a communication network characterised by features of the telemetry system
- A61B5/0022—Monitoring a patient using a global network, e.g. telephone networks, internet
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/16—Devices for psychotechnics; Testing reaction times ; Devices for evaluating the psychological state
- A61B5/165—Evaluating the state of mind, e.g. depression, anxiety
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/72—Signal processing specially adapted for physiological signals or for diagnostic purposes
- A61B5/7271—Specific aspects of physiological measurement analysis
- A61B5/7278—Artificial waveform generation or derivation, e.g. synthesizing signals from measured signals
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06Q—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
- G06Q30/00—Commerce
- G06Q30/02—Marketing; Price estimation or determination; Fundraising
- G06Q30/0201—Market modelling; Market analysis; Collecting market data
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/02—Detecting, measuring or recording for evaluating the cardiovascular system, e.g. pulse, heart rate, blood pressure or blood flow
- A61B5/021—Measuring pressure in heart or blood vessels
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/02—Detecting, measuring or recording for evaluating the cardiovascular system, e.g. pulse, heart rate, blood pressure or blood flow
- A61B5/024—Measuring pulse rate or heart rate
- A61B5/02438—Measuring pulse rate or heart rate with portable devices, e.g. worn by the patient
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/05—Detecting, measuring or recording for diagnosis by means of electric currents or magnetic fields; Measuring using microwaves or radio waves
- A61B5/053—Measuring electrical impedance or conductance of a portion of the body
- A61B5/0531—Measuring skin impedance
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/08—Measuring devices for evaluating the respiratory organs
- A61B5/0816—Measuring devices for examining respiratory frequency
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06Q—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
- G06Q30/00—Commerce
- G06Q30/02—Marketing; Price estimation or determination; Fundraising
- G06Q30/0201—Market modelling; Market analysis; Collecting market data
- G06Q30/0203—Market surveys; Market polls
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06Q—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
- G06Q30/00—Commerce
- G06Q30/02—Marketing; Price estimation or determination; Fundraising
- G06Q30/0241—Advertisements
- G06Q30/0251—Targeted advertisements
- G06Q30/0261—Targeted advertisements based on user location
Definitions
- the present invention relates to sensor and sentiment analysis for rating systems.
- conventional rating systems rate or output ratings for entertainment, such as films, movies, television shows, concerts, performances, shows, plays, and the like.
- the ratings are intended to quantify popular opinion of the entertainment so that individuals, businesses, and/or companies can utilize corresponding ratings to make decisions with respect to the entertainment (e.g., whether to attend a play, associate advertisements during movie previews, change the cost of admission to a concert, etc.).
- conventional rating systems do not capture an accurate depiction of how a user feels while experiencing the entertainment, the individuals, the businesses, and/or the companies may be making misinformed decisions.
- users that rate a movie are typically users who are on an extreme (e.g., either hate or love the movie). In other cases, users may rate the movie for a reason beyond the movie itself (e.g., the lead actor is a favorite, the popcorn was exceptional, etc.). In addition, users may delay in rating the movie, which can open their opinions to outside influences and changes. Ultimately, the rating for the movie is a skewed or wrong result with respect to a true sentiment rating.
- a processor-implemented method for analyzing sentiment data corresponding to users experiencing an entertainment to generate a sentiment rating includes collecting physical data procured by wearable devices corresponding to the users during the entertainment.
- the processor-implemented method includes analyzing the physical data to determine sentiment data and generating the sentiment rating for the entertainment based on the sentiment data.
- the sentiment rating quantifies a collective real-time emotional response for the users that experience the entertainment.
- Embodiments of the present invention can also include the above processor-implemented method implemented as a system and/or a computer program product.
- FIG. 1 depicts a system in accordance with one or more embodiments.
- FIG. 2 depicts a process flow of system in accordance with one or more embodiments
- FIG. 3 depicts a process flow of system in accordance with one or more embodiments.
- FIG. 4 depicts a processing system in accordance with one or more embodiments.
- embodiments disclosed herein may include system, method, and/or computer program product (herein a sensor and sentiment analysis system) that analyzes sentiments of users experiencing any entertainment to generate a sentiment rating.
- a sensor and sentiment analysis system that analyzes sentiments of users experiencing any entertainment to generate a sentiment rating.
- Technical effects and benefits of the sensor and sentiment analysis system include generating a sentiment rating that quantifies a collective real-time emotional response for the users that experience entertainment.
- embodiments described herein are necessarily rooted in sensor and sentiment analysis system to perform proactive operations to overcome problems specifically arising with conventional rating systems.
- the system 100 comprises one or more user devices 110 (e.g., 110 - 1 , 110 - 2 , . . . , 110 -n), at least one tethered device 112 , one or more sensor devices 114 , a local server 120 , a network 130 , and an external server 150 .
- the system 100 can be an electronic, computer framework comprising and/or employing any number and combination of computing devices and networks utilizing various communication technologies.
- the system 100 can be easily scalable, extensible, and modular, with the ability to change to different services or reconfigure some features independently of others.
- the one or more user devices 110 , the at least one tethered device 112 , the one or more sensor devices 114 , and the local server 120 are located within a venue 170 , which represents and/or comprises a local network.
- the local server 120 connects to the external server 150 , which is not located within the venue 170 via the network 130 .
- the local network and/or the external network 130 can utilize a plurality of communication technologies, such as radio technologies, cellular technologies, etc., to transmit communications between the local server 120 and the external server 150 .
- Examples of the network 130 include, but are not limited to, the Internet, a local area network, a wide area network and/or a wireless network.
- Each of the one or more user devices 110 , the at least one tethered device 112 , the one or more sensor devices 114 , the local server 120 , and the external server 150 can be a computing device with a processor and a memory configured to receive and respond to electrical communications. Further, the arrows of FIG. 1 represent the electrical communications between the elements for FIG. 1 . Note that the thick lines extending from the local server 120 represent one or more communication interfaces of the local network through which the one or more user devices 110 , the at least one tethered device 112 , and/or the one or more sensor devices 114 communicatively connect to the local server 120 ).
- the one or more user devices 110 , the at least one tethered device 112 , and the one or more sensor devices 114 can generally be referred to as mobile devices.
- a mobile device is a portable computing device comprising at least a processor and a memory. Examples of a mobile device include smartphones, tablet computers, personal digital assistants, wearable devices (e.g., watches), etc.
- Mobile devices can include one or more sensors for detecting physical data or can connect to other devices including the same.
- a sensor can comprise any electro-mechanical device for detecting conditions of an object or environment (sensor examples include biometric detectors, such as pulse sensors, moisture detectors, heat sensors, etc.).
- the system 100 utilizes can utilize one or more sensors (e.g., the one or more sensor devices 114 and/or the at least one tethered device 112 ) to capture physical data of one or more users at the venue 170 experiencing entertainment.
- the system 100 utilizes can also utilize mobile devices (e.g., the one or more user devices 110 ) to receive user inputs from the one or more users.
- the system 100 can capture the physical data and/or the user inputs during the entertainment in real-time or immediately after experiencing the entertainment.
- the physical data can comprise sensor readings that capture and identify a status of bodily functions (i.e., blood pressure, heart rate, breathing rates, and skin conductance) of the one or more users.
- the user inputs can comprise direct feedback for information submitted by the one or more users via corresponding mobile devices.
- ‘immediately after experiencing the entertainment’ can be defined as a time interval beginning at a conclusion of the entertainment and extending for a predetermined amount of time.
- the predetermined amount of time can be chosen from a range of one second to one day.
- the user device 110 - 1 and the one or more sensor devices 114 can be standalone mobile devices capable of detecting/receiving physical data, and the user device 110 - 2 can be coupled to the at least one tethered device 112 that also detects/receives physical data.
- a biometric scanner can be included in the user device 110 - 1 to determine a pulse of a user.
- the tethered device 112 can be a wearable device that includes a biometric scanner to determine a pulse of a user, which is then transmitted to the user device 110 - 2 .
- the user devices 110 - 1 and 110 - 2 can transmit the pulse, as physical data, to the local server 120 .
- the sensor device 114 case, the sensor is a wearable device that automatically transmits physical data (i.e., the pulse of a user) to the local server 120 .
- the system 100 analyzes the physical data and/or the user inputs to determine sentiment data for each of the one or more users.
- the sentiment data quantifies how a user feels (i.e., a sentimental or emotional response) with respect to the entertainment.
- the system 100 utilizes can utilize the sentiment data (for each of the one or more users who experience the entertainment) to compile a sentiment rating for the entertainment.
- the sentiment rating is a value that assesses and quantifies a user emotional response for a user population that experiences the entertainment.
- the local server 120 and/or the external server 150 can analyze the physical data and/or the user inputs to determine the sentiment data and utilize the sentiment data to compile a sentiment rating for the entertainment.
- the local server 120 and the external server 150 can also comprise a database, data repository or other data store and may include various kinds of mechanisms for storing, accessing, retrieving, and processing various kinds of data (e.g., the physical data, the user inputs, the sentiment data, and the sentiment rating), including a hierarchical database, a set of files in a file system, an application database in a proprietary format, a relational database management system (RDBMS), etc.
- RDBMS relational database management system
- FIG. 2 an operational embodiment of the system 100 of FIG. 1 is described with respect to the process flow 200 in accordance with one or more embodiments.
- the process flow 200 analyzes sentiments of users experiencing any entertainment to generate the sentiment rating.
- the process flow 200 begins at block 205 , where the system 100 receives an input indicating permission to collect sentiment data.
- the input indicating permission is representative of one or more inputs, each of which is provided by a user via their corresponding mobile device, identifying that a user has optioned-into (opted-in) participating in the collection of the sentiment data.
- the receipt of an input indicating permission by the system 100 can include receiving permission at the one or more mobile devices 110 , the local server 120 , and/or the external server 150 .
- the system 100 can incentivize the one or more users to provide both the physical data and the user inputs. For instance, if watching a movie, the system 100 can indicate to the user that they can receive a discounted movie ticket if the user grants permission to the system 100 to obtain physical data from sensors contained within their wearables and to provide a few words at an end of the movie discussing their thoughts of the movie via their mobile devices.
- the system 100 detects a location and time local to a mobile device.
- the location and the local time are detected for each mobile device that has provided the input indicated permission to collect sentiment data.
- the location can be determined utilizing location services of the mobile device and/or by the mobile device being on a local network of the venue 170 .
- the time local can be determined by the clock of the mobile device.
- the detection of the location and the local time by the system 100 can be performed by the one or more mobile devices 110 , the local server 120 , and/or the external server 150 .
- the system 100 determines an entertainment based on the location and the time.
- the system 100 can confirm that a user is at a specific location at a particular time with respect to the entertainment to verify that the user is actually experiencing the entertainment.
- the system 100 can compare the specific location and time to a schedule for the venue 117 already in the local server 120 .
- the system 100 can utilize the local network of the venue 170 to determine that the user device 110 - 1 is on the premises. In the case of a movie theater, the local network can determine which theater room the mobile device is in, and therefore can determine based on time which movie the user is watching.
- the system 100 collects physical data and user inputs during and/or immediately after the entertainment.
- the system 100 can utilize the sensor devices 114 , the tethered device 112 , and the user devices 110 to receive the physical data and the user inputs.
- the system 100 analyzes the physical data and the user inputs to determine sentiment data. That is, the physical data and the user inputs are processed to generate the sentiment data for each user who opted-in. For example, each pulse of a user is used to determine whether a heart rate has accelerated or remained the same while experiencing the entertainment. If during a horror movie, the heart rates of the users increase during the horror movie, the system 100 determines that the sentiment of the users is scared/excited.
- the analyzation of the physical data and the user inputs by the system 100 can be performed by the one or more mobile devices 110 , the local server 120 , and/or the external server 150 .
- the analyses of the physical data and the user inputs can include natural language processing, text analysis, computational linguistics, and biometrics to systematically identify, extract, quantify, and study affective states and subjective information to determine the sentiment data of the one or more users.
- the system 100 determines from the sentiment data the sentiment rating. For instance, using the sentiment data based on the collected physical data and user inputs, the system 100 determines the overall experience of the users.
- the sentiment rating is a value that assesses and quantifies a collective emotional response for users that experience the entertainment.
- the sentiment rating determination by the system 100 can be performed by the one or more mobile devices 110 , the local server 120 , and/or the external server 150 .
- FIG. 3 depicts a process flow 300 of a sensor and sentiment analysis system in accordance with one or more embodiments.
- the process flow 300 begins at block 305 , where a user purchases a ticket.
- a user arrives at an entertainment venue (e.g., a movie theater, a concert hall, etc.) and buys the ticket to experience an entertainment at the entertainment venue (e.g., the venue 170 ).
- an entertainment venue e.g., a movie theater, a concert hall, etc.
- the sensor and sentiment analysis system determines whether the user grants permission to access sensor information on a wearable device.
- the wearable device can be user provided or venue provided. For instance, upon buying the ticket at a kiosk, the user is proposed with an opportunity (via pop-up on the kiosk during the purchase) to allow the sensor and sentiment analysis system to access sensor information from the wearable device while the user experiences the entertainment. The user can also be proposed with an opportunity to provide thoughts after the entertainment. In return, the sensor and sentiment analysis system indicate that an incentive (i.e. 10% off next ticket, free popcorn, etc.) will be received by opting-in.
- an incentive i.e. 10% off next ticket, free popcorn, etc.
- the process flow 300 proceeds to decision block 315 via the YES arrow.
- the user is proposed type in a ticket code (provided via the kiosk) into an application to sync the mobile phone with the sensor and sentiment analysis system.
- the sensor and sentiment analysis system determines whether the user has an application on their mobile device (e.g., the application communicates with the sensor and sentiment analysis system in response to the ticket code being entered into the applications). If the sensor and sentiment analysis system determines that the user has the application on their mobile device, then the process flow 300 proceeds to block 320 via the YES arrow.
- the sensor and sentiment analysis system determines a location of the mobile device. For instance, the sensor and sentiment analysis system utilizes a global positioning system of the mobile device to detect a location.
- the sensor and sentiment analysis system collects sensor data (e.g., physical data). The sensor and sentiment analysis system can collect the physical data once the entertainment starts. For instance, as the user watches/listens to the entertainment, the wearable device collects physical data from sensors every few minutes. The wearable device can provide the physical data, once collected, to the sensor and sentiment analysis system.
- sensor data e.g., physical data
- the sensor and sentiment analysis system can collect the physical data once the entertainment starts. For instance, as the user watches/listens to the entertainment, the wearable device collects physical data from sensors every few minutes. The wearable device can provide the physical data, once collected, to the sensor and sentiment analysis system.
- the sensor and sentiment analysis system determines whether the user will input their thoughts about the entertainment. In accordance with one or more embodiments, when the entertainment is over, the user is prompted as to whether they want to take a few minutes to say how they felt about the entertainment. If the sensor and sentiment analysis system determines that the user agrees to input their thoughts about the entertainment, the process flow 300 proceeds to block 335 via the AGREE arrow.
- the permission received at block 310 enables the sensor and sentiment analysis system to prompt the user to enter their user input.
- the process flow 300 proceeds to block 335 .
- the user provides direct feedback. For instance, the user can speak into the mobile device, which utilizes speech to text technology to turn their thoughts into text.
- the sensor and sentiment analysis system stores the sensor data (and the user input if received). Note that if the sensor and sentiment analysis system determines that the user does not agree to input their thoughts about the entertainment at decision block 330 , the process flow 300 proceeds directly to block 340 via the DISAGREE arrow. The process flow 300 can then return to block 305 after block 340 and repeat upon the purchasing of another ticket.
- the process flow 300 proceeds to block 350 via the NO arrow.
- the sensor and sentiment analysis system requests via the kiosk for the user to download the application.
- the sensor and sentiment analysis system also can confirm that the user has granted permission to access the sensor information on the wearable device. If the sensor and sentiment analysis system determines that the user agrees to both conditions, the process flow 300 proceeds to block 320 via the AGREE arrow and the process flow 300 proceeds accordingly. If the sensor and sentiment analysis system determines that the user agrees to download the application but does not confirm their previous permission, the process flow 300 proceeds to block 330 via the DISAGREE arrow and the process flow 300 proceeds accordingly. In this case, if a user never agrees to provide physical data, the user can still provide direct feedback about the entertainment.
- the process flow 300 proceeds to decision block 360 via the NO arrow.
- the sensor and sentiment analysis system determines whether the user has an application on their mobile device. If the sensor and sentiment analysis system determines that the user has the application on their mobile device, then the process flow 300 proceeds to block 330 via the YES arrow and the process flow 300 proceeds accordingly. If the sensor and sentiment analysis system determines that the user does not have the application on their mobile device, then the process flow 300 proceeds to block 305 via the NO arrow and the process flow 300 proceeds accordingly.
- FIG. 4 depicts an example of a system 400 in accordance with one or more embodiments.
- the system 400 has one or more central processing units (CPU(s)) 401 a , 401 b , 401 c , etc. (collectively or generically referred to as processor(s) 401 ).
- the processors 401 also referred to as processing circuits, are coupled via a system bus 402 to system memory 403 and various other components.
- the system memory 403 can include a read only memory (ROM) 404 and a random access memory (RAM) 405 .
- the ROM 404 is coupled to the system bus 402 and may include a basic input/output system (BIOS), which controls certain basic functions of the system 400 .
- the RAM is read-write memory coupled to the system bus 402 for use by the processors 401 .
- BIOS basic input/output system
- FIG. 4 further depicts an input/output (I/O) adapter 406 and a communications adapter 407 coupled to the system bus 402 .
- the I/O adapter 406 may be a small computer system interface (SCSI) adapter that communicates with a hard disk 408 and/or any other similar component.
- the I/O adapter 406 and the hard disk 408 are collectively referred to herein as a mass storage 410 .
- a software 411 for execution on the system 400 may be stored in the mass storage 410 .
- the mass storage 410 is an example of a tangible storage medium readable by the processors 401 , where the software 411 is stored as instructions for execution by the processors 401 to cause the system 400 to operate, such as is described herein with reference to FIG. 2 .
- a communications adapter 407 interconnects the system bus 402 with a network 412 , which may be an outside network, enabling the system 400 to communicate with other such systems.
- a display (e.g., screen, a display monitor) 415 is connected to the system bus 402 by a display adapter 416 , which may include a graphics controller to improve the performance of graphics intensive applications and a video controller.
- the adapters 406 , 407 , and 416 may be connected to one or more I/O buses that are connected to the system bus 402 via an intermediate bus bridge (not shown).
- Suitable I/O buses for connecting peripheral devices typically include common protocols, such as the Peripheral Component Interconnect (PCI). Additional input/output devices are shown as connected to the system bus 402 via an interface adapter 420 and the display adapter 416 .
- a keyboard, a mouse, a speaker, etc. can be interconnected to the system bus 402 via the interface adapter 420 , which may include, for example, a Super I/O chip integrating multiple device adapters into a single integrated circuit.
- the system 400 includes processing capability in the form of the processors 401 , and, storage capability including the system memory 403 and the mass storage 410 , input means such as the keyboard and the mouse, and output capability including the speaker and the display 415 .
- a portion of the system memory 403 and the mass storage 410 collectively store an operating system, such as the z/OS or AIX operating system from IBM Corporation, to coordinate the functions of the various components shown in FIG. 4 .
- the present invention may be a system, a method, and/or a computer program product at any possible technical detail level of integration
- the computer program product may include a computer readable storage medium (or media) having computer readable program instructions thereon for causing a processor to carry out aspects of the present invention
- the computer readable storage medium can be a tangible device that can retain and store instructions for use by an instruction execution device.
- the computer readable storage medium may be, for example, but is not limited to, an electronic storage device, a magnetic storage device, an optical storage device, an electromagnetic storage device, a semiconductor storage device, or any suitable combination of the foregoing.
- a non-exhaustive list of more specific examples of the computer readable storage medium includes the following: a portable computer diskette, a hard disk, a random access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or Flash memory), a static random access memory (SRAM), a portable compact disc read-only memory (CD-ROM), a digital versatile disk (DVD), a memory stick, a floppy disk, a mechanically encoded device such as punch-cards or raised structures in a groove having instructions recorded thereon, and any suitable combination of the foregoing.
- RAM random access memory
- ROM read-only memory
- EPROM or Flash memory erasable programmable read-only memory
- SRAM static random access memory
- CD-ROM compact disc read-only memory
- DVD digital versatile disk
- memory stick a floppy disk
- a mechanically encoded device such as punch-cards or raised structures in a groove having instructions recorded thereon
- a computer readable storage medium is not to be construed as being transitory signals per se, such as radio waves or other freely propagating electromagnetic waves, electromagnetic waves propagating through a waveguide or other transmission media (e.g., light pulses passing through a fiber-optic cable), or electrical signals transmitted through a wire.
- Computer readable program instructions described herein can be downloaded to respective computing/processing devices from a computer readable storage medium or to an external computer or external storage device via a network, for example, the Internet, a local area network, a wide area network and/or a wireless network.
- the network may comprise copper transmission cables, optical transmission fibers, wireless transmission, routers, firewalls, switches, gateway computers and/or edge servers.
- a network adapter card or network interface in each computing/processing device receives computer readable program instructions from the network and forwards the computer readable program instructions for storage in a computer readable storage medium within the respective computing/processing device.
- Computer readable program instructions for carrying out operations of the present invention may be assembler instructions, instruction-set-architecture (ISA) instructions, machine instructions, machine dependent instructions, microcode, firmware instructions, state-setting data, configuration data for integrated circuitry, or either source code or object code written in any combination of one or more programming languages, including an object oriented programming language such as Smalltalk, C++, or the like, and procedural programming languages, such as the “C” programming language or similar programming languages.
- the computer readable program instructions may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server.
- the remote computer may be connected to the user's computer through any type of network, including a local area network (LAN) or a wide area network (WAN), or the connection may be made to an external computer (for example, through the Internet using an Internet Service Provider).
- electronic circuitry including, for example, programmable logic circuitry, field-programmable gate arrays (FPGA), or programmable logic arrays (PLA) may execute the computer readable program instructions by utilizing state information of the computer readable program instructions to personalize the electronic circuitry, in order to perform aspects of the present invention.
- These computer readable program instructions may be provided to a processor of a general purpose computer, special purpose computer, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks.
- These computer readable program instructions may also be stored in a computer readable storage medium that can direct a computer, a programmable data processing apparatus, and/or other devices to function in a particular manner, such that the computer readable storage medium having instructions stored therein comprises an article of manufacture including instructions which implement aspects of the function/act specified in the flowchart and/or block diagram block or blocks.
- the computer readable program instructions may also be loaded onto a computer, other programmable data processing apparatus, or other device to cause a series of operational steps to be performed on the computer, other programmable apparatus or other device to produce a computer implemented process, such that the instructions which execute on the computer, other programmable apparatus, or other device implement the functions/acts specified in the flowchart and/or block diagram block or blocks.
- each block in the flowchart or block diagrams may represent a module, segment, or portion of instructions, which comprises one or more executable instructions for implementing the specified logical function(s).
- the functions noted in the blocks may occur out of the order noted in the Figures.
- two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved.
Landscapes
- Health & Medical Sciences (AREA)
- Life Sciences & Earth Sciences (AREA)
- Engineering & Computer Science (AREA)
- Business, Economics & Management (AREA)
- Physics & Mathematics (AREA)
- Strategic Management (AREA)
- Finance (AREA)
- Development Economics (AREA)
- Accounting & Taxation (AREA)
- Pathology (AREA)
- Biomedical Technology (AREA)
- Heart & Thoracic Surgery (AREA)
- Medical Informatics (AREA)
- Molecular Biology (AREA)
- Surgery (AREA)
- Animal Behavior & Ethology (AREA)
- General Health & Medical Sciences (AREA)
- Public Health (AREA)
- Veterinary Medicine (AREA)
- Biophysics (AREA)
- Psychiatry (AREA)
- Entrepreneurship & Innovation (AREA)
- Cardiology (AREA)
- Physiology (AREA)
- General Physics & Mathematics (AREA)
- Game Theory and Decision Science (AREA)
- Theoretical Computer Science (AREA)
- Economics (AREA)
- Marketing (AREA)
- General Business, Economics & Management (AREA)
- Educational Technology (AREA)
- Child & Adolescent Psychology (AREA)
- Developmental Disabilities (AREA)
- Hospice & Palliative Care (AREA)
- Psychology (AREA)
- Social Psychology (AREA)
- Computer Networks & Wireless Communication (AREA)
- Artificial Intelligence (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Signal Processing (AREA)
Abstract
Description
- This application is a continuation of U.S. application Ser. No. 15/641,763 which was filed on Jul. 5, 2017. The entire disclosures of U.S. application Ser. No. 15/641,763 are incorporated herein by reference.
- The present invention relates to sensor and sentiment analysis for rating systems.
- In general, conventional rating systems rate or output ratings for entertainment, such as films, movies, television shows, concerts, performances, shows, plays, and the like. The ratings are intended to quantify popular opinion of the entertainment so that individuals, businesses, and/or companies can utilize corresponding ratings to make decisions with respect to the entertainment (e.g., whether to attend a play, associate advertisements during movie previews, change the cost of admission to a concert, etc.). However, because conventional rating systems do not capture an accurate depiction of how a user feels while experiencing the entertainment, the individuals, the businesses, and/or the companies may be making misinformed decisions.
- For instance, users that rate a movie are typically users who are on an extreme (e.g., either hate or love the movie). In other cases, users may rate the movie for a reason beyond the movie itself (e.g., the lead actor is a favorite, the popcorn was exceptional, etc.). In addition, users may delay in rating the movie, which can open their opinions to outside influences and changes. Ultimately, the rating for the movie is a skewed or wrong result with respect to a true sentiment rating.
- In accordance with one or more embodiments, a processor-implemented method for analyzing sentiment data corresponding to users experiencing an entertainment to generate a sentiment rating is provided. The processor-implemented method includes collecting physical data procured by wearable devices corresponding to the users during the entertainment. The processor-implemented method includes analyzing the physical data to determine sentiment data and generating the sentiment rating for the entertainment based on the sentiment data. The sentiment rating quantifies a collective real-time emotional response for the users that experience the entertainment.
- Embodiments of the present invention can also include the above processor-implemented method implemented as a system and/or a computer program product.
- Additional features and advantages are realized through the techniques of the present disclosure. Other embodiments and aspects of the disclosure are described in detail herein. For a better understanding of the disclosure with the advantages and the features, refer to the description and to the drawings.
-
FIG. 1 depicts a system in accordance with one or more embodiments; and -
FIG. 2 depicts a process flow of system in accordance with one or more embodiments; -
FIG. 3 depicts a process flow of system in accordance with one or more embodiments; and -
FIG. 4 depicts a processing system in accordance with one or more embodiments. - In view of the above, embodiments disclosed herein may include system, method, and/or computer program product (herein a sensor and sentiment analysis system) that analyzes sentiments of users experiencing any entertainment to generate a sentiment rating. Technical effects and benefits of the sensor and sentiment analysis system include generating a sentiment rating that quantifies a collective real-time emotional response for the users that experience entertainment. Thus, embodiments described herein are necessarily rooted in sensor and sentiment analysis system to perform proactive operations to overcome problems specifically arising with conventional rating systems.
- Turning now to
FIG. 1 , asystem 100 is shown as an example of the sensor and sentiment analysis system in accordance with an embodiment. Thesystem 100 comprises one or more user devices 110 (e.g., 110-1, 110-2, . . . , 110-n), at least onetethered device 112, one ormore sensor devices 114, alocal server 120, anetwork 130, and anexternal server 150. - The
system 100 can be an electronic, computer framework comprising and/or employing any number and combination of computing devices and networks utilizing various communication technologies. Thesystem 100 can be easily scalable, extensible, and modular, with the ability to change to different services or reconfigure some features independently of others. - The one or
more user devices 110, the at least onetethered device 112, the one ormore sensor devices 114, and thelocal server 120 are located within avenue 170, which represents and/or comprises a local network. Thelocal server 120 connects to theexternal server 150, which is not located within thevenue 170 via thenetwork 130. The local network and/or theexternal network 130 can utilize a plurality of communication technologies, such as radio technologies, cellular technologies, etc., to transmit communications between thelocal server 120 and theexternal server 150. Examples of thenetwork 130 include, but are not limited to, the Internet, a local area network, a wide area network and/or a wireless network. - Each of the one or
more user devices 110, the at least onetethered device 112, the one ormore sensor devices 114, thelocal server 120, and theexternal server 150 can be a computing device with a processor and a memory configured to receive and respond to electrical communications. Further, the arrows ofFIG. 1 represent the electrical communications between the elements forFIG. 1 . Note that the thick lines extending from thelocal server 120 represent one or more communication interfaces of the local network through which the one ormore user devices 110, the at least onetethered device 112, and/or the one ormore sensor devices 114 communicatively connect to the local server 120). - The one or
more user devices 110, the at least onetethered device 112, and the one ormore sensor devices 114 can generally be referred to as mobile devices. A mobile device is a portable computing device comprising at least a processor and a memory. Examples of a mobile device include smartphones, tablet computers, personal digital assistants, wearable devices (e.g., watches), etc. Mobile devices can include one or more sensors for detecting physical data or can connect to other devices including the same. A sensor can comprise any electro-mechanical device for detecting conditions of an object or environment (sensor examples include biometric detectors, such as pulse sensors, moisture detectors, heat sensors, etc.). - In operation, the
system 100 utilizes can utilize one or more sensors (e.g., the one ormore sensor devices 114 and/or the at least one tethered device 112) to capture physical data of one or more users at thevenue 170 experiencing entertainment. Thesystem 100 utilizes can also utilize mobile devices (e.g., the one or more user devices 110) to receive user inputs from the one or more users. Thesystem 100 can capture the physical data and/or the user inputs during the entertainment in real-time or immediately after experiencing the entertainment. - The physical data can comprise sensor readings that capture and identify a status of bodily functions (i.e., blood pressure, heart rate, breathing rates, and skin conductance) of the one or more users. The user inputs can comprise direct feedback for information submitted by the one or more users via corresponding mobile devices. Note that ‘immediately after experiencing the entertainment’ can be defined as a time interval beginning at a conclusion of the entertainment and extending for a predetermined amount of time. In accordance with one or more embodiments, the predetermined amount of time can be chosen from a range of one second to one day.
- For example, as shown in
FIG. 1 , the user device 110-1 and the one ormore sensor devices 114 can be standalone mobile devices capable of detecting/receiving physical data, and the user device 110-2 can be coupled to the at least onetethered device 112 that also detects/receives physical data. In a first user device 110-1 case, a biometric scanner can be included in the user device 110-1 to determine a pulse of a user. In a second user device 110-2 case, thetethered device 112 can be a wearable device that includes a biometric scanner to determine a pulse of a user, which is then transmitted to the user device 110-2. In both cases, the user devices 110-1 and 110-2 can transmit the pulse, as physical data, to thelocal server 120. In thesensor device 114 case, the sensor is a wearable device that automatically transmits physical data (i.e., the pulse of a user) to thelocal server 120. - The
system 100 analyzes the physical data and/or the user inputs to determine sentiment data for each of the one or more users. The sentiment data quantifies how a user feels (i.e., a sentimental or emotional response) with respect to the entertainment. Thesystem 100 utilizes can utilize the sentiment data (for each of the one or more users who experience the entertainment) to compile a sentiment rating for the entertainment. The sentiment rating is a value that assesses and quantifies a user emotional response for a user population that experiences the entertainment. - With respect to
system 100, thelocal server 120 and/or theexternal server 150 can analyze the physical data and/or the user inputs to determine the sentiment data and utilize the sentiment data to compile a sentiment rating for the entertainment. Thelocal server 120 and theexternal server 150 can also comprise a database, data repository or other data store and may include various kinds of mechanisms for storing, accessing, retrieving, and processing various kinds of data (e.g., the physical data, the user inputs, the sentiment data, and the sentiment rating), including a hierarchical database, a set of files in a file system, an application database in a proprietary format, a relational database management system (RDBMS), etc. - Turning now to
FIG. 2 , an operational embodiment of thesystem 100 ofFIG. 1 is described with respect to theprocess flow 200 in accordance with one or more embodiments. In general, theprocess flow 200 analyzes sentiments of users experiencing any entertainment to generate the sentiment rating. - The
process flow 200 begins atblock 205, where thesystem 100 receives an input indicating permission to collect sentiment data. The input indicating permission is representative of one or more inputs, each of which is provided by a user via their corresponding mobile device, identifying that a user has optioned-into (opted-in) participating in the collection of the sentiment data. The receipt of an input indicating permission by thesystem 100 can include receiving permission at the one or moremobile devices 110, thelocal server 120, and/or theexternal server 150. - The
system 100 can incentivize the one or more users to provide both the physical data and the user inputs. For instance, if watching a movie, thesystem 100 can indicate to the user that they can receive a discounted movie ticket if the user grants permission to thesystem 100 to obtain physical data from sensors contained within their wearables and to provide a few words at an end of the movie discussing their thoughts of the movie via their mobile devices. - At
block 210, thesystem 100 detects a location and time local to a mobile device. The location and the local time are detected for each mobile device that has provided the input indicated permission to collect sentiment data. The location can be determined utilizing location services of the mobile device and/or by the mobile device being on a local network of thevenue 170. The time local can be determined by the clock of the mobile device. The detection of the location and the local time by thesystem 100 can be performed by the one or moremobile devices 110, thelocal server 120, and/or theexternal server 150. - At
block 215, thesystem 100 determines an entertainment based on the location and the time. Thesystem 100 can confirm that a user is at a specific location at a particular time with respect to the entertainment to verify that the user is actually experiencing the entertainment. Thesystem 100 can compare the specific location and time to a schedule for the venue 117 already in thelocal server 120. For example, thesystem 100 can utilize the local network of thevenue 170 to determine that the user device 110-1 is on the premises. In the case of a movie theater, the local network can determine which theater room the mobile device is in, and therefore can determine based on time which movie the user is watching. - At
block 220, thesystem 100 collects physical data and user inputs during and/or immediately after the entertainment. Thesystem 100 can utilize thesensor devices 114, thetethered device 112, and theuser devices 110 to receive the physical data and the user inputs. - At
block 225, thesystem 100 analyzes the physical data and the user inputs to determine sentiment data. That is, the physical data and the user inputs are processed to generate the sentiment data for each user who opted-in. For example, each pulse of a user is used to determine whether a heart rate has accelerated or remained the same while experiencing the entertainment. If during a horror movie, the heart rates of the users increase during the horror movie, thesystem 100 determines that the sentiment of the users is scared/excited. The analyzation of the physical data and the user inputs by thesystem 100 can be performed by the one or moremobile devices 110, thelocal server 120, and/or theexternal server 150. The analyses of the physical data and the user inputs can include natural language processing, text analysis, computational linguistics, and biometrics to systematically identify, extract, quantify, and study affective states and subjective information to determine the sentiment data of the one or more users. - At
block 230, thesystem 100 determines from the sentiment data the sentiment rating. For instance, using the sentiment data based on the collected physical data and user inputs, thesystem 100 determines the overall experience of the users. The sentiment rating is a value that assesses and quantifies a collective emotional response for users that experience the entertainment. The sentiment rating determination by thesystem 100 can be performed by the one or moremobile devices 110, thelocal server 120, and/or theexternal server 150. -
FIG. 3 depicts aprocess flow 300 of a sensor and sentiment analysis system in accordance with one or more embodiments. Theprocess flow 300 begins atblock 305, where a user purchases a ticket. For instance, the user arrives at an entertainment venue (e.g., a movie theater, a concert hall, etc.) and buys the ticket to experience an entertainment at the entertainment venue (e.g., the venue 170). - At
decision block 310, the sensor and sentiment analysis system determines whether the user grants permission to access sensor information on a wearable device. The wearable device can be user provided or venue provided. For instance, upon buying the ticket at a kiosk, the user is proposed with an opportunity (via pop-up on the kiosk during the purchase) to allow the sensor and sentiment analysis system to access sensor information from the wearable device while the user experiences the entertainment. The user can also be proposed with an opportunity to provide thoughts after the entertainment. In return, the sensor and sentiment analysis system indicate that an incentive (i.e. 10% off next ticket, free popcorn, etc.) will be received by opting-in. - If the sensor and sentiment analysis system determine that the user has granted permission to access the sensor information on the wearable device, then the
process flow 300 proceeds to decision block 315 via the YES arrow. For instance, at the kiosk, the user is proposed type in a ticket code (provided via the kiosk) into an application to sync the mobile phone with the sensor and sentiment analysis system. Atdecision block 315, the sensor and sentiment analysis system determines whether the user has an application on their mobile device (e.g., the application communicates with the sensor and sentiment analysis system in response to the ticket code being entered into the applications). If the sensor and sentiment analysis system determines that the user has the application on their mobile device, then theprocess flow 300 proceeds to block 320 via the YES arrow. - At
block 320, the sensor and sentiment analysis system determines a location of the mobile device. For instance, the sensor and sentiment analysis system utilizes a global positioning system of the mobile device to detect a location. Atblock 325, the sensor and sentiment analysis system collects sensor data (e.g., physical data). The sensor and sentiment analysis system can collect the physical data once the entertainment starts. For instance, as the user watches/listens to the entertainment, the wearable device collects physical data from sensors every few minutes. The wearable device can provide the physical data, once collected, to the sensor and sentiment analysis system. - At
decision block 330, the sensor and sentiment analysis system determines whether the user will input their thoughts about the entertainment. In accordance with one or more embodiments, when the entertainment is over, the user is prompted as to whether they want to take a few minutes to say how they felt about the entertainment. If the sensor and sentiment analysis system determines that the user agrees to input their thoughts about the entertainment, theprocess flow 300 proceeds to block 335 via the AGREE arrow. - In accordance with one or more embodiments, when the entertainment is over, the permission received at
block 310 enables the sensor and sentiment analysis system to prompt the user to enter their user input. Thus, theprocess flow 300 proceeds to block 335. - At
block 335, the user provides direct feedback. For instance, the user can speak into the mobile device, which utilizes speech to text technology to turn their thoughts into text. - At
block 340, the sensor and sentiment analysis system stores the sensor data (and the user input if received). Note that if the sensor and sentiment analysis system determines that the user does not agree to input their thoughts about the entertainment atdecision block 330, theprocess flow 300 proceeds directly to block 340 via the DISAGREE arrow. The process flow 300 can then return to block 305 afterblock 340 and repeat upon the purchasing of another ticket. - Returning to decision block 315, if the sensor and sentiment analysis system determines that the user does not have the application on their mobile device, then the
process flow 300 proceeds to block 350 via the NO arrow. Atdecision block 350, the sensor and sentiment analysis system requests via the kiosk for the user to download the application. The sensor and sentiment analysis system also can confirm that the user has granted permission to access the sensor information on the wearable device. If the sensor and sentiment analysis system determines that the user agrees to both conditions, theprocess flow 300 proceeds to block 320 via the AGREE arrow and theprocess flow 300 proceeds accordingly. If the sensor and sentiment analysis system determines that the user agrees to download the application but does not confirm their previous permission, theprocess flow 300 proceeds to block 330 via the DISAGREE arrow and theprocess flow 300 proceeds accordingly. In this case, if a user never agrees to provide physical data, the user can still provide direct feedback about the entertainment. - Returning to decision block 310, if the sensor and sentiment analysis system determines that the user has not granted permission to access the sensor information on the wearable device, then the
process flow 300 proceeds to decision block 360 via the NO arrow. Atdecision block 360, the sensor and sentiment analysis system determines whether the user has an application on their mobile device. If the sensor and sentiment analysis system determines that the user has the application on their mobile device, then theprocess flow 300 proceeds to block 330 via the YES arrow and theprocess flow 300 proceeds accordingly. If the sensor and sentiment analysis system determines that the user does not have the application on their mobile device, then theprocess flow 300 proceeds to block 305 via the NO arrow and theprocess flow 300 proceeds accordingly. -
FIG. 4 depicts an example of asystem 400 in accordance with one or more embodiments. Thesystem 400 has one or more central processing units (CPU(s)) 401 a, 401 b, 401 c, etc. (collectively or generically referred to as processor(s) 401). The processors 401, also referred to as processing circuits, are coupled via a system bus 402 tosystem memory 403 and various other components. Thesystem memory 403 can include a read only memory (ROM) 404 and a random access memory (RAM) 405. TheROM 404 is coupled to the system bus 402 and may include a basic input/output system (BIOS), which controls certain basic functions of thesystem 400. The RAM is read-write memory coupled to the system bus 402 for use by the processors 401. -
FIG. 4 further depicts an input/output (I/O)adapter 406 and acommunications adapter 407 coupled to the system bus 402. The I/O adapter 406 may be a small computer system interface (SCSI) adapter that communicates with ahard disk 408 and/or any other similar component. The I/O adapter 406 and thehard disk 408 are collectively referred to herein as amass storage 410. Asoftware 411 for execution on thesystem 400 may be stored in themass storage 410. Themass storage 410 is an example of a tangible storage medium readable by the processors 401, where thesoftware 411 is stored as instructions for execution by the processors 401 to cause thesystem 400 to operate, such as is described herein with reference toFIG. 2 . Examples of computer program product and the execution of such instruction is discussed herein in more detail. Referring again toFIG. 4 , acommunications adapter 407 interconnects the system bus 402 with anetwork 412, which may be an outside network, enabling thesystem 400 to communicate with other such systems. A display (e.g., screen, a display monitor) 415 is connected to the system bus 402 by adisplay adapter 416, which may include a graphics controller to improve the performance of graphics intensive applications and a video controller. In one embodiment, theadapters interface adapter 420 and thedisplay adapter 416. A keyboard, a mouse, a speaker, etc. can be interconnected to the system bus 402 via theinterface adapter 420, which may include, for example, a Super I/O chip integrating multiple device adapters into a single integrated circuit. - Thus, as configured in
FIG. 4 , thesystem 400 includes processing capability in the form of the processors 401, and, storage capability including thesystem memory 403 and themass storage 410, input means such as the keyboard and the mouse, and output capability including the speaker and thedisplay 415. In one embodiment, a portion of thesystem memory 403 and themass storage 410 collectively store an operating system, such as the z/OS or AIX operating system from IBM Corporation, to coordinate the functions of the various components shown inFIG. 4 . - The present invention may be a system, a method, and/or a computer program product at any possible technical detail level of integration. The computer program product may include a computer readable storage medium (or media) having computer readable program instructions thereon for causing a processor to carry out aspects of the present invention.
- The computer readable storage medium can be a tangible device that can retain and store instructions for use by an instruction execution device. The computer readable storage medium may be, for example, but is not limited to, an electronic storage device, a magnetic storage device, an optical storage device, an electromagnetic storage device, a semiconductor storage device, or any suitable combination of the foregoing. A non-exhaustive list of more specific examples of the computer readable storage medium includes the following: a portable computer diskette, a hard disk, a random access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or Flash memory), a static random access memory (SRAM), a portable compact disc read-only memory (CD-ROM), a digital versatile disk (DVD), a memory stick, a floppy disk, a mechanically encoded device such as punch-cards or raised structures in a groove having instructions recorded thereon, and any suitable combination of the foregoing. A computer readable storage medium, as used herein, is not to be construed as being transitory signals per se, such as radio waves or other freely propagating electromagnetic waves, electromagnetic waves propagating through a waveguide or other transmission media (e.g., light pulses passing through a fiber-optic cable), or electrical signals transmitted through a wire.
- Computer readable program instructions described herein can be downloaded to respective computing/processing devices from a computer readable storage medium or to an external computer or external storage device via a network, for example, the Internet, a local area network, a wide area network and/or a wireless network. The network may comprise copper transmission cables, optical transmission fibers, wireless transmission, routers, firewalls, switches, gateway computers and/or edge servers. A network adapter card or network interface in each computing/processing device receives computer readable program instructions from the network and forwards the computer readable program instructions for storage in a computer readable storage medium within the respective computing/processing device.
- Computer readable program instructions for carrying out operations of the present invention may be assembler instructions, instruction-set-architecture (ISA) instructions, machine instructions, machine dependent instructions, microcode, firmware instructions, state-setting data, configuration data for integrated circuitry, or either source code or object code written in any combination of one or more programming languages, including an object oriented programming language such as Smalltalk, C++, or the like, and procedural programming languages, such as the “C” programming language or similar programming languages. The computer readable program instructions may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server. In the latter scenario, the remote computer may be connected to the user's computer through any type of network, including a local area network (LAN) or a wide area network (WAN), or the connection may be made to an external computer (for example, through the Internet using an Internet Service Provider). In some embodiments, electronic circuitry including, for example, programmable logic circuitry, field-programmable gate arrays (FPGA), or programmable logic arrays (PLA) may execute the computer readable program instructions by utilizing state information of the computer readable program instructions to personalize the electronic circuitry, in order to perform aspects of the present invention.
- Aspects of the present invention are described herein with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems), and computer program products according to embodiments of the invention. It will be understood that each block of the flowchart illustrations and/or block diagrams, and combinations of blocks in the flowchart illustrations and/or block diagrams, can be implemented by computer readable program instructions.
- These computer readable program instructions may be provided to a processor of a general purpose computer, special purpose computer, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks. These computer readable program instructions may also be stored in a computer readable storage medium that can direct a computer, a programmable data processing apparatus, and/or other devices to function in a particular manner, such that the computer readable storage medium having instructions stored therein comprises an article of manufacture including instructions which implement aspects of the function/act specified in the flowchart and/or block diagram block or blocks.
- The computer readable program instructions may also be loaded onto a computer, other programmable data processing apparatus, or other device to cause a series of operational steps to be performed on the computer, other programmable apparatus or other device to produce a computer implemented process, such that the instructions which execute on the computer, other programmable apparatus, or other device implement the functions/acts specified in the flowchart and/or block diagram block or blocks.
- The flowchart and block diagrams in the Figures illustrate the architecture, functionality, and operation of possible implementations of systems, methods, and computer program products according to various embodiments of the present invention. In this regard, each block in the flowchart or block diagrams may represent a module, segment, or portion of instructions, which comprises one or more executable instructions for implementing the specified logical function(s). In some alternative implementations, the functions noted in the blocks may occur out of the order noted in the Figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved. It will also be noted that each block of the block diagrams and/or flowchart illustration, and combinations of blocks in the block diagrams and/or flowchart illustration, can be implemented by special purpose hardware-based systems that perform the specified functions or acts or carry out combinations of special purpose hardware and computer instructions.
- The terminology used herein is for the purpose of describing particular embodiments only and is not intended to be limiting. As used herein, the singular forms “a”, “an” and “the” are intended to include the plural forms as well, unless the context clearly indicates otherwise. It will be further understood that the terms “comprises” and/or “comprising,” when used in this specification, specify the presence of stated features, integers, steps, operations, elements, and/or components, but do not preclude the presence or addition of one more other features, integers, steps, operations, element components, and/or groups thereof
- The descriptions of the various embodiments herein have been presented for purposes of illustration, but are not intended to be exhaustive or limited to the embodiments disclosed. Many modifications and variations will be apparent to those of ordinary skill in the art without departing from the scope and spirit of the described embodiments. The terminology used herein was chosen to best explain the principles of the embodiments, the practical application or technical improvement over technologies found in the marketplace, or to enable others of ordinary skill in the art to understand the embodiments disclosed herein.
Claims (7)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US15/810,328 US20190012710A1 (en) | 2017-07-05 | 2017-11-13 | Sensors and sentiment analysis for rating systems |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US15/641,763 US11010797B2 (en) | 2017-07-05 | 2017-07-05 | Sensors and sentiment analysis for rating systems |
US15/810,328 US20190012710A1 (en) | 2017-07-05 | 2017-11-13 | Sensors and sentiment analysis for rating systems |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/641,763 Continuation US11010797B2 (en) | 2017-07-05 | 2017-07-05 | Sensors and sentiment analysis for rating systems |
Publications (1)
Publication Number | Publication Date |
---|---|
US20190012710A1 true US20190012710A1 (en) | 2019-01-10 |
Family
ID=64902825
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/641,763 Expired - Fee Related US11010797B2 (en) | 2017-07-05 | 2017-07-05 | Sensors and sentiment analysis for rating systems |
US15/810,328 Abandoned US20190012710A1 (en) | 2017-07-05 | 2017-11-13 | Sensors and sentiment analysis for rating systems |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/641,763 Expired - Fee Related US11010797B2 (en) | 2017-07-05 | 2017-07-05 | Sensors and sentiment analysis for rating systems |
Country Status (1)
Country | Link |
---|---|
US (2) | US11010797B2 (en) |
Families Citing this family (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111627144A (en) * | 2019-10-13 | 2020-09-04 | 李绪臣 | Command trigger platform based on user big data and corresponding terminal |
US20220174357A1 (en) * | 2020-11-30 | 2022-06-02 | At&T Intellectual Property I, L.P. | Simulating audience feedback in remote broadcast events |
US11503090B2 (en) | 2020-11-30 | 2022-11-15 | At&T Intellectual Property I, L.P. | Remote audience feedback mechanism |
Citations (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20050289582A1 (en) * | 2004-06-24 | 2005-12-29 | Hitachi, Ltd. | System and method for capturing and using biometrics to review a product, service, creative work or thing |
US20080065468A1 (en) * | 2006-09-07 | 2008-03-13 | Charles John Berg | Methods for Measuring Emotive Response and Selection Preference |
US20080096726A1 (en) * | 2006-09-07 | 2008-04-24 | Nike, Inc. | Athletic Performance Sensing and/or Tracking Systems and Methods |
US20080117202A1 (en) * | 2006-11-22 | 2008-05-22 | Ronald Martinez | Methods, Systems and Apparatus for Delivery of Media |
US20120324494A1 (en) * | 2011-06-17 | 2012-12-20 | Microsoft Corporation | Selection of advertisements via viewer feedback |
US20130091214A1 (en) * | 2011-10-08 | 2013-04-11 | Broadcom Corporation | Media social network |
US20130245396A1 (en) * | 2010-06-07 | 2013-09-19 | Affectiva, Inc. | Mental state analysis using wearable-camera devices |
US20140347265A1 (en) * | 2013-03-15 | 2014-11-27 | Interaxon Inc. | Wearable computing apparatus and method |
US9202233B1 (en) * | 2010-09-30 | 2015-12-01 | Imdb.Com, Inc. | Event attendance determinations |
US20160021425A1 (en) * | 2013-06-26 | 2016-01-21 | Thomson Licensing | System and method for predicting audience responses to content from electro-dermal activity signals |
US20160170996A1 (en) * | 2014-08-21 | 2016-06-16 | Affectomatics Ltd. | Crowd-based scores for experiences from measurements of affective response |
US20160212466A1 (en) * | 2015-01-21 | 2016-07-21 | Krush Technologies, Llc | Automatic system and method for determining individual and/or collective intrinsic user reactions to political events |
US20160357501A1 (en) * | 2015-06-03 | 2016-12-08 | Skullcandy, Inc. | Audio devices and related methods for acquiring audio device use information |
US20180053197A1 (en) * | 2016-08-18 | 2018-02-22 | International Business Machines Corporation | Normalizing user responses to events |
US10271099B2 (en) * | 2017-02-27 | 2019-04-23 | International Business Machines Corporation | Deep movie analysis based on cognitive controls in cinematography |
Family Cites Families (14)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9326099B2 (en) * | 2008-02-11 | 2016-04-26 | Popular Metrics, Inc. | System and method for determining audience characteristics of a music concert based on mobile phone tracking and mobile data transmissions |
US10142687B2 (en) | 2010-11-07 | 2018-11-27 | Symphony Advanced Media, Inc. | Audience content exposure monitoring apparatuses, methods and systems |
US20120290910A1 (en) | 2011-05-11 | 2012-11-15 | Searchreviews LLC | Ranking sentiment-related content using sentiment and factor-based analysis of contextually-relevant user-generated data |
US20130018954A1 (en) | 2011-07-15 | 2013-01-17 | Samsung Electronics Co., Ltd. | Situation-aware user sentiment social interest models |
US8725773B2 (en) | 2011-08-30 | 2014-05-13 | E-Rewards, Inc. | System and method for generating a knowledge metric using qualitative internet data |
US9832519B2 (en) * | 2012-04-18 | 2017-11-28 | Scorpcast, Llc | Interactive video distribution system and video player utilizing a client server architecture |
US10453097B2 (en) | 2014-01-13 | 2019-10-22 | Nant Holdings Ip, Llc | Sentiments based transaction systems and methods |
US9536329B2 (en) * | 2014-05-30 | 2017-01-03 | Adobe Systems Incorporated | Method and apparatus for performing sentiment analysis based on user reactions to displayable content |
US20160055430A1 (en) * | 2014-08-20 | 2016-02-25 | Heesham Naji | Social Networking Enhanced System for Review and Sale of Music and Venue-Related Products and Services |
DE102015113942A1 (en) | 2014-08-21 | 2016-02-25 | Affectomatics Ltd. | Rating of holiday destinations based on affective response |
US20160148126A1 (en) * | 2014-11-20 | 2016-05-26 | Atom Tickets, LLC | Automated event concessions management system |
US10122483B2 (en) * | 2015-02-04 | 2018-11-06 | Ebay Inc. | Crowd sentiment detection and analysis |
US11080723B2 (en) * | 2017-03-07 | 2021-08-03 | International Business Machines Corporation | Real time event audience sentiment analysis utilizing biometric data |
CA2998249A1 (en) * | 2017-03-17 | 2018-09-17 | Edatanetworks Inc. | Artificial intelligence engine incenting merchant transaction with consumer affinity |
-
2017
- 2017-07-05 US US15/641,763 patent/US11010797B2/en not_active Expired - Fee Related
- 2017-11-13 US US15/810,328 patent/US20190012710A1/en not_active Abandoned
Patent Citations (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20050289582A1 (en) * | 2004-06-24 | 2005-12-29 | Hitachi, Ltd. | System and method for capturing and using biometrics to review a product, service, creative work or thing |
US20080065468A1 (en) * | 2006-09-07 | 2008-03-13 | Charles John Berg | Methods for Measuring Emotive Response and Selection Preference |
US20080096726A1 (en) * | 2006-09-07 | 2008-04-24 | Nike, Inc. | Athletic Performance Sensing and/or Tracking Systems and Methods |
US20080117202A1 (en) * | 2006-11-22 | 2008-05-22 | Ronald Martinez | Methods, Systems and Apparatus for Delivery of Media |
US20130245396A1 (en) * | 2010-06-07 | 2013-09-19 | Affectiva, Inc. | Mental state analysis using wearable-camera devices |
US9202233B1 (en) * | 2010-09-30 | 2015-12-01 | Imdb.Com, Inc. | Event attendance determinations |
US20120324494A1 (en) * | 2011-06-17 | 2012-12-20 | Microsoft Corporation | Selection of advertisements via viewer feedback |
US20130091214A1 (en) * | 2011-10-08 | 2013-04-11 | Broadcom Corporation | Media social network |
US20140347265A1 (en) * | 2013-03-15 | 2014-11-27 | Interaxon Inc. | Wearable computing apparatus and method |
US20160021425A1 (en) * | 2013-06-26 | 2016-01-21 | Thomson Licensing | System and method for predicting audience responses to content from electro-dermal activity signals |
US20160170996A1 (en) * | 2014-08-21 | 2016-06-16 | Affectomatics Ltd. | Crowd-based scores for experiences from measurements of affective response |
US20160212466A1 (en) * | 2015-01-21 | 2016-07-21 | Krush Technologies, Llc | Automatic system and method for determining individual and/or collective intrinsic user reactions to political events |
US20160357501A1 (en) * | 2015-06-03 | 2016-12-08 | Skullcandy, Inc. | Audio devices and related methods for acquiring audio device use information |
US20180053197A1 (en) * | 2016-08-18 | 2018-02-22 | International Business Machines Corporation | Normalizing user responses to events |
US10271099B2 (en) * | 2017-02-27 | 2019-04-23 | International Business Machines Corporation | Deep movie analysis based on cognitive controls in cinematography |
Also Published As
Publication number | Publication date |
---|---|
US11010797B2 (en) | 2021-05-18 |
US20190012709A1 (en) | 2019-01-10 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11182447B2 (en) | Customized display of emotionally filtered social media content | |
US11636367B2 (en) | Systems, apparatus, and methods for generating prediction sets based on a known set of features | |
US11044206B2 (en) | Live video anomaly detection | |
US20160048848A1 (en) | Mobile application analytics framework | |
US10778353B2 (en) | Providing real-time audience awareness to speaker | |
US11122332B2 (en) | Selective video watching by analyzing user behavior and video content | |
US20180129962A1 (en) | Database query and data mining in intelligent distributed communication networks | |
US10958973B2 (en) | Deriving and identifying view preferences of a user consuming streaming content | |
US11789980B2 (en) | Method, system, and non-transitory computer readable record medium for providing multi profile | |
CN110399505B (en) | Semantic tag generation method and device, and computer storage medium | |
US20200135206A1 (en) | Enhancing group decisions within social messaging applications | |
US11010797B2 (en) | Sensors and sentiment analysis for rating systems | |
KR20170056249A (en) | A Computer Device for Providing Psychotherapy and A Program therefor | |
US11106758B2 (en) | Customized display of filtered social media content using a private dislike button | |
US20200089812A1 (en) | Updating social media post based on subsequent related social media content | |
US10339559B2 (en) | Associating social comments with individual assets used in a campaign | |
US10595178B2 (en) | Listing service registrations through a mobile number | |
US20200257825A1 (en) | Customized display of filtered social media content using a private dislike button | |
US20220229921A1 (en) | Timing for user data erasure requests | |
US20130152157A1 (en) | Automatic filtering in social networks | |
US11301479B2 (en) | Service provider selection for queries sent to a digital assistant | |
US10755363B2 (en) | Augmented reviews with cognitive reasoning | |
KR102371028B1 (en) | Device independent re-grab based integrated content processing method and apparatus | |
KR20190109144A (en) | Text Recommendation Device and Text Recommendation Method Using Multiple Text Recommendation Engines | |
CN110555157B (en) | Content recommendation method, content recommendation device and electronic equipment |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: INTERNATIONAL BUSINESS MACHINES CORPORATION, NEW Y Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:DOW, ELI M.;TANG, YUNLI;REEL/FRAME:044110/0942 Effective date: 20170626 |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: FINAL REJECTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: FINAL REJECTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: ADVISORY ACTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: FINAL REJECTION MAILED |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |