US10142472B2 - Collection and analysis of audio during hold - Google Patents

Collection and analysis of audio during hold Download PDF

Info

Publication number
US10142472B2
US10142472B2 US14/697,436 US201514697436A US10142472B2 US 10142472 B2 US10142472 B2 US 10142472B2 US 201514697436 A US201514697436 A US 201514697436A US 10142472 B2 US10142472 B2 US 10142472B2
Authority
US
United States
Prior art keywords
audio
link
primary
selection
packets
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
US14/697,436
Other versions
US20160072949A1 (en
Inventor
Ken Kannappan
Douglas K Rosener
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Hewlett Packard Development Co LP
Original Assignee
Plantronics Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from US14/478,885 external-priority patent/US10178473B2/en
Application filed by Plantronics Inc filed Critical Plantronics Inc
Priority to US14/697,436 priority Critical patent/US10142472B2/en
Assigned to PLANTRONICS, INC. reassignment PLANTRONICS, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: ROSENER, DOUGLAS K, KANNAPPAN, KEN
Publication of US20160072949A1 publication Critical patent/US20160072949A1/en
Assigned to WELLS FARGO BANK, NATIONAL ASSOCIATION reassignment WELLS FARGO BANK, NATIONAL ASSOCIATION SECURITY AGREEMENT Assignors: PLANTRONICS, INC., POLYCOM, INC.
Application granted granted Critical
Publication of US10142472B2 publication Critical patent/US10142472B2/en
Assigned to POLYCOM, INC., PLANTRONICS, INC. reassignment POLYCOM, INC. RELEASE OF PATENT SECURITY INTERESTS Assignors: WELLS FARGO BANK, NATIONAL ASSOCIATION
Assigned to HEWLETT-PACKARD DEVELOPMENT COMPANY, L.P. reassignment HEWLETT-PACKARD DEVELOPMENT COMPANY, L.P. NUNC PRO TUNC ASSIGNMENT Assignors: PLANTRONICS, INC.
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M3/00Automatic or semi-automatic exchanges
    • H04M3/42Systems providing special services or facilities to subscribers
    • H04M3/50Centralised arrangements for answering calls; Centralised arrangements for recording messages for absent or busy subscribers ; Centralised arrangements for recording messages
    • H04M3/51Centralised call answering arrangements requiring operator intervention, e.g. call or contact centers for telemarketing
    • H04M3/5175Call or contact centers supervision arrangements
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M3/00Automatic or semi-automatic exchanges
    • H04M3/42Systems providing special services or facilities to subscribers
    • H04M3/428Arrangements for placing incoming calls on hold
    • H04M3/4283Call holding circuits
    • H04W4/008
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04WWIRELESS COMMUNICATION NETWORKS
    • H04W4/00Services specially adapted for wireless communication networks; Facilities therefor
    • H04W4/80Services using short range communication, e.g. near-field communication [NFC], radio-frequency identification [RFID] or low energy communication

Definitions

  • the present disclosure relates generally to the field of audio processing. More particularly, the present disclosure relates to analysis of audio generated by a microphone.
  • agent may employ the hold function.
  • the agent may be coughing or sneezing, and does not want the remote party to hear.
  • the agent may be having difficulty handling a call, and so is asking questions of his co-workers. Or the agent may be doing things not related to his work.
  • the behavior of the agent may indicate a problem.
  • An ill agent may spread illness to others in the call center.
  • An agent asking questions of his co-workers may need more training, or may have competency issues. Or an agent may not be providing the work desired.
  • an embodiment features an apparatus comprising: a microphone configured to produce audio; a hold control configured to select a connected selection or a hold selection; a processor configured to identify the audio produced during the connected selection as primary audio, and to identify the audio produced during the hold selection as secondary audio; and a transceiver configured to transmit the primary audio and the secondary audio.
  • the transceiver is further configured to transmit the primary audio over a first link, and to transmit the secondary audio over a second link.
  • the first link is an audio link; and the second link is a data link.
  • the first link is a Bluetooth Synchronous Connection Oriented (SCO) link; and the secondary link is a Bluetooth Asynchronous Connection-Less (ACL) link.
  • the transceiver comprises: a first transceiver configured to transmit the primary audio according to a first protocol; and a second transceiver configured to transmit the secondary audio according to a second protocol.
  • Some embodiments comprise a memory configured to store the secondary audio prior to the transceiver transmitting the secondary audio.
  • the processor is further configured to packetize the primary audio and the secondary audio, and to mark at least one of (i) packets of the primary audio and (ii) packets of the secondary audio.
  • Some embodiments comprise a headset.
  • an embodiment features a method comprising: producing audio responsive to sound; determining a selection of a hold control configured to select a connected selection or a hold selection; identifying the audio produced during the connected selection as primary audio; identifying the audio produced during the hold selection as secondary audio; and transmitting the primary audio and the secondary audio.
  • Embodiments of the method can include one or more of the following features. Some embodiments comprise transmitting the primary audio over a first link; and transmitting the secondary audio over a second link. Some embodiments comprise transmitting the primary audio according to a first protocol; and transmitting the secondary audio according to a second protocol. Some embodiments comprise packetizing the primary audio and the secondary audio; and marking at least one of (i) packets of the primary audio and (ii) packets of the secondary audio.
  • an embodiment features apparatus comprising: a receiver configured to receive audio produced by a headset, wherein the headset has a hold control configured to select a connected selection or a hold selection, and wherein the audio includes primary audio and secondary audio, wherein the primary audio is generated by a microphone of the headset during a connected selection, and wherein the secondary audio is generated by the microphone of the headset during the hold selection; and a switch configured to pass the primary audio to a communications channel, and to pass the secondary audio to an analytics engine.
  • Embodiments of the apparatus can include one or more of the following features.
  • the switch is further configured to pass the primary audio to the analytics engine. Some embodiments comprise the analytics engine.
  • the receiver is further configured to receive the primary audio over a first link, and to receive the secondary audio over a second link.
  • the first link is an audio link; and the secondary link is a data link.
  • the first link is a Bluetooth Synchronous Connection Oriented (SCO) link; and the secondary link is a Bluetooth Asynchronous Connection-Less (ACL) link.
  • the receiver comprises: a first receiver configured to receive the primary audio according to a first protocol; and a second receiver configured to receive the secondary audio according to a second protocol.
  • the audio comprises packets of the primary audio and packets of the secondary audio; at least one of (i) the packets of the primary audio and (ii) the packets of the secondary audio include marks; and the switch is further configured to distinguish the (i) the packets of the primary audio and (ii) the packets of the secondary audio based on the marks.
  • FIG. 1 shows elements of a call center audio system according to an embodiment where the primary and secondary audio are distinguished using packet marking.
  • FIG. 2 shows a process for the call center audio system of FIG. 1 according to one embodiment.
  • FIG. 3 shows elements of a call center audio system according to an embodiment where the primary and secondary audio are distinguished using different communication links.
  • FIG. 4 shows a process for the call center audio system of FIG. 3 according to one embodiment.
  • FIG. 5 shows elements of a call center audio system according to an embodiment where the primary and secondary audio are distinguished using different communication protocols.
  • FIG. 6 shows a process for the call center audio system of FIG. 5 according to one embodiment.
  • Embodiments of the present disclosure provides collection of audio during hold for analysis and the like.
  • sound received by a microphone when the call is on hold that is, the hold function is active
  • Sound received by the microphone when the call is not on hold that is, while the hold function is not active
  • Audio collected while the call is not on hold is referred to herein as “primary audio.”
  • Audio collected while the call is on hold is referred to herein as “secondary audio.”
  • various techniques are employed to distinguish the primary audio from the secondary audio.
  • packets of the primary audio and/or secondary audio may be marked, for example by setting flags in headers of the packets.
  • the primary audio and secondary audio may be transmitted over different links, using different protocols, and the like. Other features are contemplated as well.
  • Embodiments of the present disclosure are described in terms of an agent wearing a wireless headset in a call center. However, the techniques described herein are applicable to any audio device having a microphone, and in any environment.
  • FIG. 1 shows elements of a call center audio system 100 according to an embodiment where the primary and secondary audio are distinguished using packet marking.
  • elements of the call center audio system 100 are presented in one arrangement, other embodiments may feature other arrangements.
  • elements of the call center audio system 100 may be implemented in hardware, software, or combinations thereof.
  • various elements of the call center audio system 100 may be implemented as one or more digital signal processors.
  • the call center audio system 100 may include a headset 102 in communication with a host 104 over a wireless channel 106 .
  • the headset 102 may include a microphone (MIC) 108 , a hold control (HOLD) 110 , a processor (PROC) 112 , and a transceiver (TRX) 114 .
  • the host 104 may include a transceiver (TRX) 116 , a switch (SW) 118 , an audio channel 120 , and an analytics engine 122 . While in the described embodiments, the analytics engine 122 may be part of the host 104 , in other embodiments, the analytics engine 122 may not be part of the host 104 , and may be located outside the call center audio system 100 .
  • the hold control 110 may select either a connected selection or a hold selection.
  • the hold control 110 may be user-operable, automatic, or both.
  • a user-operable hold control 110 may be implemented as a button, slide switch, or the like.
  • An automatic hold control 110 may automatically select the connected selection when donned, and may automatically select the hold selection when doffed.
  • the processor 112 may include an analog-to-digital converter, a digital signal processor, a packetizer, and the like.
  • the wireless channel 106 may be a Bluetooth channel, a Digital Enhanced Cordless Telecommunications (DECT) channel, a Wi-Fi channel, or the like.
  • the audio channel 120 may be any audio channel suitable for passing packets of primary audio to a remote party.
  • the secondary audio may be routed directly to the host 104 , or via another device such as a smart phone or computer.
  • FIG. 2 shows a process 200 for the call center audio system 100 of FIG. 1 according to one embodiment.
  • the elements of process 200 are presented in one arrangement, other embodiments may feature other arrangements.
  • some or all of the elements of process 200 can be executed in a different order, concurrently, and the like.
  • some elements of process 200 may not be performed, and may not be executed immediately after each other.
  • some or all of the elements of process 200 can be performed automatically, that is, without human intervention.
  • the microphone 108 may generate audio that represents sound received by the microphone 108 .
  • the processor 112 may process the audio.
  • An analog-to-digital converter within the processor may convert the audio to digital audio.
  • the processor 112 may packetize the digital audio.
  • the hold control 110 may be operated by the agent to select either a connected selection or a hold selection. The selection may be communicated to the processor 112 by a hold signal 128 .
  • the processor 112 may determine the selection based on the hold signal 128 .
  • the processor 112 may identify the audio produced during the connected selection as primary audio, and may identify the audio produced during the hold selection as secondary audio. In the present embodiment, at 206 , the processor 112 may identify the audio by marking some or all of the packets in the audio stream. The processor 112 may mark the packets in accordance with the hold signal 128 . The processor 112 may mark the packets of the digital audio when the hold signal 128 indicates the hold selection, when the hold signal 128 indicates the connected selection, or both. The processor 112 may mark the packets, for example, by setting or clearing a flag in the header of each packet, or in the header of a packet to indicate a transition between blocks of secondary and primary audio, and the like. The processor 112 may insert control packets transition between blocks of secondary and primary audio, and the like. At 208 , the transceiver 114 of the headset 102 may transmit a signal representing the packets over the wireless channel 106 .
  • the transceiver 116 of the host 104 may receive the signal representing the packets over the wireless channel 106 .
  • the switch 118 routes the packets according to the marks in the packets. In particular, the switch 118 routes the packets of primary audio to the audio channel 120 , and routes the packets of secondary audio to the analytics engine 122 for analysis. In some embodiments, the switch 118 may also route some or all of the packets of primary audio to the analytics engine 122 for analysis.
  • FIG. 3 shows elements of a call center audio system 300 according to an embodiment where the primary and secondary audio are distinguished using different communication links.
  • elements of the call center audio system 300 are presented in one arrangement, other embodiments may feature other arrangements.
  • elements of the call center audio system 300 may be implemented in hardware, software, or combinations thereof.
  • various elements of the call center audio system 300 may be implemented as one or more digital signal processors.
  • the call center audio system 300 may include a headset 302 in communication with a host 304 over a wireless channel 306 .
  • the headset 302 may include a microphone (MIC) 308 , a hold control (HOLD) 310 , a processor (PROC) 312 , a memory 324 , and a transceiver (TRX) 314 .
  • the host 304 may include a transceiver (TRX) 316 , an audio channel 320 , and an analytics engine 322 . While in the described embodiments, the analytics engine 322 may be part of the host 304 , in other embodiments, the analytics engine 322 may not be part of the host 304 , and may be located outside the call center audio system 300 .
  • the hold control 310 may select either a connected selection or a hold selection.
  • the hold control 310 may be user-operable, automatic, or both.
  • a user-operable hold control 310 may be implemented as a button, slide switch, or the like.
  • An automatic hold control 310 may automatically select the connected selection when donned, and may automatically select the hold selection when doffed.
  • the processor 312 may include an analog-to-digital converter, a digital signal processor, a packetizer, and the like.
  • the wireless channel 306 may be a Bluetooth channel, a Digital Enhanced Cordless Telecommunications (DECT) channel, a Wi-Fi channel, or the like.
  • the audio channel 320 may be any audio channel suitable for passing packets of primary audio to a remote party. The secondary audio may be routed directly to the host 304 , or via another device such as a smart phone or computer.
  • FIG. 4 shows a process 400 for the call center audio system 300 of FIG. 3 according to one embodiment.
  • the elements of process 400 are presented in one arrangement, other embodiments may feature other arrangements.
  • some or all of the elements of process 400 can be executed in a different order, concurrently, and the like.
  • some elements of process 400 may not be performed, and may not be executed immediately after each other.
  • some or all of the elements of process 400 can be performed automatically, that is, without human intervention.
  • the microphone 308 may generate audio that represents sound received by the microphone 308 .
  • the processor 312 may process the audio.
  • An analog-to-digital converter within the processor may convert the audio to digital audio.
  • the processor 312 may packetize the digital audio.
  • the hold control 310 may be operated by the agent to select either a connected selection or a hold selection. The selection may be communicated to the processor 312 by a hold signal 328 .
  • the processor 312 may determine the selection based on the hold signal 328 .
  • the processor 312 may identify the audio produced during the connected selection as primary audio, and may identify the audio produced during the hold selection as secondary audio. In the present embodiment, the processor 312 may identify the audio by routing the primary audio to one link, and routing the secondary audio to another link. At 406 , the processor 312 may route the packets of digital audio among multiple communication links in accordance with the hold signal 328 . For example, the processor 312 may route the packets of primary audio to an audio link, and may route the packets of secondary audio to a data link.
  • the audio link may be a Bluetooth Synchronous Connection Oriented (SCO) link.
  • the data link may be a Bluetooth Asynchronous Connection-Less (ACL) link. However, other wireless protocols and links may be used.
  • the memory 324 may store the packets of the secondary audio before transmission to the host 304 .
  • the data link need not be open continuously.
  • the transceiver 314 of the headset 302 transmits one or more signals representing the packets over the wireless channel 306 .
  • the transceiver 316 of the host 304 may receive the signal representing the packets over the wireless channel 306 .
  • the transceiver 316 may pass the packets according to the communication links.
  • the transceiver 316 may route the packets of primary audio to the audio channel 320 , and may route the packets of secondary audio to the analytics engine 322 for analysis.
  • the transceiver 316 may also route some or all of the packets of primary audio to the analytics engine 322 for analysis.
  • FIG. 5 shows elements of a call center audio system 500 according to an embodiment where the primary and secondary audio are distinguished using different communication protocols.
  • elements of the call center audio system 500 are presented in one arrangement, other embodiments may feature other arrangements.
  • elements of the call center audio system 500 may be implemented in hardware, software, or combinations thereof.
  • various elements of the call center audio system 500 may be implemented as one or more digital signal processors.
  • the call center audio system 500 may include a headset 502 in communication with a host 504 over wireless channels 506 and 546 .
  • the headset 502 may include a microphone (MIC) 508 , a hold control (HOLD) 510 , a processor (PROC) 512 , a memory 524 , transceivers (TRX) 514 and 534 .
  • the host 504 may include transceivers (TRX) 516 and 536 , an audio channel 520 , and an analytics engine 522 . While in the described embodiments, the analytics engine 522 may be part of the host 504 , in other embodiments, the analytics engine 522 may not be part of the host 504 , and may be located outside the call center.
  • the hold control 510 may select either a connected selection or a hold selection.
  • the hold control 510 may be user-operable, automatic, or both.
  • a user-operable hold control 510 may be implemented as a button, slide switch, or the like.
  • An automatic hold control 510 may automatically select the connected selection when donned, and may automatically select the hold selection when doffed.
  • the processor 512 may include an analog-to-digital converter, a digital signal processor, a packetizer, and the like.
  • the wireless channels 506 and 546 may employ different wireless protocols, for example such as Bluetooth and Wi-Fi, respectively. However, any protocol may be used, for example such as Digital Enhanced Cordless Telecommunications (DECT), or the like.
  • the audio channel 520 may be any audio channel suitable for passing the packets of primary audio to a remote party.
  • the secondary audio may be routed directly to the host 504 , or via another device such as a smart phone or computer.
  • FIG. 6 shows a process 600 for the call center audio system 500 of FIG. 5 according to one embodiment.
  • the elements of process 600 are presented in one arrangement, other embodiments may feature other arrangements.
  • some or all of the elements of process 600 can be executed in a different order, concurrently, and the like.
  • some elements of process 600 may not be performed, and may not be executed immediately after each other.
  • some or all of the elements of process 600 can be performed automatically, that is, without human intervention.
  • the microphone 508 may generate audio that represents sound received by the microphone 508 .
  • the processor 512 may process the audio.
  • An analog-to-digital converter within the processor may convert the audio to digital audio.
  • the processor 512 may packetize the digital audio.
  • the hold control 510 may be operated by the agent to select either a connected selection or a hold selection. The selection may be communicated to the processor 512 by a hold signal 528 .
  • the processor 512 may determine the selection based on the hold signal 528 .
  • the processor 512 may identify the audio produced during the connected selection as primary audio, and may identify the audio produced during the hold selection as secondary audio. In the present embodiment, the processor 512 may identify the audio by routing the primary audio to one transceiver, and routing the secondary audio to another transceiver. At 606 , the processor 512 may route the packets of digital audio among multiple transceivers 514 , 534 in accordance with the hold signal 528 . For example, the processor 512 may route the packets of primary audio to one transceiver 514 , and may route the packets of secondary audio to another transceiver 534 .
  • the memory 524 may store the packets of the secondary audio before transmission to the host 504 .
  • the data link need not be open continuously.
  • the transceivers 514 , 534 of the headset 502 transmit signals representing the packets over the respective wireless channel 506 , 546 .
  • the transceivers 516 , 536 of the host 504 may receive the signals representing the packets over the respective wireless channels 506 , 546 .
  • the transceiver 516 may pass the packets of primary audio to the audio channel 520
  • the transceiver 536 may pass the packets of secondary audio to the analytics engine 522 for analysis.
  • the transceiver 516 may also route some or all of the packets of primary audio to the analytics engine 522 for analysis.
  • the analytics engines 122 , 322 , 522 described above may perform any sort of analysis on the secondary audio.
  • the analytics engines 122 , 322 , 522 may identify coughs and sneezes in the secondary audio, keeping metrics as a potential indicator of illness of individual agents and groups of agents.
  • the analytics engines 122 , 322 , 522 may detect questions, for example based on intonation, voice recognition, and the like, keeping metrics as a possible indicator of need for training of individual agents or groups of agents.
  • the analytics engines 122 , 322 , 522 may monitor the agent's speech with hold on or off and make decisions on content, keeping metrics as indicators of time spent on work communications and personal communications. In all cases, a supervisor may be alerted when a metric threshold is exceeded, making it unnecessary for a supervisor to personally monitor calls or observe agents.
  • Embodiments of the present disclosure can be implemented in digital electronic circuitry, or in computer hardware, firmware, software, or in combinations thereof.
  • Embodiments of the present disclosure can be implemented in a computer program product tangibly embodied in a computer-readable storage device for execution by a programmable processor. The described processes can be performed by a programmable processor executing a program of instructions to perform functions by operating on input data and generating output.
  • Embodiments of the present disclosure can be implemented in one or more computer programs that are executable on a programmable system including at least one programmable processor coupled to receive data and instructions from, and to transmit data and instructions to, a data storage system, at least one input device, and at least one output device.
  • Each computer program can be implemented in a high-level procedural or object-oriented programming language, or in assembly or machine language if desired; and in any case, the language can be a compiled or interpreted language.
  • Suitable processors include, by way of example, both general and special purpose microprocessors.
  • processors receive instructions and data from a read-only memory and/or a random access memory.
  • a computer includes one or more mass storage devices for storing data files. Such devices include magnetic disks, such as internal hard disks and removable disks, magneto-optical disks; optical disks, and solid-state disks.
  • Storage devices suitable for tangibly embodying computer program instructions and data include all forms of non-volatile memory, including by way of example semiconductor memory devices, such as EPROM, EEPROM, and flash memory devices; magnetic disks such as internal hard disks and removable disks; magneto-optical disks; and CD-ROM disks. Any of the foregoing can be supplemented by, or incorporated in, ASICs (application-specific integrated circuits).
  • ASICs application-specific integrated circuits.
  • module may refer to any of the above implementations.

Landscapes

  • Engineering & Computer Science (AREA)
  • Signal Processing (AREA)
  • Business, Economics & Management (AREA)
  • Marketing (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Telephone Function (AREA)
  • Computer Security & Cryptography (AREA)
  • Technology Law (AREA)
  • Circuit For Audible Band Transducer (AREA)

Abstract

Apparatus having corresponding methods comprise a microphone configured to produce audio; a hold control configured to select a connected selection or a hold selection; a processor configured to identify the audio produced during the connected selection as primary audio, and to identify the audio produced during the hold selection as secondary audio; and a transceiver configured to transmit the primary audio and the secondary audio.

Description

CROSS-REFERENCE TO RELATED APPLICATION
This application is a continuation-in-part of prior U.S. patent application Ser. No. 14/478,885, filed Sep. 5, 2014, and titled “Collection and Analysis of Muted Audio,” the entirety of which is incorporated by reference herein.
FIELD
The present disclosure relates generally to the field of audio processing. More particularly, the present disclosure relates to analysis of audio generated by a microphone.
BACKGROUND
This background section is provided for the purpose of generally describing the context of the disclosure. Work of the presently named inventor(s), to the extent the work is described in this background section, as well as aspects of the description that may not otherwise qualify as prior art at the time of filing, are neither expressly nor impliedly admitted as prior art against the present disclosure.
Currently most audio communication systems have a hold function controlled locally that prevents the remote party from hearing the local audio, and may also provide alternative audio such as music, announcements, and the like. When the hold function is active, audio generated by the microphone is not transmitted to the remote party.
In call centers, there are several reasons an agent may employ the hold function. The agent may be coughing or sneezing, and does not want the remote party to hear. The agent may be having difficulty handling a call, and so is asking questions of his co-workers. Or the agent may be doing things not related to his work.
In each of these examples, the behavior of the agent may indicate a problem. An ill agent may spread illness to others in the call center. An agent asking questions of his co-workers may need more training, or may have competency issues. Or an agent may not be providing the work desired.
Currently, these problems are generally detected by a supervisor observing the agents directly. This process costs time and resources that could be directed to more productive endeavors. An agent may be observed remotely by monitoring his calls, but such monitoring fails while the hold function is active.
SUMMARY
In general, in one aspect, an embodiment features an apparatus comprising: a microphone configured to produce audio; a hold control configured to select a connected selection or a hold selection; a processor configured to identify the audio produced during the connected selection as primary audio, and to identify the audio produced during the hold selection as secondary audio; and a transceiver configured to transmit the primary audio and the secondary audio.
Embodiments of the apparatus can include one or more of the following features. In some embodiments, the transceiver is further configured to transmit the primary audio over a first link, and to transmit the secondary audio over a second link. In some embodiments, the first link is an audio link; and the second link is a data link. In some embodiments, the first link is a Bluetooth Synchronous Connection Oriented (SCO) link; and the secondary link is a Bluetooth Asynchronous Connection-Less (ACL) link. In some embodiments, the transceiver comprises: a first transceiver configured to transmit the primary audio according to a first protocol; and a second transceiver configured to transmit the secondary audio according to a second protocol. Some embodiments comprise a memory configured to store the secondary audio prior to the transceiver transmitting the secondary audio. In some embodiments, the processor is further configured to packetize the primary audio and the secondary audio, and to mark at least one of (i) packets of the primary audio and (ii) packets of the secondary audio. Some embodiments comprise a headset.
In general, in one aspect, an embodiment features a method comprising: producing audio responsive to sound; determining a selection of a hold control configured to select a connected selection or a hold selection; identifying the audio produced during the connected selection as primary audio; identifying the audio produced during the hold selection as secondary audio; and transmitting the primary audio and the secondary audio.
Embodiments of the method can include one or more of the following features. Some embodiments comprise transmitting the primary audio over a first link; and transmitting the secondary audio over a second link. Some embodiments comprise transmitting the primary audio according to a first protocol; and transmitting the secondary audio according to a second protocol. Some embodiments comprise packetizing the primary audio and the secondary audio; and marking at least one of (i) packets of the primary audio and (ii) packets of the secondary audio.
In general, in one aspect, an embodiment features apparatus comprising: a receiver configured to receive audio produced by a headset, wherein the headset has a hold control configured to select a connected selection or a hold selection, and wherein the audio includes primary audio and secondary audio, wherein the primary audio is generated by a microphone of the headset during a connected selection, and wherein the secondary audio is generated by the microphone of the headset during the hold selection; and a switch configured to pass the primary audio to a communications channel, and to pass the secondary audio to an analytics engine.
Embodiments of the apparatus can include one or more of the following features. In some embodiments, the switch is further configured to pass the primary audio to the analytics engine. Some embodiments comprise the analytics engine. In some embodiments, the receiver is further configured to receive the primary audio over a first link, and to receive the secondary audio over a second link. In some embodiments, the first link is an audio link; and the secondary link is a data link. In some embodiments, the first link is a Bluetooth Synchronous Connection Oriented (SCO) link; and the secondary link is a Bluetooth Asynchronous Connection-Less (ACL) link. In some embodiments, the receiver comprises: a first receiver configured to receive the primary audio according to a first protocol; and a second receiver configured to receive the secondary audio according to a second protocol. In some embodiments, the audio comprises packets of the primary audio and packets of the secondary audio; at least one of (i) the packets of the primary audio and (ii) the packets of the secondary audio include marks; and the switch is further configured to distinguish the (i) the packets of the primary audio and (ii) the packets of the secondary audio based on the marks.
The details of one or more implementations are set forth in the accompanying drawings and the description below. Other features will be apparent from the description and drawings, and from the claims.
DESCRIPTION OF DRAWINGS
FIG. 1 shows elements of a call center audio system according to an embodiment where the primary and secondary audio are distinguished using packet marking.
FIG. 2 shows a process for the call center audio system of FIG. 1 according to one embodiment.
FIG. 3 shows elements of a call center audio system according to an embodiment where the primary and secondary audio are distinguished using different communication links.
FIG. 4 shows a process for the call center audio system of FIG. 3 according to one embodiment.
FIG. 5 shows elements of a call center audio system according to an embodiment where the primary and secondary audio are distinguished using different communication protocols.
FIG. 6 shows a process for the call center audio system of FIG. 5 according to one embodiment.
The leading digit(s) of each reference numeral used in this specification indicates the number of the drawing in which the reference numeral first appears.
DETAILED DESCRIPTION
Embodiments of the present disclosure provides collection of audio during hold for analysis and the like. In the described embodiments, sound received by a microphone when the call is on hold (that is, the hold function is active) is collected and analyzed. Sound received by the microphone when the call is not on hold (that is, while the hold function is not active) may be analyzed as well. Audio collected while the call is not on hold is referred to herein as “primary audio.” Audio collected while the call is on hold is referred to herein as “secondary audio.” In the described embodiments, various techniques are employed to distinguish the primary audio from the secondary audio. In some embodiments, packets of the primary audio and/or secondary audio may be marked, for example by setting flags in headers of the packets. In other embodiments, the primary audio and secondary audio may be transmitted over different links, using different protocols, and the like. Other features are contemplated as well.
Embodiments of the present disclosure are described in terms of an agent wearing a wireless headset in a call center. However, the techniques described herein are applicable to any audio device having a microphone, and in any environment.
FIG. 1 shows elements of a call center audio system 100 according to an embodiment where the primary and secondary audio are distinguished using packet marking. Although in the described embodiment elements of the call center audio system 100 are presented in one arrangement, other embodiments may feature other arrangements. For example, elements of the call center audio system 100 may be implemented in hardware, software, or combinations thereof. As another example, various elements of the call center audio system 100 may be implemented as one or more digital signal processors.
Referring to FIG. 1, the call center audio system 100 may include a headset 102 in communication with a host 104 over a wireless channel 106. The headset 102 may include a microphone (MIC) 108, a hold control (HOLD) 110, a processor (PROC) 112, and a transceiver (TRX) 114. The host 104 may include a transceiver (TRX) 116, a switch (SW) 118, an audio channel 120, and an analytics engine 122. While in the described embodiments, the analytics engine 122 may be part of the host 104, in other embodiments, the analytics engine 122 may not be part of the host 104, and may be located outside the call center audio system 100.
The hold control 110 may select either a connected selection or a hold selection. The hold control 110 may be user-operable, automatic, or both. A user-operable hold control 110 may be implemented as a button, slide switch, or the like. An automatic hold control 110 may automatically select the connected selection when donned, and may automatically select the hold selection when doffed.
The processor 112 may include an analog-to-digital converter, a digital signal processor, a packetizer, and the like. The wireless channel 106 may be a Bluetooth channel, a Digital Enhanced Cordless Telecommunications (DECT) channel, a Wi-Fi channel, or the like. The audio channel 120 may be any audio channel suitable for passing packets of primary audio to a remote party. The secondary audio may be routed directly to the host 104, or via another device such as a smart phone or computer.
FIG. 2 shows a process 200 for the call center audio system 100 of FIG. 1 according to one embodiment. Although in the described embodiments the elements of process 200 are presented in one arrangement, other embodiments may feature other arrangements. For example, in various embodiments, some or all of the elements of process 200 can be executed in a different order, concurrently, and the like. Also some elements of process 200 may not be performed, and may not be executed immediately after each other. In addition, some or all of the elements of process 200 can be performed automatically, that is, without human intervention.
Referring to FIG. 2, at 202, the microphone 108 may generate audio that represents sound received by the microphone 108. The processor 112 may process the audio. An analog-to-digital converter within the processor may convert the audio to digital audio. The processor 112 may packetize the digital audio. The hold control 110 may be operated by the agent to select either a connected selection or a hold selection. The selection may be communicated to the processor 112 by a hold signal 128. At 204, the processor 112 may determine the selection based on the hold signal 128.
The processor 112 may identify the audio produced during the connected selection as primary audio, and may identify the audio produced during the hold selection as secondary audio. In the present embodiment, at 206, the processor 112 may identify the audio by marking some or all of the packets in the audio stream. The processor 112 may mark the packets in accordance with the hold signal 128. The processor 112 may mark the packets of the digital audio when the hold signal 128 indicates the hold selection, when the hold signal 128 indicates the connected selection, or both. The processor 112 may mark the packets, for example, by setting or clearing a flag in the header of each packet, or in the header of a packet to indicate a transition between blocks of secondary and primary audio, and the like. The processor 112 may insert control packets transition between blocks of secondary and primary audio, and the like. At 208, the transceiver 114 of the headset 102 may transmit a signal representing the packets over the wireless channel 106.
At 210, the transceiver 116 of the host 104 may receive the signal representing the packets over the wireless channel 106. At 212, the switch 118 routes the packets according to the marks in the packets. In particular, the switch 118 routes the packets of primary audio to the audio channel 120, and routes the packets of secondary audio to the analytics engine 122 for analysis. In some embodiments, the switch 118 may also route some or all of the packets of primary audio to the analytics engine 122 for analysis.
FIG. 3 shows elements of a call center audio system 300 according to an embodiment where the primary and secondary audio are distinguished using different communication links. Although in the described embodiment elements of the call center audio system 300 are presented in one arrangement, other embodiments may feature other arrangements. For example, elements of the call center audio system 300 may be implemented in hardware, software, or combinations thereof. As another example, various elements of the call center audio system 300 may be implemented as one or more digital signal processors.
Referring to FIG. 3, the call center audio system 300 may include a headset 302 in communication with a host 304 over a wireless channel 306. The headset 302 may include a microphone (MIC) 308, a hold control (HOLD) 310, a processor (PROC) 312, a memory 324, and a transceiver (TRX) 314. The host 304 may include a transceiver (TRX) 316, an audio channel 320, and an analytics engine 322. While in the described embodiments, the analytics engine 322 may be part of the host 304, in other embodiments, the analytics engine 322 may not be part of the host 304, and may be located outside the call center audio system 300.
The hold control 310 may select either a connected selection or a hold selection. The hold control 310 may be user-operable, automatic, or both. A user-operable hold control 310 may be implemented as a button, slide switch, or the like. An automatic hold control 310 may automatically select the connected selection when donned, and may automatically select the hold selection when doffed.
The processor 312 may include an analog-to-digital converter, a digital signal processor, a packetizer, and the like. The wireless channel 306 may be a Bluetooth channel, a Digital Enhanced Cordless Telecommunications (DECT) channel, a Wi-Fi channel, or the like. The audio channel 320 may be any audio channel suitable for passing packets of primary audio to a remote party. The secondary audio may be routed directly to the host 304, or via another device such as a smart phone or computer.
FIG. 4 shows a process 400 for the call center audio system 300 of FIG. 3 according to one embodiment. Although in the described embodiments the elements of process 400 are presented in one arrangement, other embodiments may feature other arrangements. For example, in various embodiments, some or all of the elements of process 400 can be executed in a different order, concurrently, and the like. Also some elements of process 400 may not be performed, and may not be executed immediately after each other. In addition, some or all of the elements of process 400 can be performed automatically, that is, without human intervention.
Referring to FIG. 4, at 402, the microphone 308 may generate audio that represents sound received by the microphone 308. The processor 312 may process the audio. An analog-to-digital converter within the processor may convert the audio to digital audio. The processor 312 may packetize the digital audio. The hold control 310 may be operated by the agent to select either a connected selection or a hold selection. The selection may be communicated to the processor 312 by a hold signal 328. At 404, the processor 312 may determine the selection based on the hold signal 328.
The processor 312 may identify the audio produced during the connected selection as primary audio, and may identify the audio produced during the hold selection as secondary audio. In the present embodiment, the processor 312 may identify the audio by routing the primary audio to one link, and routing the secondary audio to another link. At 406, the processor 312 may route the packets of digital audio among multiple communication links in accordance with the hold signal 328. For example, the processor 312 may route the packets of primary audio to an audio link, and may route the packets of secondary audio to a data link. The audio link may be a Bluetooth Synchronous Connection Oriented (SCO) link. The data link may be a Bluetooth Asynchronous Connection-Less (ACL) link. However, other wireless protocols and links may be used.
At 408, the memory 324 may store the packets of the secondary audio before transmission to the host 304. In such embodiments, the data link need not be open continuously. At 410, the transceiver 314 of the headset 302 transmits one or more signals representing the packets over the wireless channel 306.
At 412, the transceiver 316 of the host 304 may receive the signal representing the packets over the wireless channel 306. At 414, the transceiver 316 may pass the packets according to the communication links. In particular, the transceiver 316 may route the packets of primary audio to the audio channel 320, and may route the packets of secondary audio to the analytics engine 322 for analysis. In some embodiments, the transceiver 316 may also route some or all of the packets of primary audio to the analytics engine 322 for analysis.
FIG. 5 shows elements of a call center audio system 500 according to an embodiment where the primary and secondary audio are distinguished using different communication protocols. Although in the described embodiment elements of the call center audio system 500 are presented in one arrangement, other embodiments may feature other arrangements. For example, elements of the call center audio system 500 may be implemented in hardware, software, or combinations thereof. As another example, various elements of the call center audio system 500 may be implemented as one or more digital signal processors.
Referring to FIG. 5, the call center audio system 500 may include a headset 502 in communication with a host 504 over wireless channels 506 and 546. The headset 502 may include a microphone (MIC) 508, a hold control (HOLD) 510, a processor (PROC) 512, a memory 524, transceivers (TRX) 514 and 534. The host 504 may include transceivers (TRX) 516 and 536, an audio channel 520, and an analytics engine 522. While in the described embodiments, the analytics engine 522 may be part of the host 504, in other embodiments, the analytics engine 522 may not be part of the host 504, and may be located outside the call center.
The hold control 510 may select either a connected selection or a hold selection. The hold control 510 may be user-operable, automatic, or both. A user-operable hold control 510 may be implemented as a button, slide switch, or the like. An automatic hold control 510 may automatically select the connected selection when donned, and may automatically select the hold selection when doffed.
The processor 512 may include an analog-to-digital converter, a digital signal processor, a packetizer, and the like. The wireless channels 506 and 546 may employ different wireless protocols, for example such as Bluetooth and Wi-Fi, respectively. However, any protocol may be used, for example such as Digital Enhanced Cordless Telecommunications (DECT), or the like. The audio channel 520 may be any audio channel suitable for passing the packets of primary audio to a remote party. The secondary audio may be routed directly to the host 504, or via another device such as a smart phone or computer.
FIG. 6 shows a process 600 for the call center audio system 500 of FIG. 5 according to one embodiment. Although in the described embodiments the elements of process 600 are presented in one arrangement, other embodiments may feature other arrangements. For example, in various embodiments, some or all of the elements of process 600 can be executed in a different order, concurrently, and the like. Also some elements of process 600 may not be performed, and may not be executed immediately after each other. In addition, some or all of the elements of process 600 can be performed automatically, that is, without human intervention.
Referring to FIG. 6, at 602, the microphone 508 may generate audio that represents sound received by the microphone 508. The processor 512 may process the audio. An analog-to-digital converter within the processor may convert the audio to digital audio. The processor 512 may packetize the digital audio. The hold control 510 may be operated by the agent to select either a connected selection or a hold selection. The selection may be communicated to the processor 512 by a hold signal 528. At 604, the processor 512 may determine the selection based on the hold signal 528.
The processor 512 may identify the audio produced during the connected selection as primary audio, and may identify the audio produced during the hold selection as secondary audio. In the present embodiment, the processor 512 may identify the audio by routing the primary audio to one transceiver, and routing the secondary audio to another transceiver. At 606, the processor 512 may route the packets of digital audio among multiple transceivers 514, 534 in accordance with the hold signal 528. For example, the processor 512 may route the packets of primary audio to one transceiver 514, and may route the packets of secondary audio to another transceiver 534.
At 608, the memory 524 may store the packets of the secondary audio before transmission to the host 504. In such embodiments, the data link need not be open continuously. At 610, the transceivers 514, 534 of the headset 502 transmit signals representing the packets over the respective wireless channel 506, 546.
At 612, the transceivers 516, 536 of the host 504 may receive the signals representing the packets over the respective wireless channels 506, 546. At 612, the transceiver 516 may pass the packets of primary audio to the audio channel 520, and the transceiver 536 may pass the packets of secondary audio to the analytics engine 522 for analysis. In some embodiments, the transceiver 516 may also route some or all of the packets of primary audio to the analytics engine 522 for analysis.
The analytics engines 122, 322, 522 described above may perform any sort of analysis on the secondary audio. The analytics engines 122, 322, 522 may identify coughs and sneezes in the secondary audio, keeping metrics as a potential indicator of illness of individual agents and groups of agents. The analytics engines 122, 322, 522 may detect questions, for example based on intonation, voice recognition, and the like, keeping metrics as a possible indicator of need for training of individual agents or groups of agents. The analytics engines 122, 322, 522 may monitor the agent's speech with hold on or off and make decisions on content, keeping metrics as indicators of time spent on work communications and personal communications. In all cases, a supervisor may be alerted when a metric threshold is exceeded, making it unnecessary for a supervisor to personally monitor calls or observe agents.
Various embodiments of the present disclosure can be implemented in digital electronic circuitry, or in computer hardware, firmware, software, or in combinations thereof. Embodiments of the present disclosure can be implemented in a computer program product tangibly embodied in a computer-readable storage device for execution by a programmable processor. The described processes can be performed by a programmable processor executing a program of instructions to perform functions by operating on input data and generating output. Embodiments of the present disclosure can be implemented in one or more computer programs that are executable on a programmable system including at least one programmable processor coupled to receive data and instructions from, and to transmit data and instructions to, a data storage system, at least one input device, and at least one output device. Each computer program can be implemented in a high-level procedural or object-oriented programming language, or in assembly or machine language if desired; and in any case, the language can be a compiled or interpreted language. Suitable processors include, by way of example, both general and special purpose microprocessors. Generally, processors receive instructions and data from a read-only memory and/or a random access memory. Generally, a computer includes one or more mass storage devices for storing data files. Such devices include magnetic disks, such as internal hard disks and removable disks, magneto-optical disks; optical disks, and solid-state disks. Storage devices suitable for tangibly embodying computer program instructions and data include all forms of non-volatile memory, including by way of example semiconductor memory devices, such as EPROM, EEPROM, and flash memory devices; magnetic disks such as internal hard disks and removable disks; magneto-optical disks; and CD-ROM disks. Any of the foregoing can be supplemented by, or incorporated in, ASICs (application-specific integrated circuits). As used herein, the term “module” may refer to any of the above implementations.
A number of implementations have been described. Nevertheless, various modifications may be made without departing from the scope of the disclosure. Accordingly, other implementations are within the scope of the following claims.

Claims (18)

What is claimed is:
1. An apparatus for audio communication between a local party and a remote party, comprising:
a microphone configured to produce audio comprising speech or sounds produced by the local party;
a hold control configured to select a connected selection or a hold selection;
a processor configured to identify the audio produced during the connected selection as primary audio, and to identify the audio produced during the hold selection as secondary audio, in use the secondary audio not being heard by the remote party;
a transceiver configured to transmit the primary audio and the secondary audio, wherein the processor is further configured to packetize the primary audio and the secondary audio, and to mark at least one of (i) packets of the primary audio and (ii) packets of the secondary audio such that the primary audio produced during the connected selection and the secondary audio produced during the hold selection can subsequently be distinguished from one another; and
an analytics engine configured to receive the primary audio and the secondary audio and to determine speech metrics therefrom.
2. The apparatus of claim 1, wherein:
the transceiver is further configured to transmit the primary audio over a first link, and to transmit the secondary audio over a second link.
3. The apparatus of claim 2, wherein:
the first link is an audio link; and
the second link is a data link.
4. The apparatus of claim 3, wherein:
the first link is a Bluetooth Synchronous Connection Oriented (SCO) link; and
the secondary link is a Bluetooth Asynchronous Connection-Less (ACL) link.
5. The apparatus of claim 1, wherein the transceiver comprises:
a first transceiver configured to transmit the primary audio according to a first protocol; and
a second transceiver configured to transmit the secondary audio according to a second protocol.
6. The apparatus of claim 1, further comprising:
a memory configured to store the secondary audio prior to the transceiver transmitting the secondary audio.
7. The apparatus of claim 1, further comprising:
a headset.
8. A method of capturing an audio communication between a local party and a remote party comprising:
producing audio comprising speech or sounds produced by the local party;
determining a selection of a hold control configured to select a connected selection or a hold selection, audio produced during the hold selection not being heard by the remote party;
identifying the audio produced during the connected selection as primary audio;
identifying the audio produced during the hold selection as secondary audio; and
transmitting both the primary audio and the secondary audio to an analytics engine; and
determining speech metrics from the primary audio and the secondary audio using the analytics engine.
9. The method of claim 8, further comprising:
transmitting the primary audio over a first link; and
transmitting the secondary audio over a second link.
10. The method of claim 8, further comprising:
transmitting the primary audio according to a first protocol; and
transmitting the secondary audio according to a second protocol.
11. The method of claim 8, further comprising:
packetizing the primary audio and the secondary audio; and
marking at least one of (i) packets of the primary audio and (ii) packets of the secondary audio.
12. An apparatus for audio communication between a local party and a remote party, comprising:
a receiver configured to receive audio produced by a headset, wherein the headset has a hold control configured to select a connected selection or a hold selection, and wherein the audio includes primary audio and secondary audio, wherein the primary audio is generated by a microphone of the headset during a connected selection, wherein the secondary audio is generated by the microphone of the headset during the hold selection, and wherein in use the secondary audio is not heard by the remote party; and
a switch configured to pass the primary audio to a communications channel, and to pass the secondary audio to an analytics engine configured to determine speech metrics from the secondary audio, wherein:
the audio comprises packets of the primary audio and packets of the secondary audio;
at least one of (i) the packets of the primary audio and (ii) the packets of the secondary audio include marks; and
the switch is further configured to distinguish the packets of the primary audio and the packets of the secondary audio from each other based on the marks.
13. The apparatus of claim 12, wherein:
the switch is further configured to pass the primary audio to the analytics engine.
14. The apparatus of claim 12, further comprising:
the analytics engine.
15. The apparatus of claim 12, wherein:
the receiver is further configured to receive the primary audio over a first link, and to receive the secondary audio over a second link.
16. The apparatus of claim 15, wherein:
the first link is an audio link; and
the secondary link is a data link.
17. The apparatus of claim 16, wherein:
the first link is a Bluetooth Synchronous Connection Oriented (SCO) link; and
the secondary link is a Bluetooth Asynchronous Connection-Less (ACL) link.
18. The apparatus of claim 12, wherein the receiver comprises:
a first receiver configured to receive the primary audio according to a first protocol; and
a second receiver configured to receive the secondary audio according to a second protocol.
US14/697,436 2014-09-05 2015-04-27 Collection and analysis of audio during hold Active US10142472B2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US14/697,436 US10142472B2 (en) 2014-09-05 2015-04-27 Collection and analysis of audio during hold

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US14/478,885 US10178473B2 (en) 2014-09-05 2014-09-05 Collection and analysis of muted audio
US14/697,436 US10142472B2 (en) 2014-09-05 2015-04-27 Collection and analysis of audio during hold

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US14/478,885 Continuation-In-Part US10178473B2 (en) 2014-09-05 2014-09-05 Collection and analysis of muted audio

Publications (2)

Publication Number Publication Date
US20160072949A1 US20160072949A1 (en) 2016-03-10
US10142472B2 true US10142472B2 (en) 2018-11-27

Family

ID=55438668

Family Applications (1)

Application Number Title Priority Date Filing Date
US14/697,436 Active US10142472B2 (en) 2014-09-05 2015-04-27 Collection and analysis of audio during hold

Country Status (1)

Country Link
US (1) US10142472B2 (en)

Families Citing this family (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10594861B2 (en) * 2017-09-28 2020-03-17 Plantronics, Inc. Forking transmit and receive call audio channels
US11510197B2 (en) 2019-02-07 2022-11-22 Plantronics, Inc. Systems and methods for managing wireless packet communications by assigning separate resources for sequential transmission attempts
US12501306B2 (en) 2020-03-30 2025-12-16 Hewlett-Packard Development Company, L.P. Systems and methods for managing an environment including both voice communications and streaming audio communications
US11219386B2 (en) 2020-06-15 2022-01-11 Cirrus Logic, Inc. Cough detection
US11134354B1 (en) 2020-06-15 2021-09-28 Cirrus Logic, Inc. Wear detection

Citations (21)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20030031327A1 (en) * 2001-08-10 2003-02-13 Ibm Corporation Method and apparatus for providing multiple output channels in a microphone
US6757361B2 (en) 1996-09-26 2004-06-29 Eyretel Limited Signal monitoring apparatus analyzing voice communication content
US6768722B1 (en) * 2000-06-23 2004-07-27 At&T Corp. Systems and methods for managing multiple communications
US20040172252A1 (en) 2003-02-28 2004-09-02 Palo Alto Research Center Incorporated Methods, apparatus, and products for identifying a conversation
US20070121824A1 (en) 2005-11-30 2007-05-31 International Business Machines Corporation System and method for call center agent quality assurance using biometric detection technologies
WO2007080517A2 (en) 2006-01-16 2007-07-19 Gregory Nathan Headset with voip capability for a cellular phone without voip capability
US20080013747A1 (en) * 2006-06-30 2008-01-17 Bao Tran Digital stethoscope and monitoring instrument
US20080167878A1 (en) 2007-01-08 2008-07-10 Motorola, Inc. Conversation outcome enhancement method and apparatus
US20080260169A1 (en) 2006-11-06 2008-10-23 Plantronics, Inc. Headset Derived Real Time Presence And Communication Systems And Methods
US20090292541A1 (en) 2008-05-25 2009-11-26 Nice Systems Ltd. Methods and apparatus for enhancing speech analytics
US20100324891A1 (en) * 2009-06-18 2010-12-23 Microsoft Corporation Mute control in audio endpoints
US20110028136A1 (en) * 2009-07-28 2011-02-03 Motorola, Inc. Method for simulating network hold
US20110208522A1 (en) 2010-02-21 2011-08-25 Nice Systems Ltd. Method and apparatus for detection of sentiment in automated transcriptions
US20120020348A1 (en) * 2010-07-21 2012-01-26 Qualcomm Incorporated Coexistence interface and arbitration for multiple radios sharing an antenna
US20120296642A1 (en) 2011-05-19 2012-11-22 Nice Systems Ltd. Method and appratus for temporal speech scoring
US20130028399A1 (en) 2011-07-28 2013-01-31 Tata Consultancy Services Limited System and method for monitoring a voice in real time
US20130208881A1 (en) 2012-02-13 2013-08-15 Tata Consultancy Services Limited System for Conversation Quality Monitoring of Call Center Conversation and a Method Thereof
US20130211567A1 (en) * 2010-10-12 2013-08-15 Armital Llc System and method for providing audio content associated with broadcasted multimedia and live entertainment events based on profiling information
US20140093091A1 (en) 2012-09-28 2014-04-03 Sorin V. Dusan System and method of detecting a user's voice activity using an accelerometer
US20150310877A1 (en) 2012-10-31 2015-10-29 Nec Corporation Conversation analysis device and conversation analysis method
US9225833B1 (en) * 2013-07-24 2015-12-29 Noble Systems Corporation Management system for using speech analytics to enhance contact center agent conformance

Patent Citations (21)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6757361B2 (en) 1996-09-26 2004-06-29 Eyretel Limited Signal monitoring apparatus analyzing voice communication content
US6768722B1 (en) * 2000-06-23 2004-07-27 At&T Corp. Systems and methods for managing multiple communications
US20030031327A1 (en) * 2001-08-10 2003-02-13 Ibm Corporation Method and apparatus for providing multiple output channels in a microphone
US20040172252A1 (en) 2003-02-28 2004-09-02 Palo Alto Research Center Incorporated Methods, apparatus, and products for identifying a conversation
US20070121824A1 (en) 2005-11-30 2007-05-31 International Business Machines Corporation System and method for call center agent quality assurance using biometric detection technologies
WO2007080517A2 (en) 2006-01-16 2007-07-19 Gregory Nathan Headset with voip capability for a cellular phone without voip capability
US20080013747A1 (en) * 2006-06-30 2008-01-17 Bao Tran Digital stethoscope and monitoring instrument
US20080260169A1 (en) 2006-11-06 2008-10-23 Plantronics, Inc. Headset Derived Real Time Presence And Communication Systems And Methods
US20080167878A1 (en) 2007-01-08 2008-07-10 Motorola, Inc. Conversation outcome enhancement method and apparatus
US20090292541A1 (en) 2008-05-25 2009-11-26 Nice Systems Ltd. Methods and apparatus for enhancing speech analytics
US20100324891A1 (en) * 2009-06-18 2010-12-23 Microsoft Corporation Mute control in audio endpoints
US20110028136A1 (en) * 2009-07-28 2011-02-03 Motorola, Inc. Method for simulating network hold
US20110208522A1 (en) 2010-02-21 2011-08-25 Nice Systems Ltd. Method and apparatus for detection of sentiment in automated transcriptions
US20120020348A1 (en) * 2010-07-21 2012-01-26 Qualcomm Incorporated Coexistence interface and arbitration for multiple radios sharing an antenna
US20130211567A1 (en) * 2010-10-12 2013-08-15 Armital Llc System and method for providing audio content associated with broadcasted multimedia and live entertainment events based on profiling information
US20120296642A1 (en) 2011-05-19 2012-11-22 Nice Systems Ltd. Method and appratus for temporal speech scoring
US20130028399A1 (en) 2011-07-28 2013-01-31 Tata Consultancy Services Limited System and method for monitoring a voice in real time
US20130208881A1 (en) 2012-02-13 2013-08-15 Tata Consultancy Services Limited System for Conversation Quality Monitoring of Call Center Conversation and a Method Thereof
US20140093091A1 (en) 2012-09-28 2014-04-03 Sorin V. Dusan System and method of detecting a user's voice activity using an accelerometer
US20150310877A1 (en) 2012-10-31 2015-10-29 Nec Corporation Conversation analysis device and conversation analysis method
US9225833B1 (en) * 2013-07-24 2015-12-29 Noble Systems Corporation Management system for using speech analytics to enhance contact center agent conformance

Non-Patent Citations (5)

* Cited by examiner, † Cited by third party
Title
International Search Report and Written Opinion dated Oct. 5, 2015, for PCT Application No. PCT/US2015/033603.
Pallotta et al., "Interaction Mining: the New Frontier of Call Center Analytics," found at URL http://www.researchgate.net/profile/Vincenzo_Pallotta/publication/265145028_Interaction_Mining_the_new_frontier_of_Call_Center_Analytics/links/542beb900cf27e39fa91bea2.pdf, Jan. 1, 2011.
Pandharipande et al. "A Language Indepedent Approach to Identify Problematic Conversations in Call Centers", Nov. 2013, ECTI Transactions on Computer and Information Technology vol. 7, No. 2. *
Pandharipande et al., "A Language Independent Approach to Identify Problematic Conversation in Call Centers," ECTI Transactions on Computer and Information Technology, 7(2):146-155.
Unknown, "Speech Analytics, Innovative Speech Technologies to Unveil Hidden Insights," found at URL http://www.nice.com/speech-analytics, on Sep. 11, 2014.

Also Published As

Publication number Publication date
US20160072949A1 (en) 2016-03-10

Similar Documents

Publication Publication Date Title
US10652652B2 (en) Collection and analysis of muted audio
US10142472B2 (en) Collection and analysis of audio during hold
US11496827B2 (en) Microphone natural speech capture voice dictation system and method
CN109246672B (en) Data transmission method, device, system and Bluetooth headset
US9363612B2 (en) Method for operating a hearing device and a hearing device
US9961446B2 (en) Earphone recognition method and apparatus, earphone control method and apparatus, and earphone
JP6402748B2 (en) Spoken dialogue apparatus and utterance control method
WO2018095035A1 (en) Earphone and speech recognition method therefor
US20160366528A1 (en) Communication system, audio server, and method for operating a communication system
US20200219515A1 (en) Audio signal processing for automatic transcription using ear-wearable device
CN106685459B (en) A wearable device operation control method and wearable device
JP2017511636A (en) Device and method for facilitating wireless communication based on an implicit user queue
CN110996308B (en) Sound playing device, control method thereof, control device thereof and readable storage medium
CN112887872B (en) Earphone voice instruction playing method, earphone and storage medium
CN108541370A (en) Audio output method, electronic device and storage medium
CN105847566A (en) Mobile terminal audio volume adjusting method and device
CN107852431A (en) Information processing device, information processing method and program
JP6476938B2 (en) Speech analysis apparatus, speech analysis system and program
CN107609371A (en) A kind of message prompt method and audio-frequence player device
JP2016005017A (en) Management device, conversation system, conversation management method and program
JP5862318B2 (en) Sound analysis apparatus, sound analysis system and program
US20240216179A1 (en) Systems and methods for sound processing in personal protective equipment
CN113196800A (en) Hybrid microphone for wireless headset
JP2024175374A (en) Call control device, call device, call control method, call method, and program
US12525217B2 (en) Audio communication between proximate devices

Legal Events

Date Code Title Description
AS Assignment

Owner name: PLANTRONICS, INC., CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:KANNAPPAN, KEN;ROSENER, DOUGLAS K;SIGNING DATES FROM 20150413 TO 20150416;REEL/FRAME:035505/0848

AS Assignment

Owner name: WELLS FARGO BANK, NATIONAL ASSOCIATION, NORTH CAROLINA

Free format text: SECURITY AGREEMENT;ASSIGNORS:PLANTRONICS, INC.;POLYCOM, INC.;REEL/FRAME:046491/0915

Effective date: 20180702

Owner name: WELLS FARGO BANK, NATIONAL ASSOCIATION, NORTH CARO

Free format text: SECURITY AGREEMENT;ASSIGNORS:PLANTRONICS, INC.;POLYCOM, INC.;REEL/FRAME:046491/0915

Effective date: 20180702

STCF Information on status: patent grant

Free format text: PATENTED CASE

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment: 4

AS Assignment

Owner name: POLYCOM, INC., CALIFORNIA

Free format text: RELEASE OF PATENT SECURITY INTERESTS;ASSIGNOR:WELLS FARGO BANK, NATIONAL ASSOCIATION;REEL/FRAME:061356/0366

Effective date: 20220829

Owner name: PLANTRONICS, INC., CALIFORNIA

Free format text: RELEASE OF PATENT SECURITY INTERESTS;ASSIGNOR:WELLS FARGO BANK, NATIONAL ASSOCIATION;REEL/FRAME:061356/0366

Effective date: 20220829

AS Assignment

Owner name: HEWLETT-PACKARD DEVELOPMENT COMPANY, L.P., TEXAS

Free format text: NUNC PRO TUNC ASSIGNMENT;ASSIGNOR:PLANTRONICS, INC.;REEL/FRAME:065549/0065

Effective date: 20231009