US20150103982A1 - Methods and systems for automatically providing an emergency service call handler with context specific emergency service protocols - Google Patents

Methods and systems for automatically providing an emergency service call handler with context specific emergency service protocols Download PDF

Info

Publication number
US20150103982A1
US20150103982A1 US14/548,082 US201414548082A US2015103982A1 US 20150103982 A1 US20150103982 A1 US 20150103982A1 US 201414548082 A US201414548082 A US 201414548082A US 2015103982 A1 US2015103982 A1 US 2015103982A1
Authority
US
United States
Prior art keywords
spoken
call handler
caller
emergency service
identified
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
US14/548,082
Other versions
US9350873B2 (en
Inventor
Stephen F. O'Conor
Richard Maw
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Synergem Technologies Inc
Original Assignee
Synergem Technologies Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Synergem Technologies Inc filed Critical Synergem Technologies Inc
Priority to US14/548,082 priority Critical patent/US9350873B2/en
Assigned to SYNERGEM TECHNOLOGIES INC reassignment SYNERGEM TECHNOLOGIES INC ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: MAW, RICHARD A, O'CONNOR, STEPHEN F
Publication of US20150103982A1 publication Critical patent/US20150103982A1/en
Application granted granted Critical
Publication of US9350873B2 publication Critical patent/US9350873B2/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M3/00Automatic or semi-automatic exchanges
    • H04M3/42Systems providing special services or facilities to subscribers
    • H04M3/50Centralised arrangements for answering calls; Centralised arrangements for recording messages for absent or busy subscribers ; Centralised arrangements for recording messages
    • H04M3/51Centralised call answering arrangements requiring operator intervention, e.g. call or contact centers for telemarketing
    • H04M3/5116Centralised call answering arrangements requiring operator intervention, e.g. call or contact centers for telemarketing for emergency applications
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M11/00Telephonic communication systems specially adapted for combination with other electrical systems
    • H04M11/04Telephonic communication systems specially adapted for combination with other electrical systems with alarm systems, e.g. fire, police or burglar alarm systems
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/08Speech classification or search
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M3/00Automatic or semi-automatic exchanges
    • H04M3/22Arrangements for supervision, monitoring or testing
    • H04M3/2281Call monitoring, e.g. for law enforcement purposes; Call tracing; Detection or prevention of malicious calls
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M2201/00Electronic components, circuits, software, systems or apparatus used in telephone systems
    • H04M2201/40Electronic components, circuits, software, systems or apparatus used in telephone systems using speech recognition
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M2203/00Aspects of automatic or semi-automatic exchanges
    • H04M2203/20Aspects of automatic or semi-automatic exchanges related to features of supplementary services
    • H04M2203/2038Call context notifications
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M2242/00Special services or facilities
    • H04M2242/04Special services or facilities for emergency applications

Definitions

  • PSAP public safety answering point
  • public safety access point is a call center responsible for answering calls to an emergency telephone number for police, firefighting, and ambulance services.
  • Trained emergency service call takers are typically responsible for obtaining relevant information from a caller and dispatching the appropriate emergency service resources to the appropriate location.
  • ESPs emergency service protocols
  • ESPs emergency service protocols
  • a caller tells the call taker someone is not breathing
  • an appropriate ESP may guide the call taker through giving the caller instructions on performing CPR or other basic first aid procedures.
  • Other protocols may be directed at how to obtain appropriate information from the caller. For example, if the call involves a bomb threat, an appropriate ESP may instruct the call taker to notify the bomb squad and fire department and give the call taker instructions on how to attempt to guide the conversation with the caller to obtain critical information.
  • Next Generation 9-1-1 (“NG9-1”) can be viewed as a system comprised of Emergency Services IP networks (“ESInets”), internet protocol (“IP”) based software services and application, and various databases and data management processes that are all interconnected to a public safety answering point (PSAP).
  • EESInets Emergency Services IP networks
  • IP internet protocol
  • PSAP public safety answering point
  • the NG9-1-1 system provides location-based routing to the appropriate emergency entity, such that a caller in need of help is automatically routed to the PSAP assigned to the caller's location.
  • NG9-1-1 also provides standardized interfaces for call and message services, processes all types of emergency calls including non-voice (multi-media) messages, acquires and integrates additional data useful to call routing and handling for appropriate emergency entities.
  • NG9-1-1 supports all legacy E9-1-1 features and functions and is intended to provide scalable solution for meeting current and emerging needs for emergency communication between callers and Public Safety entities.
  • the NG9-1-1 system architecture is defined by the National Emergency Number Association (“NENA”) i3 standard and supports end-to-end IP connectivity between a caller and a public safety answering point (PSAP).
  • NENA National Emergency Number Association
  • PSAP public safety answering point
  • the i3 standard defines an ESInet, which sits between various, non-emergency communications networks and one or more PSAPs, as well as the ESInet's various functional elements, such as a Location Information Server (LIS) and Location Validation Function (LVF), the Emergency Services Routing Proxy (ESRP) and Policy Routing Function (PRF) and the Emergency Call Routing Function and Location to Service Translation (LoST) protocol. All of these elements are designed to provide robust and secure communications between a variety of communications devices and emergency service providers.
  • LIS Location Information Server
  • LVF Location Validation Function
  • ESRP Emergency Services Routing Proxy
  • PRF Policy Routing Function
  • LoST Location to Service Translation
  • the i3 standard requires all calls presented to the ESInet from an originating network, such as a typical telecomunications service provider (“TSP”) network to use session initiation protocol (“SIP”) signaling to deliver the call and include the location with the call.
  • SIP is a signaling protocol used to start, change and end telephone and multimedia communication sessions over IP networks.
  • BCF Border Control Function
  • the first element inside the ESInet is the Emergency Services Routing Proxy (ESRP).
  • ESRP Emergency Services Routing Proxy
  • the ESRP receives the call, and passes this information to an Emergency Call Routing Function (ECRF), which determines the next hop in routing a call to the requested service.
  • ECRF Emergency Call Routing Function
  • the ECRF maps the call's location information and requested service (e.g. police, which may be routed to a city-operated PSAP or fire department, which may be routed to a county-operated PSAP) to an appropriate PSAP.
  • NENA In the event an ESInet is provisioned in an area before the regional TSPs and other originating networks or PSAPs are NG9-1-1 capable, NENA has defined a transition model.
  • the legacy E911 network has been replaced by the Emergency Services IP Network (ESInet) with all of the functional elements previously described, but on either end (originating network and/or PSAP) is a legacy environment.
  • ESInet Emergency Services IP Network
  • NG9-1-1 defines a legacy network gateway and a legacy PSAP gateway to convert the data to and from SIP messaging for transmission over the ESInet until such time as the originating networks and PSAPs become i3 capable.
  • a beneficial side effect of the transition to the NG9-1-1 environment is that all emergency service phone calls will be converted to digital data and stored for future review. This further enables new and advantageous information processing techniques to be applied to emergency service calls in real time in order to assist emergency service call takers in performing their jobs.
  • FIG. 1 is a schematic diagram depicting aspects of a non-limiting, exemplary computing architecture suitable for implementing at least some aspects and/or embodiments of the present systems and methods.
  • FIG. 2 is a functional block diagram depicting an emergency services communications network advantageously featuring aspects of the present methods and systems.
  • FIG. 3 is a functional block diagram depicting an emergency services communications network advantageously featuring additional aspects of the present methods and systems.
  • FIG. 4 is a flow chart depicting the operational steps of certain aspects of the present methods and systems.
  • Embodiments of the present methods and systems may be implemented by systems using one or more programmable digital computers.
  • Computer and computer systems in connection with embodiments of the invention may act, e.g., as workstations and/or servers, such as described below.
  • Digital voice and/or data networks such as may be used in connection with embodiments of the invention may also include components (e.g., routers, bridges, media gateways, etc.) with similar architectures, although they may be adapted, e.g., as known in the art, for their special purposes. Because of this commonality of architecture, such network components may be considered as computer systems and/or components of computer systems when consistent with the applicable context.
  • FIG. 1 depicts an example of one such computer system 100 , which includes at least one processor 110 , such as, e.g., an Intel or Advanced Micro Devices microprocessor, coupled to a communications channel or bus 112 .
  • the computer system 100 further includes at least one input device 114 such as, e.g., a keyboard, mouse, touch pad or screen, or other selection or pointing device, at least one output device 116 such as, e.g., an electronic display device, at least one communications interface 118 , at least one data storage device 120 such as a magnetic disk or an optical disk, and memory 122 such as ROM and RAM, each coupled to the communications channel 112 .
  • the communications interface 118 may be coupled to a network (not depicted) such as the Internet.
  • FIG. 1 Although the computer system 100 is shown in FIG. 1 to have only a single communications channel 112 , a person skilled in the relevant arts will recognize that a computer system may have multiple channels (not depicted), including for example one or more busses, and that such channels may be interconnected, e.g., by one or more bridges. In such a configuration, components depicted in FIG. 1 as connected by a single channel 112 may interoperate, and may thereby be considered to be coupled to one another, despite being directly connected to different communications channels.
  • data storage device 120 and memory 122 are depicted as different units, the data storage device 120 and memory 122 can be parts of the same unit or units, and that the functions of one can be shared in whole or in part by the other, e.g., as RAM disks, virtual memory, etc. It will also be appreciated that any particular computer may have multiple components of a given type, e.g., processors 110 , input devices 114 , communications interfaces 118 , etc.
  • the data storage device 120 may store instructions executable by one or more processors or kinds of processors 110 , data, or both. Some groups of instructions, possibly grouped with data, may make up one or more programs, which may include an operating system such as Microsoft Windows®, Linux®, Mac OS®, or Unix®. Other programs may be stored instead of or in addition to the operating system. It will be appreciated that a computer system may also be implemented on platforms and operating systems other than those mentioned.
  • Any operating system or other program, or any part of either, may be written using one or more programming languages such as, e.g., Java®, C, C++, C#, Visual Basic®, VB.NET®, Perl, Ruby, Python, or other programming languages, possibly using object oriented design and/or coding techniques.
  • programming languages such as, e.g., Java®, C, C++, C#, Visual Basic®, VB.NET®, Perl, Ruby, Python, or other programming languages, possibly using object oriented design and/or coding techniques.
  • the computer system 100 may also include additional components and/or systems, such as network connections, additional memory, additional processors, network interfaces, input/output busses, for example.
  • a computer-readable storage medium (CRSM) reader 136 such as, e.g., a magnetic disk drive, magneto-optical drive, optical disk drive, or flash drive, may be coupled to the communications channel 112 for reading from a CRSM 138 such as, e.g., a magnetic disk, a magneto-optical disk, an optical disk, or flash RAM.
  • CRSM computer-readable storage medium
  • one or more CRSM readers may be coupled to the rest of the computer system 100 , e.g., through a network interface (not depicted) or a communications interface 118 . In any such configuration, however, the computer system 100 may receive programs and/or data via the CRSM reader 136 .
  • the term “memory” herein is intended to include various types of suitable data storage media, whether permanent or temporary, including among other things the data storage device 120 , the memory 122 , and the CSRM 138 .
  • computer-readable storage medium and “computer-readable storage media” refer, respectively, to a medium and media capable of storing information. As such, both terms exclude transient propagating signals.
  • Two or more computer systems 100 may communicate, e.g., in one or more networks, via, e.g., their respective communications interfaces 118 and/or network interfaces (not depicted).
  • FIG. 2 depicts a communications system 200 , including an ESInet 204 connected to an origination network 208 and a PSAP 212 via BCFs 210 , 211 , suitable for use with the present methods and systems.
  • an emergency voice call 216 is routed to the PSAP 212 from the origination network 208 via ESInet 204 , the caller is connected to an emergency service call handler (not shown) via a call handling application 220 .
  • the call is also routed to a session recorder 224 for analysis, review and archival purposes.
  • an automated protocol selection function (APSF) 228 is provided. As call 216 is being recorded by session recorder 224 it is also input to the APSF 228 .
  • the APSF 228 may include a speech recognition element 232 , a comparison element 240 , a protocol selection element 244 , a keyword database 246 , and a protocol database 248 .
  • Speech recognition element 232 may monitor the digital data transmission that corresponds to the voice communication between the caller and the emergency services call taker and apply a speech recognition process to detect words and/or phrases being spoken by the caller.
  • the speech recognition element 232 may divide the caller's speech into segments, which may be on the order of magnitude of a hundredth of a second in duration and compare the segments to a set of known phonemes. The speech recognition element 232 may then perform a contextual phoneme analysis on each phoneme identified in the call to other phonemes in its temporal vicinity in order to determine the language being spoken and identify what word or phrase in that language the caller is using. This may advantageously occur substantially in real-time, as the caller is speaking.
  • Speech recognition solutions such as Microsoft Voice Command (Microsoft Corporation), Sonic Extractor (Digitial Syphon), LumenVox Speech Engine (LumenVox), Nuance Voice Control (Nuance Communications), VITO Voice2Go (Vito Technology), and Speereo Voice Translator (Speereo Software) are exemplary but non-limiting implementations of aspects of the speech recognition element 232 .
  • Comparison element 240 compares the word or phrase 252 identified by speech recognition element 232 to a set of known keywords and phrases stored in keyword database 246 . Each keyword and phrase in keyword database 246 is associated with one or more emergency service protocols 256 stored in protocol database 248 . If comparison element 240 detects a match between the spoken word or phrase 252 and one of the known keywords or phrases, it notifies the protocol selection element 244 . Protocol selection element 244 retrieves the appropriate emergency service protocol(s) 256 identified by the detected keyword or phrase and transmits the protocol 256 to the call handling application 220 where it is displayed to the emergency service call taker to assist him/her in handling the call.
  • Speech recognition element 232 will break this phrase down into the set of phonemes and, after running a contextual analysis, identify the individual words “Help,” “my,” “wife,” “isn't,” and “breathing.” These words are then passed to the comparison element 240 which may compare the individual words and sub-sets of words within the phrase to the known key words and phrases stored in keyword database 246 .
  • the one such known phrase may be “isn't breathing,” or variations thereof, and comparison element 240 will match that known phrase to the corresponding subset of words from the caller's statement.
  • the phrase “isn't breathing” may be linked to an emergency service protocol on CPR instructions.
  • the protocol selection element 244 may then retrieve the CPR protocol from protocol database 248 and display it for the emergency services call taker taking the call.
  • the emergency services call taker can seamlessly provide instructions to the caller without having to stop, mentally process the statement, and look up the appropriate protocol him/herself.
  • the protocol selection element may advantageously determine to provide the emergency service call taker with an infant specific CPR protocol.
  • Certain embodiments of the present methods and systems may advantageously filter the incoming call data to distinguish between foreground noise, i.e. the caller's voice, and background noise.
  • the background noise may be separately analyzed by a background analysis element 260 for relevant information, such as the presence of sirens, alarms, additional voices, gun shots, explosions, etc. Detection of such information may also factor into the determination of the appropriate protocol to provide to the emergency service call handler.
  • FIG. 3 depicts additional aspects of the present methods and systems, which may advantageously distinguish the caller's speech from the emergency service call taker's speech.
  • the PSAP is a legacy PSAP 312
  • the digital IP data 313 transmitted by the ESInet 303 will be converted back to a legacy format 304 by a Legacy PSAP gateway function 305 .
  • the legacy formatted data 304 may be reconverted to IP data 313 by an IP conversion function 348 prior to being input to the session recorder 324 .
  • the network transport path from the originating network to the legacy PSAP 312 is a legacy network (not depicted) rather than an ESInet
  • the data is delivered directly 360 to the IP conversion function 348 rather than the legacy PSAP gateway 305 .
  • further alternative aspects of the present methods and systems may, prior to analysis by a speech recognition element, input the call data into a parsing element 350 in order to distinguish voice-data packets originating from the PSAP's IP address from voice-data packets originating from other IP addresses, thereby distinguishing the caller's speech 354 from the call taker's speech 358 .
  • the call taker's voice may be discarded and the protocol selection process may proceed as described above with reference to FIG. 2 .
  • the separate instantiations of the speech recognition element 332 , 333 may separately process the caller's speech 354 and the call taker's speech 358 and separate instantiations of the comparison element 340 , 341 may compare identified words or phrases in the respective sides of the conversation to separate sets of keywords.
  • Such an aspect of the present method and system may, for example, give the call taker the ability to call up emergency service protocols using voice commands in the context of the conversation with the caller.
  • FIG. 4 depicts the steps of certain embodiments of the present methods and systems.
  • a caller initiates an emergency service phone call 404 via an originating communication network.
  • the origination network detects that the call is an emergency service call and routes the call to a local transport network, such as an ESInet or a legacy network, 408 .
  • the call is then routed to the appropriate PSAP 412 .
  • a two way communication channel is opened 414 between the caller and an emergency service call handler at the PSAP and the digital data corresponding to the voice communication between the caller and the call handler is monitored by a session recorder and an APSF 416 .
  • the APSF performs a speech recognition analysis on the voice communication 420 and identifies particular words and/or phrases being spoken by the caller 424 .
  • the identified words and/or phrases are then compared to a known set of keywords 428 . If a match is detected 450 , the APSF retrieves an emergency service protocol associated with the matched keyword 432 and provides the protocol to the emergency service call handler 436 .
  • present methods and systems described above can be implemented in locally installed software applications, for example, substantially running on computing hardware at the PSAP.
  • the present methods and systems could, however, also be implemented via a software as a service model, wherein the majority of computations are done at a remote location via network communications and the PSAP runs a ‘lightweight’ client application that predominately acts as an interface to the remote applications.

Abstract

Computer media and methods for providing emergency services protocols to an emergency service call taker are disclosed herein. A public safety answering point receives an emergency service phone call from a caller. The caller is placed in voice communication with an emergency call handler. The system monitors the voice communication between the caller and the emergency call handler. In response to detecting one or more known keyword in the voice communication, the system provides the emergency call handler with one or more defined protocols for guiding additional communications between the caller and the emergency call handler

Description

    CROSS-REFERENCE TO RELATED APPLICATION
  • This application claims the benefit of co-pending and commonly owned U.S. patent application Ser. No. 13/853,905 filed Mar. 29, 2013, herein incorporated by reference in its entirety for all purposes.
  • BACKGROUND
  • A public safety answering point (PSAP), sometimes called “public safety access point”, is a call center responsible for answering calls to an emergency telephone number for police, firefighting, and ambulance services. Trained emergency service call takers are typically responsible for obtaining relevant information from a caller and dispatching the appropriate emergency service resources to the appropriate location.
  • In order to assist the emergency call takers, many PSAPs utilize defined emergency service protocols (ESPs) for providing standard instructions for various types of common emergency service situations. For example, if a caller tells the call taker someone is not breathing, an appropriate ESP may guide the call taker through giving the caller instructions on performing CPR or other basic first aid procedures. Other protocols may be directed at how to obtain appropriate information from the caller. For example, if the call involves a bomb threat, an appropriate ESP may instruct the call taker to notify the bomb squad and fire department and give the call taker instructions on how to attempt to guide the conversation with the caller to obtain critical information. In conventional 9-1-1 systems, where the voice transmissions between a caller and call taker may be analog signals, and the call taker must know to recognize certain words or phrases spoken by a caller and look up any appropriate protocols. This additional step takes the call taker's attention away from dealing with the caller, and can cause delay and confusion which, in the context of an emergency services call, can lead directly to harm to individuals, damage to property, and/or additional, preventable consequences.
  • Advances in communication technology, specifically data connectivity and voice-over-IP technology, has led to the implementation of Enhanced-9-1-1 and Next Generation 9-1-1 standards. Broadly speaking, Next Generation 9-1-1 (“NG9-1-1”) can be viewed as a system comprised of Emergency Services IP networks (“ESInets”), internet protocol (“IP”) based software services and application, and various databases and data management processes that are all interconnected to a public safety answering point (PSAP). The NG9-1-1 system provides location-based routing to the appropriate emergency entity, such that a caller in need of help is automatically routed to the PSAP assigned to the caller's location. NG9-1-1 also provides standardized interfaces for call and message services, processes all types of emergency calls including non-voice (multi-media) messages, acquires and integrates additional data useful to call routing and handling for appropriate emergency entities. NG9-1-1 supports all legacy E9-1-1 features and functions and is intended to provide scalable solution for meeting current and emerging needs for emergency communication between callers and Public Safety entities.
  • The NG9-1-1 system architecture is defined by the National Emergency Number Association (“NENA”) i3 standard and supports end-to-end IP connectivity between a caller and a public safety answering point (PSAP). The i3 standard defines an ESInet, which sits between various, non-emergency communications networks and one or more PSAPs, as well as the ESInet's various functional elements, such as a Location Information Server (LIS) and Location Validation Function (LVF), the Emergency Services Routing Proxy (ESRP) and Policy Routing Function (PRF) and the Emergency Call Routing Function and Location to Service Translation (LoST) protocol. All of these elements are designed to provide robust and secure communications between a variety of communications devices and emergency service providers.
  • The i3 standard requires all calls presented to the ESInet from an originating network, such as a typical telecomunications service provider (“TSP”) network to use session initiation protocol (“SIP”) signaling to deliver the call and include the location with the call. SIP is a signaling protocol used to start, change and end telephone and multimedia communication sessions over IP networks. Upon reaching the ESInet, call traffic encounters the Border Control Function (BCF) which sits between external networks and the ESInet. An emergency service call, with location information, enters the ESInet through the BCF. After passing through the BCF, the first element inside the ESInet is the Emergency Services Routing Proxy (ESRP). The ESRP receives the call, and passes this information to an Emergency Call Routing Function (ECRF), which determines the next hop in routing a call to the requested service. The ECRF maps the call's location information and requested service (e.g. police, which may be routed to a city-operated PSAP or fire department, which may be routed to a county-operated PSAP) to an appropriate PSAP.
  • In the event an ESInet is provisioned in an area before the regional TSPs and other originating networks or PSAPs are NG9-1-1 capable, NENA has defined a transition model. In this case, the legacy E911 network has been replaced by the Emergency Services IP Network (ESInet) with all of the functional elements previously described, but on either end (originating network and/or PSAP) is a legacy environment. To provide connectivity to both the legacy networks and the legacy PSAPs, NG9-1-1 defines a legacy network gateway and a legacy PSAP gateway to convert the data to and from SIP messaging for transmission over the ESInet until such time as the originating networks and PSAPs become i3 capable.
  • A beneficial side effect of the transition to the NG9-1-1 environment is that all emergency service phone calls will be converted to digital data and stored for future review. This further enables new and advantageous information processing techniques to be applied to emergency service calls in real time in order to assist emergency service call takers in performing their jobs.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • Various embodiments in accordance with the present disclosure will be described with reference to the drawings, in which:
  • FIG. 1 is a schematic diagram depicting aspects of a non-limiting, exemplary computing architecture suitable for implementing at least some aspects and/or embodiments of the present systems and methods.
  • FIG. 2 is a functional block diagram depicting an emergency services communications network advantageously featuring aspects of the present methods and systems.
  • FIG. 3 is a functional block diagram depicting an emergency services communications network advantageously featuring additional aspects of the present methods and systems.
  • FIG. 4 is a flow chart depicting the operational steps of certain aspects of the present methods and systems.
  • DETAILED DESCRIPTION
  • This description discusses various illustrative embodiments of the present methods and systems for monitoring the content of an emergency service phone call and providing a call handler with context specific protocols (“the present methods and systems”) with reference to the accompanying drawings in order to provide a person having ordinary skill in the relevant art with a full, clear, and concise description of the subject matter defined by the claims which follow, and to enable such a person to appreciate and understand how to make and use the same. However, this description should not be read to limit the scope of the claimed subject matter, nor does the presence of an embodiment in this description imply any preference of the described embodiment over any other embodiment, unless such a preference is explicitly identified herein. It is the claims, not this description or other sections of this document or the accompanying drawings, which define the scope of the subject matter to which the inventor and/or the inventor's assignee(s) claim exclusive rights.
  • Embodiments of the present methods and systems may be implemented by systems using one or more programmable digital computers. Computer and computer systems in connection with embodiments of the invention may act, e.g., as workstations and/or servers, such as described below. Digital voice and/or data networks such as may be used in connection with embodiments of the invention may also include components (e.g., routers, bridges, media gateways, etc.) with similar architectures, although they may be adapted, e.g., as known in the art, for their special purposes. Because of this commonality of architecture, such network components may be considered as computer systems and/or components of computer systems when consistent with the applicable context.
  • FIG. 1 depicts an example of one such computer system 100, which includes at least one processor 110, such as, e.g., an Intel or Advanced Micro Devices microprocessor, coupled to a communications channel or bus 112. The computer system 100 further includes at least one input device 114 such as, e.g., a keyboard, mouse, touch pad or screen, or other selection or pointing device, at least one output device 116 such as, e.g., an electronic display device, at least one communications interface 118, at least one data storage device 120 such as a magnetic disk or an optical disk, and memory 122 such as ROM and RAM, each coupled to the communications channel 112. The communications interface 118 may be coupled to a network (not depicted) such as the Internet.
  • Although the computer system 100 is shown in FIG. 1 to have only a single communications channel 112, a person skilled in the relevant arts will recognize that a computer system may have multiple channels (not depicted), including for example one or more busses, and that such channels may be interconnected, e.g., by one or more bridges. In such a configuration, components depicted in FIG. 1 as connected by a single channel 112 may interoperate, and may thereby be considered to be coupled to one another, despite being directly connected to different communications channels.
  • One skilled in the art will recognize that, although the data storage device 120 and memory 122 are depicted as different units, the data storage device 120 and memory 122 can be parts of the same unit or units, and that the functions of one can be shared in whole or in part by the other, e.g., as RAM disks, virtual memory, etc. It will also be appreciated that any particular computer may have multiple components of a given type, e.g., processors 110, input devices 114, communications interfaces 118, etc.
  • The data storage device 120 (FIG. 1) and/or memory 122 may store instructions executable by one or more processors or kinds of processors 110, data, or both. Some groups of instructions, possibly grouped with data, may make up one or more programs, which may include an operating system such as Microsoft Windows®, Linux®, Mac OS®, or Unix®. Other programs may be stored instead of or in addition to the operating system. It will be appreciated that a computer system may also be implemented on platforms and operating systems other than those mentioned. Any operating system or other program, or any part of either, may be written using one or more programming languages such as, e.g., Java®, C, C++, C#, Visual Basic®, VB.NET®, Perl, Ruby, Python, or other programming languages, possibly using object oriented design and/or coding techniques.
  • One skilled in the art will recognize that the computer system 100 (FIG. 1) may also include additional components and/or systems, such as network connections, additional memory, additional processors, network interfaces, input/output busses, for example. One skilled in the art will also recognize that the programs and data may be received by and stored in the system in alternative ways. For example, a computer-readable storage medium (CRSM) reader 136, such as, e.g., a magnetic disk drive, magneto-optical drive, optical disk drive, or flash drive, may be coupled to the communications channel 112 for reading from a CRSM 138 such as, e.g., a magnetic disk, a magneto-optical disk, an optical disk, or flash RAM. Alternatively, one or more CRSM readers may be coupled to the rest of the computer system 100, e.g., through a network interface (not depicted) or a communications interface 118. In any such configuration, however, the computer system 100 may receive programs and/or data via the CRSM reader 136. Further, it will be appreciated that the term “memory” herein is intended to include various types of suitable data storage media, whether permanent or temporary, including among other things the data storage device 120, the memory 122, and the CSRM 138.
  • The terms “computer-readable storage medium” and “computer-readable storage media” refer, respectively, to a medium and media capable of storing information. As such, both terms exclude transient propagating signals.
  • Two or more computer systems 100 (FIG. 1) may communicate, e.g., in one or more networks, via, e.g., their respective communications interfaces 118 and/or network interfaces (not depicted).
  • FIG. 2 depicts a communications system 200, including an ESInet 204 connected to an origination network 208 and a PSAP 212 via BCFs 210, 211, suitable for use with the present methods and systems. When an emergency voice call 216 is routed to the PSAP 212 from the origination network 208 via ESInet 204, the caller is connected to an emergency service call handler (not shown) via a call handling application 220. The call is also routed to a session recorder 224 for analysis, review and archival purposes.
  • In accordance with the present methods and systems, an automated protocol selection function (APSF) 228 is provided. As call 216 is being recorded by session recorder 224 it is also input to the APSF 228. The APSF 228 may include a speech recognition element 232, a comparison element 240, a protocol selection element 244, a keyword database 246, and a protocol database 248. Speech recognition element 232 may monitor the digital data transmission that corresponds to the voice communication between the caller and the emergency services call taker and apply a speech recognition process to detect words and/or phrases being spoken by the caller. For example, the speech recognition element 232 may divide the caller's speech into segments, which may be on the order of magnitude of a hundredth of a second in duration and compare the segments to a set of known phonemes. The speech recognition element 232 may then perform a contextual phoneme analysis on each phoneme identified in the call to other phonemes in its temporal vicinity in order to determine the language being spoken and identify what word or phrase in that language the caller is using. This may advantageously occur substantially in real-time, as the caller is speaking. Commercially available speech recognition solutions such as Microsoft Voice Command (Microsoft Corporation), Sonic Extractor (Digitial Syphon), LumenVox Speech Engine (LumenVox), Nuance Voice Control (Nuance Communications), VITO Voice2Go (Vito Technology), and Speereo Voice Translator (Speereo Software) are exemplary but non-limiting implementations of aspects of the speech recognition element 232.
  • After a word or phrase 252 is identified by the speech recognition element 232 it is passed to comparison element 240. Comparison element 240 compares the word or phrase 252 identified by speech recognition element 232 to a set of known keywords and phrases stored in keyword database 246. Each keyword and phrase in keyword database 246 is associated with one or more emergency service protocols 256 stored in protocol database 248. If comparison element 240 detects a match between the spoken word or phrase 252 and one of the known keywords or phrases, it notifies the protocol selection element 244. Protocol selection element 244 retrieves the appropriate emergency service protocol(s) 256 identified by the detected keyword or phrase and transmits the protocol 256 to the call handling application 220 where it is displayed to the emergency service call taker to assist him/her in handling the call.
  • For example, a caller may state, “Help, my wife isn't breathing!” Speech recognition element 232 will break this phrase down into the set of phonemes and, after running a contextual analysis, identify the individual words “Help,” “my,” “wife,” “isn't,” and “breathing.” These words are then passed to the comparison element 240 which may compare the individual words and sub-sets of words within the phrase to the known key words and phrases stored in keyword database 246. The one such known phrase may be “isn't breathing,” or variations thereof, and comparison element 240 will match that known phrase to the corresponding subset of words from the caller's statement. The phrase “isn't breathing” may be linked to an emergency service protocol on CPR instructions. The protocol selection element 244 may then retrieve the CPR protocol from protocol database 248 and display it for the emergency services call taker taking the call. Thus, the emergency services call taker can seamlessly provide instructions to the caller without having to stop, mentally process the statement, and look up the appropriate protocol him/herself.
  • If, however, the caller states, “Help, my baby isn't breathing!” the word “baby” may be detected in addition to “isn't breathing” and the protocol selection element may advantageously determine to provide the emergency service call taker with an infant specific CPR protocol.
  • Certain embodiments of the present methods and systems may advantageously filter the incoming call data to distinguish between foreground noise, i.e. the caller's voice, and background noise. The background noise may be separately analyzed by a background analysis element 260 for relevant information, such as the presence of sirens, alarms, additional voices, gun shots, explosions, etc. Detection of such information may also factor into the determination of the appropriate protocol to provide to the emergency service call handler.
  • FIG. 3 depicts additional aspects of the present methods and systems, which may advantageously distinguish the caller's speech from the emergency service call taker's speech. For example, if the PSAP is a legacy PSAP 312, the digital IP data 313 transmitted by the ESInet 303 will be converted back to a legacy format 304 by a Legacy PSAP gateway function 305. In order to provide the functionality of the present methods and systems, the legacy formatted data 304 may be reconverted to IP data 313 by an IP conversion function 348 prior to being input to the session recorder 324. If the network transport path from the originating network to the legacy PSAP 312 is a legacy network (not depicted) rather than an ESInet, the data is delivered directly 360 to the IP conversion function 348 rather than the legacy PSAP gateway 305.
  • Still referring to FIG. 3, further alternative aspects of the present methods and systems may, prior to analysis by a speech recognition element, input the call data into a parsing element 350 in order to distinguish voice-data packets originating from the PSAP's IP address from voice-data packets originating from other IP addresses, thereby distinguishing the caller's speech 354 from the call taker's speech 358. In certain embodiments, the call taker's voice may be discarded and the protocol selection process may proceed as described above with reference to FIG. 2. Alternately, the separate instantiations of the speech recognition element 332, 333 may separately process the caller's speech 354 and the call taker's speech 358 and separate instantiations of the comparison element 340, 341 may compare identified words or phrases in the respective sides of the conversation to separate sets of keywords. Such an aspect of the present method and system may, for example, give the call taker the ability to call up emergency service protocols using voice commands in the context of the conversation with the caller.
  • FIG. 4 depicts the steps of certain embodiments of the present methods and systems. A caller initiates an emergency service phone call 404 via an originating communication network. The origination network detects that the call is an emergency service call and routes the call to a local transport network, such as an ESInet or a legacy network, 408. The call is then routed to the appropriate PSAP 412. A two way communication channel is opened 414 between the caller and an emergency service call handler at the PSAP and the digital data corresponding to the voice communication between the caller and the call handler is monitored by a session recorder and an APSF 416. The APSF performs a speech recognition analysis on the voice communication 420 and identifies particular words and/or phrases being spoken by the caller 424. The identified words and/or phrases are then compared to a known set of keywords 428. If a match is detected 450, the APSF retrieves an emergency service protocol associated with the matched keyword 432 and provides the protocol to the emergency service call handler 436.
  • It should be understood that the present methods and systems described above can be implemented in locally installed software applications, for example, substantially running on computing hardware at the PSAP. The present methods and systems could, however, also be implemented via a software as a service model, wherein the majority of computations are done at a remote location via network communications and the PSAP runs a ‘lightweight’ client application that predominately acts as an interface to the remote applications.
  • Exemplary embodiments of the present methods and systems have been described in detail above and in the accompanying figures for illustrative purposes. However, the scope of the present methods and systems are defined by the claims below and is not limited to the embodiments described above or depicted in the figures. Embodiments differing from those described and shown herein, but still within the scope of the defined methods and systems are envisioned by the inventors and will be apparent to persons having ordinary skill in the relevant art in view of this specification as a whole. The inventors intend for the defined methods and systems to be practiced other than as explicitly described herein. Accordingly, the defined methods and systems encompass all modifications and equivalents of the subject matter as permitted by applicable law.

Claims (20)

That which is claimed is:
1. A computer readable medium having computer-executable instructions stored thereon which, when executed by a computer, will cause the computer to perform a process for handling emergency services calls comprising the steps of:
monitoring digital data corresponding to a voice communication between a caller and an emergency service call handler for a digital representation of a known keyword;
responsive to detection of said digital representation of said known keyword, providing said emergency call handler with a protocol for guiding additional communications between said caller and said emergency call handler;
distinguishing between voice data corresponding to words spoken by said caller and voice data corresponding to said emergency service call handler; and
selecting a first set of instructions if said known keyword is identified as being spoken by said caller and selects a second set of instructions if said known keyword is identified as being spoken by said emergency service call handler.
2. The computer readable medium of claim 1, wherein:
the step of monitoring digital data corresponding to a voice communication between a caller and an emergency service call handler for a digital representation of a known keyword further comprises the steps of:
performing a speech recognition analysis on said digital data to identify words being spoken during said voice communication, and
comparing a set of known keywords to at least one word of the words identified in the step of performing a speech recognition analysis on said digital data to identify words being spoken during said voice communication; and
the step of providing said emergency call handler with a protocol for guiding additional communications between said caller and said emergency call handler further comprises the steps of:
selecting said protocol from a set of pre-defined protocols, and
displaying said protocol to said emergency service call handler via a call handling application.
3. The computer readable medium of claim 1, wherein:
the step of monitoring digital data corresponding to a voice communication between a caller and an emergency service call handler for a digital representation of a known keyword further comprises monitoring digital data corresponding to a voice communication between a caller and an emergency service call handler for a digital representation of a known key phrase;
the step of responsive to detection of said digital representation of said known keyword, providing said emergency call handler with a protocol for guiding additional communications between said caller and said emergency call handler further comprises, responsive to detection of said digital representation of said known key phrase, providing said emergency call handler with a protocol for guiding additional communications between said caller and said emergency call handler; and
the step of selecting a first set of instructions if said known keyword is identified as being spoken by said caller and selects a second set of instructions if said known keyword is identified as being spoken by said emergency service call handler further comprises selecting a first set of instructions if said known key phrase is identified as being spoken by said caller and selects a second set of instructions if said known key phrase is identified as being spoken by said emergency service call handler.
4. The computer readable medium of claim 1, wherein:
the step of selecting a first set of instructions if said known keyword is identified as being spoken by said caller and selects a second set of instructions if said known keyword is identified as being spoken by said emergency service call handler further includes giving priority to selecting the second set of instructions if said known keyword is identified as being spoken by said emergency service call handler.
5. The computer readable medium of claim 1, wherein:
said digital data comprises a string of digital data divided into packets, said packets include header information; and
said step of distinguishing between voice data corresponding to words spoken by said caller and voice data corresponding to said emergency service call handler further comprises comparing said header information to an IP address corresponding to said emergency service call handler to distinguish voice data packets spoken by said emergency service call handler.
6. The computer readable medium of claim 1, the method further including the step of:
comparing the digital data stream of said voice communication to a set of known background sounds; and
the step of selecting a first set of instructions if said known keyword is identified as being spoken by said caller and selects a second set of instructions if said known keyword is identified as being spoken by said emergency service call handler further includes:
selecting the first set of instructions based on said known keyword spoken by said caller and said identified background sounds, and
selecting the second set of instructions based on said known keyword spoken by said emergency service call handler and said identified background sounds.
7. The computer readable medium of claim 6, wherein the known keyword further comprises a known key phrase and the set of known keywords further includes known key phrases.
8. A computer readable medium having computer-executable instructions stored thereon which, when executed by a computer, will cause the computer to perform a process for monitoring emergency service calls comprising the steps of:
monitoring digital data corresponding to a voice communication between a caller and an emergency service call handler;
receiving said digital data stream and generating background text data, said background text data corresponding to a textual representation of sounds other than spoken words transmitted over said voice communication channel;
detecting a match between said background text data and a known background sound identifier; and
selecting a set of instructions in response to a match between said background text data and said known background sound identifier.
9. The computer readable medium of claim 8, the process further including the step of displaying the background text data to said emergency service call handler.
10. A computer readable medium having computer-executable instructions stored thereon which, when executed by a computer, will cause the computer to perform a process for emergency call handling comprising the steps of:
receiving a digital data stream corresponding to a voice communication between a caller and an emergency service call handler;
distinguishing between voice data corresponding to words spoken by said caller and voice data corresponding to said emergency service call handler; and
selecting a first set of instructions if a known word is identified as being spoken by said caller and selecting a second set of instructions if the known word is identified as being spoken by said emergency service call handler.
11. The computer readable medium of claim 10, wherein the step of distinguishing between voice data corresponding to words spoken by said caller and voice data corresponding to said emergency service call handler includes comparing said header information to an IP address corresponding to said emergency service call handler to distinguish voice data packets spoken by said emergency service call handler.
12. The computer readable medium of claim 10, wherein the step of selecting a first set of instructions if a known word is identified as being spoken by said caller and selects a second set of instructions if the known word is identified as being spoken by said emergency service call handler further comprises selecting a first set of instructions if a known phrase is identified as being spoken by said caller and selecting a second set of instructions if the known phrase is identified as being spoken by said emergency service call handler.
13. The computer readable medium of claim 10, wherein the step of selecting a first set of instructions if a known word is identified as being spoken by said caller and selecting a second set of instructions if the known word is identified as being spoken by said emergency service call handler further includes giving priority to selecting the second set of instructions if the known word is identified as being spoken by said emergency service call handler.
14. The computer readable medium of claim 10, the process further including the step of:
comparing the digital data stream of said voice communication to a set of known background sounds; and
the step of selecting a first set of instructions if said known keyword is identified as being spoken by said caller and selects a second set of instructions if said known keyword is identified as being spoken by said emergency service call handler further includes:
selecting the first set of instructions based on said known keyword spoken by said caller and said identified background sounds, and
selecting the second set of instructions based on said known keyword spoken by said emergency service call handler and said identified background sounds.
15. A computer readable medium having computer-executable instructions stored thereon which, when executed by a computer, will cause the computer to perform a process for emergency call handling comprising the steps of:
monitoring digital data corresponding to a voice communication between a caller and an emergency service call handler;
comparing the digital data stream of said voice communication to a set of known background sounds; and
in the event one or more of said known background sounds are identified in said voice communication, selecting a set of instructions for presentation to said emergency call handler, said set of instructions being related to said identified background sounds.
16. A method for handling emergency services calls comprising the steps of:
monitoring digital data corresponding to a voice communication between a caller and an emergency service call handler for a digital representation of a known key phrase;
responsive to detection of said digital representation of said known key phrase, providing said emergency call handler with a protocol for guiding additional communications between said caller and said emergency call handler;
distinguishing between voice data corresponding to words spoken by said caller and voice data corresponding to said emergency service call handler; and
selecting a first set of instructions if said known key phrase is identified as being spoken by said caller and selects a second set of instructions if said known key phrase is identified as being spoken by said emergency service call handler.
17. The method of claim 16, wherein:
the step of monitoring digital data corresponding to a voice communication between a caller and an emergency service call handler for a digital representation of a known key phrase further comprises the steps of:
performing a speech recognition analysis on said digital data to identify words being spoken during said voice communication, and
comparing a set of known key phrases to the words identified in the step of performing a speech recognition analysis on said digital data to identify words being spoken during said voice communication; and
the step of providing said emergency call handler with a protocol for guiding additional communications between said caller and said emergency call handler further comprises the steps of:
selecting said protocol from a set of pre-defined protocols, and
displaying said protocol to said emergency service call handler via a call handling application.
18. The method of claim 16, wherein:
the step of selecting a first set of instructions if said known key phrase is identified as being spoken by said caller and selects a second set of instructions if said known key phrase is identified as being spoken by said emergency service call handler further includes giving priority to selecting the second set of instructions if said known key phrase is identified as being spoken by said emergency service call handler.
19. The method of claim 16, wherein:
said digital data comprises a string of digital data divided into packets, said packets include header information; and
said step of distinguishing between voice data corresponding to words spoken by said caller and voice data corresponding to said emergency service call handler further comprises comparing said header information to an IP address corresponding to said emergency service call handler to distinguish voice data packets spoken by said emergency service call handler.
20. The method of claim 16, the process further including the step of:
comparing the digital data stream of said voice communication to a set of known background sounds; and
the step of selecting a first set of instructions if said known key phrase is identified as being spoken by said caller and selects a second set of instructions if said known key phrase is identified as being spoken by said emergency service call handler further includes:
selecting the first set of instructions based on said known key phrase spoken by said caller and said identified background sounds, and
selecting the second set of instructions based on said known key phrase spoken by said emergency service call handler and said identified background sounds.
US14/548,082 2013-03-29 2014-11-19 Methods and systems for automatically providing an emergency service call handler with context specific emergency service protocols Active US9350873B2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US14/548,082 US9350873B2 (en) 2013-03-29 2014-11-19 Methods and systems for automatically providing an emergency service call handler with context specific emergency service protocols

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US13/853,905 US8908837B2 (en) 2013-03-29 2013-03-29 Methods and systems for automatically providing an emergency service call handler with context specific emergency service protocols
US14/548,082 US9350873B2 (en) 2013-03-29 2014-11-19 Methods and systems for automatically providing an emergency service call handler with context specific emergency service protocols

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US13/853,905 Continuation US8908837B2 (en) 2013-03-29 2013-03-29 Methods and systems for automatically providing an emergency service call handler with context specific emergency service protocols

Publications (2)

Publication Number Publication Date
US20150103982A1 true US20150103982A1 (en) 2015-04-16
US9350873B2 US9350873B2 (en) 2016-05-24

Family

ID=51620864

Family Applications (2)

Application Number Title Priority Date Filing Date
US13/853,905 Active US8908837B2 (en) 2013-03-29 2013-03-29 Methods and systems for automatically providing an emergency service call handler with context specific emergency service protocols
US14/548,082 Active US9350873B2 (en) 2013-03-29 2014-11-19 Methods and systems for automatically providing an emergency service call handler with context specific emergency service protocols

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US13/853,905 Active US8908837B2 (en) 2013-03-29 2013-03-29 Methods and systems for automatically providing an emergency service call handler with context specific emergency service protocols

Country Status (1)

Country Link
US (2) US8908837B2 (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20170229125A1 (en) * 2016-02-05 2017-08-10 Honeywell International Inc. Systems and methods for contacting emergency personnel via voice recognition
WO2019045963A1 (en) * 2017-08-30 2019-03-07 T-Mobile Usa, Inc. Message transcription for emergency call prioritization
US10269372B1 (en) * 2015-09-24 2019-04-23 United Services Automobile Association (Usaa) System for sound analysis and recognition
US10616393B2 (en) * 2017-06-21 2020-04-07 Guangdong Oppo Mobile Telecommunications Corp., Ltd. Method for call control and related products

Families Citing this family (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9860722B1 (en) * 2014-01-30 2018-01-02 Sprint Communications Company L.P. Converting location information into public safety answering point (PSAP) signaling
US9799206B1 (en) * 2015-01-05 2017-10-24 Brenda Michelle Wilson Van Horn Method for automating emergency distress signals from a networked peripheral device
CN105045919B (en) * 2015-08-24 2019-08-16 北京云知声信息技术有限公司 A kind of information output method and device
US10372755B2 (en) 2015-09-23 2019-08-06 Motorola Solutions, Inc. Apparatus, system, and method for responding to a user-initiated query with a context-based response
US11868354B2 (en) 2015-09-23 2024-01-09 Motorola Solutions, Inc. Apparatus, system, and method for responding to a user-initiated query with a context-based response
GB2568013B (en) * 2016-09-21 2021-02-24 Motorola Solutions Inc Method and system for optimizing voice recognition and information searching based on talkgroup activities
US10104528B2 (en) 2017-02-01 2018-10-16 Telecommunication Systems, Inc. Supplemental location information for an emergency services call
JP7192379B2 (en) * 2018-10-11 2022-12-20 サクサ株式会社 emergency call system
US11361168B2 (en) * 2018-10-16 2022-06-14 Rovi Guides, Inc. Systems and methods for replaying content dialogue in an alternate language
US20210057055A1 (en) * 2019-08-20 2021-02-25 International Business Machines Corporation Medical Information Release Mechanism
WO2021246880A1 (en) * 2020-06-03 2021-12-09 Motorola Solutions, Inc System and method for tracking and display of compliance with instructions provided by emergency call taker

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7428002B2 (en) * 2002-06-05 2008-09-23 Monroe David A Emergency telephone with integrated surveillance system connectivity
US7646858B2 (en) * 2004-08-06 2010-01-12 Powerphone, Inc. Protocol builder for a call handling system
US10943694B2 (en) * 2008-04-30 2021-03-09 Vitalclick Llc Automated incident response method and system
US8817952B2 (en) * 2012-12-13 2014-08-26 Avaya Inc. Method, apparatus, and system for providing real-time PSAP call analysis

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10269372B1 (en) * 2015-09-24 2019-04-23 United Services Automobile Association (Usaa) System for sound analysis and recognition
US20170229125A1 (en) * 2016-02-05 2017-08-10 Honeywell International Inc. Systems and methods for contacting emergency personnel via voice recognition
US10062387B2 (en) * 2016-02-05 2018-08-28 Honeywell International Inc. Systems and methods for contacting emergency personnel via voice recognition
US10616393B2 (en) * 2017-06-21 2020-04-07 Guangdong Oppo Mobile Telecommunications Corp., Ltd. Method for call control and related products
WO2019045963A1 (en) * 2017-08-30 2019-03-07 T-Mobile Usa, Inc. Message transcription for emergency call prioritization
US10264121B2 (en) 2017-08-30 2019-04-16 T-Mobile Usa, Inc. Message transcription for emergency call prioritization

Also Published As

Publication number Publication date
US9350873B2 (en) 2016-05-24
US8908837B2 (en) 2014-12-09
US20140294161A1 (en) 2014-10-02

Similar Documents

Publication Publication Date Title
US9350873B2 (en) Methods and systems for automatically providing an emergency service call handler with context specific emergency service protocols
US11528772B2 (en) Apparatus and method for obtaining emergency data related to emergency sessions
US9420099B1 (en) Merging multiple emergency calls and information therefrom at emergency systems
US9798722B2 (en) System and method for transmitting multiple text streams of a communication in different languages
US10715662B2 (en) System and method for artificial intelligence on hold call handling
US8478229B2 (en) Method and apparatus for notifying registered devices of an emergency call
US20150229766A1 (en) Methods and Systems for Routing Emergency Service Calls Background
US8712757B2 (en) Methods and apparatus for monitoring communication through identification of priority-ranked keywords
US11947592B2 (en) Systems and method of generating custom messages based on rule-based database queries in a cloud platform
US9167090B2 (en) Public safety answering point language detection
US11889398B2 (en) Artificial intelligence for emergency assistance with human feedback for machine learning
US9767802B2 (en) Methods and apparatus for conducting internet protocol telephony communications
US10255919B2 (en) Identifying speaker roles in a streaming environment
WO2016132631A1 (en) Communication system and communication method
KR102504076B1 (en) Methods, systems, mobile communication devices, packet-switched communication systems, programs and computer program products for enhanced initiation and/or routing of emergency sessions in packet-switched communication systems
US20220201456A1 (en) Emergency session translation and transcription via audio forking and machine learning
US11570403B2 (en) Automated recording highlights for conferences
US9374465B1 (en) Multi-channel and multi-modal language interpretation system utilizing a gated or non-gated configuration
US20160295016A1 (en) Splitting a call for an emergent event into multiple devices using data channels
US20230136309A1 (en) Virtual Assistant For Task Identification
EP3585039B1 (en) System and method for recording and reviewing mixed-media communications
US20230239397A1 (en) 911 Call Enhancement
KR20220157141A (en) Method and Apparatus for Managing on Reporting data
US11909914B2 (en) Forwarding emergency messages from IoT devices to PSAPs
CN115174750B (en) DTMF signal transmission method and electronic equipment

Legal Events

Date Code Title Description
AS Assignment

Owner name: SYNERGEM TECHNOLOGIES INC, NORTH CAROLINA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:O'CONNOR, STEPHEN F;MAW, RICHARD A;REEL/FRAME:034213/0418

Effective date: 20130329

STCF Information on status: patent grant

Free format text: PATENTED CASE

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YR, SMALL ENTITY (ORIGINAL EVENT CODE: M2551); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY

Year of fee payment: 4

FEPP Fee payment procedure

Free format text: MAINTENANCE FEE REMINDER MAILED (ORIGINAL EVENT CODE: REM.); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY