US20150103982A1 - Methods and systems for automatically providing an emergency service call handler with context specific emergency service protocols - Google Patents
Methods and systems for automatically providing an emergency service call handler with context specific emergency service protocols Download PDFInfo
- Publication number
- US20150103982A1 US20150103982A1 US14/548,082 US201414548082A US2015103982A1 US 20150103982 A1 US20150103982 A1 US 20150103982A1 US 201414548082 A US201414548082 A US 201414548082A US 2015103982 A1 US2015103982 A1 US 2015103982A1
- Authority
- US
- United States
- Prior art keywords
- spoken
- call handler
- caller
- emergency service
- identified
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 title claims abstract description 46
- 238000004891 communication Methods 0.000 claims abstract description 55
- 230000008569 process Effects 0.000 claims description 13
- 238000004458 analytical method Methods 0.000 claims description 10
- 238000012544 monitoring process Methods 0.000 claims description 10
- 238000001514 detection method Methods 0.000 claims description 5
- 230000006870 function Effects 0.000 description 11
- 230000029058 respiratory gaseous exchange Effects 0.000 description 7
- 238000013500 data storage Methods 0.000 description 6
- 108010007100 Pulmonary Surfactant-Associated Protein A Proteins 0.000 description 5
- 102100027773 Pulmonary surfactant-associated protein A2 Human genes 0.000 description 5
- 230000005540 biological transmission Effects 0.000 description 3
- 238000010586 diagram Methods 0.000 description 3
- 238000005516 engineering process Methods 0.000 description 3
- 230000003287 optical effect Effects 0.000 description 3
- 235000015854 Heliotropium curassavicum Nutrition 0.000 description 2
- 244000301682 Heliotropium curassavicum Species 0.000 description 2
- 238000006243 chemical reaction Methods 0.000 description 2
- 238000012552 review Methods 0.000 description 2
- 230000011664 signaling Effects 0.000 description 2
- 230000007704 transition Effects 0.000 description 2
- 235000019801 trisodium phosphate Nutrition 0.000 description 2
- 239000004165 Methyl ester of fatty acids Substances 0.000 description 1
- 241000269400 Sirenidae Species 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 230000008901 benefit Effects 0.000 description 1
- 238000001651 catalytic steam reforming of methanol Methods 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 238000013523 data management Methods 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 238000004880 explosion Methods 0.000 description 1
- 230000010365 information processing Effects 0.000 description 1
- 230000000977 initiatory effect Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000001902 propagating effect Effects 0.000 description 1
- 230000002123 temporal effect Effects 0.000 description 1
- 230000001052 transient effect Effects 0.000 description 1
- 238000013519 translation Methods 0.000 description 1
- 238000010200 validation analysis Methods 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M3/00—Automatic or semi-automatic exchanges
- H04M3/42—Systems providing special services or facilities to subscribers
- H04M3/50—Centralised arrangements for answering calls; Centralised arrangements for recording messages for absent or busy subscribers ; Centralised arrangements for recording messages
- H04M3/51—Centralised call answering arrangements requiring operator intervention, e.g. call or contact centers for telemarketing
- H04M3/5116—Centralised call answering arrangements requiring operator intervention, e.g. call or contact centers for telemarketing for emergency applications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M11/00—Telephonic communication systems specially adapted for combination with other electrical systems
- H04M11/04—Telephonic communication systems specially adapted for combination with other electrical systems with alarm systems, e.g. fire, police or burglar alarm systems
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/08—Speech classification or search
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M3/00—Automatic or semi-automatic exchanges
- H04M3/22—Arrangements for supervision, monitoring or testing
- H04M3/2281—Call monitoring, e.g. for law enforcement purposes; Call tracing; Detection or prevention of malicious calls
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M2201/00—Electronic components, circuits, software, systems or apparatus used in telephone systems
- H04M2201/40—Electronic components, circuits, software, systems or apparatus used in telephone systems using speech recognition
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M2203/00—Aspects of automatic or semi-automatic exchanges
- H04M2203/20—Aspects of automatic or semi-automatic exchanges related to features of supplementary services
- H04M2203/2038—Call context notifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M2242/00—Special services or facilities
- H04M2242/04—Special services or facilities for emergency applications
Definitions
- PSAP public safety answering point
- public safety access point is a call center responsible for answering calls to an emergency telephone number for police, firefighting, and ambulance services.
- Trained emergency service call takers are typically responsible for obtaining relevant information from a caller and dispatching the appropriate emergency service resources to the appropriate location.
- ESPs emergency service protocols
- ESPs emergency service protocols
- a caller tells the call taker someone is not breathing
- an appropriate ESP may guide the call taker through giving the caller instructions on performing CPR or other basic first aid procedures.
- Other protocols may be directed at how to obtain appropriate information from the caller. For example, if the call involves a bomb threat, an appropriate ESP may instruct the call taker to notify the bomb squad and fire department and give the call taker instructions on how to attempt to guide the conversation with the caller to obtain critical information.
- Next Generation 9-1-1 (“NG9-1”) can be viewed as a system comprised of Emergency Services IP networks (“ESInets”), internet protocol (“IP”) based software services and application, and various databases and data management processes that are all interconnected to a public safety answering point (PSAP).
- EESInets Emergency Services IP networks
- IP internet protocol
- PSAP public safety answering point
- the NG9-1-1 system provides location-based routing to the appropriate emergency entity, such that a caller in need of help is automatically routed to the PSAP assigned to the caller's location.
- NG9-1-1 also provides standardized interfaces for call and message services, processes all types of emergency calls including non-voice (multi-media) messages, acquires and integrates additional data useful to call routing and handling for appropriate emergency entities.
- NG9-1-1 supports all legacy E9-1-1 features and functions and is intended to provide scalable solution for meeting current and emerging needs for emergency communication between callers and Public Safety entities.
- the NG9-1-1 system architecture is defined by the National Emergency Number Association (“NENA”) i3 standard and supports end-to-end IP connectivity between a caller and a public safety answering point (PSAP).
- NENA National Emergency Number Association
- PSAP public safety answering point
- the i3 standard defines an ESInet, which sits between various, non-emergency communications networks and one or more PSAPs, as well as the ESInet's various functional elements, such as a Location Information Server (LIS) and Location Validation Function (LVF), the Emergency Services Routing Proxy (ESRP) and Policy Routing Function (PRF) and the Emergency Call Routing Function and Location to Service Translation (LoST) protocol. All of these elements are designed to provide robust and secure communications between a variety of communications devices and emergency service providers.
- LIS Location Information Server
- LVF Location Validation Function
- ESRP Emergency Services Routing Proxy
- PRF Policy Routing Function
- LoST Location to Service Translation
- the i3 standard requires all calls presented to the ESInet from an originating network, such as a typical telecomunications service provider (“TSP”) network to use session initiation protocol (“SIP”) signaling to deliver the call and include the location with the call.
- SIP is a signaling protocol used to start, change and end telephone and multimedia communication sessions over IP networks.
- BCF Border Control Function
- the first element inside the ESInet is the Emergency Services Routing Proxy (ESRP).
- ESRP Emergency Services Routing Proxy
- the ESRP receives the call, and passes this information to an Emergency Call Routing Function (ECRF), which determines the next hop in routing a call to the requested service.
- ECRF Emergency Call Routing Function
- the ECRF maps the call's location information and requested service (e.g. police, which may be routed to a city-operated PSAP or fire department, which may be routed to a county-operated PSAP) to an appropriate PSAP.
- NENA In the event an ESInet is provisioned in an area before the regional TSPs and other originating networks or PSAPs are NG9-1-1 capable, NENA has defined a transition model.
- the legacy E911 network has been replaced by the Emergency Services IP Network (ESInet) with all of the functional elements previously described, but on either end (originating network and/or PSAP) is a legacy environment.
- ESInet Emergency Services IP Network
- NG9-1-1 defines a legacy network gateway and a legacy PSAP gateway to convert the data to and from SIP messaging for transmission over the ESInet until such time as the originating networks and PSAPs become i3 capable.
- a beneficial side effect of the transition to the NG9-1-1 environment is that all emergency service phone calls will be converted to digital data and stored for future review. This further enables new and advantageous information processing techniques to be applied to emergency service calls in real time in order to assist emergency service call takers in performing their jobs.
- FIG. 1 is a schematic diagram depicting aspects of a non-limiting, exemplary computing architecture suitable for implementing at least some aspects and/or embodiments of the present systems and methods.
- FIG. 2 is a functional block diagram depicting an emergency services communications network advantageously featuring aspects of the present methods and systems.
- FIG. 3 is a functional block diagram depicting an emergency services communications network advantageously featuring additional aspects of the present methods and systems.
- FIG. 4 is a flow chart depicting the operational steps of certain aspects of the present methods and systems.
- Embodiments of the present methods and systems may be implemented by systems using one or more programmable digital computers.
- Computer and computer systems in connection with embodiments of the invention may act, e.g., as workstations and/or servers, such as described below.
- Digital voice and/or data networks such as may be used in connection with embodiments of the invention may also include components (e.g., routers, bridges, media gateways, etc.) with similar architectures, although they may be adapted, e.g., as known in the art, for their special purposes. Because of this commonality of architecture, such network components may be considered as computer systems and/or components of computer systems when consistent with the applicable context.
- FIG. 1 depicts an example of one such computer system 100 , which includes at least one processor 110 , such as, e.g., an Intel or Advanced Micro Devices microprocessor, coupled to a communications channel or bus 112 .
- the computer system 100 further includes at least one input device 114 such as, e.g., a keyboard, mouse, touch pad or screen, or other selection or pointing device, at least one output device 116 such as, e.g., an electronic display device, at least one communications interface 118 , at least one data storage device 120 such as a magnetic disk or an optical disk, and memory 122 such as ROM and RAM, each coupled to the communications channel 112 .
- the communications interface 118 may be coupled to a network (not depicted) such as the Internet.
- FIG. 1 Although the computer system 100 is shown in FIG. 1 to have only a single communications channel 112 , a person skilled in the relevant arts will recognize that a computer system may have multiple channels (not depicted), including for example one or more busses, and that such channels may be interconnected, e.g., by one or more bridges. In such a configuration, components depicted in FIG. 1 as connected by a single channel 112 may interoperate, and may thereby be considered to be coupled to one another, despite being directly connected to different communications channels.
- data storage device 120 and memory 122 are depicted as different units, the data storage device 120 and memory 122 can be parts of the same unit or units, and that the functions of one can be shared in whole or in part by the other, e.g., as RAM disks, virtual memory, etc. It will also be appreciated that any particular computer may have multiple components of a given type, e.g., processors 110 , input devices 114 , communications interfaces 118 , etc.
- the data storage device 120 may store instructions executable by one or more processors or kinds of processors 110 , data, or both. Some groups of instructions, possibly grouped with data, may make up one or more programs, which may include an operating system such as Microsoft Windows®, Linux®, Mac OS®, or Unix®. Other programs may be stored instead of or in addition to the operating system. It will be appreciated that a computer system may also be implemented on platforms and operating systems other than those mentioned.
- Any operating system or other program, or any part of either, may be written using one or more programming languages such as, e.g., Java®, C, C++, C#, Visual Basic®, VB.NET®, Perl, Ruby, Python, or other programming languages, possibly using object oriented design and/or coding techniques.
- programming languages such as, e.g., Java®, C, C++, C#, Visual Basic®, VB.NET®, Perl, Ruby, Python, or other programming languages, possibly using object oriented design and/or coding techniques.
- the computer system 100 may also include additional components and/or systems, such as network connections, additional memory, additional processors, network interfaces, input/output busses, for example.
- a computer-readable storage medium (CRSM) reader 136 such as, e.g., a magnetic disk drive, magneto-optical drive, optical disk drive, or flash drive, may be coupled to the communications channel 112 for reading from a CRSM 138 such as, e.g., a magnetic disk, a magneto-optical disk, an optical disk, or flash RAM.
- CRSM computer-readable storage medium
- one or more CRSM readers may be coupled to the rest of the computer system 100 , e.g., through a network interface (not depicted) or a communications interface 118 . In any such configuration, however, the computer system 100 may receive programs and/or data via the CRSM reader 136 .
- the term “memory” herein is intended to include various types of suitable data storage media, whether permanent or temporary, including among other things the data storage device 120 , the memory 122 , and the CSRM 138 .
- computer-readable storage medium and “computer-readable storage media” refer, respectively, to a medium and media capable of storing information. As such, both terms exclude transient propagating signals.
- Two or more computer systems 100 may communicate, e.g., in one or more networks, via, e.g., their respective communications interfaces 118 and/or network interfaces (not depicted).
- FIG. 2 depicts a communications system 200 , including an ESInet 204 connected to an origination network 208 and a PSAP 212 via BCFs 210 , 211 , suitable for use with the present methods and systems.
- an emergency voice call 216 is routed to the PSAP 212 from the origination network 208 via ESInet 204 , the caller is connected to an emergency service call handler (not shown) via a call handling application 220 .
- the call is also routed to a session recorder 224 for analysis, review and archival purposes.
- an automated protocol selection function (APSF) 228 is provided. As call 216 is being recorded by session recorder 224 it is also input to the APSF 228 .
- the APSF 228 may include a speech recognition element 232 , a comparison element 240 , a protocol selection element 244 , a keyword database 246 , and a protocol database 248 .
- Speech recognition element 232 may monitor the digital data transmission that corresponds to the voice communication between the caller and the emergency services call taker and apply a speech recognition process to detect words and/or phrases being spoken by the caller.
- the speech recognition element 232 may divide the caller's speech into segments, which may be on the order of magnitude of a hundredth of a second in duration and compare the segments to a set of known phonemes. The speech recognition element 232 may then perform a contextual phoneme analysis on each phoneme identified in the call to other phonemes in its temporal vicinity in order to determine the language being spoken and identify what word or phrase in that language the caller is using. This may advantageously occur substantially in real-time, as the caller is speaking.
- Speech recognition solutions such as Microsoft Voice Command (Microsoft Corporation), Sonic Extractor (Digitial Syphon), LumenVox Speech Engine (LumenVox), Nuance Voice Control (Nuance Communications), VITO Voice2Go (Vito Technology), and Speereo Voice Translator (Speereo Software) are exemplary but non-limiting implementations of aspects of the speech recognition element 232 .
- Comparison element 240 compares the word or phrase 252 identified by speech recognition element 232 to a set of known keywords and phrases stored in keyword database 246 . Each keyword and phrase in keyword database 246 is associated with one or more emergency service protocols 256 stored in protocol database 248 . If comparison element 240 detects a match between the spoken word or phrase 252 and one of the known keywords or phrases, it notifies the protocol selection element 244 . Protocol selection element 244 retrieves the appropriate emergency service protocol(s) 256 identified by the detected keyword or phrase and transmits the protocol 256 to the call handling application 220 where it is displayed to the emergency service call taker to assist him/her in handling the call.
- Speech recognition element 232 will break this phrase down into the set of phonemes and, after running a contextual analysis, identify the individual words “Help,” “my,” “wife,” “isn't,” and “breathing.” These words are then passed to the comparison element 240 which may compare the individual words and sub-sets of words within the phrase to the known key words and phrases stored in keyword database 246 .
- the one such known phrase may be “isn't breathing,” or variations thereof, and comparison element 240 will match that known phrase to the corresponding subset of words from the caller's statement.
- the phrase “isn't breathing” may be linked to an emergency service protocol on CPR instructions.
- the protocol selection element 244 may then retrieve the CPR protocol from protocol database 248 and display it for the emergency services call taker taking the call.
- the emergency services call taker can seamlessly provide instructions to the caller without having to stop, mentally process the statement, and look up the appropriate protocol him/herself.
- the protocol selection element may advantageously determine to provide the emergency service call taker with an infant specific CPR protocol.
- Certain embodiments of the present methods and systems may advantageously filter the incoming call data to distinguish between foreground noise, i.e. the caller's voice, and background noise.
- the background noise may be separately analyzed by a background analysis element 260 for relevant information, such as the presence of sirens, alarms, additional voices, gun shots, explosions, etc. Detection of such information may also factor into the determination of the appropriate protocol to provide to the emergency service call handler.
- FIG. 3 depicts additional aspects of the present methods and systems, which may advantageously distinguish the caller's speech from the emergency service call taker's speech.
- the PSAP is a legacy PSAP 312
- the digital IP data 313 transmitted by the ESInet 303 will be converted back to a legacy format 304 by a Legacy PSAP gateway function 305 .
- the legacy formatted data 304 may be reconverted to IP data 313 by an IP conversion function 348 prior to being input to the session recorder 324 .
- the network transport path from the originating network to the legacy PSAP 312 is a legacy network (not depicted) rather than an ESInet
- the data is delivered directly 360 to the IP conversion function 348 rather than the legacy PSAP gateway 305 .
- further alternative aspects of the present methods and systems may, prior to analysis by a speech recognition element, input the call data into a parsing element 350 in order to distinguish voice-data packets originating from the PSAP's IP address from voice-data packets originating from other IP addresses, thereby distinguishing the caller's speech 354 from the call taker's speech 358 .
- the call taker's voice may be discarded and the protocol selection process may proceed as described above with reference to FIG. 2 .
- the separate instantiations of the speech recognition element 332 , 333 may separately process the caller's speech 354 and the call taker's speech 358 and separate instantiations of the comparison element 340 , 341 may compare identified words or phrases in the respective sides of the conversation to separate sets of keywords.
- Such an aspect of the present method and system may, for example, give the call taker the ability to call up emergency service protocols using voice commands in the context of the conversation with the caller.
- FIG. 4 depicts the steps of certain embodiments of the present methods and systems.
- a caller initiates an emergency service phone call 404 via an originating communication network.
- the origination network detects that the call is an emergency service call and routes the call to a local transport network, such as an ESInet or a legacy network, 408 .
- the call is then routed to the appropriate PSAP 412 .
- a two way communication channel is opened 414 between the caller and an emergency service call handler at the PSAP and the digital data corresponding to the voice communication between the caller and the call handler is monitored by a session recorder and an APSF 416 .
- the APSF performs a speech recognition analysis on the voice communication 420 and identifies particular words and/or phrases being spoken by the caller 424 .
- the identified words and/or phrases are then compared to a known set of keywords 428 . If a match is detected 450 , the APSF retrieves an emergency service protocol associated with the matched keyword 432 and provides the protocol to the emergency service call handler 436 .
- present methods and systems described above can be implemented in locally installed software applications, for example, substantially running on computing hardware at the PSAP.
- the present methods and systems could, however, also be implemented via a software as a service model, wherein the majority of computations are done at a remote location via network communications and the PSAP runs a ‘lightweight’ client application that predominately acts as an interface to the remote applications.
Abstract
Description
- This application claims the benefit of co-pending and commonly owned U.S. patent application Ser. No. 13/853,905 filed Mar. 29, 2013, herein incorporated by reference in its entirety for all purposes.
- A public safety answering point (PSAP), sometimes called “public safety access point”, is a call center responsible for answering calls to an emergency telephone number for police, firefighting, and ambulance services. Trained emergency service call takers are typically responsible for obtaining relevant information from a caller and dispatching the appropriate emergency service resources to the appropriate location.
- In order to assist the emergency call takers, many PSAPs utilize defined emergency service protocols (ESPs) for providing standard instructions for various types of common emergency service situations. For example, if a caller tells the call taker someone is not breathing, an appropriate ESP may guide the call taker through giving the caller instructions on performing CPR or other basic first aid procedures. Other protocols may be directed at how to obtain appropriate information from the caller. For example, if the call involves a bomb threat, an appropriate ESP may instruct the call taker to notify the bomb squad and fire department and give the call taker instructions on how to attempt to guide the conversation with the caller to obtain critical information. In conventional 9-1-1 systems, where the voice transmissions between a caller and call taker may be analog signals, and the call taker must know to recognize certain words or phrases spoken by a caller and look up any appropriate protocols. This additional step takes the call taker's attention away from dealing with the caller, and can cause delay and confusion which, in the context of an emergency services call, can lead directly to harm to individuals, damage to property, and/or additional, preventable consequences.
- Advances in communication technology, specifically data connectivity and voice-over-IP technology, has led to the implementation of Enhanced-9-1-1 and Next Generation 9-1-1 standards. Broadly speaking, Next Generation 9-1-1 (“NG9-1-1”) can be viewed as a system comprised of Emergency Services IP networks (“ESInets”), internet protocol (“IP”) based software services and application, and various databases and data management processes that are all interconnected to a public safety answering point (PSAP). The NG9-1-1 system provides location-based routing to the appropriate emergency entity, such that a caller in need of help is automatically routed to the PSAP assigned to the caller's location. NG9-1-1 also provides standardized interfaces for call and message services, processes all types of emergency calls including non-voice (multi-media) messages, acquires and integrates additional data useful to call routing and handling for appropriate emergency entities. NG9-1-1 supports all legacy E9-1-1 features and functions and is intended to provide scalable solution for meeting current and emerging needs for emergency communication between callers and Public Safety entities.
- The NG9-1-1 system architecture is defined by the National Emergency Number Association (“NENA”) i3 standard and supports end-to-end IP connectivity between a caller and a public safety answering point (PSAP). The i3 standard defines an ESInet, which sits between various, non-emergency communications networks and one or more PSAPs, as well as the ESInet's various functional elements, such as a Location Information Server (LIS) and Location Validation Function (LVF), the Emergency Services Routing Proxy (ESRP) and Policy Routing Function (PRF) and the Emergency Call Routing Function and Location to Service Translation (LoST) protocol. All of these elements are designed to provide robust and secure communications between a variety of communications devices and emergency service providers.
- The i3 standard requires all calls presented to the ESInet from an originating network, such as a typical telecomunications service provider (“TSP”) network to use session initiation protocol (“SIP”) signaling to deliver the call and include the location with the call. SIP is a signaling protocol used to start, change and end telephone and multimedia communication sessions over IP networks. Upon reaching the ESInet, call traffic encounters the Border Control Function (BCF) which sits between external networks and the ESInet. An emergency service call, with location information, enters the ESInet through the BCF. After passing through the BCF, the first element inside the ESInet is the Emergency Services Routing Proxy (ESRP). The ESRP receives the call, and passes this information to an Emergency Call Routing Function (ECRF), which determines the next hop in routing a call to the requested service. The ECRF maps the call's location information and requested service (e.g. police, which may be routed to a city-operated PSAP or fire department, which may be routed to a county-operated PSAP) to an appropriate PSAP.
- In the event an ESInet is provisioned in an area before the regional TSPs and other originating networks or PSAPs are NG9-1-1 capable, NENA has defined a transition model. In this case, the legacy E911 network has been replaced by the Emergency Services IP Network (ESInet) with all of the functional elements previously described, but on either end (originating network and/or PSAP) is a legacy environment. To provide connectivity to both the legacy networks and the legacy PSAPs, NG9-1-1 defines a legacy network gateway and a legacy PSAP gateway to convert the data to and from SIP messaging for transmission over the ESInet until such time as the originating networks and PSAPs become i3 capable.
- A beneficial side effect of the transition to the NG9-1-1 environment is that all emergency service phone calls will be converted to digital data and stored for future review. This further enables new and advantageous information processing techniques to be applied to emergency service calls in real time in order to assist emergency service call takers in performing their jobs.
- Various embodiments in accordance with the present disclosure will be described with reference to the drawings, in which:
-
FIG. 1 is a schematic diagram depicting aspects of a non-limiting, exemplary computing architecture suitable for implementing at least some aspects and/or embodiments of the present systems and methods. -
FIG. 2 is a functional block diagram depicting an emergency services communications network advantageously featuring aspects of the present methods and systems. -
FIG. 3 is a functional block diagram depicting an emergency services communications network advantageously featuring additional aspects of the present methods and systems. -
FIG. 4 is a flow chart depicting the operational steps of certain aspects of the present methods and systems. - This description discusses various illustrative embodiments of the present methods and systems for monitoring the content of an emergency service phone call and providing a call handler with context specific protocols (“the present methods and systems”) with reference to the accompanying drawings in order to provide a person having ordinary skill in the relevant art with a full, clear, and concise description of the subject matter defined by the claims which follow, and to enable such a person to appreciate and understand how to make and use the same. However, this description should not be read to limit the scope of the claimed subject matter, nor does the presence of an embodiment in this description imply any preference of the described embodiment over any other embodiment, unless such a preference is explicitly identified herein. It is the claims, not this description or other sections of this document or the accompanying drawings, which define the scope of the subject matter to which the inventor and/or the inventor's assignee(s) claim exclusive rights.
- Embodiments of the present methods and systems may be implemented by systems using one or more programmable digital computers. Computer and computer systems in connection with embodiments of the invention may act, e.g., as workstations and/or servers, such as described below. Digital voice and/or data networks such as may be used in connection with embodiments of the invention may also include components (e.g., routers, bridges, media gateways, etc.) with similar architectures, although they may be adapted, e.g., as known in the art, for their special purposes. Because of this commonality of architecture, such network components may be considered as computer systems and/or components of computer systems when consistent with the applicable context.
-
FIG. 1 depicts an example of onesuch computer system 100, which includes at least oneprocessor 110, such as, e.g., an Intel or Advanced Micro Devices microprocessor, coupled to a communications channel orbus 112. Thecomputer system 100 further includes at least oneinput device 114 such as, e.g., a keyboard, mouse, touch pad or screen, or other selection or pointing device, at least oneoutput device 116 such as, e.g., an electronic display device, at least onecommunications interface 118, at least onedata storage device 120 such as a magnetic disk or an optical disk, andmemory 122 such as ROM and RAM, each coupled to thecommunications channel 112. Thecommunications interface 118 may be coupled to a network (not depicted) such as the Internet. - Although the
computer system 100 is shown inFIG. 1 to have only asingle communications channel 112, a person skilled in the relevant arts will recognize that a computer system may have multiple channels (not depicted), including for example one or more busses, and that such channels may be interconnected, e.g., by one or more bridges. In such a configuration, components depicted inFIG. 1 as connected by asingle channel 112 may interoperate, and may thereby be considered to be coupled to one another, despite being directly connected to different communications channels. - One skilled in the art will recognize that, although the
data storage device 120 andmemory 122 are depicted as different units, thedata storage device 120 andmemory 122 can be parts of the same unit or units, and that the functions of one can be shared in whole or in part by the other, e.g., as RAM disks, virtual memory, etc. It will also be appreciated that any particular computer may have multiple components of a given type, e.g.,processors 110,input devices 114,communications interfaces 118, etc. - The data storage device 120 (
FIG. 1 ) and/ormemory 122 may store instructions executable by one or more processors or kinds ofprocessors 110, data, or both. Some groups of instructions, possibly grouped with data, may make up one or more programs, which may include an operating system such as Microsoft Windows®, Linux®, Mac OS®, or Unix®. Other programs may be stored instead of or in addition to the operating system. It will be appreciated that a computer system may also be implemented on platforms and operating systems other than those mentioned. Any operating system or other program, or any part of either, may be written using one or more programming languages such as, e.g., Java®, C, C++, C#, Visual Basic®, VB.NET®, Perl, Ruby, Python, or other programming languages, possibly using object oriented design and/or coding techniques. - One skilled in the art will recognize that the computer system 100 (
FIG. 1 ) may also include additional components and/or systems, such as network connections, additional memory, additional processors, network interfaces, input/output busses, for example. One skilled in the art will also recognize that the programs and data may be received by and stored in the system in alternative ways. For example, a computer-readable storage medium (CRSM)reader 136, such as, e.g., a magnetic disk drive, magneto-optical drive, optical disk drive, or flash drive, may be coupled to thecommunications channel 112 for reading from a CRSM 138 such as, e.g., a magnetic disk, a magneto-optical disk, an optical disk, or flash RAM. Alternatively, one or more CRSM readers may be coupled to the rest of thecomputer system 100, e.g., through a network interface (not depicted) or acommunications interface 118. In any such configuration, however, thecomputer system 100 may receive programs and/or data via the CRSMreader 136. Further, it will be appreciated that the term “memory” herein is intended to include various types of suitable data storage media, whether permanent or temporary, including among other things thedata storage device 120, thememory 122, and the CSRM 138. - The terms “computer-readable storage medium” and “computer-readable storage media” refer, respectively, to a medium and media capable of storing information. As such, both terms exclude transient propagating signals.
- Two or more computer systems 100 (
FIG. 1 ) may communicate, e.g., in one or more networks, via, e.g., theirrespective communications interfaces 118 and/or network interfaces (not depicted). -
FIG. 2 depicts acommunications system 200, including anESInet 204 connected to anorigination network 208 and aPSAP 212 viaBCFs emergency voice call 216 is routed to thePSAP 212 from theorigination network 208 viaESInet 204, the caller is connected to an emergency service call handler (not shown) via acall handling application 220. The call is also routed to asession recorder 224 for analysis, review and archival purposes. - In accordance with the present methods and systems, an automated protocol selection function (APSF) 228 is provided. As
call 216 is being recorded bysession recorder 224 it is also input to theAPSF 228. TheAPSF 228 may include aspeech recognition element 232, acomparison element 240, aprotocol selection element 244, akeyword database 246, and aprotocol database 248.Speech recognition element 232 may monitor the digital data transmission that corresponds to the voice communication between the caller and the emergency services call taker and apply a speech recognition process to detect words and/or phrases being spoken by the caller. For example, thespeech recognition element 232 may divide the caller's speech into segments, which may be on the order of magnitude of a hundredth of a second in duration and compare the segments to a set of known phonemes. Thespeech recognition element 232 may then perform a contextual phoneme analysis on each phoneme identified in the call to other phonemes in its temporal vicinity in order to determine the language being spoken and identify what word or phrase in that language the caller is using. This may advantageously occur substantially in real-time, as the caller is speaking. Commercially available speech recognition solutions such as Microsoft Voice Command (Microsoft Corporation), Sonic Extractor (Digitial Syphon), LumenVox Speech Engine (LumenVox), Nuance Voice Control (Nuance Communications), VITO Voice2Go (Vito Technology), and Speereo Voice Translator (Speereo Software) are exemplary but non-limiting implementations of aspects of thespeech recognition element 232. - After a word or
phrase 252 is identified by thespeech recognition element 232 it is passed tocomparison element 240.Comparison element 240 compares the word orphrase 252 identified byspeech recognition element 232 to a set of known keywords and phrases stored inkeyword database 246. Each keyword and phrase inkeyword database 246 is associated with one or moreemergency service protocols 256 stored inprotocol database 248. Ifcomparison element 240 detects a match between the spoken word orphrase 252 and one of the known keywords or phrases, it notifies theprotocol selection element 244.Protocol selection element 244 retrieves the appropriate emergency service protocol(s) 256 identified by the detected keyword or phrase and transmits theprotocol 256 to thecall handling application 220 where it is displayed to the emergency service call taker to assist him/her in handling the call. - For example, a caller may state, “Help, my wife isn't breathing!”
Speech recognition element 232 will break this phrase down into the set of phonemes and, after running a contextual analysis, identify the individual words “Help,” “my,” “wife,” “isn't,” and “breathing.” These words are then passed to thecomparison element 240 which may compare the individual words and sub-sets of words within the phrase to the known key words and phrases stored inkeyword database 246. The one such known phrase may be “isn't breathing,” or variations thereof, andcomparison element 240 will match that known phrase to the corresponding subset of words from the caller's statement. The phrase “isn't breathing” may be linked to an emergency service protocol on CPR instructions. Theprotocol selection element 244 may then retrieve the CPR protocol fromprotocol database 248 and display it for the emergency services call taker taking the call. Thus, the emergency services call taker can seamlessly provide instructions to the caller without having to stop, mentally process the statement, and look up the appropriate protocol him/herself. - If, however, the caller states, “Help, my baby isn't breathing!” the word “baby” may be detected in addition to “isn't breathing” and the protocol selection element may advantageously determine to provide the emergency service call taker with an infant specific CPR protocol.
- Certain embodiments of the present methods and systems may advantageously filter the incoming call data to distinguish between foreground noise, i.e. the caller's voice, and background noise. The background noise may be separately analyzed by a
background analysis element 260 for relevant information, such as the presence of sirens, alarms, additional voices, gun shots, explosions, etc. Detection of such information may also factor into the determination of the appropriate protocol to provide to the emergency service call handler. -
FIG. 3 depicts additional aspects of the present methods and systems, which may advantageously distinguish the caller's speech from the emergency service call taker's speech. For example, if the PSAP is alegacy PSAP 312, thedigital IP data 313 transmitted by the ESInet 303 will be converted back to alegacy format 304 by a LegacyPSAP gateway function 305. In order to provide the functionality of the present methods and systems, the legacy formatteddata 304 may be reconverted toIP data 313 by anIP conversion function 348 prior to being input to thesession recorder 324. If the network transport path from the originating network to thelegacy PSAP 312 is a legacy network (not depicted) rather than an ESInet, the data is delivered directly 360 to theIP conversion function 348 rather than thelegacy PSAP gateway 305. - Still referring to
FIG. 3 , further alternative aspects of the present methods and systems may, prior to analysis by a speech recognition element, input the call data into aparsing element 350 in order to distinguish voice-data packets originating from the PSAP's IP address from voice-data packets originating from other IP addresses, thereby distinguishing the caller'sspeech 354 from the call taker'sspeech 358. In certain embodiments, the call taker's voice may be discarded and the protocol selection process may proceed as described above with reference toFIG. 2 . Alternately, the separate instantiations of thespeech recognition element speech 354 and the call taker'sspeech 358 and separate instantiations of thecomparison element -
FIG. 4 depicts the steps of certain embodiments of the present methods and systems. A caller initiates an emergencyservice phone call 404 via an originating communication network. The origination network detects that the call is an emergency service call and routes the call to a local transport network, such as an ESInet or a legacy network, 408. The call is then routed to theappropriate PSAP 412. A two way communication channel is opened 414 between the caller and an emergency service call handler at the PSAP and the digital data corresponding to the voice communication between the caller and the call handler is monitored by a session recorder and anAPSF 416. The APSF performs a speech recognition analysis on thevoice communication 420 and identifies particular words and/or phrases being spoken by thecaller 424. The identified words and/or phrases are then compared to a known set ofkeywords 428. If a match is detected 450, the APSF retrieves an emergency service protocol associated with the matchedkeyword 432 and provides the protocol to the emergencyservice call handler 436. - It should be understood that the present methods and systems described above can be implemented in locally installed software applications, for example, substantially running on computing hardware at the PSAP. The present methods and systems could, however, also be implemented via a software as a service model, wherein the majority of computations are done at a remote location via network communications and the PSAP runs a ‘lightweight’ client application that predominately acts as an interface to the remote applications.
- Exemplary embodiments of the present methods and systems have been described in detail above and in the accompanying figures for illustrative purposes. However, the scope of the present methods and systems are defined by the claims below and is not limited to the embodiments described above or depicted in the figures. Embodiments differing from those described and shown herein, but still within the scope of the defined methods and systems are envisioned by the inventors and will be apparent to persons having ordinary skill in the relevant art in view of this specification as a whole. The inventors intend for the defined methods and systems to be practiced other than as explicitly described herein. Accordingly, the defined methods and systems encompass all modifications and equivalents of the subject matter as permitted by applicable law.
Claims (20)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/548,082 US9350873B2 (en) | 2013-03-29 | 2014-11-19 | Methods and systems for automatically providing an emergency service call handler with context specific emergency service protocols |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US13/853,905 US8908837B2 (en) | 2013-03-29 | 2013-03-29 | Methods and systems for automatically providing an emergency service call handler with context specific emergency service protocols |
US14/548,082 US9350873B2 (en) | 2013-03-29 | 2014-11-19 | Methods and systems for automatically providing an emergency service call handler with context specific emergency service protocols |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/853,905 Continuation US8908837B2 (en) | 2013-03-29 | 2013-03-29 | Methods and systems for automatically providing an emergency service call handler with context specific emergency service protocols |
Publications (2)
Publication Number | Publication Date |
---|---|
US20150103982A1 true US20150103982A1 (en) | 2015-04-16 |
US9350873B2 US9350873B2 (en) | 2016-05-24 |
Family
ID=51620864
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/853,905 Active US8908837B2 (en) | 2013-03-29 | 2013-03-29 | Methods and systems for automatically providing an emergency service call handler with context specific emergency service protocols |
US14/548,082 Active US9350873B2 (en) | 2013-03-29 | 2014-11-19 | Methods and systems for automatically providing an emergency service call handler with context specific emergency service protocols |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/853,905 Active US8908837B2 (en) | 2013-03-29 | 2013-03-29 | Methods and systems for automatically providing an emergency service call handler with context specific emergency service protocols |
Country Status (1)
Country | Link |
---|---|
US (2) | US8908837B2 (en) |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20170229125A1 (en) * | 2016-02-05 | 2017-08-10 | Honeywell International Inc. | Systems and methods for contacting emergency personnel via voice recognition |
WO2019045963A1 (en) * | 2017-08-30 | 2019-03-07 | T-Mobile Usa, Inc. | Message transcription for emergency call prioritization |
US10269372B1 (en) * | 2015-09-24 | 2019-04-23 | United Services Automobile Association (Usaa) | System for sound analysis and recognition |
US10616393B2 (en) * | 2017-06-21 | 2020-04-07 | Guangdong Oppo Mobile Telecommunications Corp., Ltd. | Method for call control and related products |
Families Citing this family (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9860722B1 (en) * | 2014-01-30 | 2018-01-02 | Sprint Communications Company L.P. | Converting location information into public safety answering point (PSAP) signaling |
US9799206B1 (en) * | 2015-01-05 | 2017-10-24 | Brenda Michelle Wilson Van Horn | Method for automating emergency distress signals from a networked peripheral device |
CN105045919B (en) * | 2015-08-24 | 2019-08-16 | 北京云知声信息技术有限公司 | A kind of information output method and device |
US10372755B2 (en) | 2015-09-23 | 2019-08-06 | Motorola Solutions, Inc. | Apparatus, system, and method for responding to a user-initiated query with a context-based response |
US11868354B2 (en) | 2015-09-23 | 2024-01-09 | Motorola Solutions, Inc. | Apparatus, system, and method for responding to a user-initiated query with a context-based response |
GB2568013B (en) * | 2016-09-21 | 2021-02-24 | Motorola Solutions Inc | Method and system for optimizing voice recognition and information searching based on talkgroup activities |
US10104528B2 (en) | 2017-02-01 | 2018-10-16 | Telecommunication Systems, Inc. | Supplemental location information for an emergency services call |
JP7192379B2 (en) * | 2018-10-11 | 2022-12-20 | サクサ株式会社 | emergency call system |
US11361168B2 (en) * | 2018-10-16 | 2022-06-14 | Rovi Guides, Inc. | Systems and methods for replaying content dialogue in an alternate language |
US20210057055A1 (en) * | 2019-08-20 | 2021-02-25 | International Business Machines Corporation | Medical Information Release Mechanism |
WO2021246880A1 (en) * | 2020-06-03 | 2021-12-09 | Motorola Solutions, Inc | System and method for tracking and display of compliance with instructions provided by emergency call taker |
Family Cites Families (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7428002B2 (en) * | 2002-06-05 | 2008-09-23 | Monroe David A | Emergency telephone with integrated surveillance system connectivity |
US7646858B2 (en) * | 2004-08-06 | 2010-01-12 | Powerphone, Inc. | Protocol builder for a call handling system |
US10943694B2 (en) * | 2008-04-30 | 2021-03-09 | Vitalclick Llc | Automated incident response method and system |
US8817952B2 (en) * | 2012-12-13 | 2014-08-26 | Avaya Inc. | Method, apparatus, and system for providing real-time PSAP call analysis |
-
2013
- 2013-03-29 US US13/853,905 patent/US8908837B2/en active Active
-
2014
- 2014-11-19 US US14/548,082 patent/US9350873B2/en active Active
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10269372B1 (en) * | 2015-09-24 | 2019-04-23 | United Services Automobile Association (Usaa) | System for sound analysis and recognition |
US20170229125A1 (en) * | 2016-02-05 | 2017-08-10 | Honeywell International Inc. | Systems and methods for contacting emergency personnel via voice recognition |
US10062387B2 (en) * | 2016-02-05 | 2018-08-28 | Honeywell International Inc. | Systems and methods for contacting emergency personnel via voice recognition |
US10616393B2 (en) * | 2017-06-21 | 2020-04-07 | Guangdong Oppo Mobile Telecommunications Corp., Ltd. | Method for call control and related products |
WO2019045963A1 (en) * | 2017-08-30 | 2019-03-07 | T-Mobile Usa, Inc. | Message transcription for emergency call prioritization |
US10264121B2 (en) | 2017-08-30 | 2019-04-16 | T-Mobile Usa, Inc. | Message transcription for emergency call prioritization |
Also Published As
Publication number | Publication date |
---|---|
US9350873B2 (en) | 2016-05-24 |
US8908837B2 (en) | 2014-12-09 |
US20140294161A1 (en) | 2014-10-02 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US9350873B2 (en) | Methods and systems for automatically providing an emergency service call handler with context specific emergency service protocols | |
US11528772B2 (en) | Apparatus and method for obtaining emergency data related to emergency sessions | |
US9420099B1 (en) | Merging multiple emergency calls and information therefrom at emergency systems | |
US9798722B2 (en) | System and method for transmitting multiple text streams of a communication in different languages | |
US10715662B2 (en) | System and method for artificial intelligence on hold call handling | |
US8478229B2 (en) | Method and apparatus for notifying registered devices of an emergency call | |
US20150229766A1 (en) | Methods and Systems for Routing Emergency Service Calls Background | |
US8712757B2 (en) | Methods and apparatus for monitoring communication through identification of priority-ranked keywords | |
US11947592B2 (en) | Systems and method of generating custom messages based on rule-based database queries in a cloud platform | |
US9167090B2 (en) | Public safety answering point language detection | |
US11889398B2 (en) | Artificial intelligence for emergency assistance with human feedback for machine learning | |
US9767802B2 (en) | Methods and apparatus for conducting internet protocol telephony communications | |
US10255919B2 (en) | Identifying speaker roles in a streaming environment | |
WO2016132631A1 (en) | Communication system and communication method | |
KR102504076B1 (en) | Methods, systems, mobile communication devices, packet-switched communication systems, programs and computer program products for enhanced initiation and/or routing of emergency sessions in packet-switched communication systems | |
US20220201456A1 (en) | Emergency session translation and transcription via audio forking and machine learning | |
US11570403B2 (en) | Automated recording highlights for conferences | |
US9374465B1 (en) | Multi-channel and multi-modal language interpretation system utilizing a gated or non-gated configuration | |
US20160295016A1 (en) | Splitting a call for an emergent event into multiple devices using data channels | |
US20230136309A1 (en) | Virtual Assistant For Task Identification | |
EP3585039B1 (en) | System and method for recording and reviewing mixed-media communications | |
US20230239397A1 (en) | 911 Call Enhancement | |
KR20220157141A (en) | Method and Apparatus for Managing on Reporting data | |
US11909914B2 (en) | Forwarding emergency messages from IoT devices to PSAPs | |
CN115174750B (en) | DTMF signal transmission method and electronic equipment |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: SYNERGEM TECHNOLOGIES INC, NORTH CAROLINA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:O'CONNOR, STEPHEN F;MAW, RICHARD A;REEL/FRAME:034213/0418 Effective date: 20130329 |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YR, SMALL ENTITY (ORIGINAL EVENT CODE: M2551); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY Year of fee payment: 4 |
|
FEPP | Fee payment procedure |
Free format text: MAINTENANCE FEE REMINDER MAILED (ORIGINAL EVENT CODE: REM.); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY |