US20150213304A1 - Verifying Presence of a Person During an Electronic Visitation - Google Patents
Verifying Presence of a Person During an Electronic Visitation Download PDFInfo
- Publication number
- US20150213304A1 US20150213304A1 US14/683,655 US201514683655A US2015213304A1 US 20150213304 A1 US20150213304 A1 US 20150213304A1 US 201514683655 A US201514683655 A US 201514683655A US 2015213304 A1 US2015213304 A1 US 2015213304A1
- Authority
- US
- United States
- Prior art keywords
- image
- user
- captured image
- corporeal
- resident
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 claims abstract description 115
- 230000008569 process Effects 0.000 claims abstract description 70
- 230000001815 facial effect Effects 0.000 claims abstract description 58
- 238000001514 detection method Methods 0.000 claims abstract description 54
- 230000004044 response Effects 0.000 claims abstract description 41
- 238000012545 processing Methods 0.000 claims description 27
- 238000013500 data storage Methods 0.000 claims description 13
- 230000009471 action Effects 0.000 claims description 12
- 238000012795 verification Methods 0.000 claims description 9
- 238000004891 communication Methods 0.000 description 26
- 238000010586 diagram Methods 0.000 description 12
- 230000006870 function Effects 0.000 description 4
- 230000002093 peripheral effect Effects 0.000 description 4
- 230000005236 sound signal Effects 0.000 description 4
- 230000002452 interceptive effect Effects 0.000 description 3
- 238000004519 manufacturing process Methods 0.000 description 3
- 238000005259 measurement Methods 0.000 description 3
- 239000000203 mixture Substances 0.000 description 3
- 238000012544 monitoring process Methods 0.000 description 3
- 230000000295 complement effect Effects 0.000 description 2
- 238000012790 confirmation Methods 0.000 description 2
- 238000005516 engineering process Methods 0.000 description 2
- 238000003384 imaging method Methods 0.000 description 2
- 239000004973 liquid crystal related substance Substances 0.000 description 2
- 238000007726 management method Methods 0.000 description 2
- 230000003287 optical effect Effects 0.000 description 2
- 230000008520 organization Effects 0.000 description 2
- 239000004065 semiconductor Substances 0.000 description 2
- LFQSCWFLJHTTHZ-UHFFFAOYSA-N Ethanol Chemical compound CCO LFQSCWFLJHTTHZ-UHFFFAOYSA-N 0.000 description 1
- 229920005372 Plexiglas® Polymers 0.000 description 1
- 230000004075 alteration Effects 0.000 description 1
- 230000005540 biological transmission Effects 0.000 description 1
- 239000003795 chemical substances by application Substances 0.000 description 1
- 238000010276 construction Methods 0.000 description 1
- 238000013501 data transformation Methods 0.000 description 1
- 230000001066 destructive effect Effects 0.000 description 1
- 239000003814 drug Substances 0.000 description 1
- 229940079593 drug Drugs 0.000 description 1
- 239000000835 fiber Substances 0.000 description 1
- 238000005286 illumination Methods 0.000 description 1
- 230000000366 juvenile effect Effects 0.000 description 1
- 239000000463 material Substances 0.000 description 1
- 230000004630 mental health Effects 0.000 description 1
- 230000000474 nursing effect Effects 0.000 description 1
- 229920003023 plastic Polymers 0.000 description 1
- 239000004926 polymethyl methacrylate Substances 0.000 description 1
- 230000001902 propagating effect Effects 0.000 description 1
- 238000009877 rendering Methods 0.000 description 1
- 230000000284 resting effect Effects 0.000 description 1
- 238000012552 review Methods 0.000 description 1
- 230000003068 static effect Effects 0.000 description 1
- 229920000638 styrene acrylonitrile Polymers 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
- 230000001360 synchronised effect Effects 0.000 description 1
- 239000004416 thermosoftening plastic Substances 0.000 description 1
- 229920006352 transparent thermoplastic Polymers 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
- 230000002747 voluntary effect Effects 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M3/00—Automatic or semi-automatic exchanges
- H04M3/42—Systems providing special services or facilities to subscribers
- H04M3/42025—Calling or Called party identification service
- H04M3/42034—Calling party identification service
- H04M3/42059—Making use of the calling party identifier
-
- G06K9/00255—
-
- G06K9/00268—
-
- G06K9/00288—
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/161—Detection; Localisation; Normalisation
- G06V40/166—Detection; Localisation; Normalisation using acquisition arrangements
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N7/00—Television systems
- H04N7/14—Systems for two-way working
- H04N7/141—Systems for two-way working between two video terminals, e.g. videophone
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M2203/00—Aspects of automatic or semi-automatic exchanges
- H04M2203/60—Aspects of automatic or semi-automatic exchanges related to security aspects in telephonic communication systems
- H04M2203/6054—Biometric subscriber identification
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M3/00—Automatic or semi-automatic exchanges
- H04M3/22—Arrangements for supervision, monitoring or testing
- H04M3/2281—Call monitoring, e.g. for law enforcement purposes; Call tracing; Detection or prevention of malicious calls
Definitions
- Embodiments of the present invention are directed, in general, to interactive computing devices, more specifically to methods and systems for verifying the presence of an actual person during an electronic visitation, and in particular embodiments, to methods and systems for verifying the presence of an authorized person or persons during an electronic visitation.
- Controlled-environment facilities such as correctional facilities provide various options for visitation with residents (inmates). Options include in-person visitation, room-to-room visitation through a divider, telephone visitation, and video visitation. residents of controlled-environment facilities are typically restricted to receiving visitation from approved persons.
- the individual's identity may be determined by providing identification documents to staff of the controlled-environment facility for verification.
- the identification documents may include a picture of the individual.
- the staff member may cross-reference the individual's name with a list of individuals on the inmate's approved visitor list.
- Identification of visitors is more difficult with telephone or video visitation.
- the individual may provide a personal identification number, phrase, or pass code, but it is often difficult to ascertain whether the person providing the identifying information is in fact the authorized visitor without visual confirmation of the person's identity.
- an authorized visitor may pass identification information to unauthorized individuals so that they may pose as an authorized visitor for the electronic visitation.
- it may be difficult to confirm that an actual person is taking part in a video visitation.
- Embodiments of the present invention are directed to methods, apparatuses, and systems, which verify the presence of a person, or persons, during an electronic visitation.
- a method may include receiving a request to initiate an electronic visitation session and capturing an image, with an image capture device, of a user to provide a captured image in response to the request.
- the image capture device may be a component of a video visitation system located within a controlled-environment facility, and/or a component of an electronic device located outside of the controlled-environment facility.
- the user may be prompted to face the image capture device in response to receiving the request in accordance with some embodiments, and/or the image capture device may be configured to capture a still-frame image and/or a video frame to provide the captured image of the user in accordance with various embodiments.
- This captured image may be stored in a data storage device.
- a feature detection process may be performed, with a processor, on the captured image, to verify that images of features of a corporeal human face are present in the captured image to thereby determine whether an image of a corporeal human face is present in the captured image.
- This feature detection process may utilize three-dimensional (3D) feature detection process.
- the electronic visitation session may be connected in response to a determination that an image of a corporeal human face is present in the captured image.
- Various embodiments may call for performing a responsive action in response to a determination that an image of a corporeal human face is not present in the captured image. Further embodiments may include preforming a facial recognition process on the captured image, following performance of the feature detection process and/or in response to a determination that an image of a corporeal human face is present in the captured image, to identify the user and/or to confirm an identity of the user.
- an embodiment of a video visitation system or apparatus may include an image capture device configured to capture an image of a user, to provide a captured image of the user, such as in response to a request to initiate an electronic visitation session, and a processing device coupled to the image capture device.
- This processing device may be configured to perform a feature detection process on the captured image, to detect images of corporeal human face features present in the captured image, to thereby verify that an image of a corporeal human face is present in the captured image.
- the processing device may be further configured to connect the electronic visitation session in response to verification that that an image of a corporeal human face is present in the captured image.
- the processing device initiates a responsive action in response to a failure to verify that an image of a corporeal human face is present in the captured image.
- the image capture device may be a component of a video visitation system located within a controlled-environment facility and/or an image capture device may be a component of an electronic device located outside of the controlled-environment facility.
- the processing device in some embodiments may be further configured to prompt the user to face the image capture device in response to receiving the request, and/or the image capture device may be further configured to capture a still-frame image and/or a video frame to provide the captured image of the user.
- the processing device may also be configured to store the captured image in a data storage device.
- the processing device may utilize a 3D feature detection process to carry out the feature detection process. Further, the processing device may be configured to perform a facial recognition process on the captured image, following performance of the feature detection process and/or in response to a failure to verify that an image of a corporeal human face is present in the captured image, to identify the user and/or to confirm an identity of the user.
- an embodiment of a computer implemented method for verifying presence of authorized person(s) during an electronic visitation may include capturing images of residents of a controlled-environment facility at the time of intake or registration of the residents into the controlled-environment facility.
- a feature detection process may be performed on each of the images of each of the residents to verify that face detection-useable and/or facial recognition-usable images of corporeal facial features are present in each of the images of each of the residents.
- a copy of each image of each resident in which face detection-useable and/or facial recognition-usable images of corporeal facial features are present may be stored to a data storage device in response to verification that images of corporeal facial features were present.
- a request to initiate an electronic visitation session by a user may be received and user identification information may be received from the user.
- an image of the user may be captured, with an image capture device, such as in response to the request, to thereby provide a captured user image.
- a facial recognition process may be performed on the captured user image to identify the user.
- the user may be a resident of the controlled-environment facility and a saved copy of an image of the resident, in which face detection-useable and/or facial recognition-usable images of corporeal facial features are present, may be used in performing a facial recognition process on the captured user image to identify the resident.
- This saved copy of the image of the resident may be one saved in response to verification that images of corporeal facial features were present at the time of intake or registration, such as discussed above.
- the user may be a non-resident of the controlled-environment facility.
- an image of the non-resident may be captured to provide a captured image of the non-resident.
- a feature detection process may be performed on the captured image of the non-resident to verify that images of features of a corporeal human face are present in the captured image, so as to determine whether an image of a corporeal human face is present in the captured image of the non-resident.
- a facial recognition process may be performed on the captured image of the non-resident to identify the non-resident.
- a tangible computer-readable storage medium may have program instructions stored thereon that, upon execution by one or more computer systems, cause the one or more computer systems to execute one or more operations disclosed herein.
- a system may include at least one processor and memory coupled to the at least one processor, the memory configured to store program instructions executable by the at least one processor to cause the system to execute one or more operations disclosed herein.
- FIG. 1 is a schematic block diagram illustrating one embodiment of an example system for verifying presence of a person and/or authorized person(s) during an electronic visitation.
- FIG. 2 is a schematic block diagram illustrating one embodiment of an example apparatus for verifying presence of a person and/or authorized person(s) during an electronic visitation.
- FIG. 3 is a schematic block diagram illustrating one embodiment of an example processing device configurable as an apparatus for verifying presence of a person and/or authorized person(s) during video visitation, in accordance with some embodiments.
- FIG. 4 is a schematic flowchart diagram illustrating one embodiment of a method for verifying presence of a person during an electronic visitation.
- FIG. 7 is a mockup diagram illustrating one embodiment of an example image frame captured by an example image capture device.
- a controlled-environment facility may be referred to as a jail or prison, and its residents may be referred to as residents, arrestees, detainees, or inmates. It should be understood, however, that the systems and methods described herein may be similarly applicable to other types of controlled-environment facilities and their respective residents (e.g., a hospital and its patients). Other embodiments may be suitable for use in other public places, for example at ATM devices or at airport check-in kiosks.
- embodiments of the present invention are directed, in general to interactive computing devices, more specifically to methods and systems for verifying the presence of an actual person during an electronic visitation, and in particular embodiments to methods and systems for verifying the presence of an authorized person or persons during an electronic visitation.
- images of residents of a controlled-environment facility may be captured at the time of intake into the facility.
- a feature detection process may be performed on each image to verify that corporeal facial features are present in each resident image and a copy of such an image of each resident may be saved.
- a request to initiate an electronic visitation session may be received and an image of a user is captured.
- Face detection may be used alone, without the use of facial recognition, in fields such as video surveillance, human computer interface, image database management, and/or in accordance with various embodiments of the present systems and methods. Face detection may be a fast, less computational-intensive process, which may be performed by a small local computer system, without the need to necessarily access databases, more powerful computational resources, or the like, which may be necessary for facial recognition.
- devices 102 , 103 and 104 may be capable of connecting to a non-resident's (i.e., a person not committed to a controlled-environment facility) device 105 or telephone 106 across a publicly switched telephone network (PSTN) 107 .
- PSTN publicly switched telephone network
- device 105 may be a mobile phone, whereas telephone 106 may be located at a non-resident's home, inmate visitation center, etc.
- Switch 108 in communication processing system 101 may be used to connect calls across PSTN 107 .
- the non-resident may be at telephone 109 or device 112 , which is on Voice-over-IP (VoIP), or packet data network 110 , such as, for example the Internet.
- Router 111 of communication system 101 is used to route data packets associated with a call connection to destination telephone 109 or device 112 .
- communication processing system 101 may attempt to ensure that a resident's calls, video conferences, online chats, etc. are performed only with non-residents whose identities, devices, email addresses, phone numbers, etc. are listed in that resident's Pre-Approved Contact (PAC) list.
- PAC Pre-Approved Contact
- Each resident's PAC list may be stored, for example, in database 115 maintained by Administration and Management System (AMS) 116 .
- AMS 116 may also store inmate or resident profile data (RPD), as well as resident financial data, such as commissary accounts.
- Visitation system 130 may be configured to provide, schedule, and manage visitation services to residents and non-residents of a controlled-environment facility.
- visitation system 130 may be coupled to PSTN 107 and/or network 110 in a manner similar to communication processing system 101 by also including one or more gateways, switches and/or routers (not shown).
- visitation system 130 may be configured to communicate with one or more residents of the controlled-environment facility via devices 102 - 104 and with one or more non-residents via devices 105 , 106 , 109 , 112 , and/or 113 .
- visitation system 130 is shown in FIG. 1 as being remotely located with respect to communication processing system 101 (and the controlled-environment facility), in other cases visitation system 130 may be co-located with the facility and/or integrated within system 101 .
- communication system 101 may be configured to perform video visitation monitoring operations configured to monitor and or record video visitations (e.g., as electronic video files). In scenarios where communication system 101 is located within the controlled-environment facility, it may have direct access to AMS 116 . In other embodiments, however, communication system 101 may be located remotely with respect to the controlled-environment facility, and access to AMS 116 may be obtained via a computer network such as, for example, network 110 .
- intelligent facility devices 103 may be configured to capture a video image of a resident to be transmitted to a non-resident using the camera, and to display a video image of the non-resident to the resident using the display. Intelligent facility devices 103 may also be configured to capture an audio signal from the resident to be transmitted to a non-resident using the mouthpiece portion of the handset, and to provide an audio signal from the non-resident to the resident using the earpiece portion of the handset. Additionally or alternatively, audio received from the non-resident may be reproduced via a loudspeaker, and audio provided by the resident may be captured via a microphone.
- intelligent facility devices 103 may assume the form of any computer, tablet computer, smart phone, etc., or any other consumer device or appliance with videoconferencing capabilities.
- FIG. 2 is a diagram of a video visitation device, such as an intelligent facility device 103 .
- device 103 may be implemented as a computer-based system such as described in connection with FIG. 3 .
- device 103 includes display 204 , camera 205 , and handset 201 coupled to device 103 via wire 209 .
- Display 204 may be any suitable electronic display such as, for example, a Liquid Crystal Display (LCD), a touchscreen display (e.g., resistive, capacitive, etc.), or the like, whereas camera 205 may be any suitable imaging device such as, for instance, a video camera or webcam equipped with Charge-Coupled Devices (CCDs), Complementary Metal-Oxide-Semiconductor (CMOS) active pixel sensors, etc.
- Handset 201 may be similar to a traditional telephone handset including an earpiece portion (with a speaker), a handle portion, and a mouthpiece portion (with a microphone).
- handset holder or holster 208 may be configured to allow a user to securely rest handset 201 against device 103 (e.g., in an “off” position) when device 103 is not in use.
- handset holder 208 may include a mechanical or electromechanical switch or sensor (not shown) that senses when handset 201 is not resting against it (e.g. in its “on” position).
- Lighting or light sensor 207 may include one or more suitable sensors such as, for example, photoresistors, photodiodes, etc.
- camera 205 may serve as light sensor, and a (separate) light sensor may be absent.
- video visitation device 103 may include one or more tampering/impact resistant or hardened elements configured to protect them from vandalism or otherwise destructive acts.
- one or more of devices 202 - 207 may be protected by a transparent plastic or thermoplastic (e.g., Plexiglas, etc.) material.
- the video visitation device 103 may include a keypad 210 or other user input device to allow a user to enter information.
- the user may dial a telephone number associated with a recipient of a video visitation call, enter an account number for billing purposes, or enter a Personal Identification Number (PIN), inmate number, or the like, for authentication and/or identification purposes.
- PIN Personal Identification Number
- FIG. 3 is a schematic block diagram illustrating one embodiment of a processing device configurable for use according to the present embodiments.
- system 300 may be a server, a workstation, a kiosk, an intelligent facility device, a desktop computer, a laptop, a tablet computer, a mobile device, a smart phone, or the like.
- system 300 may be used to implement various ones of controlled-environment facility communication processing system 101 , intelligent facility devise 103 , AMS 116 , visitation system 130 , and/or the like.
- computer system 300 includes one or more processors 310 A-N coupled to a system memory 320 via an input/output (I/O) interface 330 .
- Computer system 300 further includes a network interface 340 coupled to I/O interface 330 , and one or more input/output devices 350 , such as cursor control device 360 , keyboard 370 , display(s) 380 , or mobile device(s) 390 .
- I/O interface 330 may include a touch-sensitive screen.
- Other devices may include, for example, image capture devices 395 , surveillance cameras, microphones, antennas/wireless transducers, phone detection modules, etc.
- each of visitation system 130 , AMS 116 , communication processing system 101 , devices 102 through 104 , and/or devices 105 , 106 , 109 , 112 , and 113 may be implemented using a single instance of computer system 300 , while in other embodiments multiple such systems, or multiple nodes making up computer system 300 , may be configured to host different portions of a given electronic device or system.
- computer system 300 may be a single-processor system including one processor 310 , or a multi-processor system including two or more processors 310 (e.g., two, four, eight, or another suitable number).
- Processors 310 may be any processor capable of executing program instructions.
- processors 310 may be general-purpose or embedded processors implementing any of a variety of instruction set architectures (ISAs), such as the x86, PowerPC®, ARM®, SPARC®, or MIPS® ISAs, or any other suitable ISA.
- ISAs instruction set architectures
- each of processors 310 may commonly, but not necessarily, implement the same ISA.
- at least one processor 310 may be a graphics processing unit (GPU) or other dedicated graphics-rendering device.
- GPU graphics processing unit
- System memory 320 may be configured to store program instructions and/or data accessible by processor 310 .
- system memory 320 may be implemented using any suitable memory technology, such as static random access memory (SRAM), synchronous dynamic RAM (SDRAM), nonvolatile/Flash-type memory, or any other type of memory.
- SRAM static random access memory
- SDRAM synchronous dynamic RAM
- program instructions and data implementing certain operations such as, for example, those described below in connection with FIGS. 4-7 , may be stored within system memory 320 as program instructions 325 and data storage 335 , respectively.
- program instructions and/or data may be received, sent or stored upon different types of computer-accessible media or on similar media separate from system memory 320 or computer system 300 .
- Program instructions and data stored on a tangible computer-accessible storage medium in non-transitory form may further be transmitted by transmission media or signals such as electrical, electromagnetic, or digital signals, which may be conveyed via a communication medium such as a network and/or a wireless link.
- transmission media or signals such as electrical, electromagnetic, or digital signals, which may be conveyed via a communication medium such as a network and/or a wireless link.
- I/O interface 330 may be configured to coordinate I/O traffic between processor 310 , system memory 320 , and any peripheral devices in the device, including network interface 340 or other peripheral interfaces, such as input/output devices 350 .
- I/O interface 330 may perform any necessary protocol, timing or other data transformations to convert data signals from one component (e.g., system memory 320 ) into a format suitable for use by another component (e.g., processor 310 ).
- I/O interface 330 may include support for devices attached through various types of peripheral buses, such as a variant of the Peripheral Component Interconnect (PCI) bus standard or the Universal Serial Bus (USB) standard, for example.
- PCI Peripheral Component Interconnect
- USB Universal Serial Bus
- memory 320 may include program instructions 325 , configured to implement certain embodiments described herein, and data storage 335 , comprising various data, which may be accessible, by program instructions 325 .
- program instructions 325 may include software elements of embodiments illustrated in the above figures.
- program instructions 325 may be implemented in various embodiments using any desired programming language, scripting language, or combination of programming languages and/or scripting languages (e.g., C, C++, C#, JavaTM, JavaScriptTM, Perl, etc.).
- Data storage 335 may include data that may be used in these embodiments (e.g., recorded communications, profiles for different modes of operations, etc.). In other embodiments, other or different software elements and data may be included.
- computer system 300 is merely illustrative and is not intended to limit the scope of the disclosure described herein.
- the computer system and devices may include any combination of hardware or software that can perform the indicated operations.
- the operations performed by the illustrated components may, in some embodiments, be performed by fewer components or distributed across additional components.
- the operations of some of the illustrated components may not be provided and/or other additional operations may be available. Accordingly, systems and methods described herein may be implemented or executed with other computer system configurations.
- FIG. 4 is a schematic flowchart diagram illustrating one embodiment of a method 400 for verifying presence of a person during an electronic visitation.
- the visitation system 130 receives a request to initiate an electronic visitation session as shown at block 401 .
- the request may be received from either an intelligent facility device 103 a telephone 106 , 109 , a device 105 , 112 , a laptop computer 113 or other suitable communication device. Accordingly, the request may be initiated by either a resident in the controlled-environment facility or a remote user.
- the image may be captured by camera 114 , webcam 205 , image capture device 395 , or the like, in various embodiments, inside or outside the controlled-environment facility, accordingly, and may be saved to a storage device, such as AMS database 115 , data storage 335 , and/or the like.
- a processor 310 may perform a face detection process at 403 .
- a feature detection process may be performed on the captured image, such as to verify that images of features of a corporeal human face are present in the captured image, to thereby determine whether an image of a corporeal human face is present in the captured image.
- a three-dimensional (3D) feature detection process may be performed at 403 on the captured image to verify that an actual face is present in the image.
- Such a 3D feature detection process may identify three-dimensional characteristics of an image, including measurements of features of a face at a plurality of points on the image.
- changes in the measurements from frame to frame of a video image may indicate that the person is an actual person and not simply a photograph presented to trick the system.
- a plurality of still frame photograms may be captured and differences in measurements may be calculated to determine if the presented person is an actual person or a photograph.
- face detection may be used alone, without the use of facial recognition. As described above, such a face detection process differs from a facial recognition process.
- Face detection may be a fast, less computational-intensive process, which may be performed by small local computer systems, such as intelligent facility device 103 , nonresident devices 112 , 113 , or the like, without the need to necessarily access (outside) databases, more powerful computational resources, or the like, which may be necessary for (full) facial recognition.
- Assorted embodiments may call for performing a responsive action in answer to a determination at 403 that an image of a corporeal human face is not present in the captured image. Further embodiments may further include preforming a facial recognition process on the captured image, following performance of the feature detection process at 403 and/or in response to a determination that an image of a corporeal human face is present in the captured image at 403 , such as to identify the user and/or to confirm an identity of the user.
- the processing device may be configured to perform a facial recognition process on the captured image, to identify the user and/or to confirm an identity of the user, such as in a situation where a video visitation system user has attempted to deceive the system, or the like.
- an image capture guide such as an oval overlaid on a video image, may be displayed to instruct the user regarding a position for presenting his/her face for image capture.
- the image capture device 395 may then capture an image or a series of images or video frames as shown at block 503 .
- Processor 310 of visitation system 130 may then perform a (3D) facial recognition process on the image(s) at 504 , to affirmatively identify the user, rather than only performing a face detection process using feature detection, such as discussed above with respect to FIG. 4 .
- This facial recognition process may be used to confirm that the user is an authorized user.
- the visitation system 130 may verify that the resident present matches a PIN entered by the resident and that the remote user is a member of the resident's PAC list. If the face matches an authorized user as shown at block 505 , then the visitation system 130 may connect the parties on an electronic visitation session as shown at block 506 .
- an electronic visitation session could include a telephone call, a video visitation, an instant message visitation, etc.
- the visitation session may be denied or disconnected as shown at block 509 .
- the captured images may be stored, as shown in block 510 , in a data storage device, such as database 115 , for further review.
- Additional responsive actions may include sounding an alert, generating a notification to the users and/or an administrator of the visitation system 130 . Additionally, responsive action may include monitoring of the visitation session by a live agent, recording of the visitation session, etc.
- One of ordinary skill in the art may recognize additional responsive actions which may be taken.
- FIG. 6 is a mockup diagram illustrating one embodiment of an apparatus for verifying presence of a person or authorized persons during an electronic visitation during use. This embodiment may also further demonstrate methods 400 and 500 described in FIGS. 4 and 5 .
- the method is carried out, at least in part, by an intelligent facility device 103 .
- the intelligent facility device activates the webcam 205 and requires that the resident show his/her face 601 .
- the visitation system 130 may monitor webcam 205 to ensure that the actual face 601 of the (authorized) person is still present, for example, while interacting with content 602 on intelligent facility device 103 . If the monitoring application no longer detects an actual (authorized) face 601 within camera field of view 603 , the existing login session and display information may be closed.
- This solution may utilizes the integrated webcam 205 of a resident intelligent facility device 103 and utilizes algorithms for detecting (3D) facial features to verify that a human face is presented to the camera 205 .
- face detection may be required before and after logging into select functions. While logged into one or more select functions, the resident may maintain their face 601 within the view of the webcam 205 .
- a small window may be presented, into which is displayed the image being captured by the webcam 205 to assist the resident in properly placing their face in view of the webcam 205 .
- facial identification processes may be performed.
- face detection may be employed, but the face may be matched against a known likeness of the resident, non-resident, or others, such as known associates of the resident, or the like, using a facial recognition process.
- a match to the known face of the resident may act not only as a means of maintaining the electronic visitation session, but also as an additional level of confirmation of the resident's identity so as to preclude unauthorized use.
- FIG. 7 illustrates a mockup of a captured image 700 .
- the captured image 700 may include a face 601 of a user.
- the image may include a second face on a second person 703 in the background 702 of the image 700 .
- the visitation system 130 may perform facial recognition on any faces present in the image 700 to ensure that all parties to the visitation are authorized to participate and to ensure that no other residents are eavesdropping on the electronic visitation session.
- FIG. 8 is a flowchart of another embodiment of a process ( 800 ) for verifying presence of a person and/or authorized persons during an electronic visitation.
- images of residents of a controlled-environment facility are captured at 801 , such as at the time of intake or registration of the residents into the controlled-environment facility, for example during booking at a jail, or the like.
- a feature detection process may be performed at 802 on each of the images of each of the residents to verify that face detection-useable and/or facial recognition-usable images of corporeal facial features are present in each of the images of each the residents. That is, a feature detection process may be performed at 802 on each image to identify images where corporeal facial features are present for each the residents.
- a copy of at least one image of each resident in which face detection-useable and/or facial recognition-usable images of corporeal facial features are present may be stored to a data storage device, in response to verification at 802 that images of corporeal facial features were present.
- a request to initiate an electronic visitation session by a user may be received at 804 and user identification information may be received from the user at 805 .
- an image may be captured at 806 , with an image capture device, of the user to provide a captured user image in response to the request.
- a face detection process and/or a facial recognition process may be performed on the captured user image to identify the user at 807 , such as the party identified at 805 .
- this user may be a resident of the controlled-environment facility.
- a saved copy of an image of the resident saved at 803 in response to verification at 802 that images of corporeal facial features were present at the time of intake or registration may be used in performing a facial recognition process on the captured user image to identify the resident at 807 , for example as the party identified at 805 .
- the user may be a non-resident of the controlled-environment facility.
- an image of the non-resident may be captured at 806 to provide a captured image of the non-resident, whereupon a feature detection process may be performed on the captured image of the non-resident at 808 to verify that images of features of a corporeal human face are present in the captured image so as to determine whether an image of a corporeal human face is present in the captured image of the non-resident.
- a facial recognition process may be performed on the captured image of the non-resident at 807 to identify the non-resident.
- a facial recognition process at 807 may be carried out on the captured image, following performance of the feature detection process at 808 and/or in response to a determination that an image of a corporeal human face is present in the captured image at 808 , such as to identify the user and/or to confirm an identity of the user, such as the party identified at 805 .
- Various embodiments may call for performing a responsive action in response to a determination at 808 that an image of a corporeal human face is not present in the image captured at 806 and/or as a result of recognition of a face at 807 other than the party identified at 805 .
Landscapes
- Engineering & Computer Science (AREA)
- Signal Processing (AREA)
- Multimedia (AREA)
- Health & Medical Sciences (AREA)
- General Health & Medical Sciences (AREA)
- Oral & Maxillofacial Surgery (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Image Analysis (AREA)
Abstract
Description
- This is a continuation-in-part of U.S. patent application Ser. No. 14/152,831, entitled Verifying Presence of Authorized Persons During an Electronic Visitation, filed Jan. 10, 2014, which is incorporated herein by reference.
- Embodiments of the present invention are directed, in general, to interactive computing devices, more specifically to methods and systems for verifying the presence of an actual person during an electronic visitation, and in particular embodiments, to methods and systems for verifying the presence of an authorized person or persons during an electronic visitation.
- Controlled-environment facilities, such as correctional facilities provide various options for visitation with residents (inmates). Options include in-person visitation, room-to-room visitation through a divider, telephone visitation, and video visitation. Residents of controlled-environment facilities are typically restricted to receiving visitation from approved persons.
- When an individual visits an inmate in person, the individual's identity may be determined by providing identification documents to staff of the controlled-environment facility for verification. The identification documents may include a picture of the individual. The staff member may cross-reference the individual's name with a list of individuals on the inmate's approved visitor list.
- Identification of visitors is more difficult with telephone or video visitation. The individual may provide a personal identification number, phrase, or pass code, but it is often difficult to ascertain whether the person providing the identifying information is in fact the authorized visitor without visual confirmation of the person's identity. For example, an authorized visitor may pass identification information to unauthorized individuals so that they may pose as an authorized visitor for the electronic visitation. Moreover, it may be difficult to confirm that an actual person is taking part in a video visitation.
- Embodiments of the present invention are directed to methods, apparatuses, and systems, which verify the presence of a person, or persons, during an electronic visitation. In one embodiment, a method may include receiving a request to initiate an electronic visitation session and capturing an image, with an image capture device, of a user to provide a captured image in response to the request. The image capture device may be a component of a video visitation system located within a controlled-environment facility, and/or a component of an electronic device located outside of the controlled-environment facility. The user may be prompted to face the image capture device in response to receiving the request in accordance with some embodiments, and/or the image capture device may be configured to capture a still-frame image and/or a video frame to provide the captured image of the user in accordance with various embodiments. This captured image may be stored in a data storage device. A feature detection process may be performed, with a processor, on the captured image, to verify that images of features of a corporeal human face are present in the captured image to thereby determine whether an image of a corporeal human face is present in the captured image. This feature detection process may utilize three-dimensional (3D) feature detection process. The electronic visitation session may be connected in response to a determination that an image of a corporeal human face is present in the captured image. Various embodiments may call for performing a responsive action in response to a determination that an image of a corporeal human face is not present in the captured image. Further embodiments may include preforming a facial recognition process on the captured image, following performance of the feature detection process and/or in response to a determination that an image of a corporeal human face is present in the captured image, to identify the user and/or to confirm an identity of the user.
- Hence, an embodiment of a video visitation system or apparatus may include an image capture device configured to capture an image of a user, to provide a captured image of the user, such as in response to a request to initiate an electronic visitation session, and a processing device coupled to the image capture device. This processing device may be configured to perform a feature detection process on the captured image, to detect images of corporeal human face features present in the captured image, to thereby verify that an image of a corporeal human face is present in the captured image. The processing device may be further configured to connect the electronic visitation session in response to verification that that an image of a corporeal human face is present in the captured image.
- In accordance with some embodiments the processing device initiates a responsive action in response to a failure to verify that an image of a corporeal human face is present in the captured image. In various embodiments the image capture device may be a component of a video visitation system located within a controlled-environment facility and/or an image capture device may be a component of an electronic device located outside of the controlled-environment facility. The processing device, in some embodiments may be further configured to prompt the user to face the image capture device in response to receiving the request, and/or the image capture device may be further configured to capture a still-frame image and/or a video frame to provide the captured image of the user. The processing device may also be configured to store the captured image in a data storage device. The processing device may utilize a 3D feature detection process to carry out the feature detection process. Further, the processing device may be configured to perform a facial recognition process on the captured image, following performance of the feature detection process and/or in response to a failure to verify that an image of a corporeal human face is present in the captured image, to identify the user and/or to confirm an identity of the user.
- Additionally or alternatively, an embodiment of a computer implemented method for verifying presence of authorized person(s) during an electronic visitation may include capturing images of residents of a controlled-environment facility at the time of intake or registration of the residents into the controlled-environment facility. A feature detection process may be performed on each of the images of each of the residents to verify that face detection-useable and/or facial recognition-usable images of corporeal facial features are present in each of the images of each of the residents. Whereupon, a copy of each image of each resident in which face detection-useable and/or facial recognition-usable images of corporeal facial features are present may be stored to a data storage device in response to verification that images of corporeal facial features were present.
- Thereafter, a request to initiate an electronic visitation session by a user may be received and user identification information may be received from the user. Thereupon, an image of the user may be captured, with an image capture device, such as in response to the request, to thereby provide a captured user image. A facial recognition process may be performed on the captured user image to identify the user. In accordance with some embodiments, the user may be a resident of the controlled-environment facility and a saved copy of an image of the resident, in which face detection-useable and/or facial recognition-usable images of corporeal facial features are present, may be used in performing a facial recognition process on the captured user image to identify the resident. This saved copy of the image of the resident may be one saved in response to verification that images of corporeal facial features were present at the time of intake or registration, such as discussed above.
- In some embodiments the user may be a non-resident of the controlled-environment facility. In some such embodiments, an image of the non-resident may be captured to provide a captured image of the non-resident. Whereupon, a feature detection process may be performed on the captured image of the non-resident to verify that images of features of a corporeal human face are present in the captured image, so as to determine whether an image of a corporeal human face is present in the captured image of the non-resident. Additionally or alternatively, a facial recognition process may be performed on the captured image of the non-resident to identify the non-resident.
- In various embodiments, one or more of the techniques described herein may be performed by one or more computer systems. In other various embodiments, a tangible computer-readable storage medium may have program instructions stored thereon that, upon execution by one or more computer systems, cause the one or more computer systems to execute one or more operations disclosed herein. In yet other various embodiments, a system may include at least one processor and memory coupled to the at least one processor, the memory configured to store program instructions executable by the at least one processor to cause the system to execute one or more operations disclosed herein.
- The foregoing has outlined rather broadly the features and technical advantages of the present invention in order that the detailed description of the invention that follows may be better understood. Additional features and advantages of the invention will be described hereinafter which form the subject of the claims of the invention. It should be appreciated that the conception and specific embodiment disclosed may be readily utilized as a basis for modifying or designing other structures for carrying out the same purposes of the present invention. It should also be realized that such equivalent constructions do not depart from the invention as set forth in the appended claims. The novel features which are believed to be characteristic of the invention, both as to its organization and method of operation, together with further objects and advantages will be better understood from the following description when considered in connection with the accompanying figures. It is to be expressly understood, however, that each of the figures is provided for the purpose of illustration and description only and is not intended as a definition of the limits of the present invention.
- Having thus described the invention in general terms, reference will now be made to the accompanying drawings, which are not necessarily drawn to scale, and wherein:
-
FIG. 1 is a schematic block diagram illustrating one embodiment of an example system for verifying presence of a person and/or authorized person(s) during an electronic visitation. -
FIG. 2 is a schematic block diagram illustrating one embodiment of an example apparatus for verifying presence of a person and/or authorized person(s) during an electronic visitation. -
FIG. 3 is a schematic block diagram illustrating one embodiment of an example processing device configurable as an apparatus for verifying presence of a person and/or authorized person(s) during video visitation, in accordance with some embodiments. -
FIG. 4 is a schematic flowchart diagram illustrating one embodiment of a method for verifying presence of a person during an electronic visitation. -
FIG. 5 is a schematic flowchart diagram illustrating an embodiment of a method for verifying authorized person(s) during an electronic visitation. -
FIG. 6 is a mockup diagram illustrating one embodiment of an apparatus for verifying presence of a person and/or authorized person(s) during an electronic visitation during use. -
FIG. 7 is a mockup diagram illustrating one embodiment of an example image frame captured by an example image capture device. -
FIG. 8 is a flowchart of another embodiment of a process for verifying presence of a person and/or authorized person(s) during an electronic visitation. - The invention now will be described more fully hereinafter with reference to the accompanying drawings. This invention may, however, be embodied in many different forms and should not be construed as limited to the embodiments set forth herein. Rather, these embodiments are provided so that this disclosure will be thorough and complete, and will fully convey the scope of the invention to those skilled in the art. One skilled in the art may be able to use the various embodiments of the invention.
- This specification discloses methods and systems for verifying the presence of one or more actual and/or authorized persons during an electronic visitation. The various techniques described herein may find applicability in a wide variety of controlled-environment facilities (as well as outside of controlled-environment facilities). Various types of controlled-environment facilities are present in today's society, and persons may be voluntary or involuntary residents of such facilities, whether temporarily or permanently. Examples of controlled-environment facilities may include correctional institutions or facilities (e.g., municipal jails, county jails, state prisons, federal prisons, military stockades, juvenile facilities, detention camps, home incarceration environments, etc.), healthcare facilities (e.g., hospitals, nursing homes, mental health facilities, rehabilitation clinics, such as drug and alcohol rehabilitation facilities, etc.), restricted living quarters (e.g., hotels, resorts, camps, dormitories, barracks, etc.), and the like.
- For convenience of explanation, various examples discussed herein are presented in the context of correctional facilities. For instance, in some of the embodiments discussed below, a controlled-environment facility may be referred to as a jail or prison, and its residents may be referred to as residents, arrestees, detainees, or inmates. It should be understood, however, that the systems and methods described herein may be similarly applicable to other types of controlled-environment facilities and their respective residents (e.g., a hospital and its patients). Other embodiments may be suitable for use in other public places, for example at ATM devices or at airport check-in kiosks.
- As noted, embodiments of the present invention are directed, in general to interactive computing devices, more specifically to methods and systems for verifying the presence of an actual person during an electronic visitation, and in particular embodiments to methods and systems for verifying the presence of an authorized person or persons during an electronic visitation. In accordance with embodiments of the present systems and methods, images of residents of a controlled-environment facility may be captured at the time of intake into the facility. A feature detection process may be performed on each image to verify that corporeal facial features are present in each resident image and a copy of such an image of each resident may be saved. Sometime thereafter, and/or in other embodiments, a request to initiate an electronic visitation session may be received and an image of a user is captured. A feature detection process may be performed on the captured image to verify that images of features of a corporeal human face are present in the captured image, determining whether an image of a corporeal human face is present in the captured image. The electronic visitation session may then be connected in response to a determination that an image of a corporeal human face is present. In accordance with these and/or further embodiments, a facial recognition process may also be performed on the captured image, such as through use of the saved resident intake image.
- Hence, embodiments of the present systems and methods may make use of “face detection” processes, and/or further or various other embodiments may make use of “facial recognition” processes. Face detection is a computer technology that identifies human faces in digital images. Face detection can be regarded as a specific case of object-class detection. In object-class detection, the task is to find the locations and sizes of all objects in an image that belong to a given class. Face detection algorithms focus on the detection of frontal human faces. Face detection is used in biometrics, often as a part of, or together with, a facial recognition system. Hence, face detection may be used to detect human faces, which might then be used for recognizing a particular face in a facial recognition process. However, such face detection may be used alone, without the use of facial recognition, in fields such as video surveillance, human computer interface, image database management, and/or in accordance with various embodiments of the present systems and methods. Face detection may be a fast, less computational-intensive process, which may be performed by a small local computer system, without the need to necessarily access databases, more powerful computational resources, or the like, which may be necessary for facial recognition.
- Turning now to
FIG. 1 , a block diagram of an illustrative environment where verifying presence of a person and/or authorized persons during an electronic visitation may be employed is depicted according to some embodiments. As shown,communication processing system 101 may provide telephone services, videoconferencing, online chat, and other communication services to a controlled-environment facility. For example, in some cases,communication system 101 may be co-located with a controlled-environment facility. Alternatively,communication system 101 may be centrally or remotely located with respect to one or more controlled-environment facilities and/or may provide communication services to multiple controlled-environment facilities. More generally, however, it should be noted thatcommunication system 101 may assume a variety of forms, and may be configured to serve a variety of facilities and/or users, whether within or outside of a controlled-environment facility. - In the context of a correctional facility, for instance, inmates may use
telephones 102 to access certain communication services. In some facilities, inmates may also use a personalcomputer wireless device 104 to access such services. For example, an inmate may initiate telephone services by lifting the receiver ontelephone 102, at which time the inmate may be prompted to provide a personal identification number (PIN), other identifying information or biometrics. An interactive voice response (IVR) unit (not shown) may generate and play prompt or other messages to inmates ondevices 102 through 104. - Under the control of
communication processing system 101,devices device 105 ortelephone 106 across a publicly switched telephone network (PSTN) 107. For example,device 105 may be a mobile phone, whereastelephone 106 may be located at a non-resident's home, inmate visitation center, etc.Switch 108 incommunication processing system 101 may be used to connect calls acrossPSTN 107. Additionally or alternatively, the non-resident may be attelephone 109 ordevice 112, which is on Voice-over-IP (VoIP), orpacket data network 110, such as, for example the Internet.Router 111 ofcommunication system 101 is used to route data packets associated with a call connection todestination telephone 109 ordevice 112. -
Intelligent facility devices 103 a-n (each collectively referred to as “Intelligent facility device 103”) may have video conferencing capabilities to enable residents to participate in video visitation sessions with non-residents of the controlled-environment facility via video call, secure online chat, etc. In one embodiment,intelligent facility devices 103 may include touch-screen devices. In some embodiments,intelligent facility devices 103 may include an image capture device, such as a webcam. For example, a non-resident party may have a personal orlaptop computer 113 with camera 114 (or a cell phone, tablet computer, etc.). Additionally or alternatively,device 112 may have an integrated camera and touch-screen display (e.g., a smart phone, tablet, etc.). A network connection between the parties may be established and supported by an organization or commercial service that provides computer services and software for use in telecommunications and/or VOIP, such as SKYPE®. Additionally or alternatively, the correctional facility and/or the destination may use videoconferencing equipment compatible with ITU H.323, H.320, H.264, and/or V.80, or other suitable standards. Generally speaking, eachintelligent facility device 103 may be disposed in a visitation room, in a pod, kiosk, etc. - In addition to providing certain visitation and communication operations,
communication processing system 101 may attempt to ensure that a resident's calls, video conferences, online chats, etc. are performed only with non-residents whose identities, devices, email addresses, phone numbers, etc. are listed in that resident's Pre-Approved Contact (PAC) list. Each resident's PAC list may be stored, for example, indatabase 115 maintained by Administration and Management System (AMS) 116. In addition to PAC list(s),AMS 116 may also store inmate or resident profile data (RPD), as well as resident financial data, such as commissary accounts. - As an example, in the context of a correctional facility,
database 115 may include information such as balances for inmate trust and calling accounts; trial schedule; conviction data; criminal record; sentencing data, such as time served, time remaining to be served, and release date; cell and cellmate assignments; inmate restrictions and warnings; commissary order history; telephone call history; call recordings; known or suspected gang or criminal affiliations; known or suspected affiliates, accomplices, or gang members; and any other information that may be relevant or useful to correctional facility staff to house and maintain inmates. -
Visitation system 130 may be configured to provide, schedule, and manage visitation services to residents and non-residents of a controlled-environment facility. To that end,visitation system 130 may be coupled toPSTN 107 and/ornetwork 110 in a manner similar tocommunication processing system 101 by also including one or more gateways, switches and/or routers (not shown). Accordingly,visitation system 130 may be configured to communicate with one or more residents of the controlled-environment facility via devices 102-104 and with one or more non-residents viadevices visitation system 130 is shown inFIG. 1 as being remotely located with respect to communication processing system 101 (and the controlled-environment facility), in othercases visitation system 130 may be co-located with the facility and/or integrated withinsystem 101. - In some implementations,
communication system 101 may be configured to perform video visitation monitoring operations configured to monitor and or record video visitations (e.g., as electronic video files). In scenarios wherecommunication system 101 is located within the controlled-environment facility, it may have direct access toAMS 116. In other embodiments, however,communication system 101 may be located remotely with respect to the controlled-environment facility, and access toAMS 116 may be obtained via a computer network such as, for example,network 110. - In some embodiments,
intelligent facility devices 103 may be implemented as a computer-based system. For example, each ofintelligent facility devices 103 may include a display, camera, and handset. The display may be any suitable electronic display such as, for example, a Liquid Crystal Display (LCD), a touchscreen display (e.g., resistive, capacitive, etc.), or the like, whereas the camera may be a suitable imaging device such as, for instance, a video camera or webcam equipped with Charge-Coupled Devices (CCDs), Complementary Metal-Oxide-Semiconductor (CMOS) active pixel sensors, etc. A handset may be similar to a traditional telephone handset including an earpiece portion (with a loudspeaker), a handle portion, and a mouthpiece portion (with a microphone). - During a video visitation session,
intelligent facility devices 103 may be configured to capture a video image of a resident to be transmitted to a non-resident using the camera, and to display a video image of the non-resident to the resident using the display.Intelligent facility devices 103 may also be configured to capture an audio signal from the resident to be transmitted to a non-resident using the mouthpiece portion of the handset, and to provide an audio signal from the non-resident to the resident using the earpiece portion of the handset. Additionally or alternatively, audio received from the non-resident may be reproduced via a loudspeaker, and audio provided by the resident may be captured via a microphone. - In addition to video visitation,
intelligent facility devices 103 may also be used by the resident to access a resident portal, which allows access tocommunication processing system 101. The access available to the resident through the resident portal may be limited or restricted according to the resident's permissions on thesystem 101. In some embodiments, the resident may be able to access his/her personal account information, make purchases, check electronic communications such as email, and the like. - In some cases,
intelligent facility devices 103 may assume the form of any computer, tablet computer, smart phone, etc., or any other consumer device or appliance with videoconferencing capabilities. -
FIG. 2 is a diagram of a video visitation device, such as anintelligent facility device 103. In some embodiments,device 103 may be implemented as a computer-based system such as described in connection withFIG. 3 . As illustrated,device 103 includesdisplay 204,camera 205, andhandset 201 coupled todevice 103 viawire 209.Display 204 may be any suitable electronic display such as, for example, a Liquid Crystal Display (LCD), a touchscreen display (e.g., resistive, capacitive, etc.), or the like, whereascamera 205 may be any suitable imaging device such as, for instance, a video camera or webcam equipped with Charge-Coupled Devices (CCDs), Complementary Metal-Oxide-Semiconductor (CMOS) active pixel sensors, etc.Handset 201 may be similar to a traditional telephone handset including an earpiece portion (with a speaker), a handle portion, and a mouthpiece portion (with a microphone). - During a video visitation session,
video visitation device 103 may be configured to capture a video image of a resident to be transmitted to anon-resident using camera 205, and to display a video image of the non-resident to theresident using display 204.Video visitation device 103 may also be configured to capture an audio signal from the resident to be transmitted to the non-resident using the mouthpiece portion ofhandset 201, and to provide an audio signal from the non-resident to the resident using the earpiece portion ofhandset 201. Additionally or alternatively, audio received from the non-resident may be reproduced vialoudspeaker 202, and audio provided by the resident may be captured viamicrophone 203. In some embodiments, handset holder orholster 208 may be configured to allow a user to securely resthandset 201 against device 103 (e.g., in an “off” position) whendevice 103 is not in use. Furthermore,handset holder 208 may include a mechanical or electromechanical switch or sensor (not shown) that senses whenhandset 201 is not resting against it (e.g. in its “on” position). -
Video visitation device 103 may also includelighting device 206 andlighting sensor 207. In some embodiments,lighting device 206 may include one or more incandescent light sources (e.g., bulbs, lamps, etc.), fluorescent lamps, Compact Fluorescent Lamps (CFL), Cold Cathode Fluorescent Lamps (CCFL), high-intensity discharge sources, Light-Emitting Diodes (LEDs), or the like. Additionally or alternatively,lighting device 206 may include one or more adjustable lighting devices with variable intensity, brightness, luminance, color, tone, etc. For example, in certain embodiments,lighting device 206 may include an infrared (IR) light source (e.g. IR LED(s)), which may be used as a covert means to provide light to capture an image of a resident and/or the surroundings, such as without the resident or other around knowing the image is being captured. Lighting orlight sensor 207 may include one or more suitable sensors such as, for example, photoresistors, photodiodes, etc. In various implementations,camera 205 may serve as light sensor, and a (separate) light sensor may be absent. - Furthermore, it should be noted that
video visitation device 103 may assume the form of any computer, tablet computer, smart phone, etc., or any other consumer device or appliance with videoconferencing capabilities. For example, in the case wherevideo visitation device 103 is a camera-enabled television (or smart phone, etc.) that does not include a dedicated lightning apparatus, the brightness ofdisplay 204 may be adjusted so thatdisplay 204 may itself act as lighting device 206 (e.g., the brightness ofdisplay 204 may be increased to also increase the illumination of the person(s) “watching” the teleconference through the television, for example, upon detection that a video connection has been established). - In the context of correctional facilities or the like,
video visitation device 103 may include one or more tampering/impact resistant or hardened elements configured to protect them from vandalism or otherwise destructive acts. For example, one or more of devices 202-207 may be protected by a transparent plastic or thermoplastic (e.g., Plexiglas, etc.) material. - In one embodiment, the
video visitation device 103 may include akeypad 210 or other user input device to allow a user to enter information. For example, the user may dial a telephone number associated with a recipient of a video visitation call, enter an account number for billing purposes, or enter a Personal Identification Number (PIN), inmate number, or the like, for authentication and/or identification purposes. -
FIG. 3 is a schematic block diagram illustrating one embodiment of a processing device configurable for use according to the present embodiments. In various embodiments,system 300 may be a server, a workstation, a kiosk, an intelligent facility device, a desktop computer, a laptop, a tablet computer, a mobile device, a smart phone, or the like. In some cases,system 300 may be used to implement various ones of controlled-environment facilitycommunication processing system 101, intelligent facility devise 103,AMS 116,visitation system 130, and/or the like. - As illustrated,
computer system 300 includes one ormore processors 310A-N coupled to asystem memory 320 via an input/output (I/O)interface 330.Computer system 300 further includes anetwork interface 340 coupled to I/O interface 330, and one or more input/output devices 350, such ascursor control device 360,keyboard 370, display(s) 380, or mobile device(s) 390. In an embodiment, I/O interface 330 may include a touch-sensitive screen. Other devices may include, for example,image capture devices 395, surveillance cameras, microphones, antennas/wireless transducers, phone detection modules, etc. In some embodiments, each ofvisitation system 130,AMS 116,communication processing system 101,devices 102 through 104, and/ordevices computer system 300, while in other embodiments multiple such systems, or multiple nodes making upcomputer system 300, may be configured to host different portions of a given electronic device or system. - In various embodiments,
computer system 300 may be a single-processor system including one processor 310, or a multi-processor system including two or more processors 310 (e.g., two, four, eight, or another suitable number). Processors 310 may be any processor capable of executing program instructions. For example, in various embodiments, processors 310 may be general-purpose or embedded processors implementing any of a variety of instruction set architectures (ISAs), such as the x86, PowerPC®, ARM®, SPARC®, or MIPS® ISAs, or any other suitable ISA. In multi-processor systems, each of processors 310 may commonly, but not necessarily, implement the same ISA. Also, in some embodiments, at least one processor 310 may be a graphics processing unit (GPU) or other dedicated graphics-rendering device. -
System memory 320 may be configured to store program instructions and/or data accessible by processor 310. In various embodiments,system memory 320 may be implemented using any suitable memory technology, such as static random access memory (SRAM), synchronous dynamic RAM (SDRAM), nonvolatile/Flash-type memory, or any other type of memory. As illustrated, program instructions and data implementing certain operations such as, for example, those described below in connection withFIGS. 4-7 , may be stored withinsystem memory 320 asprogram instructions 325 anddata storage 335, respectively. In other embodiments, program instructions and/or data may be received, sent or stored upon different types of computer-accessible media or on similar media separate fromsystem memory 320 orcomputer system 300. - Generally speaking, a computer-accessible medium may include any tangible or non-transitory storage media or memory media such as electronic, magnetic, or optical media—e.g., disk or CD/DVD-ROM coupled to
computer system 300 via I/O interface 330. The terms “tangible” and “non-transitory,” as used herein, are intended to describe a computer-readable storage medium (or “memory”) excluding propagating electromagnetic signals, but are not intended to otherwise limit the type of physical computer-readable storage device that is encompassed by the phrase computer-readable medium or memory. For instance, the terms “non-transitory computer-readable medium” or “tangible memory” are intended to encompass types of storage devices that do not necessarily store information permanently, including for example, random access memory (RAM). Program instructions and data stored on a tangible computer-accessible storage medium in non-transitory form may further be transmitted by transmission media or signals such as electrical, electromagnetic, or digital signals, which may be conveyed via a communication medium such as a network and/or a wireless link. - In an embodiment, I/
O interface 330 may be configured to coordinate I/O traffic between processor 310,system memory 320, and any peripheral devices in the device, includingnetwork interface 340 or other peripheral interfaces, such as input/output devices 350. In some embodiments, I/O interface 330 may perform any necessary protocol, timing or other data transformations to convert data signals from one component (e.g., system memory 320) into a format suitable for use by another component (e.g., processor 310). In some embodiments, I/O interface 330 may include support for devices attached through various types of peripheral buses, such as a variant of the Peripheral Component Interconnect (PCI) bus standard or the Universal Serial Bus (USB) standard, for example. In some embodiments, the function of I/O interface 330 may be split into two or more separate components, such as a north bridge and a south bridge, for example. In addition, in some embodiments some or all of the functionality of I/O interface 330, such as an interface tosystem memory 320, may be incorporated directly into processor 310. -
Network interface 340 may be configured to allow data to be exchanged betweencomputer system 300 and other devices attached to a network, such as other computer systems, or between nodes ofcomputer system 300. In various embodiments,network interface 340 may support communication via wired or wireless general data networks, such as any suitable type of Ethernet network, for example; via telecommunications/telephony networks such as analog voice networks or digital fiber communications networks; via storage area networks such as Fibre-Channel SANs, or via any other suitable type of network and/or protocol. - Input/
output devices 350 may, in some embodiments, include one or more display terminals, keyboards, keypads, touchpads, touch screens, scanning devices, voice or optical recognition devices, mobile devices, or any other devices suitable for entering or retrieving data by one ormore computer system 300. Multiple input/output devices 350 may be present incomputer system 300 or may be distributed on various nodes ofcomputer system 300. In some embodiments, similar input/output devices may be separate fromcomputer system 300 and may interact with one or more nodes ofcomputer system 300 through a wired or wireless connection, such as overnetwork interface 340. - As shown in
FIG. 3 ,memory 320 may includeprogram instructions 325, configured to implement certain embodiments described herein, anddata storage 335, comprising various data, which may be accessible, byprogram instructions 325. In an embodiment,program instructions 325 may include software elements of embodiments illustrated in the above figures. For example,program instructions 325 may be implemented in various embodiments using any desired programming language, scripting language, or combination of programming languages and/or scripting languages (e.g., C, C++, C#, Java™, JavaScript™, Perl, etc.).Data storage 335 may include data that may be used in these embodiments (e.g., recorded communications, profiles for different modes of operations, etc.). In other embodiments, other or different software elements and data may be included. - A person of ordinary skill in the art will appreciate that
computer system 300 is merely illustrative and is not intended to limit the scope of the disclosure described herein. In particular, the computer system and devices may include any combination of hardware or software that can perform the indicated operations. In addition, the operations performed by the illustrated components may, in some embodiments, be performed by fewer components or distributed across additional components. Similarly, in other embodiments, the operations of some of the illustrated components may not be provided and/or other additional operations may be available. Accordingly, systems and methods described herein may be implemented or executed with other computer system configurations. -
FIG. 4 is a schematic flowchart diagram illustrating one embodiment of a method 400 for verifying presence of a person during an electronic visitation. In an embodiment, thevisitation system 130 receives a request to initiate an electronic visitation session as shown atblock 401. In such an embodiment, the request may be received from either an intelligent facility device 103 atelephone device laptop computer 113 or other suitable communication device. Accordingly, the request may be initiated by either a resident in the controlled-environment facility or a remote user. - In an embodiment, the
visitation system 130 may instruct the user's device to capture an image of the user in response to the request as shown atblock 402. To do so, thevisitation system 130 may prompt the user to present his/her face for image capture. In a further embodiment, thevisitation system 130 may instruct each users' device to capture an image, so that each user may be authenticated and authorized to participate in the electronic visitation. The captured image may include either a still frame photograph or a video image. A single video frame may be captured or a plurality of video frames may be recorded. The image may be captured bycamera 114,webcam 205,image capture device 395, or the like, in various embodiments, inside or outside the controlled-environment facility, accordingly, and may be saved to a storage device, such asAMS database 115,data storage 335, and/or the like. - In an embodiment, a processor 310, such as of
visitation system 130, may perform a face detection process at 403. As shown atblock 403, a feature detection process may be performed on the captured image, such as to verify that images of features of a corporeal human face are present in the captured image, to thereby determine whether an image of a corporeal human face is present in the captured image. For example, a three-dimensional (3D) feature detection process, as illustrated, or the like, may be performed at 403 on the captured image to verify that an actual face is present in the image. Such a 3D feature detection process may identify three-dimensional characteristics of an image, including measurements of features of a face at a plurality of points on the image. In certain embodiments, changes in the measurements from frame to frame of a video image may indicate that the person is an actual person and not simply a photograph presented to trick the system. In still another embodiment, a plurality of still frame photograms may be captured and differences in measurements may be calculated to determine if the presented person is an actual person or a photograph. Hence, in accordance with embodiments of the present systems and methods, such face detection may be used alone, without the use of facial recognition. As described above, such a face detection process differs from a facial recognition process. Face detection may be a fast, less computational-intensive process, which may be performed by small local computer systems, such asintelligent facility device 103,nonresident devices - In a further embodiment, the
visitation system 130 may connect the requesting parties in an electronic visitation in response to a determination that an actual face was presented in the image as shown atblock 404. That is, the electronic visitation session may be connected at 404 in response to a determination, at 403, that an image of a corporeal human face is present in the captured image, in accordance with various embodiments of the present systems and methods. - Assorted embodiments may call for performing a responsive action in answer to a determination at 403 that an image of a corporeal human face is not present in the captured image. Further embodiments may further include preforming a facial recognition process on the captured image, following performance of the feature detection process at 403 and/or in response to a determination that an image of a corporeal human face is present in the captured image at 403, such as to identify the user and/or to confirm an identity of the user. Additionally or alternatively, in response to a failure to verify that an image of a corporeal human face is present in the captured image, the processing device may be configured to perform a facial recognition process on the captured image, to identify the user and/or to confirm an identity of the user, such as in a situation where a video visitation system user has attempted to deceive the system, or the like.
-
FIG. 5 illustrates a different embodiment of a method 500 for verifying authorized persons during an electronic visitation. In this embodiment, thevisitation system 130, for example, receives a request to initiate an electronic visitation session, such as a video visitation session, as shown atblock 501. Thevisitation system 130 may then prompt the user to present his/her face for image capture as shown atblock 502. For example, thevisitation system 130 may play an audio prompt instruction the user to position his/her face in front of the camera. In an alternative embodiment, a video visitation device, such asintelligent facility device 103 orlaptop 113 may display a textual prompt to the user. In still a further embodiment, an image capture guide, such as an oval overlaid on a video image, may be displayed to instruct the user regarding a position for presenting his/her face for image capture. Theimage capture device 395 may then capture an image or a series of images or video frames as shown atblock 503. - Processor 310 of
visitation system 130 may then perform a (3D) facial recognition process on the image(s) at 504, to affirmatively identify the user, rather than only performing a face detection process using feature detection, such as discussed above with respect toFIG. 4 . This facial recognition process may be used to confirm that the user is an authorized user. For example, thevisitation system 130 may verify that the resident present matches a PIN entered by the resident and that the remote user is a member of the resident's PAC list. If the face matches an authorized user as shown atblock 505, then thevisitation system 130 may connect the parties on an electronic visitation session as shown atblock 506. For example, an electronic visitation session could include a telephone call, a video visitation, an instant message visitation, etc. If the faces don't match authorized persons, then the visitation session may be denied or disconnected as shown atblock 509. In such embodiments, the captured images may be stored, as shown inblock 510, in a data storage device, such asdatabase 115, for further review. - If the visitation session is connected at
block 506, then a second image of the users may be captured at a second time as shown atblock 507. The second time may be at a predetermined time after connection of the visitation session. Alternatively, the second time may be randomly selected to avoid predictability. The second image may be processed according to the (3D) facial recognition process, and it may be further determined whether the face in the second image matches an authorized user as shown atblock 508. If the face does not match an authorized user, then a responsive action may be taken. For example, the visitation session may be disconnected byvisitation system 130 as shown atblock 509. In such an embodiment, the second captured image may also be stored in a data storage device as shown atblock 510. Additional responsive actions may include sounding an alert, generating a notification to the users and/or an administrator of thevisitation system 130. Additionally, responsive action may include monitoring of the visitation session by a live agent, recording of the visitation session, etc. One of ordinary skill in the art may recognize additional responsive actions which may be taken. -
FIG. 6 is a mockup diagram illustrating one embodiment of an apparatus for verifying presence of a person or authorized persons during an electronic visitation during use. This embodiment may also further demonstrate methods 400 and 500 described inFIGS. 4 and 5 . In this embodiment, the method is carried out, at least in part, by anintelligent facility device 103. When a resident attempts to participate in an electronic visitation session, the intelligent facility device activates thewebcam 205 and requires that the resident show his/herface 601. During the use of the electronic visitation session, thevisitation system 130 may monitorwebcam 205 to ensure that theactual face 601 of the (authorized) person is still present, for example, while interacting withcontent 602 onintelligent facility device 103. If the monitoring application no longer detects an actual (authorized) face 601 within camera field ofview 603, the existing login session and display information may be closed. - This solution may utilizes the
integrated webcam 205 of a residentintelligent facility device 103 and utilizes algorithms for detecting (3D) facial features to verify that a human face is presented to thecamera 205. In such an embodiment, face detection may be required before and after logging into select functions. While logged into one or more select functions, the resident may maintain theirface 601 within the view of thewebcam 205. - In a further embodiment, a small window may be presented, into which is displayed the image being captured by the
webcam 205 to assist the resident in properly placing their face in view of thewebcam 205. - In still a further embodiment, facial identification processes may be performed. Within this variation, not only may face detection be employed, but the face may be matched against a known likeness of the resident, non-resident, or others, such as known associates of the resident, or the like, using a facial recognition process. A match to the known face of the resident may act not only as a means of maintaining the electronic visitation session, but also as an additional level of confirmation of the resident's identity so as to preclude unauthorized use.
-
FIG. 7 illustrates a mockup of a capturedimage 700. As illustrated, the capturedimage 700 may include aface 601 of a user. In a further embodiment, the image may include a second face on asecond person 703 in thebackground 702 of theimage 700. In an embodiment, thevisitation system 130 may perform facial recognition on any faces present in theimage 700 to ensure that all parties to the visitation are authorized to participate and to ensure that no other residents are eavesdropping on the electronic visitation session. -
FIG. 8 is a flowchart of another embodiment of a process (800) for verifying presence of a person and/or authorized persons during an electronic visitation. Therein, images of residents of a controlled-environment facility are captured at 801, such as at the time of intake or registration of the residents into the controlled-environment facility, for example during booking at a jail, or the like. A feature detection process may be performed at 802 on each of the images of each of the residents to verify that face detection-useable and/or facial recognition-usable images of corporeal facial features are present in each of the images of each the residents. That is, a feature detection process may be performed at 802 on each image to identify images where corporeal facial features are present for each the residents. Whereupon, at 803, a copy of at least one image of each resident in which face detection-useable and/or facial recognition-usable images of corporeal facial features are present may be stored to a data storage device, in response to verification at 802 that images of corporeal facial features were present. - Thereafter, in accordance with some embodiments, a request to initiate an electronic visitation session by a user may be received at 804 and user identification information may be received from the user at 805. Thereupon, an image may be captured at 806, with an image capture device, of the user to provide a captured user image in response to the request. A face detection process and/or a facial recognition process may be performed on the captured user image to identify the user at 807, such as the party identified at 805.
- In accordance with some embodiments, this user may be a resident of the controlled-environment facility. In such embodiments a saved copy of an image of the resident saved at 803 in response to verification at 802 that images of corporeal facial features were present at the time of intake or registration may be used in performing a facial recognition process on the captured user image to identify the resident at 807, for example as the party identified at 805.
- In some additional or alternative embodiments, the user may be a non-resident of the controlled-environment facility. In some such embodiments, an image of the non-resident may be captured at 806 to provide a captured image of the non-resident, whereupon a feature detection process may be performed on the captured image of the non-resident at 808 to verify that images of features of a corporeal human face are present in the captured image so as to determine whether an image of a corporeal human face is present in the captured image of the non-resident. Additionally or alternatively, a facial recognition process may be performed on the captured image of the non-resident at 807 to identify the non-resident. For example, a facial recognition process at 807 may be carried out on the captured image, following performance of the feature detection process at 808 and/or in response to a determination that an image of a corporeal human face is present in the captured image at 808, such as to identify the user and/or to confirm an identity of the user, such as the party identified at 805.
- Various embodiments may call for performing a responsive action in response to a determination at 808 that an image of a corporeal human face is not present in the image captured at 806 and/or as a result of recognition of a face at 807 other than the party identified at 805.
- Although the present invention and its advantages have been described in detail, it should be understood that various changes, substitutions and alterations can be made herein without departing from the spirit and scope of the invention as defined by the appended claims. Moreover, the scope of the present application is not intended to be limited to the particular embodiments of the process, machine, manufacture, composition of matter, means, methods and steps described in the specification. As one of ordinary skill in the art will readily appreciate from the disclosure of the present invention, processes, machines, manufacture, compositions of matter, means, methods, or steps, presently existing or later to be developed that perform substantially the same function or achieve substantially the same result as the corresponding embodiments described herein may be utilized according to the present invention. Accordingly, the appended claims are intended to include within their scope such processes, machines, manufacture, compositions of matter, means, methods, or steps.
Claims (27)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/683,655 US10296784B2 (en) | 2014-01-10 | 2015-04-10 | Verifying presence of a person during an electronic visitation |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/152,831 US9007420B1 (en) | 2014-01-10 | 2014-01-10 | Verifying presence of authorized persons during an electronic visitation |
US14/683,655 US10296784B2 (en) | 2014-01-10 | 2015-04-10 | Verifying presence of a person during an electronic visitation |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/152,831 Continuation-In-Part US9007420B1 (en) | 2014-01-10 | 2014-01-10 | Verifying presence of authorized persons during an electronic visitation |
Publications (2)
Publication Number | Publication Date |
---|---|
US20150213304A1 true US20150213304A1 (en) | 2015-07-30 |
US10296784B2 US10296784B2 (en) | 2019-05-21 |
Family
ID=53679358
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/683,655 Active 2034-12-03 US10296784B2 (en) | 2014-01-10 | 2015-04-10 | Verifying presence of a person during an electronic visitation |
Country Status (1)
Country | Link |
---|---|
US (1) | US10296784B2 (en) |
Cited By (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10097538B1 (en) * | 2017-08-12 | 2018-10-09 | Growpath, Inc. | User authentication systems and methods |
US10699810B2 (en) * | 2017-07-20 | 2020-06-30 | Kenneth W. STEWART | Employee visit verification system |
CN111835531A (en) * | 2020-07-30 | 2020-10-27 | 腾讯科技(深圳)有限公司 | Session processing method, device, computer equipment and storage medium |
WO2020233071A1 (en) * | 2019-05-22 | 2020-11-26 | 深圳壹账通智能科技有限公司 | Person-certificate verification method and device, computer device and computer readable storage medium |
US20210201610A1 (en) * | 2017-11-03 | 2021-07-01 | Sensormatic Electronics, LLC | Methods and System for Distributed Cameras and Demographics Analysis |
US11157601B2 (en) * | 2017-08-03 | 2021-10-26 | Morphotrust Usa, Llc | Electronic identity verification |
US11201899B1 (en) * | 2015-09-02 | 2021-12-14 | Confinement Telephony Technology, Llc | Systems and methods for secure, controlled virtual visitation with confinement institution inmates |
US11561580B1 (en) * | 2018-03-06 | 2023-01-24 | Securus Technologies, Llc | Controlled-environment facility communication terminal and personal computer wireless device docking station with integral keypads |
CN115801480A (en) * | 2022-11-10 | 2023-03-14 | 杭州乐舜信息科技有限公司 | Network telephone terminal charging method and device considering individual differences of students |
US11757958B1 (en) | 2015-09-02 | 2023-09-12 | Confinement Telephony Technology, Llc | Systems and methods for secure, controlled virtual visitation with confinement institution inmates |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20080000966A1 (en) * | 2006-06-30 | 2008-01-03 | Evercom Systems, Inc. | System and method for electronic visitation registration |
US20120019620A1 (en) * | 2010-07-20 | 2012-01-26 | Hon Hai Precision Industry Co., Ltd. | Image capture device and control method |
US20120262271A1 (en) * | 2011-04-18 | 2012-10-18 | Richard Torgersrud | Interactive audio/video system and device for use in a secure facility |
US20120281058A1 (en) * | 2011-05-04 | 2012-11-08 | Eyconix, Llc | Virtual Inmate Visitation |
US20140279789A1 (en) * | 2013-03-15 | 2014-09-18 | Telmate, Llc | Predicting an identity of a person based on an activity history |
US9007420B1 (en) * | 2014-01-10 | 2015-04-14 | Securus Technologies, Inc. | Verifying presence of authorized persons during an electronic visitation |
Family Cites Families (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7650291B2 (en) | 1999-06-23 | 2010-01-19 | Koninklijke Philips Electronics N.V. | Video visitation system and method for a health care location |
US6219640B1 (en) | 1999-08-06 | 2001-04-17 | International Business Machines Corporation | Methods and apparatus for audio-visual speaker recognition and utterance verification |
US8218829B2 (en) | 2001-08-20 | 2012-07-10 | Polycom, Inc. | System and method for using biometrics technology in conferencing |
US7436988B2 (en) | 2004-06-03 | 2008-10-14 | Arizona Board Of Regents | 3D face authentication and recognition based on bilateral symmetry analysis |
US7256816B2 (en) | 2004-10-25 | 2007-08-14 | 3V Technologies Incorporated | Systems and processes for scheduling and conducting audio/video communications |
US20080201158A1 (en) | 2007-02-15 | 2008-08-21 | Johnson Mark D | System and method for visitation management in a controlled-access environment |
US9064257B2 (en) | 2010-11-02 | 2015-06-23 | Homayoon Beigi | Mobile device transaction using multi-factor authentication |
US9124763B2 (en) | 2012-02-01 | 2015-09-01 | Homewav, Llc | Method and apparatus for providing both audio/video visitation and VOIP telephonic visitation originated either by an inmate or by an outside visitor directly between inmates of a prison and an outside visitor without need of intervention by prison personnel |
-
2015
- 2015-04-10 US US14/683,655 patent/US10296784B2/en active Active
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20080000966A1 (en) * | 2006-06-30 | 2008-01-03 | Evercom Systems, Inc. | System and method for electronic visitation registration |
US20120019620A1 (en) * | 2010-07-20 | 2012-01-26 | Hon Hai Precision Industry Co., Ltd. | Image capture device and control method |
US20120262271A1 (en) * | 2011-04-18 | 2012-10-18 | Richard Torgersrud | Interactive audio/video system and device for use in a secure facility |
US20120281058A1 (en) * | 2011-05-04 | 2012-11-08 | Eyconix, Llc | Virtual Inmate Visitation |
US20140279789A1 (en) * | 2013-03-15 | 2014-09-18 | Telmate, Llc | Predicting an identity of a person based on an activity history |
US9007420B1 (en) * | 2014-01-10 | 2015-04-14 | Securus Technologies, Inc. | Verifying presence of authorized persons during an electronic visitation |
Cited By (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11201899B1 (en) * | 2015-09-02 | 2021-12-14 | Confinement Telephony Technology, Llc | Systems and methods for secure, controlled virtual visitation with confinement institution inmates |
US11757958B1 (en) | 2015-09-02 | 2023-09-12 | Confinement Telephony Technology, Llc | Systems and methods for secure, controlled virtual visitation with confinement institution inmates |
US12120160B1 (en) | 2015-09-02 | 2024-10-15 | Confinement Telephony Technology, Llc | Systems and methods for secure, controlled virtual visitation with confinement institution inmates |
US10699810B2 (en) * | 2017-07-20 | 2020-06-30 | Kenneth W. STEWART | Employee visit verification system |
US11157601B2 (en) * | 2017-08-03 | 2021-10-26 | Morphotrust Usa, Llc | Electronic identity verification |
US10097538B1 (en) * | 2017-08-12 | 2018-10-09 | Growpath, Inc. | User authentication systems and methods |
US11924197B1 (en) | 2017-08-12 | 2024-03-05 | Growpath, Llc | User authentication systems and methods |
US20210201610A1 (en) * | 2017-11-03 | 2021-07-01 | Sensormatic Electronics, LLC | Methods and System for Distributed Cameras and Demographics Analysis |
US11561580B1 (en) * | 2018-03-06 | 2023-01-24 | Securus Technologies, Llc | Controlled-environment facility communication terminal and personal computer wireless device docking station with integral keypads |
WO2020233071A1 (en) * | 2019-05-22 | 2020-11-26 | 深圳壹账通智能科技有限公司 | Person-certificate verification method and device, computer device and computer readable storage medium |
CN111835531A (en) * | 2020-07-30 | 2020-10-27 | 腾讯科技(深圳)有限公司 | Session processing method, device, computer equipment and storage medium |
CN115801480A (en) * | 2022-11-10 | 2023-03-14 | 杭州乐舜信息科技有限公司 | Network telephone terminal charging method and device considering individual differences of students |
Also Published As
Publication number | Publication date |
---|---|
US10296784B2 (en) | 2019-05-21 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US9007420B1 (en) | Verifying presence of authorized persons during an electronic visitation | |
US10296784B2 (en) | Verifying presence of a person during an electronic visitation | |
US9083850B1 (en) | Video blurring in a secure environment | |
US10284820B2 (en) | Covert monitoring and recording of audio and video in controlled-environment facilities | |
US10049543B1 (en) | Covert infrared monitoring and recording of images and video in controlled-environment facilities | |
US10455084B2 (en) | Inbound calls to intelligent controlled-environment facility resident media and/or communications devices | |
US9007425B1 (en) | Software-controlled lighting for video visitation devices | |
US9106789B1 (en) | Videoconference and video visitation security | |
US10142462B2 (en) | Controlled-environment facility resident communication detection | |
US8929525B1 (en) | On-demand video communication for controlled-environment facility residents | |
US10404941B1 (en) | Overt and covert capture of images of controlled-environment facility residents using intelligent controlled-environment facility resident communications and/or media devices | |
US9473497B1 (en) | Exclusion engine for electronic communications in controlled-environment facilities | |
US9380082B1 (en) | Remote multi-party and/or event-driven visitation for residents of controlled-environment facilities | |
US11100739B1 (en) | Two factor identification verification of controlled-environment facility residents and associated non-residents | |
US8855280B1 (en) | Communication detail records (CDRs) containing media for communications in controlled-environment facilities | |
US20110074911A1 (en) | Gesture-Based Actions in a Video Communication Session | |
US9667763B1 (en) | Family communications in a controlled-environment facility | |
US10972701B1 (en) | One-way video conferencing | |
US9736428B1 (en) | Providing remote visitation and other services to non-residents of controlled-environment facilities via display devices | |
US10645341B1 (en) | Three-way video visitation detection using video cues | |
US11194540B1 (en) | Controlled-environment facility digital bulletin board | |
US11509695B1 (en) | Management of controlled-environment facility resident image and/or background during video visitation | |
US9930171B1 (en) | Systems and methods for identification and action on attempted controlled-environment facility resident communications with internet-protocol phone numbers | |
US10205826B1 (en) | Inbound calls to intelligent controlled-environment facility resident media and/or communications devices | |
US20170142368A1 (en) | Video mail between residents of controlled-environment facilities and non-residents |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: SECURUS TECHNOLOGIES, INC., TEXAS Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:PASSE, SCOTT;REEL/FRAME:035382/0751 Effective date: 20150410 |
|
AS | Assignment |
Owner name: DEUTSCHE BANK AG NEW YORK BRANCH, AS COLLATERAL AGENT, NEW YORK Free format text: FIRST LIEN SECURITY AGREEMENT;ASSIGNORS:SECURUS TECHNOLOGIES HOLDINGS, INC.;SECURUS TECHNOLOGIES, INC.;SATELLITE TRACKING OF PEOPLE LLC;REEL/FRAME:044779/0573 Effective date: 20171101 Owner name: DEUTSCHE BANK AG NEW YORK BRANCH, AS COLLATERAL AG Free format text: FIRST LIEN SECURITY AGREEMENT;ASSIGNORS:SECURUS TECHNOLOGIES HOLDINGS, INC.;SECURUS TECHNOLOGIES, INC.;SATELLITE TRACKING OF PEOPLE LLC;REEL/FRAME:044779/0573 Effective date: 20171101 |
|
AS | Assignment |
Owner name: DEUTSCHE BANK AG NEW YORK BRANCH, AS COLLATERAL AGENT, NEW YORK Free format text: SECOND LIEN SECURITY AGREEMENT;ASSIGNORS:SECURUS TECHNOLOGIES HOLDINGS, INC.;SECURUS TECHNOLOGIES, INC.;SATELLITE TRACKING OF PEOPLE LLC;REEL/FRAME:044368/0969 Effective date: 20171101 Owner name: DEUTSCHE BANK AG NEW YORK BRANCH, AS COLLATERAL AG Free format text: SECOND LIEN SECURITY AGREEMENT;ASSIGNORS:SECURUS TECHNOLOGIES HOLDINGS, INC.;SECURUS TECHNOLOGIES, INC.;SATELLITE TRACKING OF PEOPLE LLC;REEL/FRAME:044368/0969 Effective date: 20171101 |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT VERIFIED |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
AS | Assignment |
Owner name: SECURUS TECHNOLOGIES, LLC, TEXAS Free format text: CHANGE OF NAME;ASSIGNOR:SECURUS TECHNOLOGIES, INC.;REEL/FRAME:053476/0289 Effective date: 20191213 |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 4 |
|
AS | Assignment |
Owner name: DEUTSCHE BANK AG NEW YORK BRANCH, NEW YORK Free format text: SECURITY INTEREST;ASSIGNORS:SECURUS TECHNOLOGIES, LLC;SATELLITE TRACKING OF PEOPLE LLC;REEL/FRAME:064600/0235 Effective date: 20230804 Owner name: DEUTSCHE BANK AG NEW YORK BRANCH, NEW YORK Free format text: SECURITY INTEREST;ASSIGNORS:SECURUS TECHNOLOGIES, LLC;SATELLITE TRACKING OF PEOPLE LLC;REEL/FRAME:064600/0451 Effective date: 20230804 |
|
AS | Assignment |
Owner name: THE BANK OF NEW YORK MELLON TRUST COMPANY, N.A., ILLINOIS Free format text: SECURITY INTEREST;ASSIGNORS:ALLPAID, INC.;SATELLITE TRACKING OF PEOPLE LLC;SECURUS TECHNOLOGIES, LLC;REEL/FRAME:064699/0842 Effective date: 20230626 |
|
AS | Assignment |
Owner name: DEUTSCHE BANK AG NEW YORK BRANCH, AS COLLATERAL AGENT, NEW YORK Free format text: PRIORITY FIRST LIEN PATENT SECURITY AGREEMENT;ASSIGNORS:SECURUS TECHNOLOGIES, LLC;ALLPAID, INC.;SATELLITE TRACKING OF PEOPLE LLC;REEL/FRAME:066567/0620 Effective date: 20231229 |
|
AS | Assignment |
Owner name: WILMINGTON SAVINGS FUND SOCIETY, FSB AS COLLATERAL AGENT, DELAWARE Free format text: PRIORITY SECOND LIEN PATENT SECURITY AGREEMENT;ASSIGNORS:SECURUS TECHNOLOGIES, LLC;SATELLITE TRACKING OF PEOPLE LLC;ALLPAID, INC.;REEL/FRAME:066436/0420 Effective date: 20231229 |
|
AS | Assignment |
Owner name: DEUTSCHE BANK AG NEW YORK BRANCH AS COLLATERAL AGENT, NEW YORK Free format text: SUPER-PRIORITY FIRST LIEN PATENT SECURITY AGREEMENT;ASSIGNORS:SECURUS TECHNOLOGIES, LLC;SATELLITE TRACKING OF PEOPLE LLC;ALLPAID, INC.;REEL/FRAME:066945/0310 Effective date: 20240328 |
|
AS | Assignment |
Owner name: WILMINGTON SAVINGS FUND SOCIETY, FSB, DELAWARE Free format text: SECURITY INTEREST;ASSIGNORS:AVENTIV TECHNOLOGIES, LLC;SECURUS TECHNOLOGIES, LLC;SATELLITE TRACKING OF PEOPLE LLC;AND OTHERS;REEL/FRAME:066951/0054 Effective date: 20240328 Owner name: ALLPAID, INC., CALIFORNIA Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:ALTER DOMUS (US) LLC;REEL/FRAME:066952/0914 Effective date: 20240328 Owner name: SECURUS TECHNOLOGIES, LLC, CALIFORNIA Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:ALTER DOMUS (US) LLC;REEL/FRAME:066952/0914 Effective date: 20240328 Owner name: AVENTIV TECHNOLOGIES, LLC, CALIFORNIA Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:ALTER DOMUS (US) LLC;REEL/FRAME:066952/0914 Effective date: 20240328 Owner name: ALLPAID, INC., CALIFORNIA Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:ALTER DOMUS (US) LLC;REEL/FRAME:066951/0630 Effective date: 20240328 Owner name: SECURUS TECHNOLOGIES, LLC, CALIFORNIA Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:ALTER DOMUS (US) LLC;REEL/FRAME:066951/0630 Effective date: 20240328 Owner name: AVENTIV TECHNOLOGIES, LLC, CALIFORNIA Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:ALTER DOMUS (US) LLC;REEL/FRAME:066951/0630 Effective date: 20240328 Owner name: ALLPAID, INC., CALIFORNIA Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:ALTER DOMUS (US) LLC;REEL/FRAME:066951/0561 Effective date: 20240328 Owner name: SECURUS TECHNOLOGIES, LLC, CALIFORNIA Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:ALTER DOMUS (US) LLC;REEL/FRAME:066951/0561 Effective date: 20240328 Owner name: AVENTIV TECHNOLOGIES, LLC, CALIFORNIA Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:ALTER DOMUS (US) LLC;REEL/FRAME:066951/0561 Effective date: 20240328 Owner name: ALLPAID, INC., CALIFORNIA Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:ALTER DOMUS (US) LLC;REEL/FRAME:066951/0514 Effective date: 20240328 Owner name: SECURUS TECHNOLOGIES, LLC, CALIFORNIA Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:ALTER DOMUS (US) LLC;REEL/FRAME:066951/0514 Effective date: 20240328 Owner name: AVENTIV TECHNOLOGIES, LLC, CALIFORNIA Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:ALTER DOMUS (US) LLC;REEL/FRAME:066951/0514 Effective date: 20240328 Owner name: ALLPAID, INC., CALIFORNIA Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:ALTER DOMUS (US) LLC;REEL/FRAME:066951/0385 Effective date: 20240328 Owner name: SECURUS TECHNOLOGIES, LLC, CALIFORNIA Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:ALTER DOMUS (US) LLC;REEL/FRAME:066951/0385 Effective date: 20240328 Owner name: AVENTIV TECHNOLOGIES, LLC, CALIFORNIA Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:ALTER DOMUS (US) LLC;REEL/FRAME:066951/0385 Effective date: 20240328 |