WO2014035372A1 - Détection de visage pour la commande de mode discrétion - Google Patents

Détection de visage pour la commande de mode discrétion Download PDF

Info

Publication number
WO2014035372A1
WO2014035372A1 PCT/US2012/052628 US2012052628W WO2014035372A1 WO 2014035372 A1 WO2014035372 A1 WO 2014035372A1 US 2012052628 W US2012052628 W US 2012052628W WO 2014035372 A1 WO2014035372 A1 WO 2014035372A1
Authority
WO
WIPO (PCT)
Prior art keywords
facial
image
face
facial orientation
communication device
Prior art date
Application number
PCT/US2012/052628
Other languages
English (en)
Inventor
John H. Yoakum
Original Assignee
Rockstar Bidco, LP
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Rockstar Bidco, LP filed Critical Rockstar Bidco, LP
Priority to PCT/US2012/052628 priority Critical patent/WO2014035372A1/fr
Publication of WO2014035372A1 publication Critical patent/WO2014035372A1/fr

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
    • G06F3/012Head tracking input arrangements
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/70Determining position or orientation of objects or cameras
    • G06T7/73Determining position or orientation of objects or cameras using feature-based methods
    • G06T7/74Determining position or orientation of objects or cameras using feature-based methods involving reference images or patches
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/72Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
    • H04M1/724User interfaces specially adapted for cordless or mobile telephones
    • H04M1/72403User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality
    • H04M1/7243User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages
    • H04M1/72439User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages for image or video messaging
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/72Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
    • H04M1/724User interfaces specially adapted for cordless or mobile telephones
    • H04M1/72448User interfaces specially adapted for cordless or mobile telephones with means for adapting the functionality of the device according to specific conditions
    • H04M1/72463User interfaces specially adapted for cordless or mobile telephones with means for adapting the functionality of the device according to specific conditions to restrict the functionality of the device
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30196Human being; Person
    • G06T2207/30201Face
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M2250/00Details of telephonic subscriber devices
    • H04M2250/52Details of telephonic subscriber devices including functional features of a camera

Definitions

  • the disclosure relates generally to systems and methods of operating a user communication device having a mute function.
  • User communication devices generally are configured with a mute function.
  • the mute function can be turned on to prevent audio into the user communication device from being communicated to other user communication devices.
  • the mute function can therefore allow for more efficient communications between the users as audio not intended for the communication session is not communicated to the other user communication devices when the mute function is activated.
  • a multitude of different users may be taking part in a communication session.
  • users not currently talking to other users can activate the mute function. In this manner, audio from those users not intended for the conference call does not clutter the conference call.
  • the mute function is often cumbersome to operate.
  • the user has to manually manipulate the user communication device in order to activate the mute function.
  • the user needs to manually manipulate the user communication device to deactivate the mute function and talk again to the other users.
  • a user taking part in the communication session may be trying to juggle several tasks and simply forget to operate the user communication device to activate or deactivate the mute function.
  • these mistakes often lead to uncomfortable, clumsy, and even embarrassing situations.
  • the user is prone to talking when the mute function is unintentionally activated.
  • other users may hear the user speaking about sensitive or personal matters with colleagues, assistants, or family members in the room with the user when the mute function unintentionally remains deactivated.
  • the disclosure relates generally to systems, methods, and devices for operating a user communication device having a mute function.
  • the systems, methods, and devices allow for the mute function to be controlled in accordance with a facial orientation of the user. In this manner, the user is not required to manually manipulate the user communication device in order to control the mute function.
  • a controller is operably associated with a camera and a user communication device.
  • the controller may be internal to the user communication device or may be external in an ancillary device.
  • the controller is configured to obtain a visual representation that includes a facial image of a face of a user captured with the camera.
  • the controller implements an image processing function on the visual representation.
  • the controller controls the mute function of the user
  • Another exemplary embodiment relates to a method of operating a user communication device having a mute function.
  • this exemplary embodiment relates to a method of operating a user communication device having a mute function.
  • a visual representation of the user is captured with a camera.
  • the visual representation includes a facial image of a face of the user.
  • an image processing function is implemented on the visual
  • the mute function of the user communication device is controlled based on the facial orientation data extracted from the visual representation.
  • a non-transitory computer readable medium stores computer executable instructions for causing at least one microprocessor to implement a method. More particularly, the computer executable instructions stored on the non-transitory computer readable medium causes the microprocessor(s) to capture a visual representation that includes a facial image of a user with a camera. To determine facial orientation data that indicates a facial orientation of the face from the facial image, the computer executable instructions further cause the microprocessor(s) to implement an image processing function on the visual representation. The computer executable instructions also cause the microprocessor(s) to control a mute function of a user communication device based on the facial orientation data.
  • Figure 1 illustrates one embodiment of a system that may be utilized by users to communicate through user communication devices via a network.
  • Figure 2 relates to exemplary procedures that may be implemented to control a mute function of a user communication device in accordance with a facial orientation of the user.
  • Figure 3 is a visual representation stream captured with a camera where each visual representation in the visual representation stream includes a facial image of a face of the user.
  • Figure 4 relates to exemplary procedures for extracting facial orientation data indicating a facial orientation of the face from the facial images in the visual representations shown in Figure 3 in order to control the mute function of the user communication device.
  • Figure 5 illustrates geometrical relationships based on measurements made on a visual representation to calculate facial orientation data.
  • Figure 6 illustrates one embodiment of a user communication device configured to control the mute function of the user communication device in accordance with a facial orientation of a user.
  • Figure 7 illustrates one embodiment of a personal computer configured to control a mute function of a user communication device in accordance with a facial orientation of a user.
  • This disclosure relates generally to systems, devices, and methods of operating a user communication device (UCD) having a mute function.
  • the systems, methods, and devices allow for the mute function to be controlled in accordance with a facial orientation of a user. In this manner, the user is not required to manually manipulate the user communication device in order to control the mute function.
  • the mute function may be any function that mutes audio inputs into the user communication device during an audio-based communication session with the UCD.
  • a mute function may be implemented all of which are within the scope of this disclosure.
  • a microphone of the UCD may simply be disabled when a mute function is in an on state.
  • the mute function may be implemented in call service layers so that audio inputs from the UCD are not transmitted via a network as communication signals while the mute function is in an on state.
  • the UCD may be any type of user device that allows users to establish
  • the UCD may be a desktop telephone device, a wireless telephone, a desktop computer, or a mobile communication device, such as a cellular telephone, a laptop, a smartphone, a tablet, and/or the like.
  • FIG 1 illustrates one embodiment of a system 10 in which user communication devices (referred to generically as elements 12 and specifically as elements 12-1 through 12-N).
  • each UCD 12 may be associated with one or more users, which are referred to generically with reference numeral 14 and individually with reference numerals 14-1 through 14- N.
  • the UCDs 12 may be communicatively coupled to each other through the network 16. In this manner, audio from the users 14 may be input into the corresponding UCD 12 and communicated via the network 16 in digital and/or analog form.
  • Each UCD 12 may have a mute function that allows the UCD 12 to remain connected with other UCD 12 during a communication session but prevents audio inputs from the corresponding users 14 from being communicated to the other UCDs 12.
  • the network 16 may be any type of network or any combinations of networks.
  • the network 16 may include a distributed public network such as the internet, a circuit switched telephonic network, a packet based telephonic network, local area networks (LAN), a wide area network (WAN) such as the Internet, one or more mobile communications networks, one or more personal area networks (PANs), and/or the like.
  • the network 16 may include gateways, or the like, to facilitate communication between the different types of networks.
  • the network may include wired components, wireless components, or both wired and wireless components.
  • the communication sessions that may be established by the UCD 12 through the network 16 may include telephone calls, conference calls, or any other type of audio-based communication session or communication system that has an audio component, such as for example a video conference.
  • the UCD 12-1 is a mobile communication device and in particular is a cellular phone.
  • Some examples of cellular phones that may be configured to operate in accordance to this disclosure are the Apple iPhone, Apple iPad, Apple iPod Touch device, a Smartphone programmed to have Google's Android operating system, Palm Pre, Samsung Rogue, and Blackberry Storm. Note, this listing is simply illustrative and is not intended to limit the scope of the present disclosure.
  • the UCD 12-2 is another type of mobile communication device and in particular is a laptop.
  • the UCDs 12-1 and 12-2 may each include a camera 18-1 and 18-2 respectively.
  • the cameras 18-1 and 18-2 may capture one or more visual representations that include a facial image of a face of the users 14-1 and 14-2, respectively.
  • the visual representation captured by the cameras 18-1 and 18-2 may be used to control the mute function provided by the UCD 12-1 and UCD 12-2, respectively.
  • the UCD 12-1 is a cellular telephone and the camera 18-1 is integrated into the cellular telephone.
  • the UCD 12-1 implements a communication client 19-1 that controls the
  • the communication client 19-1 may also generate graphical indicators on the display of the UCD 12-1 to allow the user 14-1 to provide input for the communication functions of the UCD 12-1 .
  • the UCD 12-2 is a laptop and the camera 18-2 is integrated into the laptop.
  • the UCD 12-2 implements a communication client 19-2 that controls the communication functions of the UCD 12-2 and allows the UCD 12-2 to establish communications sessions over the network 16.
  • the communication client 19-2 may also generate graphical indicators on the display of the UCD 12-2 to allow the user 14-2 to provide input for the communication functions of the UCD 12-2.
  • the UCD 12-2 may be another type of personal computer such as a desktop computer.
  • the desktop computer may implement the communication client 19-2, like the laptop.
  • the UCD 12-3 and the UCD 12-N are each desktop telephones (TDM or IP).
  • the UCD 12-3 also includes a camera 18-3 that captures one or more visual representations that include a facial image of a face of the user 14-3.
  • the visual representation captured by the camera 18-3 may be used to control a mute function provided by the UCD 12-3, which again, in this case, is one of the desktop telephones.
  • the camera 18-3 is integrated into the desktop telephone.
  • the camera 18-3 may be connected to a port (such as a USB port) of the UCD 12-3.
  • the UCD 12-3 may implement a communication client 19-3 that controls the communication functions of the UCD 12-3 and allows the UCD 12-3 to establish communications sessions over the network 16.
  • the communication client 19-3 may also generate graphical indicators on the display of the UCD 12-3 to allow the user 14-3 to provide input for the communication functions of the UCD 12-3.
  • the desktop telephone and a camera 20 are connected to a personal computer 22, such as the desktop computer shown in Figure 1 .
  • the camera 20 is connected to the personal computer 22 and is operable to capture a visual representation of the user 14-N that includes a facial image of a face of the user 14-N.
  • the personal computer 22 may use the visual representation to operate a mute function of the UCD 12-N.
  • the personal computer 22 implements a communication client 19-N.
  • the personal computer 22 and the UCD 12-N are operably associated with one another so that the UCD 12-N can receive commands from the communication client 19-N.
  • the communication client 19-N controls the communication functions of the UCD 12- N and allows the UCD 12-N to establish communications sessions over the network 16.
  • the communication client 19-N may also generate graphical indicators on the display of the personal computer 22 to allow the user 14-N to provide input for the communication functions of the UCD 12-N.
  • Figure 2 illustrates exemplary procedures for operating the UCDs 12 shown in Figure 1 . It should be noted that different embodiments of the exemplary procedures shown in Figure 2 may be implemented by the various types of UCDs 12 shown in Figure 1 . This may depend on the particular functionality and design configuration of the particular UCD 12.
  • a visual representation that includes a facial image of a face of the user 14 is captured with the camera 18 (procedure 100).
  • a facial orientation of the face of the user 14 can be indicative of whether the user 14 desires to speak to other users 14 or whether the speech from the user 14 is not directed to the
  • the users 14-1 and 14-2 may intend for their speech to be communicated through the network 16 to the other users 14.
  • the facial orientation of the users 14-1 and 14-2 is directed away from the UCD 12-1 and 12-2, the user 14-1 and 14-2 may not be intending for speech to be communicated over the network 16.
  • the personal computer 22 when the facial orientation of the user 14-N is directed towards the UCD 12-N, the user 14-N may be intending for speech to be communicated to other users 14 during a communication session on the network 16.
  • an image processing function may be implemented on the visual representation to determine facial orientation data that indicates a facial orientation of the face of the user 14 from the facial image in the visual representation (procedure 102).
  • the image processing function may be performed through facial recognition software that is capable of recognizing the facial image and extract positions within the facial image of certain facial features. The facial orientation data can then be extracted from these positions to describe the facial orientations of the face of the user 14.
  • the mute function of the UCD 12 is controlled based on the facial orientation data (procedure 104).
  • the facial orientation data may indicate that the facial orientation of the face is directed substantially away from the UCD 12.
  • the mute function of the UCD 12 is switched from an "off" state to an "on” state in response to the facial orientation data.
  • the mute function prevents audio from the UCD 12 from being communicated over the network 16 to other UCDs 12.
  • the facial orientation data may indicate that the facial orientation of the face from the facial image is directed substantially towards the UCD 12.
  • the mute function is switched from the "on" state to the "off” state in response to the facial orientation data.
  • audio inputs are communicated from the UCD 12 via the network 16 to other UCDs 12 taking part in an established communication session.
  • the mute function may be switched to the on state if the facial orientation data indicates that the facial orientation of the face is directed substantially towards a selected area.
  • the mute function may be switched to the off state if the facial orientation data indicates that the facial orientation of the face is directed substantially away from the selected area.
  • This area may be arbitrarily selected and may not involve the UCD 12 or may be related to a different device.
  • the selected area may be a graphical indicator generated by the communication client 19-N and rendered on the display of the personal computer 22.
  • the selected area may refer to a more specific location on the UCD 12.
  • the selected area may be an area selected through user configuration.
  • Figure 3 illustrates various embodiments of visual representations (referred to generically as elements 24 and individually as elements 24A through 24D) that include a facial image (referred to generically as elements 26 and specifically as elements 26A and 26D) of a face of the user 14.
  • the visual representations 24 illustrated in Figure 3 are provided as part of a visual representation stream that has been captured with the camera 18.
  • the visual representations 24 may have been captured with the camera 18 through a viewfinder application implemented on the UCD 12.
  • Each visual representation 24 shown in Figure 3 is thus a frame from a viewfinder stream. The frames thus have a position within the viewfinder stream based on a time of capture. Accordingly, as the user 14 changes their facial orientation over time, the mute function of the UCD 12 activated and deactivated based on these changes.
  • Visual representation 24A of the user 14 has a facial image 26A of the user 14 where the facial orientation of the face from the facial image 26A is directed substantially towards the camera 18.
  • UCD 12-1 and 12- 2 having a facial orientation directed towards the camera 18-1 and 18-2 indicates that the facial orientation is directed towards the UCD 12-1 and 12-2, respectively.
  • the remainder of the discussion with regards to Figure 3 will make this assumption for the sake of explanation and clarity.
  • the user 14-N does not have a facial orientation directed towards the UCD 12-N when facing the camera 20. In this case, some range of facial orientations may describe when the user 14-N is facing the UCD 12-N.
  • the facial orientation data extracted from the visual representation 24A indicates that the facial orientation of the face from the facial image 26A is directed substantially towards the UCD 12.
  • the mute function of the UCD 12 is switched from an "on" state to an "off” state in response to the facial orientation data indicating that the facial orientation of the face from the facial image is directed substantially towards the UCD 12. Audio inputs from the user 14 are thus communicated over the network 16 to other UCDs 12.
  • another visual representation 24B is captured with the camera 18 that includes another facial image 26B of the face of the user 14 after capturing the visual representation 24A.
  • the facial orientation of the face from the facial image 26B is directed substantially away from the UCD 12.
  • the facial orientation data extracted from the facial image 26B indicates that the facial orientation of the face from the facial image 26B is directed substantially away from the UCD 12.
  • the mute function is thus switched from the "off" state to the "on” state in response to the facial orientation data indicating that the facial orientation of the face from the facial image 26B is directed substantially towards the UCD 12.
  • some embodiments of the mute function may allow for audio inputs from the other UCDs 12 to be received by the UCD 12 while the mute function is in the "on" state.
  • the user 14 may be able to listen to the audio inputs from the other users 14 while preventing audio inputs from themselves from being transmitted to the other UCDs 12.
  • the user may direct their face again towards the UCD 12 as shown by the facial image 26C and visual representation 24C.
  • the facial orientation data will indicate that the facial orientation of the face from the facial image 26C is directed substantially towards the UCD 12.
  • the mute function of the UCD 12 is switched from an "on” state to an "off” state in response to the facial orientation data extracted from the facial image 26C indicating that the facial orientation of the face from the facial image 26C is directed substantially towards the UCD 12.
  • the user 14 again turns their face away from the UCD 12.
  • the facial orientation data will indicate that the facial orientation of the face from the facial image 26D is directed substantially away from the UCD 12.
  • the mute function is thus switched from the "off" state to the "on” state in response to the facial orientation data extracted from the visual representation 24D indicating that the facial orientation of the face from the facial image 26D is directed
  • the user 14 may control the mute function without having to manually manipulate the UCD 12.
  • Figure 4 illustrates one embodiment of exemplary procedures for implementing the image processing function on the visual representations 24 shown in Figure 3 to determine the facial orientation data of the face from the facial images 26.
  • the exemplary procedures illustrated in Figure 4 extract images of facial features from the facial images 26 to determine facial orientation data.
  • the facial feature images that are relevant are eye regional images of the eye regions on the face of the user 14 and another facial feature image of a reference facial feature of the face of the user 14.
  • the reference facial feature may be the mouth of the user 14.
  • edge image detection masks can be utilized to easily detect these facial feature images within the facial image 26 of the face.
  • the position of the eye region images and the reference feature images are used to determine a facial image plain of the facial image 26.
  • Certain assumptions are made about the geometry of the face such as that the positions of the eye region images and the reference facial feature image (i.e., in this case the mouth feature image) form an isosceles triangle when the user 14 is looking directly toward the camera 18.
  • the facial orientation data can be extracted.
  • a first displacement in the facial image 26 is measured wherein the displacement is between a first eye region image of a first eye region on the face of the user 14 and a second eye region image of a second eye region on the face of the user 14 within the facial image 26 (procedure 200).
  • This displacement can thus be thought of as one side of a triangle.
  • a second displacement is measured between the first eye region image and a reference facial feature image of the reference facial feature on the face of the user
  • the reference facial feature image is of the mouth on the face of the user 14. This second
  • a third displacement is measured between the second eye region image and the reference facial feature image (procedure 204). Finally, the facial orientation data is calculated based on the first displacement, the second displacement, and the third displacement
  • Figure 5 illustrates one embodiment of the isosceles triangle
  • may be provided based on measurements from a previously obtained visual representation.
  • This previously obtained visual representation i.e., base image
  • the second displacement between the first eye region image Ei and the reference (i.e., the mouth feature image in this embodiment) Ri is equal to the displacement
  • the second eye region image E 2 and the reference facial feature image Ri have a displacement
  • the image processing function may determine positions and displacements for the triangle for visual representations captured in real time.
  • the triangle has a first displacement between the first eye region image ⁇ and second eye region image E' 2 equal to approximately to the displacement
  • the second displacement between the first eye region E'i and the reference facial feature image R'i is approximately equal to the displacement
  • the third displacement between the second eye region image E' 2 and the reference facial feature image R'i is equal to the displacement
  • an angle FO can be calculated in accordance with the equation below.
  • the angle FO calculated by from the equation describes the facial orientation of the user 14.
  • the angle FO is thus calculated in accordance with the displacements extracted from of the triangle
  • the angle ⁇ is the angle between lines
  • may be out of scale with one another.
  • scaling factors may be determined from zooming parameters, or the like, to provide these distances at the same scale.
  • the facial orientation data may represent the angle FO to indicate the facial orientation of the user 14.
  • the facial orientation data may represent a unit vector extracted from the facial image plane of the triangle
  • the facial orientation data may indicate the facial orientation in any manner suitable to describe the orientation of an object from an image.
  • the angle FO may or may not indicate that the facial orientation of the user is directed towards the camera when the facial orientation of the user is directed substantial towards the UCD 12.
  • the facial orientation of the user 14 is directed substantially towards the cameras 18-1 through 18-3, respectively, when the users 14-1 through 14-3 have a facial orientation directed towards the UCD 12-1 through 12-3.
  • the angle FO may be relatively small when the user is facing their respective UCD 12-1 through 12-3. If the angle FO is relatively large, the users 14-1 through 14-3 have a facial orientation that is directed substantially away from the UCDs UCD 12-1 through 12-3, respectively.
  • the user 14-N is going to have a facial orientation directed towards the UCD 12-N (i.e., the respective desktop telephone) and not the camera 20 when the mute function is deactivated.
  • FIG 6 illustrates one embodiment of the UCD 12, which may be the UCDs 12-1 , 12-2, or the 12-3 shown in Figure 1 .
  • the UCD 12 includes a controller 28, one or more communication interface devices 30, a display 32, and the camera 18. Other user input and output devices may also be provided with the UCD 12.
  • the communication interface device(s) 30 is operable to communicatively couple the UCD 18 to the network 16.
  • the network 16 may include various different types of mobile communications networks, packet-switched networks, and circuit switched networks.
  • the communication interface device(s) 30 may be adapted to facilitate communications with one or more communication services on the network 16. In this example, the communication interface device 30 may facilitate
  • communications for any number of communication services provided by mobile communications networks, packet-switched networks, circuit switched networks, wireless local area networks and/or the like.
  • the controller 28 has general purpose computer hardware, in this case, one or more microprocessors 34, a non-transitory computer readable medium, such as a memory device 36, and a system bus 38.
  • the system bus 38 is operably associated with the microprocessors 34 so that microprocessors 34 can exchange information with the communication interface device 30, the display 32, the camera 18, and other user input and output devices.
  • the controller 28 may also include other hardware such as, control logic, other processing devices, additional non-transitory computer readable mediums, and the like.
  • the memory device 36 may store computer executable instructions 40 for execution by the microprocessors 34.
  • the display 32 may be any suitable display for UCDs.
  • the display 32 may be a touch screen, a monitor, an LCD display, a plasma display, and/or the like.
  • Other user input and output devices may be a keyboard, a microphone, a head-set, a mouse, and/or an input button, and may depend on the particular configuration of the UCD 12.
  • the computer executable instructions 40 configure the operation of the microprocessors 34 so that the microprocessors 34 implement the software applications for the UCD 12 previously described.
  • the controller 28 that operates the mute function of the UCD 12 is a mobile communications device controller within the UCD 12.
  • the UCD 12 has the camera 18 and thus can capture visual representations of the user 14 to determine facial orientation data and thereby control the mute function of the UCD 12 based on the facial orientation data.
  • Figure 7 illustrates one embodiment of the personal computer 22 connected to the camera 20 and the UCD 12-N (i.e., the desktop telephone) shown in Figure 1 .
  • the camera 20 and the UCD 12-N are connected to the personal computer 22 through input/output ports 42 provided by the personal computer 22.
  • the personal computer 22 includes a controller 44, one or more communication interface devices 46, and a display 48. Other user input and output devices may also be provided with the personal computer 22.
  • the communication interface device(s) 46 is operable to communicatively couple the personal computer 22 and the UCD 12-N to the network 16.
  • the network 16 may include various different types of mobile
  • the communication interface device(s) 46 may be adapted to facilitate communications with one or more communication services on the network 16. In this example, the communication interface device 46 may facilitate
  • UCD 12- N may also include communication interface devices that permit for it to independently connect to the network 16.
  • the controller 44 has general purpose computer hardware, in this case, one or more microprocessors 50, a non-transitory computer readable medium, such as a memory device 52, and a system bus 54.
  • the system bus 54 is operably associated with the microprocessors 50 so that microprocessors 50 can exchange information with the communication interface device 46, the display 48, the camera 20, the UCD 12-N and other user input and output devices.
  • the UCD 12-N may include its own separate controller and
  • the controller 44 may also include other hardware such as, control logic, other processing devices, additional non-transitory computer readable mediums, and the like.
  • the memory device 52 may store computer executable instructions 56 for execution by the microprocessors 50.
  • the display 48 may be any suitable display for the personal computer 22.
  • the display 48 may be a touch screen, a monitor, an LCD display, a plasma display, and/or the like.
  • Other user input and output devices may be a keyboard, a microphone, a head-set, a mouse, and/or an input button, and may depend on the particular configuration of the personal computer 22.
  • the computer executable instructions 56 configure the operation of the microprocessors 50 so that the microprocessors 50 implement the software applications previously described to control the mute function of the UCD 12-N.
  • the controller 44 that operates the mute function of the UCD 12-N is a desktop computer controller.
  • the personal computer 22 can capture visual representations of the user 14-N to determine facial orientation data and thereby control the mute function of the UCD 12-N based on the facial orientation data.

Landscapes

  • Engineering & Computer Science (AREA)
  • Human Computer Interaction (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Signal Processing (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • General Engineering & Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • General Health & Medical Sciences (AREA)
  • Business, Economics & Management (AREA)
  • General Business, Economics & Management (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Image Processing (AREA)

Abstract

La présente invention concerne des systèmes, procédés, et dispositifs pour le fonctionnement d'un dispositif de communication utilisateur (12) comportant une fonction discrétion. Les systèmes, procédés, et dispositifs permettent la commande de la fonction discrétion selon une orientation faciale de l'utilisateur. Selon un mode de réalisation représentatif, un contrôleur (28) est associé en fonctionnement à une caméra (18) et à un dispositif de communication utilisateur (12). Le contrôleur (28) peut être incorporé dans le dispositif de communication utilisateur (12) ou peut être externe dans un dispositif auxiliaire (20). Le contrôleur (28) est configuré pour obtenir une représentation visuelle qui comporte une image faciale d'un visage d'un utilisateur capturée avec la caméra (18). En vue de la détermination de données d'orientation faciale indiquant une orientation faciale du visage à partir de l'image faciale, le contrôleur (28) utilise une fonction de traitement d'images sur le dispositif de communication utilisateur (12) en fonction des données d'orientation faciale extraites à partir de la représentation visuelle.
PCT/US2012/052628 2012-08-28 2012-08-28 Détection de visage pour la commande de mode discrétion WO2014035372A1 (fr)

Priority Applications (1)

Application Number Priority Date Filing Date Title
PCT/US2012/052628 WO2014035372A1 (fr) 2012-08-28 2012-08-28 Détection de visage pour la commande de mode discrétion

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/US2012/052628 WO2014035372A1 (fr) 2012-08-28 2012-08-28 Détection de visage pour la commande de mode discrétion

Publications (1)

Publication Number Publication Date
WO2014035372A1 true WO2014035372A1 (fr) 2014-03-06

Family

ID=46801651

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2012/052628 WO2014035372A1 (fr) 2012-08-28 2012-08-28 Détection de visage pour la commande de mode discrétion

Country Status (1)

Country Link
WO (1) WO2014035372A1 (fr)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103870824A (zh) * 2014-03-28 2014-06-18 海信集团有限公司 一种在人脸检测跟踪过程中的人脸捕捉方法及装置
US20170180557A1 (en) * 2015-12-18 2017-06-22 Lenovo Enterprise Solutions (Singapore) Pte. Ltd. Conference call unmuting

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6144755A (en) * 1996-10-11 2000-11-07 Mitsubishi Electric Information Technology Center America, Inc. (Ita) Method and apparatus for determining poses
US20040180690A1 (en) * 2002-12-16 2004-09-16 Lg Electronics Inc. Apparatus for operating a mobile communication terminal with integrated photographic apparatus and method thereof
US20060233422A1 (en) * 2005-04-16 2006-10-19 Microsoft Corporation Machine vision system and method for estimating and tracking facial pose
EP2131266A2 (fr) * 2008-06-02 2009-12-09 Sony Ericsson Mobile Communications Japan, Inc. Dispositif de traitement d'affichage, procédé de traitement d'affichage, programme de traitement d'affichage et dispositif terminal mobile
US20100149343A1 (en) * 2008-12-16 2010-06-17 Samsung Digital Imaging Co., Ltd. Photographing method and apparatus using face pose estimation of face

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6144755A (en) * 1996-10-11 2000-11-07 Mitsubishi Electric Information Technology Center America, Inc. (Ita) Method and apparatus for determining poses
US20040180690A1 (en) * 2002-12-16 2004-09-16 Lg Electronics Inc. Apparatus for operating a mobile communication terminal with integrated photographic apparatus and method thereof
US20060233422A1 (en) * 2005-04-16 2006-10-19 Microsoft Corporation Machine vision system and method for estimating and tracking facial pose
EP2131266A2 (fr) * 2008-06-02 2009-12-09 Sony Ericsson Mobile Communications Japan, Inc. Dispositif de traitement d'affichage, procédé de traitement d'affichage, programme de traitement d'affichage et dispositif terminal mobile
US20100149343A1 (en) * 2008-12-16 2010-06-17 Samsung Digital Imaging Co., Ltd. Photographing method and apparatus using face pose estimation of face

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103870824A (zh) * 2014-03-28 2014-06-18 海信集团有限公司 一种在人脸检测跟踪过程中的人脸捕捉方法及装置
CN103870824B (zh) * 2014-03-28 2017-10-20 海信集团有限公司 一种在人脸检测跟踪过程中的人脸捕捉方法及装置
US20170180557A1 (en) * 2015-12-18 2017-06-22 Lenovo Enterprise Solutions (Singapore) Pte. Ltd. Conference call unmuting
US9936078B2 (en) * 2015-12-18 2018-04-03 Lenovo Enterprise Solutions (Singapore) Pte. Ltd. Conference call unmuting

Similar Documents

Publication Publication Date Title
US8706162B1 (en) Automatic routing of call audio at incoming call
US8254984B2 (en) Speaker activation for mobile communication device
US20160234260A1 (en) Methods and apparaturses for binding with device
WO2016192323A1 (fr) Procédé et dispositif de communication vidéo
EP3068115A1 (fr) Procédé et dispositif de rappel d'appel entrant
EP3136699A1 (fr) Procédé et dispositif permettant de connecter un équipement externe
US9491371B2 (en) Method and device for configuring photographing parameters
EP3091753A1 (fr) Procédé et dispositif d'optimisation de signal sonore
WO2019129020A1 (fr) Procédé de mise au point automatique d'un appareil photo, dispositif de stockage, et terminal mobile
WO2015131664A1 (fr) Procédé, dispositif et terminal d'ajustement de mise au point
US10009283B2 (en) Method and device for processing information
US20140254832A1 (en) Volume adjusting system and method
WO2015117347A1 (fr) Procédé et dispositif de réglage pour mode de scène de terminal
CN111049984A (zh) 音频播放控制方法和设备以及存储介质
TW201637423A (zh) 來電鈴聲控制系統及方法
EP2981054B1 (fr) Procédé, appareil et terminal pour le transfert d'appel
KR20170120707A (ko) 터치 스크린의 포인트 보고를 처리하는 방법, 장치, 프로그램 및 기록매체
JP2022133267A (ja) 端末装置、画像切り出し方法、および画像切り出しプログラム
WO2014035372A1 (fr) Détection de visage pour la commande de mode discrétion
EP2670135B1 (fr) Appareil d'entrée visiophonique mettant en oeuvre des fonctions de commande d'appel vidéo basées sur des positions ouvertes et fermées de partie de couverture
US20170147134A1 (en) Method and apparatus for controlling touch-screen sensitivity
JP7377614B2 (ja) ウェアラブル装置および送信制御方法
US9930467B2 (en) Sound recording method and device
CN108401518B (zh) 信道协调方法及装置
US20200059845A1 (en) Wireless communication method and apparatus

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 12756311

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 03/09/2015)

122 Ep: pct application non-entry in european phase

Ref document number: 12756311

Country of ref document: EP

Kind code of ref document: A1