WO2011069668A1 - Konferenzsystem und zugeordnetes signalgebungsverfahren - Google Patents

Konferenzsystem und zugeordnetes signalgebungsverfahren Download PDF

Info

Publication number
WO2011069668A1
WO2011069668A1 PCT/EP2010/007534 EP2010007534W WO2011069668A1 WO 2011069668 A1 WO2011069668 A1 WO 2011069668A1 EP 2010007534 W EP2010007534 W EP 2010007534W WO 2011069668 A1 WO2011069668 A1 WO 2011069668A1
Authority
WO
WIPO (PCT)
Prior art keywords
conference
signaling device
conference system
output signal
motion signal
Prior art date
Application number
PCT/EP2010/007534
Other languages
German (de)
English (en)
French (fr)
Inventor
Florian Agsteiner
Tobias Haeberle
Matthias Hanel
Adnane Jadid
Thomas Lederer
Maximilian Reiss
Andreas Tai
Original Assignee
Siemens Enterprise Communications Gmbh & Co.Kg
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Siemens Enterprise Communications Gmbh & Co.Kg filed Critical Siemens Enterprise Communications Gmbh & Co.Kg
Priority to EP10793164.4A priority Critical patent/EP2389753B1/de
Priority to CN2010800153346A priority patent/CN102369718A/zh
Priority to US13/379,447 priority patent/US8965350B2/en
Publication of WO2011069668A1 publication Critical patent/WO2011069668A1/de
Priority to US14/589,329 priority patent/US9397850B2/en

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L12/00Data switching networks
    • H04L12/02Details
    • H04L12/16Arrangements for providing special services to substations
    • H04L12/18Arrangements for providing special services to substations for broadcast or conference, e.g. multicast
    • H04L12/189Arrangements for providing special services to substations for broadcast or conference, e.g. multicast in combination with wireless systems
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/017Gesture based interaction, e.g. based on a set of recognized hand gestures
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/253Telephone sets using digital voice transmission
    • H04M1/2535Telephone sets using digital voice transmission adapted for voice communication over an Internet Protocol [IP] network
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/72Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
    • H04M1/724User interfaces specially adapted for cordless or mobile telephones
    • H04M1/72403User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality
    • H04M1/72427User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality for supporting games or graphical animations
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/72Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
    • H04M1/724User interfaces specially adapted for cordless or mobile telephones
    • H04M1/72448User interfaces specially adapted for cordless or mobile telephones with means for adapting the functionality of the device according to specific conditions
    • H04M1/72454User interfaces specially adapted for cordless or mobile telephones with means for adapting the functionality of the device according to specific conditions according to context-related or environment-related conditions
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04WWIRELESS COMMUNICATION NETWORKS
    • H04W4/00Services specially adapted for wireless communication networks; Facilities therefor
    • H04W4/16Communication-related supplementary services, e.g. call-transfer or call-hold
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04WWIRELESS COMMUNICATION NETWORKS
    • H04W84/00Network topologies
    • H04W84/02Hierarchically pre-organised networks, e.g. paging networks, cellular networks, WLAN [Wireless Local Area Network] or WLL [Wireless Local Loop]
    • H04W84/10Small scale networks; Flat hierarchical networks
    • H04W84/12WLAN [Wireless Local Area Networks]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M2250/00Details of telephonic subscriber devices
    • H04M2250/12Details of telephonic subscriber devices including a sensor for measuring a physical value, e.g. temperature or motion
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M2250/00Details of telephonic subscriber devices
    • H04M2250/62Details of telephonic subscriber devices user interface aspects of conference calls
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M3/00Automatic or semi-automatic exchanges
    • H04M3/42Systems providing special services or facilities to subscribers
    • H04M3/56Arrangements for connecting several subscribers to a common circuit, i.e. affording conference facilities

Definitions

  • the invention relates to a conference system with at least two mobile
  • Terminals each with a signaling device and an associated signaling method.
  • mobile terminals serving interpersonal communication e.g. Mobile phones, smartphones, PDAs, tablet PCs and laptops can be used in various ways.
  • the user input can be made via keyboard, with input devices, via voice control, by controller, or in some other way. It is also known, by movements of an input device, e.g. the finger to control the device functions on a touch-sensitive user interface. It is also known to move the terminal in order, for example, to influence the presentation mode of the mobile terminal, for example to switch from a portrait reproduction mode to a landscape format, see for example: Apple iPhone (http://www.apple.com/phone/ ).
  • control and input devices and methods known in the context of mobile terminals for interpersonal communication have the disadvantage that they are not accessible to gestures of interpersonal non-verbal communication, for control or input.
  • Such gestures of interpersonal nonverbal communication in particular non-specific such gestures, remain closed to the mobile communication terminals in the prior art.
  • WO 2006/107463 A1 discloses a method and a device for visual presentation, which determines and displays predetermined stored speech characteristics from a speech connection of a communication device.
  • WO 2008/138407 A1 discloses a mobile terminal with a WLAN transceiver and a motion sensor, wherein a multimedia object is generated by the determined movement.
  • telephone or video conferences are increasingly being conducted for joint coordination or for the efficient and timely dissemination of information. It may well happen that the telephone conferences reach a larger number of participants, eg 10 or 20 participants in larger companies, possibly even much more.
  • conference bridges allow different modes of operation - e.g. For example, all subscribers can communicate bidirectionally, i. Speaking at the same time as listening to all other participants. It may happen that participants bring in disturbing ambient noise that disturbs the quality of the conference acoustics. Therefore, the use of the mode of a so-called moderated conference may be useful,
  • connection or disconnection (muting) of the participants is done by moderators.
  • the moderator has an overview of all participants and can then perform this function via a control application connected to the conference system and, so to speak, give the floor.
  • the subscriber In order to then speak up, the subscriber must call the presenter via an additional telephone call, e.g. Call the presenter's mobile phone to hear that he is "loudly" dialing in to speak at the conference.
  • the conference system comprises at least two mobile terminals, each with a signaling device containing a sensor which detects movement of the signaling device and a
  • the conference system includes a transmitting device for supplying the output signal to a transmission channel, which is in each case connected to a local area network.
  • a transmitting device for supplying the output signal to a transmission channel, which is in each case connected to a local area network.
  • This is a telephone preferably an IP phone (11) for a moderator and a screen for conference control connected.
  • a signaling device for mobile terminals which is suitable for interpersonal communication, which includes a sensor which detects a movement of the signaling device and outputs a corresponding motion signal.
  • the signaling device includes a motion signal processing device that detects the motion signal of a gesture of non-verbal interpersonal communication (including, but not limited to, gestures due to the movement of body parts for the purpose of interpersonal communication
  • the signaling device further includes a transmitting device for
  • a transmission channel This may be a wireless or wireline transmission channel or a
  • the motion signal processing device assigns the motion signal to a motion signal pattern. This can be one
  • the motion signal processing device generates an output signal which contains information about the type of detected gesture of the nonverbal interpersonal
  • the output signal contains additional information, in particular those which give an indication of the causer of the motion signal.
  • the motion signal processing apparatus is capable of detecting the non-verbal human gestures of noticing, in particular the winking and / or shaking and / or a message, due to movement of the signaling device.
  • the motion signal processing device can recognize user-specific motion signal patterns and / or store them in the signaling device, in particular for reassignment.
  • the senor and / or the motion signal processing device is designed so that unconscious gestures of the non-verbal interpersonal communication are suppressed, so that they no motion signal and / or
  • the signaling device processes and / or stores motion signal patterns from an external data source which is not part of the signaling device.
  • the transmission channel is between the
  • the transmission channel may be formed in the form of a radio link. Transmitting and receiving device are spaced from each other.
  • the conference system includes a conference control signaling device that includes an output device that outputs the information of the output corresponding to the gesture of the non-verbal interpersonal communication.
  • the signaling device further includes an output signal processing device, which the
  • the signaling device includes a receiving device which receives the output signal. This may for example be part of the output signal processing device.
  • Information also give an indication of the cause of the motion signal.
  • the information to be output is generated on the basis of signal patterns stored in the output signal processing device, in particular by assignment of such to the output signal.
  • the output of the information takes place visually on a suitable means of the signaling device, in particular a screen, acoustically by a suitable means of the signaling device, and / or by means of the signaling device which generates haptic stimuli, in particular vibration.
  • the signaling device processes signal patterns from an external data source which is not part of the signaling device.
  • a signaling method for the conference system is proposed, which as the first step the
  • Communication is assigned and in a third step an output signal in Trap of a successful assignment generated containing information with the gesture of non-verbal interpersonal communication
  • the fourth step is to send the output signal over a transmission channel.
  • the fifth step is to receive the 5 output signal transmitted over the transmission channel.
  • a signaling method which as a first step provides for the generation of information on the basis of the output signal, wherein the information to with the gesture of non-verbal interpersonal communication
  • FIG. 1 shows an embodiment of a mobile terminal with a signaling device for explaining the invention
  • Fig. 2 shows schematically a further embodiment of a mobile terminal with a signaling device for explaining the invention
  • FIG. 3 schematically shows a signal of a sensor of a signaling device according to FIG. 1 or FIG. 2 for explaining the invention
  • Fig. 5 shows another embodiment of a mobile terminal with a
  • FIG. 6 shows a first embodiment of a conference system according to the invention using the signaling device
  • Fig. 7 is a block diagram of a mobile terminal used in the embodiment of the present invention of the conference system shown in Fig. 6;
  • FIG. 8 shows an IP telephone with display of a user interface for use in a further conference system according to the invention
  • FIG. 9 shows a second embodiment of a conference system according to the invention for carrying out an auction using mobile terminals shown in FIG. 7.
  • gesture refers in the context of the present invention, the symbolic movement of one or more specific body parts of a person, which is mainly used for communication, contact and the like purposes.
  • the gesture as part of human gestures is to be distinguished here from the movements of facial expressions.
  • the term gesture in its meaning largely approaches the meaning of the term gesture (movement of a body part) and differs, if necessary, in that gestures sometimes have a certain symbolism.
  • the term "interpersonal communication” refers to communication between people, that is to say, in particular, the mutual interaction of individuals.
  • FIG. 1 shows by way of example a mobile terminal 1 with a screen 1a and icons 18 displayed thereon, which contains a signaling device 2.
  • the mobile terminal for example a personal digital
  • PDA Personal Computer Assistant
  • a mobile phone a smartphone, a tablet PC or similar device, is capable of communicating over one
  • Such communication links include, for example, telephone connections, in particular wireless and
  • these devices can connect to multiple devices
  • the signaling device 2 is preferably integrated in a mobile terminal 1, for example by its technical equipment such as e.g. the antenna, the power supply, a display device, etc. use.
  • the device 2 may be a separate, separate
  • Be device that is not part of a mobile device.
  • the signaling device 2 is preferably suitable for movement, in particular during operation, by a user, for example by holding by hand and moving the same, without difficulty.
  • the signaling device 2 or the mobile terminal 1 receiving it has e.g. low weight and handy design and is therefore preferably designed as a handheld device or as a portable device.
  • movements in the direction of arrow 3 about an axis A are possible as well as those in the direction of arrow 4 about an axis B or those about an axis C.
  • movements in the direction of arrow 3 about an axis A are possible as well as those in the direction of arrow 4 about an axis B or those about an axis C.
  • the signaling device 2 includes a sensor 5, for example an acceleration sensor, which provides movement of the signalers
  • the sensor 5 On the basis of a detected movement, the sensor 5 provides or delivers a movement signal BS at an output 6, see FIG. 2.
  • This movement signal BS corresponding to the movement becomes a
  • Motion signal processing device 7 supplied to or retrieved from this.
  • the motion signal processing device 7 comprises, for example, a computing unit, for example a computerized computing unit, such as e.g. a signal processor 8.
  • the motion signal processing device 7 interprets the motion signal BS of the sensor 5 as to whether in the
  • Motion signal BS is a motion signal pattern BSM, which is a gesture of non-verbal interpersonal communication, in particular the non-verbal interpersonal communication can be assigned.
  • the motion signal processing device 7 can compare, for example, the motion signal BS with defined motion signal patterns BSM of such gestures, which are stored and retrievable, for example, in a memory device 9 or the motion signal processing device 7 are known.
  • the storage device 9 may be part of the signalers
  • device 2 in particular part of the motion signal processing device 7, may also be e.g. be an external data source.
  • the motion signal processing apparatus 7 interprets the
  • Motion signal BS preferably in terms of what kind the the
  • Motion signal BS is assignable gesture.
  • Such gestures of non-verbal interpersonal communication are, for example, those of human noticing as a result of a movement of a body part, such as: hand waving, hand shaking, hand / arm pivoting, or, for example, raising an arm (motion signal BS) Gesture is generated by moving the held signaling device 2 as described above).
  • any other types of such gestures are conceivable.
  • Such interpersonal gestures described above are expected, for example, where users of the signaling device 2 want to draw attention to themselves, eg in conferences, in emergency situations, etc.
  • a gesture could be, for example, waving a participant in a conference, making a bid in an auction by raising a hand, knocking on a party who muted their communication device, reporting message to the moderator of a conference, reporting in a conference to signal a request to talk, the interpretation of the intensity of a Scrissains as emotion to the evaluation function (as for example similar to the clapping), and for example the mixing of a sorting.
  • the movement signal processing device 7 outputs an output signal AS containing information corresponding to the assignment.
  • the output signal AS contains, for example, the information that a gesture of non-verbal interpersonal communication due to the
  • Moving signal BS has been detected by assignment or is present.
  • the output signal AS may preferably also contain information as to which type of gesture has been recognized.
  • the signaling device 2 further includes, in an exemplary preferred embodiment, a transmitting device 10 that generates the generated signal
  • Transmission channel 11 which, for example, a connection to another mobile terminal or for example another, the
  • the transmission channel may be, for example, a connection with one described in connection with FIG. 4 below
  • the transmitting device 10 may include conventional means for conditioning the output signal for transmission purposes, such as means for encoding (e.g., D / A or A / D) and / or amplifying the output signal, filtering, etc.
  • the transmitting device 10 may have, for example, an antenna 11a or a connection for a cable-bound transmission channel 11.
  • the transmission channel 11 can be a transmission channel which is customary in communications technology, for example a wireless in the form of a radio link or conducted or both.
  • the transmission channel 1 can be implemented in many ways and formed, for example, using conventional protocols and methods. Both analog and digital
  • Transmission channel shapes are conceivable.
  • the transmission channel 11 is formed between a transmitting device 10 and a receiving device, in particular if they are spatially separated from each other or removed.
  • a possible movement signal BS is shown as an example in FIG. 3 as amplitude a (t) over time t.
  • the sensor 5 responds to a specific movement of the signaling device 2, for example with a frequency change (range B1) and an altered (in this case larger) amplitude a (range B2) in the waveform 12 Changes in the waveform 12 are reproducible upon repeating the particular motion.
  • the reproducible changed waveform 12 can be
  • a movement signal pattern BSM for example as a movement signal pattern BSM, e.g. stored as a curve or in digitized form.
  • the signal components lying above a threshold value S are detected (clipping, cutting off of the signal peaks) and these are used, for example. as a digital motion signal BS
  • Moving signal BS corresponds, assignable motion signal pattern BSM can be obtained. Shown is also a threshold S in the form of a dashed line. If the movement signal amplitudes a (t) are below, no movement signal BS is output, for example.
  • the movement signal patterns BSM generated from movement signals BS can be stored in the signaling device 2. However, it is also conceivable that the signaling device. 2
  • the external data source may e.g. in the form of a memory card, a memory unit of another signaling device 2, an internal web-based or a network memory unit, etc. It is also within the scope of the invention that the
  • Signaling device 2 is self-learning. This can be realized, for example by making a gesture of nonverbal interpersonal
  • Communication performs while the signaling device 2 detects the motion signal BS generated thereby by means of the sensor 5 and as
  • Motion signal pattern BSM e.g. in the signaling device 2, stores. In this way, user-specific gestures can be detected.
  • the signaling device 2 deliberately differentiates from unconscious gestures.
  • motion signals BS of unconscious gestures for example, can be suppressed.
  • the intensity of a detected gesture for example, the intensity of a detected gesture.
  • the sensor 5 usually detects a weak signal, since the movement amplitudes or the
  • Threshold S (range B2).
  • the output signal AS can also be supplemented with further information before it can be output - e.g. via a transmission channel - is made available. So it is e.g. make sense to add a user ID in order to output information on the sender of the signal at a receiver of the output signal.
  • This functionality can be achieved, for example, by an extended supply of functions of the motion signal processing device 7 or, for example, the
  • Arithmetic unit are provided. Also conceivable are additional, e.g. signaling or signal shaping elements associated with the signaling device 2 for this purpose.
  • the output signal AS which was generated by the signaling device 2 and preferably contains the information corresponding to the detected gesture, which was generated on the basis of the successful assignment, is used, for example, for signaling according to the invention
  • the output signal AS at a signal output for retrieval by the signaling device 13 provides. This is conceivable, for example, when the signaling device 13 and the signaling device 2 in a common device unit, for example a mobile terminal. 1
  • the information corresponding thereto for example, can also be displayed.
  • the signaling device 13 is preferably accommodated in a mobile terminal 1 of the type described above.
  • the signaling device 13 preferably comprises an output signal processing device 14, which processes the output signal AS, which at a signal input of the
  • Output signal processing device 14 has been detected.
  • the signaling device 13 may make use of a receiving device 15, e.g. from one arranged in the signaling device 13
  • the receiving device 15 may be formed in the usual way and depending on e.g. used transmission channel filters, amplifier elements, antenna elements, A / D, D / A converter, etc. have.
  • the output signal AS detected by the receiving device 15 is preferably provided to the output signal processing device 14.
  • the output signal processing device 14 comprises, for example, a computing unit, e.g. in the form of a signal processor 14a, which contains the information contained in the output signal AS, with the assignment of the
  • Motion signal BS to a gesture corresponding information from the output signal AS wins.
  • Output signal AS in addition to the arithmetic unit, for example, a memory device 16 may be provided, e.g. in the signaling device 13 or the output signal processing device 14, the memory content of which defines an assignment of the output signal AS obtained
  • output signal AS compares the output signal processing device 14 with the output signal patterns stored in the memory device 16, and if the output signal and the stored signal pattern are successfully matched, it retrieves the stored information associated with the stored signal pattern Signaling or output, whereupon the output signal AS is decoded on the side of the signaling device.
  • Information by the output signal processing device 14 is preferably conditioned as a signal such that the information can be output via a dedicated output device, e.g. to a human user or communication participant.
  • the signaling device for this purpose preferably contains an output device 17 for outputting the information to be signaled.
  • the conditioned information signal is provided to the output device from an output of the output signal processing device 14.
  • the dispenser 17 may conveniently incorporate an imaging device, such as an optical device. a screen, a sounding element such as a speaker, a haptic stimulus triggering element such as a speaker. a vibration device or a common device suitable for signaling to a human user or a combination of several such elements.
  • a mobile terminal 1 upon receiving an output signal AS which contains the information of the presence of a message gesture such as the extension of an arm of a user, can pivot an icon 18 ( see Fig. 5). Furthermore, it is provided to cause the signaling device 13 to vibrate as a result of such a gesture, for example, in order to notify the user of the signaling device 13 of the presence of such a gesture.
  • a message gesture such as the extension of an arm of a user
  • Information reproduction may e.g. are also to respond to an incoming message gesture information with a light-dark keying of the screen of the signaling device, e.g. in combination with the display of a subscriber ID.
  • the output signal processing device 14 may output, for example, a conditioned signal containing the additional Information also for the output, in particular by means of the output device 17 brings, for example, a signal which gives an indication to the user, which performs the gesture or causes the output signal AS / the movement signal BS.
  • the signaling device 13 is preferably designed such that it is based on an external data source, which has gesture information with underlying data
  • the external data source can be in the form of a memory card, a memory unit of another
  • Signaling device 2 an internal web-based or a network
  • Memory unit etc. be realized. It is also within the scope of the invention that the signaling device 2 is self-learning.
  • the signaling device 13 it does not contain any output signal processing device 14. If, for example, at the output of the motion signal processing device 7 already for output by means of output device 17 suitably processed output signal AS for
  • Signaling device 13 in a common device unit e.g. in a common mobile terminal 1, are housed.
  • a transmission via a transmission channel 11 could be omitted here.
  • a movement of the signaling device 2 is detected in a first step, for
  • Example via sensor 5, and a corresponding motion signal BS output The movement signal BS is generated as described above by moving the signaling device 2 accommodated, for example, in a mobile terminal 1, e.g. about an axis A or B or toward one of them etc.
  • the movement signal BS is assigned to a gesture, which is the gesture of nonverbal interpersonal communication belongs. This can be done by comparing motion signal patterns BSM with the motion signal BS as described above.
  • an output signal AS is generated, which contains information which corresponds to the gesture detected as a result of the assignment.
  • This output signal AS can be sent in a further step via a transmission channel 11. This is done on the above
  • the signaling method according to the invention comprises the step of generating information based on the output signal AS and the output, e.g. via an output device 17, the information which in the
  • Output signal AS is included and with the gesture of nonverbal
  • the step of generating information may be omitted if e.g. the output signal AS can be forwarded directly to an output device 17 in a suitably prepared manner.
  • the output signal AS which has been transmitted via a transmission channel, is received, for example, from a receiving device 15, as discussed above.
  • FIG. 6 shows a first embodiment of a moderated conference system KS with three mobile terminals M1, M2 and M3, each of which has the same structure, as described using the example of the mobile terminal M2 below in conjunction with FIG.
  • the two mobile terminals M1 and M2 are connected via a radio transmission link F1 or F2 to a first WLAN access point AP1, which is connected to a Local Area Network LAN.
  • the third mobile terminal M3 is connected via the radio transmission path F3 to a second WLAN access point AP2, which is also connected to the same Local Area Network LAN.
  • On the Local Area Network LAN are still a SIP server SS, a
  • Conference server MCU the z. B. to the standard ITU-T H.323 (09/99), sections 6.7 and 6.8 is formed, a conference control application KA, an IP telephone 11 and a personal computer PC with a keyboard T.
  • the conference server is responsible for mixing the media
  • the personal computer PC has a touch screen as a screen, which is controlled as a touch user interface Ul for conference control by the moderator.
  • the display on the touch screen Ul uses the Web protocol according to the lETF standard Hypertext Transfer Protocol http / 1.1 RFC 2616 (1999-06).
  • the respective switching states "mute” (mute) and “unmute” (loud) as well as possible word messages are displayed on special display fields A1, A2 and A3 by means of a special symbol W for the mobile terminals - see larger display on the touch Screen Ul of an alternative Internet phone 12, shown in FIG.
  • an alternative embodiment of an Internet phone 12 is shown (in which a personal computer PC is omitted), the telephone set TA with a dial pad WT and a touch screen with a touch user interface
  • a sensor S e.g. as movement or
  • Acceleration sensor S (corresponding to the sensor 5 in the mobile terminal 1 of FIG. 1) is formed.
  • An evaluation logic or detection logic EL (corresponding to the motion signal processing device 7 of the signaling device 2 in FIGS. 1 and 2) evaluates the detected movements accordingly, whose output signal (corresponding to AS in FIG. 2) from the transmitting unit SEND (corresponds to 10 in FIG Fig. 2) as INFO message is transmitted in addition to the audio A / ideosignal over the radio link F2.
  • connection or disconnection (sound or mute) of the mobile terminals M1, M2 and M3 of the participants is carried out by the moderator on the personal computer PC.
  • the moderator has the overview of all participants and can then use one with the
  • Conference system KS connected control application to perform this function and, so to speak, give the floor.
  • conference system KS according to the invention, there is now the advantage that muted participants of the conference can now speak.
  • Such a message can now be sent by each mobile terminal M1, M2 or M3 with a gesture e.g. Shaking or jerking, e.g. as a hand lift, to be made.
  • a movement is determined by means of the motion sensor S in the mobile communication terminal M2 (see FIG. 7) and by means of a
  • Evaluation logic EL evaluates the gesture to recognize the gesture, e.g. a hand lift as a fast movement in the direction of the arrow P1 - as shown in Fig. 1.
  • This information will now be e.g. in the form of an INFO message (standard RFC 2976) in the direction of the arrow P2 to the WLAN access point AP2 and then sent via the SIP signaling of the SIP server SS to the conference control application KA (arrow P3).
  • This can now semantically interpret M2's "hand lift" as a spoken word message and forward it to the touch user interface U1 on the personal computer PC of the moderator in the direction of the arrow P4 and represent it as symbol W.
  • IPv4 Internet protocol
  • WLAN wireless LAN
  • SIP session initiation Protocol
  • the moderator can after the message (W) within the
  • Conference user interface on the touch user interface Ul by means of an action by touching the screen on the touch user interface U1 in the area of the display panel A2, a change of the muting for sound switching or in reverse direction of the mobile terminal M2 of a participant of the conference effect.
  • a transmission from the mobile terminals M1, M2 and M3 to the local area network LAN can also take place via a UMTS network.
  • the information "lifting hands" can also be modeled in the audio signal (if not inaudible).
  • a vote of the participants can be carried out in a simple manner.
  • the arrangement is similar to that shown in Fig. 6 - however, in a vote, e.g. by means of a position sensor an upright holding of the mobile terminal then from the conference
  • Control application KA interpreted as "YES”
  • a hold in landscape mode is interpreted as "NO” and in turn displayed in all the participants M1 to M3 within the conference user interface on the touch user interface Ul in the display areas A1 to A3.
  • FIG. 9 - A second embodiment of the conference system KS according to the invention shown in FIG. 9 - similar to FIG. 6 - represents an auction.
  • KS Within the conference system KS, e.g. by a moderator described an object if necessary. Also with video transmission to the participants. Alternatively, this description can also be from a vending machine with electronic
  • Speech output (synthesized speech).
  • participants can increase the bid - much like in a real auction - by hand signals, i. E. e.g. by quickly lifting the mobile terminal M1 in the direction of arrow P5.
  • a corresponding info message "Hand lift" to the UMTS data network in the direction of arrow P6 and continue in the direction of arrow P7 to the
  • Auction application AA with a "text to speech” component implements and interprets the message “hand raise” as a bid, creating a
  • Price increase is made by a certain amount, the announcement of the new bid by means of a "text to speech" component through a transfer from the auctioneer, who is the moderator of the conference system KS, in
  • the conference control application KA will take over all the functions of the auctioneer, i. also the voice announcements - a visual display would not be necessary then.
  • gestures within a videoconference are used, e.g. by means of the gesture "Hand Lifting", which is now analyzed by the evaluation recognition logic EL connected to the camera in the video terminal. If this gesture is recognized, an INFO message “Hand Lifting” is also generated. Further
  • gestures detected by video shaking the head, nodding, fingering, e.g. Thumbs up, thumbs down.
  • the recognized gesture is in turn transmitted via the SIP server to the conference control application KA, which carries out a further interpretation and further processing, depending on the application, e.g. Interpretation of thumbs up or nods as approval in a vote.
  • Another possibility for an action after interpretation of a hand sign is e.g. switching the video signal seen by all subscribers, which is changing the presentation mode in an associated one
  • Collaboration connection e.g., viewing screen contents

Landscapes

  • Engineering & Computer Science (AREA)
  • Signal Processing (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Human Computer Interaction (AREA)
  • General Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Environmental & Geological Engineering (AREA)
  • User Interface Of Digital Computer (AREA)
  • Telephonic Communication Services (AREA)
PCT/EP2010/007534 2009-12-10 2010-12-10 Konferenzsystem und zugeordnetes signalgebungsverfahren WO2011069668A1 (de)

Priority Applications (4)

Application Number Priority Date Filing Date Title
EP10793164.4A EP2389753B1 (de) 2009-12-10 2010-12-10 Konferenzsystem und zugeordnetes signalgebungsverfahren
CN2010800153346A CN102369718A (zh) 2009-12-10 2010-12-10 会议系统和相关的发信方法
US13/379,447 US8965350B2 (en) 2009-12-10 2010-12-10 Conference system and associated signalling method
US14/589,329 US9397850B2 (en) 2009-12-10 2015-01-05 Conference system and associated signalling method

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
DE102009057725A DE102009057725A1 (de) 2009-12-10 2009-12-10 Signalgebende Vorrichtung, Signalisiervorrichtung, Signalgebungsverfahren sowie Signalisierverfahren
DE102009057725.4 2009-12-10

Related Child Applications (2)

Application Number Title Priority Date Filing Date
US13/379,447 A-371-Of-International US8965350B2 (en) 2009-12-10 2010-12-10 Conference system and associated signalling method
US14/589,329 Continuation US9397850B2 (en) 2009-12-10 2015-01-05 Conference system and associated signalling method

Publications (1)

Publication Number Publication Date
WO2011069668A1 true WO2011069668A1 (de) 2011-06-16

Family

ID=43618826

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/EP2010/007534 WO2011069668A1 (de) 2009-12-10 2010-12-10 Konferenzsystem und zugeordnetes signalgebungsverfahren

Country Status (5)

Country Link
US (2) US8965350B2 (zh)
EP (1) EP2389753B1 (zh)
CN (1) CN102369718A (zh)
DE (1) DE102009057725A1 (zh)
WO (1) WO2011069668A1 (zh)

Families Citing this family (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20130061153A1 (en) * 2011-09-07 2013-03-07 Avaya Inc. System and Method for Inserting a Control System Into a Conference
EP2812772A4 (en) * 2012-02-06 2015-10-07 Ericsson Telefon Ab L M USER DEVICE WITH IMPROVED FEEDBACK POSSIBILITIES
US9661269B2 (en) * 2013-10-03 2017-05-23 Polycom, Inc. System for enabling communications and conferencing between dissimilar computing devices including mobile computing devices
KR20160035884A (ko) * 2014-09-24 2016-04-01 삼성전자주식회사 회의 진행 장치 및 회의 진행 방법
US12022021B2 (en) 2015-03-06 2024-06-25 Apple Inc. Band adjustment for electronic devices
US9885627B2 (en) 2015-04-24 2018-02-06 Apple Inc. Drift compensation for force sensing devices
US10316884B2 (en) * 2015-06-18 2019-06-11 Matthew C. Prestwich Motion activated switch and method
US9865281B2 (en) * 2015-09-02 2018-01-09 International Business Machines Corporation Conversational analytics
TWI631507B (zh) * 2016-03-04 2018-08-01 德凡特未來股份有限公司 動作判識裝置及其控制方法
CN111355919B (zh) * 2018-12-24 2021-05-25 中移(杭州)信息技术有限公司 一种通信会话控制方法及装置
US10999555B1 (en) * 2020-06-01 2021-05-04 Fuze, Inc. Meeting room control via mobile device
US20220129668A1 (en) * 2020-10-27 2022-04-28 Cisco Technology, Inc., Applause gesture detection for video conferences

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20060098085A1 (en) * 2004-11-05 2006-05-11 Nichols Paul H Display management during a multi-party conversation
WO2006107463A1 (en) 2005-03-31 2006-10-12 Motorola, Inc. Method and apparatus for representing communication attributes
DE102007058586A1 (de) * 2006-12-04 2008-06-05 Kempf, Fabian, Dr. Vorrichtung und Verfahren zur Durchführung von Fernbesprechungen
US20080158334A1 (en) * 2006-12-29 2008-07-03 Nokia Corporation Visual Effects For Video Calls
WO2008138407A1 (en) 2007-05-11 2008-11-20 Sony Ericsson Mobile Communications Ab Methods and devices for generating multimedia content in response to simultaneous inputs from related portable devices
US20090009588A1 (en) * 2007-07-02 2009-01-08 Cisco Technology, Inc. Recognition of human gestures by a mobile phone

Family Cites Families (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH1141578A (ja) * 1997-07-18 1999-02-12 Ricoh Co Ltd 電子会議システム
CN1272932C (zh) * 2003-04-28 2006-08-30 王生安 基于ip网络的分布式多媒体会议系统
JP2007104354A (ja) 2005-10-05 2007-04-19 Aruze Corp テレビ会議システム、テレビ会議方法及びテレビ会議端末装置
JP5559691B2 (ja) * 2007-09-24 2014-07-23 クアルコム,インコーポレイテッド 音声及びビデオ通信のための機能向上したインタフェース
US8676224B2 (en) 2008-02-19 2014-03-18 Apple Inc. Speakerphone control for mobile device
ES2648049T3 (es) * 2008-07-25 2017-12-28 Qualcomm Incorporated Detección mejorada de gesto ondulatorio
US8503932B2 (en) * 2008-11-14 2013-08-06 Sony Mobile Comminications AB Portable communication device and remote motion input device
US8605039B2 (en) * 2009-03-06 2013-12-10 Zimpl Ab Text input
US9571625B2 (en) * 2009-08-11 2017-02-14 Lg Electronics Inc. Electronic device and control method thereof
KR20110123933A (ko) * 2010-05-10 2011-11-16 삼성전자주식회사 휴대단말의 기능 제공 방법 및 장치
KR20110125358A (ko) * 2010-05-13 2011-11-21 삼성전자주식회사 휴대 단말기의 표시부 제어 방법 및 장치
US9431027B2 (en) * 2011-01-26 2016-08-30 Honda Motor Co., Ltd. Synchronized gesture and speech production for humanoid robots using random numbers
WO2013170383A1 (en) * 2012-05-16 2013-11-21 Xtreme Interactions Inc. System, device and method for processing interlaced multimodal user input

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20060098085A1 (en) * 2004-11-05 2006-05-11 Nichols Paul H Display management during a multi-party conversation
WO2006107463A1 (en) 2005-03-31 2006-10-12 Motorola, Inc. Method and apparatus for representing communication attributes
DE102007058586A1 (de) * 2006-12-04 2008-06-05 Kempf, Fabian, Dr. Vorrichtung und Verfahren zur Durchführung von Fernbesprechungen
US20080158334A1 (en) * 2006-12-29 2008-07-03 Nokia Corporation Visual Effects For Video Calls
WO2008138407A1 (en) 2007-05-11 2008-11-20 Sony Ericsson Mobile Communications Ab Methods and devices for generating multimedia content in response to simultaneous inputs from related portable devices
US20090009588A1 (en) * 2007-07-02 2009-01-08 Cisco Technology, Inc. Recognition of human gestures by a mobile phone

Also Published As

Publication number Publication date
EP2389753A1 (de) 2011-11-30
US20150140979A1 (en) 2015-05-21
CN102369718A (zh) 2012-03-07
EP2389753B1 (de) 2014-07-30
US8965350B2 (en) 2015-02-24
DE102009057725A1 (de) 2011-06-16
DE102009057725A8 (de) 2013-04-11
US20130165088A1 (en) 2013-06-27
US9397850B2 (en) 2016-07-19

Similar Documents

Publication Publication Date Title
EP2389753B1 (de) Konferenzsystem und zugeordnetes signalgebungsverfahren
DE102006001607B4 (de) Verfahren und Systeme zur Übertragung von Ton- und Bilddaten
DE60311608T2 (de) Verfahren und Vorrichtung zum Schalten der Steuerung zwischen verbundenen Vorrichtungen
EP1912474B1 (de) Verfahren zum Betreiben einer Hörhilfe, sowie Hörhilfe
DE202013012457U1 (de) Digitale Vorrichtung
DE102008055180A1 (de) Steuersystem, Hörer und Steuerungsverfahren
Goggin Disability, mobiles, and social policy: New modes of communication and governance
EP1848245A2 (de) Hörgerät mit Quellentrennung und entsprechendes Verfahren
DE102021207673A1 (de) Interrupt für geräuschunterdrückende audiovorrichtungen
DE112022001291T5 (de) Aufzeichnen eines aus einem gemisch von tonströmen getrennten tons auf einer persönlichen einheit
EP4013043A1 (de) Videokonferenzsystem, verfahren zum übertragen von informationen und computerprogrammprodukt
DE102006058758A1 (de) Verfahren und Vorrichtung zum Steuern einer Telekommunikationsendeinrichtung
DE102009010449B4 (de) Verfahren zur Bereitstellung von Informationen und Kommunikationswahlgerät zur Durchführung eines Verfahrens zur Bereitstellung von Informationen
DE102019135799A1 (de) Verfahren zum Verbessern von Sprachverständlichkeit einer elektronischen Sprechverbindung und Headset zur Durchführung des Verfahrens
DE102007058585B4 (de) Vorrichtung und Verfahren zur Durchführung von Fernbesprechungen
EP2285626B1 (de) Kommunikationssystem und verfahren zum kommunizieren zwischen kommunikationseinrichtungen und zumindest einem nutzer eines kommunikationssystems
DE102023201075B3 (de) Verfahren zum Betrieb eines Hörinstruments und Hörsystem mit einem solchen Hörinstrument
DE102018200431A1 (de) Verfahren zur Ausgabe eines Signals einer Drahtloskommunikationsverbindung zwischen einem ersten Anwender in einem Fortbewegungsmittel und einem externen Kommunikationsteilnehmer
EP3371026B1 (de) Verfahren und vorrichtung zur kopplung von elektronischen geräten mit einem kraftfahrzeug
DE10135837A1 (de) Bedienen und Beobachten einer Anlage
DE10348149B4 (de) Verfahren zur Durchführung einer Telefonkonferenz
DE102013001771B4 (de) Verfahren zum Betreiben einer Funktionseinheit sowie Funktionseinheit
DE102007058586A1 (de) Vorrichtung und Verfahren zur Durchführung von Fernbesprechungen
DE102006004248A1 (de) Testverfahren für ein eine Benutzerschnittstelle aufweisendes technisches Gerät
EP2648348A1 (de) Verfahren zur Erfassung des Nutzerverhaltens und der Vorlieben von Radiohörern

Legal Events

Date Code Title Description
WWE Wipo information: entry into national phase

Ref document number: 201080015334.6

Country of ref document: CN

WWE Wipo information: entry into national phase

Ref document number: 2010793164

Country of ref document: EP

121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 10793164

Country of ref document: EP

Kind code of ref document: A1

WWE Wipo information: entry into national phase

Ref document number: 3825/KOLNP/2011

Country of ref document: IN

NENP Non-entry into the national phase

Ref country code: DE

WWE Wipo information: entry into national phase

Ref document number: 13379447

Country of ref document: US