WO2011069668A1 - Konferenzsystem und zugeordnetes signalgebungsverfahren - Google Patents
Konferenzsystem und zugeordnetes signalgebungsverfahren Download PDFInfo
- Publication number
- WO2011069668A1 WO2011069668A1 PCT/EP2010/007534 EP2010007534W WO2011069668A1 WO 2011069668 A1 WO2011069668 A1 WO 2011069668A1 EP 2010007534 W EP2010007534 W EP 2010007534W WO 2011069668 A1 WO2011069668 A1 WO 2011069668A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- conference
- signaling device
- conference system
- output signal
- motion signal
- Prior art date
Links
- 230000011664 signaling Effects 0.000 title claims abstract description 107
- 238000000034 method Methods 0.000 title claims abstract description 20
- 230000033001 locomotion Effects 0.000 claims abstract description 123
- 238000004891 communication Methods 0.000 claims abstract description 44
- 238000012545 processing Methods 0.000 claims abstract description 39
- 230000005540 biological transmission Effects 0.000 claims abstract description 38
- 230000001755 vocal effect Effects 0.000 claims description 18
- 208000003443 Unconsciousness Diseases 0.000 claims description 5
- 230000008569 process Effects 0.000 claims description 5
- 239000003550 marker Substances 0.000 claims 1
- 230000006870 function Effects 0.000 description 7
- 238000011156 evaluation Methods 0.000 description 4
- 230000008901 benefit Effects 0.000 description 3
- 230000001143 conditioned effect Effects 0.000 description 3
- 210000003813 thumb Anatomy 0.000 description 3
- 230000001133 acceleration Effects 0.000 description 2
- 230000009471 action Effects 0.000 description 2
- 230000008859 change Effects 0.000 description 2
- 238000013461 design Methods 0.000 description 2
- 238000001514 detection method Methods 0.000 description 2
- 238000010295 mobile communication Methods 0.000 description 2
- 229920001690 polydopamine Polymers 0.000 description 2
- 238000012546 transfer Methods 0.000 description 2
- 230000000007 visual effect Effects 0.000 description 2
- 238000013459 approach Methods 0.000 description 1
- 230000003750 conditioning effect Effects 0.000 description 1
- 230000001419 dependent effect Effects 0.000 description 1
- 238000010586 diagram Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 230000008451 emotion Effects 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 230000008921 facial expression Effects 0.000 description 1
- 238000001914 filtration Methods 0.000 description 1
- 210000003811 finger Anatomy 0.000 description 1
- 210000004247 hand Anatomy 0.000 description 1
- 238000003384 imaging method Methods 0.000 description 1
- 230000000977 initiatory effect Effects 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 238000007493 shaping process Methods 0.000 description 1
- 230000005236 sound signal Effects 0.000 description 1
- 238000001228 spectrum Methods 0.000 description 1
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L12/00—Data switching networks
- H04L12/02—Details
- H04L12/16—Arrangements for providing special services to substations
- H04L12/18—Arrangements for providing special services to substations for broadcast or conference, e.g. multicast
- H04L12/189—Arrangements for providing special services to substations for broadcast or conference, e.g. multicast in combination with wireless systems
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/017—Gesture based interaction, e.g. based on a set of recognized hand gestures
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M1/00—Substation equipment, e.g. for use by subscribers
- H04M1/253—Telephone sets using digital voice transmission
- H04M1/2535—Telephone sets using digital voice transmission adapted for voice communication over an Internet Protocol [IP] network
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M1/00—Substation equipment, e.g. for use by subscribers
- H04M1/72—Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
- H04M1/724—User interfaces specially adapted for cordless or mobile telephones
- H04M1/72403—User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality
- H04M1/72427—User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality for supporting games or graphical animations
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M1/00—Substation equipment, e.g. for use by subscribers
- H04M1/72—Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
- H04M1/724—User interfaces specially adapted for cordless or mobile telephones
- H04M1/72448—User interfaces specially adapted for cordless or mobile telephones with means for adapting the functionality of the device according to specific conditions
- H04M1/72454—User interfaces specially adapted for cordless or mobile telephones with means for adapting the functionality of the device according to specific conditions according to context-related or environment-related conditions
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04W—WIRELESS COMMUNICATION NETWORKS
- H04W4/00—Services specially adapted for wireless communication networks; Facilities therefor
- H04W4/16—Communication-related supplementary services, e.g. call-transfer or call-hold
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04W—WIRELESS COMMUNICATION NETWORKS
- H04W84/00—Network topologies
- H04W84/02—Hierarchically pre-organised networks, e.g. paging networks, cellular networks, WLAN [Wireless Local Area Network] or WLL [Wireless Local Loop]
- H04W84/10—Small scale networks; Flat hierarchical networks
- H04W84/12—WLAN [Wireless Local Area Networks]
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M2250/00—Details of telephonic subscriber devices
- H04M2250/12—Details of telephonic subscriber devices including a sensor for measuring a physical value, e.g. temperature or motion
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M2250/00—Details of telephonic subscriber devices
- H04M2250/62—Details of telephonic subscriber devices user interface aspects of conference calls
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M3/00—Automatic or semi-automatic exchanges
- H04M3/42—Systems providing special services or facilities to subscribers
- H04M3/56—Arrangements for connecting several subscribers to a common circuit, i.e. affording conference facilities
Definitions
- the invention relates to a conference system with at least two mobile
- Terminals each with a signaling device and an associated signaling method.
- mobile terminals serving interpersonal communication e.g. Mobile phones, smartphones, PDAs, tablet PCs and laptops can be used in various ways.
- the user input can be made via keyboard, with input devices, via voice control, by controller, or in some other way. It is also known, by movements of an input device, e.g. the finger to control the device functions on a touch-sensitive user interface. It is also known to move the terminal in order, for example, to influence the presentation mode of the mobile terminal, for example to switch from a portrait reproduction mode to a landscape format, see for example: Apple iPhone (http://www.apple.com/phone/ ).
- control and input devices and methods known in the context of mobile terminals for interpersonal communication have the disadvantage that they are not accessible to gestures of interpersonal non-verbal communication, for control or input.
- Such gestures of interpersonal nonverbal communication in particular non-specific such gestures, remain closed to the mobile communication terminals in the prior art.
- WO 2006/107463 A1 discloses a method and a device for visual presentation, which determines and displays predetermined stored speech characteristics from a speech connection of a communication device.
- WO 2008/138407 A1 discloses a mobile terminal with a WLAN transceiver and a motion sensor, wherein a multimedia object is generated by the determined movement.
- telephone or video conferences are increasingly being conducted for joint coordination or for the efficient and timely dissemination of information. It may well happen that the telephone conferences reach a larger number of participants, eg 10 or 20 participants in larger companies, possibly even much more.
- conference bridges allow different modes of operation - e.g. For example, all subscribers can communicate bidirectionally, i. Speaking at the same time as listening to all other participants. It may happen that participants bring in disturbing ambient noise that disturbs the quality of the conference acoustics. Therefore, the use of the mode of a so-called moderated conference may be useful,
- connection or disconnection (muting) of the participants is done by moderators.
- the moderator has an overview of all participants and can then perform this function via a control application connected to the conference system and, so to speak, give the floor.
- the subscriber In order to then speak up, the subscriber must call the presenter via an additional telephone call, e.g. Call the presenter's mobile phone to hear that he is "loudly" dialing in to speak at the conference.
- the conference system comprises at least two mobile terminals, each with a signaling device containing a sensor which detects movement of the signaling device and a
- the conference system includes a transmitting device for supplying the output signal to a transmission channel, which is in each case connected to a local area network.
- a transmitting device for supplying the output signal to a transmission channel, which is in each case connected to a local area network.
- This is a telephone preferably an IP phone (11) for a moderator and a screen for conference control connected.
- a signaling device for mobile terminals which is suitable for interpersonal communication, which includes a sensor which detects a movement of the signaling device and outputs a corresponding motion signal.
- the signaling device includes a motion signal processing device that detects the motion signal of a gesture of non-verbal interpersonal communication (including, but not limited to, gestures due to the movement of body parts for the purpose of interpersonal communication
- the signaling device further includes a transmitting device for
- a transmission channel This may be a wireless or wireline transmission channel or a
- the motion signal processing device assigns the motion signal to a motion signal pattern. This can be one
- the motion signal processing device generates an output signal which contains information about the type of detected gesture of the nonverbal interpersonal
- the output signal contains additional information, in particular those which give an indication of the causer of the motion signal.
- the motion signal processing apparatus is capable of detecting the non-verbal human gestures of noticing, in particular the winking and / or shaking and / or a message, due to movement of the signaling device.
- the motion signal processing device can recognize user-specific motion signal patterns and / or store them in the signaling device, in particular for reassignment.
- the senor and / or the motion signal processing device is designed so that unconscious gestures of the non-verbal interpersonal communication are suppressed, so that they no motion signal and / or
- the signaling device processes and / or stores motion signal patterns from an external data source which is not part of the signaling device.
- the transmission channel is between the
- the transmission channel may be formed in the form of a radio link. Transmitting and receiving device are spaced from each other.
- the conference system includes a conference control signaling device that includes an output device that outputs the information of the output corresponding to the gesture of the non-verbal interpersonal communication.
- the signaling device further includes an output signal processing device, which the
- the signaling device includes a receiving device which receives the output signal. This may for example be part of the output signal processing device.
- Information also give an indication of the cause of the motion signal.
- the information to be output is generated on the basis of signal patterns stored in the output signal processing device, in particular by assignment of such to the output signal.
- the output of the information takes place visually on a suitable means of the signaling device, in particular a screen, acoustically by a suitable means of the signaling device, and / or by means of the signaling device which generates haptic stimuli, in particular vibration.
- the signaling device processes signal patterns from an external data source which is not part of the signaling device.
- a signaling method for the conference system is proposed, which as the first step the
- Communication is assigned and in a third step an output signal in Trap of a successful assignment generated containing information with the gesture of non-verbal interpersonal communication
- the fourth step is to send the output signal over a transmission channel.
- the fifth step is to receive the 5 output signal transmitted over the transmission channel.
- a signaling method which as a first step provides for the generation of information on the basis of the output signal, wherein the information to with the gesture of non-verbal interpersonal communication
- FIG. 1 shows an embodiment of a mobile terminal with a signaling device for explaining the invention
- Fig. 2 shows schematically a further embodiment of a mobile terminal with a signaling device for explaining the invention
- FIG. 3 schematically shows a signal of a sensor of a signaling device according to FIG. 1 or FIG. 2 for explaining the invention
- Fig. 5 shows another embodiment of a mobile terminal with a
- FIG. 6 shows a first embodiment of a conference system according to the invention using the signaling device
- Fig. 7 is a block diagram of a mobile terminal used in the embodiment of the present invention of the conference system shown in Fig. 6;
- FIG. 8 shows an IP telephone with display of a user interface for use in a further conference system according to the invention
- FIG. 9 shows a second embodiment of a conference system according to the invention for carrying out an auction using mobile terminals shown in FIG. 7.
- gesture refers in the context of the present invention, the symbolic movement of one or more specific body parts of a person, which is mainly used for communication, contact and the like purposes.
- the gesture as part of human gestures is to be distinguished here from the movements of facial expressions.
- the term gesture in its meaning largely approaches the meaning of the term gesture (movement of a body part) and differs, if necessary, in that gestures sometimes have a certain symbolism.
- the term "interpersonal communication” refers to communication between people, that is to say, in particular, the mutual interaction of individuals.
- FIG. 1 shows by way of example a mobile terminal 1 with a screen 1a and icons 18 displayed thereon, which contains a signaling device 2.
- the mobile terminal for example a personal digital
- PDA Personal Computer Assistant
- a mobile phone a smartphone, a tablet PC or similar device, is capable of communicating over one
- Such communication links include, for example, telephone connections, in particular wireless and
- these devices can connect to multiple devices
- the signaling device 2 is preferably integrated in a mobile terminal 1, for example by its technical equipment such as e.g. the antenna, the power supply, a display device, etc. use.
- the device 2 may be a separate, separate
- Be device that is not part of a mobile device.
- the signaling device 2 is preferably suitable for movement, in particular during operation, by a user, for example by holding by hand and moving the same, without difficulty.
- the signaling device 2 or the mobile terminal 1 receiving it has e.g. low weight and handy design and is therefore preferably designed as a handheld device or as a portable device.
- movements in the direction of arrow 3 about an axis A are possible as well as those in the direction of arrow 4 about an axis B or those about an axis C.
- movements in the direction of arrow 3 about an axis A are possible as well as those in the direction of arrow 4 about an axis B or those about an axis C.
- the signaling device 2 includes a sensor 5, for example an acceleration sensor, which provides movement of the signalers
- the sensor 5 On the basis of a detected movement, the sensor 5 provides or delivers a movement signal BS at an output 6, see FIG. 2.
- This movement signal BS corresponding to the movement becomes a
- Motion signal processing device 7 supplied to or retrieved from this.
- the motion signal processing device 7 comprises, for example, a computing unit, for example a computerized computing unit, such as e.g. a signal processor 8.
- the motion signal processing device 7 interprets the motion signal BS of the sensor 5 as to whether in the
- Motion signal BS is a motion signal pattern BSM, which is a gesture of non-verbal interpersonal communication, in particular the non-verbal interpersonal communication can be assigned.
- the motion signal processing device 7 can compare, for example, the motion signal BS with defined motion signal patterns BSM of such gestures, which are stored and retrievable, for example, in a memory device 9 or the motion signal processing device 7 are known.
- the storage device 9 may be part of the signalers
- device 2 in particular part of the motion signal processing device 7, may also be e.g. be an external data source.
- the motion signal processing apparatus 7 interprets the
- Motion signal BS preferably in terms of what kind the the
- Motion signal BS is assignable gesture.
- Such gestures of non-verbal interpersonal communication are, for example, those of human noticing as a result of a movement of a body part, such as: hand waving, hand shaking, hand / arm pivoting, or, for example, raising an arm (motion signal BS) Gesture is generated by moving the held signaling device 2 as described above).
- any other types of such gestures are conceivable.
- Such interpersonal gestures described above are expected, for example, where users of the signaling device 2 want to draw attention to themselves, eg in conferences, in emergency situations, etc.
- a gesture could be, for example, waving a participant in a conference, making a bid in an auction by raising a hand, knocking on a party who muted their communication device, reporting message to the moderator of a conference, reporting in a conference to signal a request to talk, the interpretation of the intensity of a Scrissains as emotion to the evaluation function (as for example similar to the clapping), and for example the mixing of a sorting.
- the movement signal processing device 7 outputs an output signal AS containing information corresponding to the assignment.
- the output signal AS contains, for example, the information that a gesture of non-verbal interpersonal communication due to the
- Moving signal BS has been detected by assignment or is present.
- the output signal AS may preferably also contain information as to which type of gesture has been recognized.
- the signaling device 2 further includes, in an exemplary preferred embodiment, a transmitting device 10 that generates the generated signal
- Transmission channel 11 which, for example, a connection to another mobile terminal or for example another, the
- the transmission channel may be, for example, a connection with one described in connection with FIG. 4 below
- the transmitting device 10 may include conventional means for conditioning the output signal for transmission purposes, such as means for encoding (e.g., D / A or A / D) and / or amplifying the output signal, filtering, etc.
- the transmitting device 10 may have, for example, an antenna 11a or a connection for a cable-bound transmission channel 11.
- the transmission channel 11 can be a transmission channel which is customary in communications technology, for example a wireless in the form of a radio link or conducted or both.
- the transmission channel 1 can be implemented in many ways and formed, for example, using conventional protocols and methods. Both analog and digital
- Transmission channel shapes are conceivable.
- the transmission channel 11 is formed between a transmitting device 10 and a receiving device, in particular if they are spatially separated from each other or removed.
- a possible movement signal BS is shown as an example in FIG. 3 as amplitude a (t) over time t.
- the sensor 5 responds to a specific movement of the signaling device 2, for example with a frequency change (range B1) and an altered (in this case larger) amplitude a (range B2) in the waveform 12 Changes in the waveform 12 are reproducible upon repeating the particular motion.
- the reproducible changed waveform 12 can be
- a movement signal pattern BSM for example as a movement signal pattern BSM, e.g. stored as a curve or in digitized form.
- the signal components lying above a threshold value S are detected (clipping, cutting off of the signal peaks) and these are used, for example. as a digital motion signal BS
- Moving signal BS corresponds, assignable motion signal pattern BSM can be obtained. Shown is also a threshold S in the form of a dashed line. If the movement signal amplitudes a (t) are below, no movement signal BS is output, for example.
- the movement signal patterns BSM generated from movement signals BS can be stored in the signaling device 2. However, it is also conceivable that the signaling device. 2
- the external data source may e.g. in the form of a memory card, a memory unit of another signaling device 2, an internal web-based or a network memory unit, etc. It is also within the scope of the invention that the
- Signaling device 2 is self-learning. This can be realized, for example by making a gesture of nonverbal interpersonal
- Communication performs while the signaling device 2 detects the motion signal BS generated thereby by means of the sensor 5 and as
- Motion signal pattern BSM e.g. in the signaling device 2, stores. In this way, user-specific gestures can be detected.
- the signaling device 2 deliberately differentiates from unconscious gestures.
- motion signals BS of unconscious gestures for example, can be suppressed.
- the intensity of a detected gesture for example, the intensity of a detected gesture.
- the sensor 5 usually detects a weak signal, since the movement amplitudes or the
- Threshold S (range B2).
- the output signal AS can also be supplemented with further information before it can be output - e.g. via a transmission channel - is made available. So it is e.g. make sense to add a user ID in order to output information on the sender of the signal at a receiver of the output signal.
- This functionality can be achieved, for example, by an extended supply of functions of the motion signal processing device 7 or, for example, the
- Arithmetic unit are provided. Also conceivable are additional, e.g. signaling or signal shaping elements associated with the signaling device 2 for this purpose.
- the output signal AS which was generated by the signaling device 2 and preferably contains the information corresponding to the detected gesture, which was generated on the basis of the successful assignment, is used, for example, for signaling according to the invention
- the output signal AS at a signal output for retrieval by the signaling device 13 provides. This is conceivable, for example, when the signaling device 13 and the signaling device 2 in a common device unit, for example a mobile terminal. 1
- the information corresponding thereto for example, can also be displayed.
- the signaling device 13 is preferably accommodated in a mobile terminal 1 of the type described above.
- the signaling device 13 preferably comprises an output signal processing device 14, which processes the output signal AS, which at a signal input of the
- Output signal processing device 14 has been detected.
- the signaling device 13 may make use of a receiving device 15, e.g. from one arranged in the signaling device 13
- the receiving device 15 may be formed in the usual way and depending on e.g. used transmission channel filters, amplifier elements, antenna elements, A / D, D / A converter, etc. have.
- the output signal AS detected by the receiving device 15 is preferably provided to the output signal processing device 14.
- the output signal processing device 14 comprises, for example, a computing unit, e.g. in the form of a signal processor 14a, which contains the information contained in the output signal AS, with the assignment of the
- Motion signal BS to a gesture corresponding information from the output signal AS wins.
- Output signal AS in addition to the arithmetic unit, for example, a memory device 16 may be provided, e.g. in the signaling device 13 or the output signal processing device 14, the memory content of which defines an assignment of the output signal AS obtained
- output signal AS compares the output signal processing device 14 with the output signal patterns stored in the memory device 16, and if the output signal and the stored signal pattern are successfully matched, it retrieves the stored information associated with the stored signal pattern Signaling or output, whereupon the output signal AS is decoded on the side of the signaling device.
- Information by the output signal processing device 14 is preferably conditioned as a signal such that the information can be output via a dedicated output device, e.g. to a human user or communication participant.
- the signaling device for this purpose preferably contains an output device 17 for outputting the information to be signaled.
- the conditioned information signal is provided to the output device from an output of the output signal processing device 14.
- the dispenser 17 may conveniently incorporate an imaging device, such as an optical device. a screen, a sounding element such as a speaker, a haptic stimulus triggering element such as a speaker. a vibration device or a common device suitable for signaling to a human user or a combination of several such elements.
- a mobile terminal 1 upon receiving an output signal AS which contains the information of the presence of a message gesture such as the extension of an arm of a user, can pivot an icon 18 ( see Fig. 5). Furthermore, it is provided to cause the signaling device 13 to vibrate as a result of such a gesture, for example, in order to notify the user of the signaling device 13 of the presence of such a gesture.
- a message gesture such as the extension of an arm of a user
- Information reproduction may e.g. are also to respond to an incoming message gesture information with a light-dark keying of the screen of the signaling device, e.g. in combination with the display of a subscriber ID.
- the output signal processing device 14 may output, for example, a conditioned signal containing the additional Information also for the output, in particular by means of the output device 17 brings, for example, a signal which gives an indication to the user, which performs the gesture or causes the output signal AS / the movement signal BS.
- the signaling device 13 is preferably designed such that it is based on an external data source, which has gesture information with underlying data
- the external data source can be in the form of a memory card, a memory unit of another
- Signaling device 2 an internal web-based or a network
- Memory unit etc. be realized. It is also within the scope of the invention that the signaling device 2 is self-learning.
- the signaling device 13 it does not contain any output signal processing device 14. If, for example, at the output of the motion signal processing device 7 already for output by means of output device 17 suitably processed output signal AS for
- Signaling device 13 in a common device unit e.g. in a common mobile terminal 1, are housed.
- a transmission via a transmission channel 11 could be omitted here.
- a movement of the signaling device 2 is detected in a first step, for
- Example via sensor 5, and a corresponding motion signal BS output The movement signal BS is generated as described above by moving the signaling device 2 accommodated, for example, in a mobile terminal 1, e.g. about an axis A or B or toward one of them etc.
- the movement signal BS is assigned to a gesture, which is the gesture of nonverbal interpersonal communication belongs. This can be done by comparing motion signal patterns BSM with the motion signal BS as described above.
- an output signal AS is generated, which contains information which corresponds to the gesture detected as a result of the assignment.
- This output signal AS can be sent in a further step via a transmission channel 11. This is done on the above
- the signaling method according to the invention comprises the step of generating information based on the output signal AS and the output, e.g. via an output device 17, the information which in the
- Output signal AS is included and with the gesture of nonverbal
- the step of generating information may be omitted if e.g. the output signal AS can be forwarded directly to an output device 17 in a suitably prepared manner.
- the output signal AS which has been transmitted via a transmission channel, is received, for example, from a receiving device 15, as discussed above.
- FIG. 6 shows a first embodiment of a moderated conference system KS with three mobile terminals M1, M2 and M3, each of which has the same structure, as described using the example of the mobile terminal M2 below in conjunction with FIG.
- the two mobile terminals M1 and M2 are connected via a radio transmission link F1 or F2 to a first WLAN access point AP1, which is connected to a Local Area Network LAN.
- the third mobile terminal M3 is connected via the radio transmission path F3 to a second WLAN access point AP2, which is also connected to the same Local Area Network LAN.
- On the Local Area Network LAN are still a SIP server SS, a
- Conference server MCU the z. B. to the standard ITU-T H.323 (09/99), sections 6.7 and 6.8 is formed, a conference control application KA, an IP telephone 11 and a personal computer PC with a keyboard T.
- the conference server is responsible for mixing the media
- the personal computer PC has a touch screen as a screen, which is controlled as a touch user interface Ul for conference control by the moderator.
- the display on the touch screen Ul uses the Web protocol according to the lETF standard Hypertext Transfer Protocol http / 1.1 RFC 2616 (1999-06).
- the respective switching states "mute” (mute) and “unmute” (loud) as well as possible word messages are displayed on special display fields A1, A2 and A3 by means of a special symbol W for the mobile terminals - see larger display on the touch Screen Ul of an alternative Internet phone 12, shown in FIG.
- an alternative embodiment of an Internet phone 12 is shown (in which a personal computer PC is omitted), the telephone set TA with a dial pad WT and a touch screen with a touch user interface
- a sensor S e.g. as movement or
- Acceleration sensor S (corresponding to the sensor 5 in the mobile terminal 1 of FIG. 1) is formed.
- An evaluation logic or detection logic EL (corresponding to the motion signal processing device 7 of the signaling device 2 in FIGS. 1 and 2) evaluates the detected movements accordingly, whose output signal (corresponding to AS in FIG. 2) from the transmitting unit SEND (corresponds to 10 in FIG Fig. 2) as INFO message is transmitted in addition to the audio A / ideosignal over the radio link F2.
- connection or disconnection (sound or mute) of the mobile terminals M1, M2 and M3 of the participants is carried out by the moderator on the personal computer PC.
- the moderator has the overview of all participants and can then use one with the
- Conference system KS connected control application to perform this function and, so to speak, give the floor.
- conference system KS according to the invention, there is now the advantage that muted participants of the conference can now speak.
- Such a message can now be sent by each mobile terminal M1, M2 or M3 with a gesture e.g. Shaking or jerking, e.g. as a hand lift, to be made.
- a movement is determined by means of the motion sensor S in the mobile communication terminal M2 (see FIG. 7) and by means of a
- Evaluation logic EL evaluates the gesture to recognize the gesture, e.g. a hand lift as a fast movement in the direction of the arrow P1 - as shown in Fig. 1.
- This information will now be e.g. in the form of an INFO message (standard RFC 2976) in the direction of the arrow P2 to the WLAN access point AP2 and then sent via the SIP signaling of the SIP server SS to the conference control application KA (arrow P3).
- This can now semantically interpret M2's "hand lift" as a spoken word message and forward it to the touch user interface U1 on the personal computer PC of the moderator in the direction of the arrow P4 and represent it as symbol W.
- IPv4 Internet protocol
- WLAN wireless LAN
- SIP session initiation Protocol
- the moderator can after the message (W) within the
- Conference user interface on the touch user interface Ul by means of an action by touching the screen on the touch user interface U1 in the area of the display panel A2, a change of the muting for sound switching or in reverse direction of the mobile terminal M2 of a participant of the conference effect.
- a transmission from the mobile terminals M1, M2 and M3 to the local area network LAN can also take place via a UMTS network.
- the information "lifting hands" can also be modeled in the audio signal (if not inaudible).
- a vote of the participants can be carried out in a simple manner.
- the arrangement is similar to that shown in Fig. 6 - however, in a vote, e.g. by means of a position sensor an upright holding of the mobile terminal then from the conference
- Control application KA interpreted as "YES”
- a hold in landscape mode is interpreted as "NO” and in turn displayed in all the participants M1 to M3 within the conference user interface on the touch user interface Ul in the display areas A1 to A3.
- FIG. 9 - A second embodiment of the conference system KS according to the invention shown in FIG. 9 - similar to FIG. 6 - represents an auction.
- KS Within the conference system KS, e.g. by a moderator described an object if necessary. Also with video transmission to the participants. Alternatively, this description can also be from a vending machine with electronic
- Speech output (synthesized speech).
- participants can increase the bid - much like in a real auction - by hand signals, i. E. e.g. by quickly lifting the mobile terminal M1 in the direction of arrow P5.
- a corresponding info message "Hand lift" to the UMTS data network in the direction of arrow P6 and continue in the direction of arrow P7 to the
- Auction application AA with a "text to speech” component implements and interprets the message “hand raise” as a bid, creating a
- Price increase is made by a certain amount, the announcement of the new bid by means of a "text to speech" component through a transfer from the auctioneer, who is the moderator of the conference system KS, in
- the conference control application KA will take over all the functions of the auctioneer, i. also the voice announcements - a visual display would not be necessary then.
- gestures within a videoconference are used, e.g. by means of the gesture "Hand Lifting", which is now analyzed by the evaluation recognition logic EL connected to the camera in the video terminal. If this gesture is recognized, an INFO message “Hand Lifting” is also generated. Further
- gestures detected by video shaking the head, nodding, fingering, e.g. Thumbs up, thumbs down.
- the recognized gesture is in turn transmitted via the SIP server to the conference control application KA, which carries out a further interpretation and further processing, depending on the application, e.g. Interpretation of thumbs up or nods as approval in a vote.
- Another possibility for an action after interpretation of a hand sign is e.g. switching the video signal seen by all subscribers, which is changing the presentation mode in an associated one
- Collaboration connection e.g., viewing screen contents
Landscapes
- Engineering & Computer Science (AREA)
- Signal Processing (AREA)
- Computer Networks & Wireless Communication (AREA)
- Human Computer Interaction (AREA)
- General Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Environmental & Geological Engineering (AREA)
- User Interface Of Digital Computer (AREA)
- Telephonic Communication Services (AREA)
Priority Applications (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
EP10793164.4A EP2389753B1 (de) | 2009-12-10 | 2010-12-10 | Konferenzsystem und zugeordnetes signalgebungsverfahren |
CN2010800153346A CN102369718A (zh) | 2009-12-10 | 2010-12-10 | 会议系统和相关的发信方法 |
US13/379,447 US8965350B2 (en) | 2009-12-10 | 2010-12-10 | Conference system and associated signalling method |
US14/589,329 US9397850B2 (en) | 2009-12-10 | 2015-01-05 | Conference system and associated signalling method |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
DE102009057725A DE102009057725A1 (de) | 2009-12-10 | 2009-12-10 | Signalgebende Vorrichtung, Signalisiervorrichtung, Signalgebungsverfahren sowie Signalisierverfahren |
DE102009057725.4 | 2009-12-10 |
Related Child Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/379,447 A-371-Of-International US8965350B2 (en) | 2009-12-10 | 2010-12-10 | Conference system and associated signalling method |
US14/589,329 Continuation US9397850B2 (en) | 2009-12-10 | 2015-01-05 | Conference system and associated signalling method |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2011069668A1 true WO2011069668A1 (de) | 2011-06-16 |
Family
ID=43618826
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/EP2010/007534 WO2011069668A1 (de) | 2009-12-10 | 2010-12-10 | Konferenzsystem und zugeordnetes signalgebungsverfahren |
Country Status (5)
Country | Link |
---|---|
US (2) | US8965350B2 (zh) |
EP (1) | EP2389753B1 (zh) |
CN (1) | CN102369718A (zh) |
DE (1) | DE102009057725A1 (zh) |
WO (1) | WO2011069668A1 (zh) |
Families Citing this family (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20130061153A1 (en) * | 2011-09-07 | 2013-03-07 | Avaya Inc. | System and Method for Inserting a Control System Into a Conference |
EP2812772A4 (en) * | 2012-02-06 | 2015-10-07 | Ericsson Telefon Ab L M | USER DEVICE WITH IMPROVED FEEDBACK POSSIBILITIES |
US9661269B2 (en) * | 2013-10-03 | 2017-05-23 | Polycom, Inc. | System for enabling communications and conferencing between dissimilar computing devices including mobile computing devices |
KR20160035884A (ko) * | 2014-09-24 | 2016-04-01 | 삼성전자주식회사 | 회의 진행 장치 및 회의 진행 방법 |
US12022021B2 (en) | 2015-03-06 | 2024-06-25 | Apple Inc. | Band adjustment for electronic devices |
US9885627B2 (en) | 2015-04-24 | 2018-02-06 | Apple Inc. | Drift compensation for force sensing devices |
US10316884B2 (en) * | 2015-06-18 | 2019-06-11 | Matthew C. Prestwich | Motion activated switch and method |
US9865281B2 (en) * | 2015-09-02 | 2018-01-09 | International Business Machines Corporation | Conversational analytics |
TWI631507B (zh) * | 2016-03-04 | 2018-08-01 | 德凡特未來股份有限公司 | 動作判識裝置及其控制方法 |
CN111355919B (zh) * | 2018-12-24 | 2021-05-25 | 中移(杭州)信息技术有限公司 | 一种通信会话控制方法及装置 |
US10999555B1 (en) * | 2020-06-01 | 2021-05-04 | Fuze, Inc. | Meeting room control via mobile device |
US20220129668A1 (en) * | 2020-10-27 | 2022-04-28 | Cisco Technology, Inc., | Applause gesture detection for video conferences |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20060098085A1 (en) * | 2004-11-05 | 2006-05-11 | Nichols Paul H | Display management during a multi-party conversation |
WO2006107463A1 (en) | 2005-03-31 | 2006-10-12 | Motorola, Inc. | Method and apparatus for representing communication attributes |
DE102007058586A1 (de) * | 2006-12-04 | 2008-06-05 | Kempf, Fabian, Dr. | Vorrichtung und Verfahren zur Durchführung von Fernbesprechungen |
US20080158334A1 (en) * | 2006-12-29 | 2008-07-03 | Nokia Corporation | Visual Effects For Video Calls |
WO2008138407A1 (en) | 2007-05-11 | 2008-11-20 | Sony Ericsson Mobile Communications Ab | Methods and devices for generating multimedia content in response to simultaneous inputs from related portable devices |
US20090009588A1 (en) * | 2007-07-02 | 2009-01-08 | Cisco Technology, Inc. | Recognition of human gestures by a mobile phone |
Family Cites Families (13)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH1141578A (ja) * | 1997-07-18 | 1999-02-12 | Ricoh Co Ltd | 電子会議システム |
CN1272932C (zh) * | 2003-04-28 | 2006-08-30 | 王生安 | 基于ip网络的分布式多媒体会议系统 |
JP2007104354A (ja) | 2005-10-05 | 2007-04-19 | Aruze Corp | テレビ会議システム、テレビ会議方法及びテレビ会議端末装置 |
JP5559691B2 (ja) * | 2007-09-24 | 2014-07-23 | クアルコム,インコーポレイテッド | 音声及びビデオ通信のための機能向上したインタフェース |
US8676224B2 (en) | 2008-02-19 | 2014-03-18 | Apple Inc. | Speakerphone control for mobile device |
ES2648049T3 (es) * | 2008-07-25 | 2017-12-28 | Qualcomm Incorporated | Detección mejorada de gesto ondulatorio |
US8503932B2 (en) * | 2008-11-14 | 2013-08-06 | Sony Mobile Comminications AB | Portable communication device and remote motion input device |
US8605039B2 (en) * | 2009-03-06 | 2013-12-10 | Zimpl Ab | Text input |
US9571625B2 (en) * | 2009-08-11 | 2017-02-14 | Lg Electronics Inc. | Electronic device and control method thereof |
KR20110123933A (ko) * | 2010-05-10 | 2011-11-16 | 삼성전자주식회사 | 휴대단말의 기능 제공 방법 및 장치 |
KR20110125358A (ko) * | 2010-05-13 | 2011-11-21 | 삼성전자주식회사 | 휴대 단말기의 표시부 제어 방법 및 장치 |
US9431027B2 (en) * | 2011-01-26 | 2016-08-30 | Honda Motor Co., Ltd. | Synchronized gesture and speech production for humanoid robots using random numbers |
WO2013170383A1 (en) * | 2012-05-16 | 2013-11-21 | Xtreme Interactions Inc. | System, device and method for processing interlaced multimodal user input |
-
2009
- 2009-12-10 DE DE102009057725A patent/DE102009057725A1/de active Pending
-
2010
- 2010-12-10 EP EP10793164.4A patent/EP2389753B1/de active Active
- 2010-12-10 WO PCT/EP2010/007534 patent/WO2011069668A1/de active Application Filing
- 2010-12-10 US US13/379,447 patent/US8965350B2/en active Active
- 2010-12-10 CN CN2010800153346A patent/CN102369718A/zh active Pending
-
2015
- 2015-01-05 US US14/589,329 patent/US9397850B2/en active Active
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20060098085A1 (en) * | 2004-11-05 | 2006-05-11 | Nichols Paul H | Display management during a multi-party conversation |
WO2006107463A1 (en) | 2005-03-31 | 2006-10-12 | Motorola, Inc. | Method and apparatus for representing communication attributes |
DE102007058586A1 (de) * | 2006-12-04 | 2008-06-05 | Kempf, Fabian, Dr. | Vorrichtung und Verfahren zur Durchführung von Fernbesprechungen |
US20080158334A1 (en) * | 2006-12-29 | 2008-07-03 | Nokia Corporation | Visual Effects For Video Calls |
WO2008138407A1 (en) | 2007-05-11 | 2008-11-20 | Sony Ericsson Mobile Communications Ab | Methods and devices for generating multimedia content in response to simultaneous inputs from related portable devices |
US20090009588A1 (en) * | 2007-07-02 | 2009-01-08 | Cisco Technology, Inc. | Recognition of human gestures by a mobile phone |
Also Published As
Publication number | Publication date |
---|---|
EP2389753A1 (de) | 2011-11-30 |
US20150140979A1 (en) | 2015-05-21 |
CN102369718A (zh) | 2012-03-07 |
EP2389753B1 (de) | 2014-07-30 |
US8965350B2 (en) | 2015-02-24 |
DE102009057725A1 (de) | 2011-06-16 |
DE102009057725A8 (de) | 2013-04-11 |
US20130165088A1 (en) | 2013-06-27 |
US9397850B2 (en) | 2016-07-19 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
EP2389753B1 (de) | Konferenzsystem und zugeordnetes signalgebungsverfahren | |
DE102006001607B4 (de) | Verfahren und Systeme zur Übertragung von Ton- und Bilddaten | |
DE60311608T2 (de) | Verfahren und Vorrichtung zum Schalten der Steuerung zwischen verbundenen Vorrichtungen | |
EP1912474B1 (de) | Verfahren zum Betreiben einer Hörhilfe, sowie Hörhilfe | |
DE202013012457U1 (de) | Digitale Vorrichtung | |
DE102008055180A1 (de) | Steuersystem, Hörer und Steuerungsverfahren | |
Goggin | Disability, mobiles, and social policy: New modes of communication and governance | |
EP1848245A2 (de) | Hörgerät mit Quellentrennung und entsprechendes Verfahren | |
DE102021207673A1 (de) | Interrupt für geräuschunterdrückende audiovorrichtungen | |
DE112022001291T5 (de) | Aufzeichnen eines aus einem gemisch von tonströmen getrennten tons auf einer persönlichen einheit | |
EP4013043A1 (de) | Videokonferenzsystem, verfahren zum übertragen von informationen und computerprogrammprodukt | |
DE102006058758A1 (de) | Verfahren und Vorrichtung zum Steuern einer Telekommunikationsendeinrichtung | |
DE102009010449B4 (de) | Verfahren zur Bereitstellung von Informationen und Kommunikationswahlgerät zur Durchführung eines Verfahrens zur Bereitstellung von Informationen | |
DE102019135799A1 (de) | Verfahren zum Verbessern von Sprachverständlichkeit einer elektronischen Sprechverbindung und Headset zur Durchführung des Verfahrens | |
DE102007058585B4 (de) | Vorrichtung und Verfahren zur Durchführung von Fernbesprechungen | |
EP2285626B1 (de) | Kommunikationssystem und verfahren zum kommunizieren zwischen kommunikationseinrichtungen und zumindest einem nutzer eines kommunikationssystems | |
DE102023201075B3 (de) | Verfahren zum Betrieb eines Hörinstruments und Hörsystem mit einem solchen Hörinstrument | |
DE102018200431A1 (de) | Verfahren zur Ausgabe eines Signals einer Drahtloskommunikationsverbindung zwischen einem ersten Anwender in einem Fortbewegungsmittel und einem externen Kommunikationsteilnehmer | |
EP3371026B1 (de) | Verfahren und vorrichtung zur kopplung von elektronischen geräten mit einem kraftfahrzeug | |
DE10135837A1 (de) | Bedienen und Beobachten einer Anlage | |
DE10348149B4 (de) | Verfahren zur Durchführung einer Telefonkonferenz | |
DE102013001771B4 (de) | Verfahren zum Betreiben einer Funktionseinheit sowie Funktionseinheit | |
DE102007058586A1 (de) | Vorrichtung und Verfahren zur Durchführung von Fernbesprechungen | |
DE102006004248A1 (de) | Testverfahren für ein eine Benutzerschnittstelle aufweisendes technisches Gerät | |
EP2648348A1 (de) | Verfahren zur Erfassung des Nutzerverhaltens und der Vorlieben von Radiohörern |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
WWE | Wipo information: entry into national phase |
Ref document number: 201080015334.6 Country of ref document: CN |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2010793164 Country of ref document: EP |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 10793164 Country of ref document: EP Kind code of ref document: A1 |
|
WWE | Wipo information: entry into national phase |
Ref document number: 3825/KOLNP/2011 Country of ref document: IN |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
WWE | Wipo information: entry into national phase |
Ref document number: 13379447 Country of ref document: US |