WO1999054822A1 - Computer network - Google Patents

Computer network Download PDF

Info

Publication number
WO1999054822A1
WO1999054822A1 PCT/GB1999/001149 GB9901149W WO9954822A1 WO 1999054822 A1 WO1999054822 A1 WO 1999054822A1 GB 9901149 W GB9901149 W GB 9901149W WO 9954822 A1 WO9954822 A1 WO 9954822A1
Authority
WO
WIPO (PCT)
Prior art keywords
terminal
user
terminals
attributes
display
Prior art date
Application number
PCT/GB1999/001149
Other languages
French (fr)
Other versions
WO1999054822B1 (en
Inventor
Andrew Mcgrath
Peter Mark Platt
Jeremy Michael Bowskill
Jason Morphett
Alison Jane Willard
Original Assignee
British Telecommunications Public Limited Company
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by British Telecommunications Public Limited Company filed Critical British Telecommunications Public Limited Company
Priority to US09/646,384 priority Critical patent/US6934747B1/en
Priority to EP99915949A priority patent/EP1071995B1/en
Priority to DE69903933T priority patent/DE69903933T2/en
Priority to JP2000545102A priority patent/JP4391691B2/en
Publication of WO1999054822A1 publication Critical patent/WO1999054822A1/en
Publication of WO1999054822B1 publication Critical patent/WO1999054822B1/en

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/50Network services
    • H04L67/535Tracking the activity of the user
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F11/00Error detection; Error correction; Monitoring
    • G06F11/30Monitoring
    • G06F11/34Recording or statistical evaluation of computer activity, e.g. of down time, of input/output operation ; Recording or statistical evaluation of user activity, e.g. usability assessment
    • G06F11/3409Recording or statistical evaluation of computer activity, e.g. of down time, of input/output operation ; Recording or statistical evaluation of user activity, e.g. usability assessment for performance assessment
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q10/00Administration; Management
    • G06Q10/10Office automation; Time management
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L41/00Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks
    • H04L41/22Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks comprising specially adapted graphical user interfaces [GUI]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L9/00Cryptographic mechanisms or cryptographic arrangements for secret or secure communications; Network security protocols
    • H04L9/40Network security protocols
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M3/00Automatic or semi-automatic exchanges
    • H04M3/42Systems providing special services or facilities to subscribers
    • H04M3/56Arrangements for connecting several subscribers to a common circuit, i.e. affording conference facilities
    • H04M3/567Multimedia conference systems
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F11/00Error detection; Error correction; Monitoring
    • G06F11/30Monitoring
    • G06F11/34Recording or statistical evaluation of computer activity, e.g. of down time, of input/output operation ; Recording or statistical evaluation of user activity, e.g. usability assessment
    • G06F11/3466Performance evaluation by tracing or monitoring
    • G06F11/349Performance evaluation by tracing or monitoring for interfaces, buses
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2201/00Indexing scheme relating to error detection, to error correction, and to monitoring
    • G06F2201/815Virtual
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2201/00Indexing scheme relating to error detection, to error correction, and to monitoring
    • G06F2201/86Event-based monitoring
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L69/00Network arrangements, protocols or services independent of the application payload and not provided for in the other groups of this subclass
    • H04L69/30Definitions, standards or architectural aspects of layered protocol stacks
    • H04L69/32Architecture of open systems interconnection [OSI] 7-layer type protocol stacks, e.g. the interfaces between the data link level and the physical level
    • H04L69/322Intralayer communication protocols among peer entities or protocol data unit [PDU] definitions
    • H04L69/329Intralayer communication protocols among peer entities or protocol data unit [PDU] definitions in the application layer [OSI layer 7]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M2203/00Aspects of automatic or semi-automatic exchanges
    • H04M2203/20Aspects of automatic or semi-automatic exchanges related to features of supplementary services
    • H04M2203/2038Call context notifications

Definitions

  • This invention relates to computer networks and their operation.
  • the use of computers is becoming increasingly common in all types of business.
  • private networks so-called Local Area Networks, or LANs
  • a public network such as the "Internet”
  • teleworking Remote working in this way is known as "teleworking".
  • a problem with teleworking is that the creative atmosphere of an office full of people is lost. While working in isolation can be productive for certain activities demanding concentration, the ambience of an office or laboratory is better for generating informal and serendipitous associations.
  • the problem of teleworking is that workers lose awareness of what their colleagues are doing, which can lead to a breakdown in company or group culture.
  • a computer network comprising a plurality of user terminals, each having means for indicating a plurality of attributes relating to the terminal or a user of that terminal, means for monitoring one or more of the terminals, and identifying the attributes of the terminals so monitored, means for displaying on at least one displaying terminal information relating to the attributes of the terminals monitored thereby, wherein the displaying terminal is arranged to display an indication of the relatedness of the attributes of the respective monitored terminals to those currently indicated by the displaying terminal.
  • the attributes of a given terminal may relate to the activities currently performed by the user of that terminal.
  • the display on a first terminal is selected to be indicative of the relatedness of the current attributes of the terminals, to the current attributes of the first terminal.
  • the display may be visual, by means of a screen image, or audible, by means of sounds, or both.
  • a computer terminal may be configured to record as an attribute the identity of a specified user, and track activities of the user when working elsewhere, for example by monitoring the network and identifying whether that user's identity is being used on another terminal, and what activities are currently in use at that terminal.
  • the system operates sing agent-based operators.
  • Each terminal operates one or more agent applications, which identify the attributes of the terminal or its user.
  • Terminals requiring to interact with other terminals have further agents which identify the nature of the interaction required, and two terminals' agents interact to determine an appropriate manner in which the interaction may take place, and to carry it out.
  • Figure 1 illustrates a typical computer network, illustrating one terminal in detail.
  • Figures 2 to 7 illustrate various screen images displayed to a user of the system.
  • Figures 8 and 9 are schematic diagrams illustrating a process for adding an audio input to the system.
  • Figure 1 shows a network 100 serving several computers
  • Links between the individual computers may be permanent hard-wired connections, virtual connections (permanent as perceived by the user, but provided over shared lines by the telecommunications provider), or dial-up connections (available on demand, and provided on a pay-per-use basis), and may include radio links, for example to a mobile terminal 7.
  • One of the computers (9) acts as a "server", co-ordinating the interactions of the other computers.
  • This computer 9 may also have similar functionality to the others, but as shown it is dedicated to the server function only.
  • One of the computers ( 1 ) is shown in more detail. It comprises one or more input devices, such as a keyboard (1 1 ), mouse (1 2), and audio input (1 3), and means ( 14) for monitoring the activity of associated devices such as a telephone (1 5). It also comprises one or more output devices, such as a screen
  • the computer ( 1 ) also comprises a memory (18) and a central processor (19).
  • computer 2 has equivalent devices 21 ,22,23 etc , (not shown in detail), and similarly for computers 3,4,5,6,7,8
  • the computers 1 ,2,3,4,5,6,7,8,9 do not all need to have all the input devices and output devices specified above.
  • Each computer 1 ,2,3,4,5,6,7,8 communicates with the server computer 9 to arrange user access rights, and to define user groups. Any user 1 may instruct the server 9 to re-define his own user group to add or delete users.
  • a computer 1 When a computer 1 is switched on, the user is requested to identify himself and the user identity is passed to the server computer 9.
  • the user 1 accesses a computer application, whether this requires the use of data from the server or only operations requiring the use of the computer's own memory 1 8 and central processor 1 9, the user's operation of this application is communicated to the server 9
  • the computer 1 is also arranged to communicate to the server 9 certain other activities, such as keyboard strokes or the accessing of specified computer files, indicative of the subject the individual user is currently working on.
  • the activities to be communicated to the server 9 by the computer 1 may be defined in the operating system of the computer 1 , or by the server 9.
  • the server computer 9 returns data to each user 1 relating to which other members 2,3,4,5,6,7,8 of the pre-defined user group are currently on-line, and information about the applications they currently have running, and their current subjects of interest This is represented to the user 1 on a display device, the representation being related also to the user's own current status, such that users 2,3,4,5,6,7,8 working on similar subjects to the user 1 can be readily identified.
  • the representation may be generated either by the server 9 or by the user's own computer 1 . Example representations will be described later.
  • a terminal 8 may be arranged to indicate the activity of a specified user. In general, it is envisaged that this will be the terminal at the user's own permanent workstation.
  • the server recognises the user's identity and transmits a signal to the user's own terminal 8, causing a display to be generated indicating the location and activities of that user This allows colleagues at the user's home location, where the terminal 8 is located, to monitor his activity and choose whether it is appropriate to contact him.
  • the attributes stored may relate to users' interests (which may be made to vary with time of day to reflect work time and leisure time), geographical location (which may change if the terminal is portable), and the user's position in a company or other organisation They may also relate to activities directly related to the user's operation of the computer 1 , such as what subject the user is working on (identified by the computer files the user is working on, the content of the document, "Internet" page currently in use, keywords identified in the last few characters typed, and the applications the user has running), other activities which may take place at a work station, such as which input devices 1 1 , 1 2, 1 3 are in use, or whether a screensaver is being displayed (a lack of activity suggesting that the user may be away from the computer), whether an associated telephone 1 5 is in use, or whether an audio input 1 3 is detecting voices, suggesting that the user is in conference
  • the display on the screen 1 6 of the computer terminal 1 may take the form of a "virtual environment", (illustrated in Figures 2 to 7) in which users (including the user of the computer 1 himself) using associated applications are represented as being closer together than users working on disparate subjects.
  • the viewpoint in the virtual environment will depend on the position of the user within it, which in turn depends on the application the user is working on.
  • the server 9 interprets these activities to control the user's position and viewpoint.
  • Sound effects may either supplement or replace the visual display to give an ambient sound effect, which varies according to the context the user 1 is working in
  • These may be artificially generated sounds, pre-recorded, or live sound transmitted from the other users' terminals 2,3,4,5,6,7,8.
  • These sounds may be varied, for example in amplitude or pitch, according to the proximity in the virtual space of the users represented by them.
  • the use of sound has particular benefit in virtual environments in which part of the environment is out of the field of view of the user.
  • Spatialised sound may be used to facilitate the location, by the user, of the sound sources in the virtual environment Suitable spatialised systems are described for example in United States Patents 55331 29 (Gefvert), 530741 5 (Fosgate), article "Spatial Sound for Telepresence” by M.Hollier, D. Burraston, and A. Rimell in the British Telecom Technology Journal, October 1 997, and International Patent Specification W098/58523
  • the server 9 loads the virtual environment application into the user's terminal 1 , for display on the screen 1 6, and updates it in response to the activities of the individual users 1 ,2,3,4,5,6,7,8.
  • the virtual environment is represented by a number of levels 400, 401 , 402, 403 ( Figure 2) .
  • Each user 1 is represented on the corresponding screens 26, 36, 46 etc of the other users 2, 3, 4 etc.
  • the user 1 can navigate freely about the world, including moving between levels, and correspond with other users, using electronic mail, videoconference, or other available means
  • the representation of the user's position, both on his own display device and those of the other users, is driven by the server 9 interpreting the activities of the user, for example operation of the mouse 1 2, keyboard 1 1 , telephone 1 5, etc.
  • the user is generally presented with a viewpoint of this virtual environment as seen from a ground level 400 ( Figure 3), representative of his own designated working area.
  • Other users 2, 3, 4, 5, 6, 7, 8 in his designated user group are represented by "avatars" positioned in the virtual space according to their current activities
  • the user 1 is initially represented in an open space, and arranged to be looking towards certain other users 2, 3, 4 in the user group selected to be those with whom the user shares some common interest, those currently working in similar subjects to the user 1 being arranged closest to the viewpoint of the user 1 , whilst those who, although having the same interest, are currently working in some other activity, are more remote.
  • a user carries out an activity such as document downloading, searching the Internet, or having a meeting, his avatar is moved to one of the other levels 401 , 402, 403, each representing one such activity.
  • Figure 2 Users from other groups engaged in similar activities will be found there. This allows interactions to take place between members of different interest groups when they are engaged on similar activities. This emulates the behaviour of personnel in a real office environment, where workers from different working groups may discover matters of common interest during chance encounters in common areas of the office such as in the corridor or by the photocopier, but would not wish to disturb each other when concentrating on work specific to their group. These matters of common interest may relate to the activity represented by the level the users meet on, but may be unconnected to it.
  • Sound may be used to support the visual information represented on the screen 1 6, and create a more lifelike ambience.
  • hearing is the primary awareness cue. Consequently, in an open-plan office, occupants are usually aware of each other's presence even when they are not in view.
  • the system is therefore arranged to provide auditory awareness of specific people - for example the members of a specified group, and/or those nearby in the virtual environment.
  • the users can navigate freely about the virtual environment. If the user 1 wishes to discuss matters relating to another interest group, he can navigate to the relevant interest group 301 on the ground plane 400. The user can then arrange to converse with another user 5 in that interest group, for example using electronic mail, by identifying the relevant avatar and transmitting the required text.
  • the user 1 may wish to enter a different mode of communication, e.g. an audio link, with the other user 5.
  • a different mode of communication e.g. an audio link
  • the user 1 selects the two avatars 1 ,5 (for example by using a "mouse” 1 2 to "click” on the respective avatars, and entering a command to cause the server 9 to set up a telephone connection between the user's respective telephones 1 5, 55, or the audio input/output devices 1 3, 1 7, 53, 57.
  • the other people 6,7 nearby in the virtual environment receive a visual indication that the two users 1 ,5 are in audio communication, by the representation of an "audio zone" 101 surrounding their avatars 1 ,5.
  • the users 1 ,5 may select the discussion to be private, or alternatively may allow the server 9 to arrange for part of the conversation may be fed (at low bit rate and amplitude) into the audio outputs 27, 37, 47 etc of the other users, so that another user e.g. user 6, can "overhear", the conversation and join in if he wishes to contribute. To do so, the user 6 moves into the 'audio zone' 101 , for example by moving his avatar under control of his mouse 62. The server 9 responds to this movement by establishing a three-way conference call, connecting user 6, by means of his telephone 65 or other audio devices 63, 67, to enter the conference at full bandwidth.
  • Any of the users 1 , 5, 6 can, if desired, cause the discussion to continue in a more formal "virtual meeting space" 102, for example of the type described in International Patent Specification PCT/GB98/03555 filed on November 27th 1 998, and illustrated in Figure 6.
  • This system has facilities such as a shared whiteboard, document distribution, document sharing, etc.
  • Any of the users 1 , 5, or 6 can create the "meeting space" 102 by entering an appropriate command using an input device (e.g.
  • the respective "mouse" 1 2, 52, 62) to instruct the server 9 (or one of the individual user equipments 1 ,5,6) to activate the meeting space application, and to instruct the individual user equipments 1 ,5,6 to interact with the meeting space application to display to each user, using their respective display devices 1 6, 1 7, 56, 57, 66, 67, the activities of the users 1 ,5,6 within the meeting space 1 02, as indicated by the individual users' use of their respective input devices ( 1 1 , 1 2, 1 3, 51 , 52, 53, 61 , 62, 63) .
  • the meeting space application 102 As the meeting space application 102 is activated, various conferencing tools appear such as a table 1 31 , whiteboard 1 32 and so on and the avatars 1 ,5,6 start to 'act out' the interactions between each other. As shown in Figure 6, the meeting space 102 is displayed to each user 1 , 5, 6 from the viewpoint of that respective user, so the user to whom it is displayed (user 1 in this example) does not appear. Other people 2,3,4 in the social space, not involved in the meeting, are given an indication that the users 1 ,5,6 are in a private meeting and cannot overhear any of the discussion (Figure 7). These other users 2,3,4 are not represented in the meeting space 102, although the overall virtual environment 100 may be represented in a "window".
  • Any user 5 in the meeting space 102 can show a document to the other users 1 ,6, in the space by entering the details of the document, and a representation 133, 1 33a of the document then appears on the table in front of each avatar 1 ,5,6
  • a user can select the document (by "clicking" on its representation 1 33 and the document appears in a two dimensional "window” in the respective display 1 6, 56, 66 for him to read.
  • a user 6 may take notes by "clicking" on a representation of a whiteboard 1 32, causing the respective avatar 6 to move to the whiteboard in the viewpoint of the other users 1 ,5, whilst the user 6 is presented with a full-screen image of the whiteboard on which text, etc, can be entered
  • Users may leave the meeting space 102, or request admission, at any time - their departure from and entry to the meeting space 102 being represented by appropriate movement of the avatars in the representations of the overall virtual environment 1 00 or meeting space 1 02 displayed to each user 1 ,2,3,4,5,6,7,8, in the same way as the audio zone 101 can be entered and left, except that access to the meeting zone 102 may be refused by the current occupants of the meeting zone.
  • the meeting space 102 is replaced by an informal audio zone 1 01 similar to that of Figure 5, surrounding any participants 1 ,5 present in the meeting zone 102 at the time of closure
  • a first user 1 wishes to contact another specific user 2, but does not know when the other user 2 is available, the first user 1 selects the avatar of the other user 2, and the server 9 generates a sound output representative of the current activity of the other user 2.
  • This sound output changes when the other user 2 becomes available. For example when the other user is available a 'background noise' is generated, representative of the other user 2.
  • This background noise could be a sample of speech, or a low bit-rate audio feed could be used.
  • the server 9 infers the availability of the user 2 based on the applications running on the respective computer, the frequency of key presses on the keyboard 21 , whether the telephone 25 is in use, the level 400, 401 , 402 etc the user 2 is currently assigned to etc. The degree of availability may be used to vary audio amplitude or other characteristics, to indicate how likely the user 2 is to welcome an interruption. If the user 2 is not currently logged on, the sound level is zero.
  • the system may also be used as an "intelligent" screen saver.
  • a computer terminal 3 is allocated to a respective user. When the user is working elsewhere the system is configured such that it reflects his availability. So for example, if the user is currently working at another terminal 4, for example at home, but is willing to be contacted, he can select an 'at work' option, causing the terminal 4 to log on to the server 9 in the conventional way.
  • the server 9 transmits an instruction to the terminal 3 to display a graphic or avatar of the user at a computer.
  • the server 9, by monitoring the operation of the terminal 4, can cause the graphic on the display 37 to indicate the applications being run.
  • Spatialised audio of the type which uses a plurality of audio outputs 17, in connection with a movable seat and screen, whose position is monitored by the computer equipment 1 , (as described for example in United States Patent 551 5078 (Greschler) may be used to track the user 2 in the virtual environment, and to make audio connections based on which sample or low bit rate audio awareness channel the user 1 is currently facing. For example, the user 1 , hearing the voice of user 2, can turn in his seat towards the sound, and hit a button on the keyboard 1 1 . The current position of the seat is used by the computer 1 to identify the required user 2, causing the telephone 1 5 to ring that user's telephone 25.
  • the audible output may be conventional audible icons - individual "beeps" to indicate events such as electronic mail arriving, or continuous sound representative of individual users, such as the background sound representative of each user currently logged on. Auditory icons are ideal for alerting users to particular events, but they are less good at conveying continuous states. While it is useful to have a sound to indicate that some new email has arrived in a virtual office environment, it may become distracting if that sound was played continuously until the email is read. If sounds were used to convey other events and states in the office environment such as forthcoming meetings, changes in important documents, the availability of colleagues, etc., then it may become difficult to extract any information at all from the mixture of noises.
  • the presence of mail in a mailbox, the approach of a scheduled appointment, and the whereabouts of nearby users in the virtual environment who are outside the user's field of vision, or too distant to identify readily given the picture resolution of a typical computer monitor, are all characteristics which may require low-level continuous indications, rather than a single sound to indicate an event.
  • the sound may be modified according to the significance of the event - for example the number or urgency of messages, the proximity of the appointment, etc, may be indicated by the amplitude, pitch, or repetition rate of a sound.
  • music which can be considered as a system for organising sounds so that the combination of sounds is capable of conveying meaning, and its use in a virtual environment is comparable to the way music is used to support the action in theatre, opera and particularly cinema films.
  • Music as a display device is ideally suited to a virtual environment as, unlike a purely graphical system, it can subtly convey non-specific information without requiring the listener's full attention. This means that as long as the main focus of the work is visually based (e.g. word processing, programming etc.) the music is able to exist alongside the mam work activity without causing interruption.
  • music is used that does not rely on segments being prepared in advance, but can generate music in real time based on the states and events taking place
  • “Generative music” is a term used to describe music that is composed by a computer in real time based on a set of static or changing rules and parameters. The music is defined (written) by setting parameters and rules that affect the system's random choices of note pitch, length, rhythm, instrument etc.
  • One particular preferred implementation uses computer-generated music, such as "Koan", discussed on the SSEYO website at http://www.sseyo.com.
  • This application runs on a PC and uses a midi soundcard to produce the music
  • Koan has built into it many of the rules and organising structures that allow recognisable musical styles to be generated easily, including common western scales as well as Indian and Asian forms. This means that pieces of music can be created that will always conform to these styles but which can have any variety of component parts
  • the Koan system allows the composer to select from over 200 different musical parameters ranging from note length/pitch through to advanced features of midi sound cards, which can be altered to produce a huge variety of different sounds and musical styles.
  • the system to be described here replaces the human composer by input from the virtual environment 100, which selects the parameters to be input to the system.
  • a generative music system allows musical parts to be defined for each level of activity for each person.
  • the parts are not explicitly created, but parameters are set so that the system produces music for each person that will sound similar each time it is played.
  • the advantage of the generative method is that it is possible to associate the different levels of activity directly with the parameter that controls the pitch of the notes. This means that the number of levels of activity does not need to be known in advance, and separate pieces don't need to be created for each possible state.
  • the generative system is able to cope with any number of different states and translate this into a continuous variation in the pitch (or other parameter) of the parts.
  • the 901 is used to set up communication between the virtual environment 902 and the music application 903 Events in the virtual environment 902 are passed to the applet 901 , which then sends the appropriate messages to alter the behaviour of the music application 903. Likewise, changes occurring in the music application 903 may be passed back to the virtual environment 902 using the applet 901 , for example to control the direction of a sound source
  • An event occurring in the virtual environment system causes a program to be run (either in the server 9 or in an individual terminal 1 ) that consults a look-up table 910 to determine which musical segment 91 1 , 91 2, 91 N should be used to represent this event.
  • the selected segment 91 N is then added to any other segments (92N, 93N, etc) already selected to be part of the music to be played by the music application 903.
  • the music application 903 combines the selected segments to generate the musical output.
  • the interactions between the segments may be quite complex - the addition of one segment may modify the way in which others are presented.
  • the selected segments 91 , 92N, 93N continue to play, continuously or periodically, until the system determines that another event has occurred requiring a segment to be added or deleted. Certain events may require a segment to exhibit a temporal characteristic, for example getting louder or softer or changing in pitch, according to how recently the associated event occurred.

Landscapes

  • Engineering & Computer Science (AREA)
  • Signal Processing (AREA)
  • Theoretical Computer Science (AREA)
  • Business, Economics & Management (AREA)
  • General Engineering & Computer Science (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Entrepreneurship & Innovation (AREA)
  • Strategic Management (AREA)
  • General Physics & Mathematics (AREA)
  • Quality & Reliability (AREA)
  • Computer Hardware Design (AREA)
  • Multimedia (AREA)
  • Physics & Mathematics (AREA)
  • Human Resources & Organizations (AREA)
  • Economics (AREA)
  • Computer Security & Cryptography (AREA)
  • Marketing (AREA)
  • Operations Research (AREA)
  • Tourism & Hospitality (AREA)
  • General Business, Economics & Management (AREA)
  • Data Mining & Analysis (AREA)
  • Human Computer Interaction (AREA)
  • Information Transfer Between Computers (AREA)
  • User Interface Of Digital Computer (AREA)
  • Processing Or Creating Images (AREA)
  • Telephonic Communication Services (AREA)

Abstract

Each terminal (e.g. terminal 8) of a computer network 100 can monitor the activities of the other terminals (1, 2, 3, 4, 5, 6, 7); or associated equipment such as a telephone (15); and display these activities in visual and/or audible form. The display may be represented as a 'virtual world', and/or a generative music system such as 'Koan'. The nature of the activities performed by the terminal (8) itself also control the display, for example giving greatest prominence to those terminals engaged in similar or related acivities. A terminal (3) may be configured to act as a specified user's 'base' terminal, to display the activity of the specified user, by identifying whether the specified user's identity is being used on another terminal (e.g. terminal 4). The activity of this terminal can then be displayed on the 'base' terminal (3). The base terminal (3) can therefore indicate the absent user's availability, level of activity etc. to colleagues working nearby.

Description

COMPUTER NETWORK
This invention relates to computer networks and their operation. The use of computers is becoming increasingly common in all types of business. By linking computers together through private networks (so-called Local Area Networks, or LANs) or a public network such as the "Internet" it is possible for many people to co-operate on a project without being in each other's physical presence. The individuals may be anywhere in the world, and still work together. Remote working in this way is known as "teleworking". A problem with teleworking is that the creative atmosphere of an office full of people is lost. While working in isolation can be productive for certain activities demanding concentration, the ambience of an office or laboratory is better for generating informal and serendipitous associations. The problem of teleworking is that workers lose awareness of what their colleagues are doing, which can lead to a breakdown in company or group culture.
According to one aspect of the invention, there is provided a computer network comprising a plurality of user terminals, each having means for indicating a plurality of attributes relating to the terminal or a user of that terminal, means for monitoring one or more of the terminals, and identifying the attributes of the terminals so monitored, means for displaying on at least one displaying terminal information relating to the attributes of the terminals monitored thereby, wherein the displaying terminal is arranged to display an indication of the relatedness of the attributes of the respective monitored terminals to those currently indicated by the displaying terminal. The attributes of a given terminal may relate to the activities currently performed by the user of that terminal. This invention allows distributed groups of people to maintain an awareness of each other. In one implementation the display on a first terminal is selected to be indicative of the relatedness of the current attributes of the terminals, to the current attributes of the first terminal. The display may be visual, by means of a screen image, or audible, by means of sounds, or both.
In another implementation, a computer terminal may be configured to record as an attribute the identity of a specified user, and track activities of the user when working elsewhere, for example by monitoring the network and identifying whether that user's identity is being used on another terminal, and what activities are currently in use at that terminal.
In a preferred embodiment, the system operates sing agent-based operators. Each terminal operates one or more agent applications, which identify the attributes of the terminal or its user. Terminals requiring to interact with other terminals have further agents which identify the nature of the interaction required, and two terminals' agents interact to determine an appropriate manner in which the interaction may take place, and to carry it out.
Various embodiments are envisaged within the scope of the invention. Several of these will be discussed below, with reference to the accompanying drawings, in which:
Figure 1 illustrates a typical computer network, illustrating one terminal in detail.
Figures 2 to 7 illustrate various screen images displayed to a user of the system.
Figures 8 and 9 are schematic diagrams illustrating a process for adding an audio input to the system.
Figure 1 shows a network 100 serving several computers
1 ,2,3,4,5,6,7,8,9. (The same reference numerals will also be used to represent the users of these terminals, and avatars representing them in the "virtual environment" representation displayed by these computers). Links between the individual computers may be permanent hard-wired connections, virtual connections (permanent as perceived by the user, but provided over shared lines by the telecommunications provider), or dial-up connections (available on demand, and provided on a pay-per-use basis), and may include radio links, for example to a mobile terminal 7. One of the computers (9) acts as a "server", co-ordinating the interactions of the other computers. This computer 9 may also have similar functionality to the others, but as shown it is dedicated to the server function only.
One of the computers ( 1 ) is shown in more detail. It comprises one or more input devices, such as a keyboard (1 1 ), mouse (1 2), and audio input (1 3), and means ( 14) for monitoring the activity of associated devices such as a telephone (1 5). It also comprises one or more output devices, such as a screen
( 1 6) and audio output ( 1 7) . As is conventional, the computer ( 1 ) also comprises a memory (18) and a central processor (19). As will be understood, computer 2 has equivalent devices 21 ,22,23 etc , (not shown in detail), and similarly for computers 3,4,5,6,7,8 The computers 1 ,2,3,4,5,6,7,8,9 do not all need to have all the input devices and output devices specified above.
Each computer 1 ,2,3,4,5,6,7,8 communicates with the server computer 9 to arrange user access rights, and to define user groups. Any user 1 may instruct the server 9 to re-define his own user group to add or delete users.
When a computer 1 is switched on, the user is requested to identify himself and the user identity is passed to the server computer 9. Whenever the user 1 accesses a computer application, whether this requires the use of data from the server or only operations requiring the use of the computer's own memory 1 8 and central processor 1 9, the user's operation of this application is communicated to the server 9 The computer 1 is also arranged to communicate to the server 9 certain other activities, such as keyboard strokes or the accessing of specified computer files, indicative of the subject the individual user is currently working on. The activities to be communicated to the server 9 by the computer 1 may be defined in the operating system of the computer 1 , or by the server 9.
The server computer 9 returns data to each user 1 relating to which other members 2,3,4,5,6,7,8 of the pre-defined user group are currently on-line, and information about the applications they currently have running, and their current subjects of interest This is represented to the user 1 on a display device, the representation being related also to the user's own current status, such that users 2,3,4,5,6,7,8 working on similar subjects to the user 1 can be readily identified. The representation may be generated either by the server 9 or by the user's own computer 1 . Example representations will be described later. In one particular embodiment a terminal 8 may be arranged to indicate the activity of a specified user. In general, it is envisaged that this will be the terminal at the user's own permanent workstation. When that user "signs in" at another terminal 7, the server recognises the user's identity and transmits a signal to the user's own terminal 8, causing a display to be generated indicating the location and activities of that user This allows colleagues at the user's home location, where the terminal 8 is located, to monitor his activity and choose whether it is appropriate to contact him.
The attributes stored may relate to users' interests (which may be made to vary with time of day to reflect work time and leisure time), geographical location (which may change if the terminal is portable), and the user's position in a company or other organisation They may also relate to activities directly related to the user's operation of the computer 1 , such as what subject the user is working on (identified by the computer files the user is working on, the content of the document, "Internet" page currently in use, keywords identified in the last few characters typed, and the applications the user has running), other activities which may take place at a work station, such as which input devices 1 1 , 1 2, 1 3 are in use, or whether a screensaver is being displayed (a lack of activity suggesting that the user may be away from the computer), whether an associated telephone 1 5 is in use, or whether an audio input 1 3 is detecting voices, suggesting that the user is in conference
The display on the screen 1 6 of the computer terminal 1 may take the form of a "virtual environment", (illustrated in Figures 2 to 7) in which users (including the user of the computer 1 himself) using associated applications are represented as being closer together than users working on disparate subjects. The viewpoint in the virtual environment will depend on the position of the user within it, which in turn depends on the application the user is working on. As the user operates the various input devices 1 1 , 1 2, 1 5 of his terminal 1 , the server 9 interprets these activities to control the user's position and viewpoint. Sound effects may either supplement or replace the visual display to give an ambient sound effect, which varies according to the context the user 1 is working in These may be artificially generated sounds, pre-recorded, or live sound transmitted from the other users' terminals 2,3,4,5,6,7,8. These sounds may be varied, for example in amplitude or pitch, according to the proximity in the virtual space of the users represented by them. The use of sound has particular benefit in virtual environments in which part of the environment is out of the field of view of the user. Spatialised sound may be used to facilitate the location, by the user, of the sound sources in the virtual environment Suitable spatialised systems are described for example in United States Patents 55331 29 (Gefvert), 530741 5 (Fosgate), article "Spatial Sound for Telepresence" by M.Hollier, D. Burraston, and A. Rimell in the British Telecom Technology Journal, October 1 997, and International Patent Specification W098/58523
Several modes of operation of the system will now be described with reference to the drawings. In the first mode, represented in Figures 2 to 7, the user moves through a "virtual environment" which has a number of predetermined activities, and various levels at which the user can be operating on those activities. Users working on the same subject will appear near to each other in the world.
The server 9 loads the virtual environment application into the user's terminal 1 , for display on the screen 1 6, and updates it in response to the activities of the individual users 1 ,2,3,4,5,6,7,8. The virtual environment is represented by a number of levels 400, 401 , 402, 403 (Figure 2) . Each user 1 is represented on the corresponding screens 26, 36, 46 etc of the other users 2, 3, 4 etc. The user 1 can navigate freely about the world, including moving between levels, and correspond with other users, using electronic mail, videoconference, or other available means
The representation of the user's position, both on his own display device and those of the other users, is driven by the server 9 interpreting the activities of the user, for example operation of the mouse 1 2, keyboard 1 1 , telephone 1 5, etc. The user is generally presented with a viewpoint of this virtual environment as seen from a ground level 400 (Figure 3), representative of his own designated working area. Other users 2, 3, 4, 5, 6, 7, 8 in his designated user group are represented by "avatars" positioned in the virtual space according to their current activities The user 1 is initially represented in an open space, and arranged to be looking towards certain other users 2, 3, 4 in the user group selected to be those with whom the user shares some common interest, those currently working in similar subjects to the user 1 being arranged closest to the viewpoint of the user 1 , whilst those who, although having the same interest, are currently working in some other activity, are more remote. As shown in Figure 4, as the user 1 works on different subjects (identified by the server 9 according to the computer files and applications in use, key strokes, etc), his respective avatar is moved to a part of the virtual environment representing an "interest area" 201 relating to that subject, and changed in appearance (for example colour) to reflect the fact that the user is busy. Avatars of other users 7,8 who share an interest in this subject will be located in the same interest area of the virtual environment, at an appropriate level to their current activities.
If a user carries out an activity such as document downloading, searching the Internet, or having a meeting, his avatar is moved to one of the other levels 401 , 402, 403, each representing one such activity. (Figure 2) Users from other groups engaged in similar activities will be found there. This allows interactions to take place between members of different interest groups when they are engaged on similar activities. This emulates the behaviour of personnel in a real office environment, where workers from different working groups may discover matters of common interest during chance encounters in common areas of the office such as in the corridor or by the photocopier, but would not wish to disturb each other when concentrating on work specific to their group. These matters of common interest may relate to the activity represented by the level the users meet on, but may be unconnected to it.
When a user first "logs on" to his terminal 1 to start a session, the user first appears on an "entry" level 403, to allow such encounters to take place before the users get involved in more specialised activities at the "ground level" 400, and to gain an "overview" of who is already present in the virtual environment. (Figure 2) . After a short interval, if the user does not initiate an interaction at this entry level, the server 9 automatically moves the user 1 to the ground level 400, as shown in Figure 3.
Sound may be used to support the visual information represented on the screen 1 6, and create a more lifelike ambience. In a real office environment hearing is the primary awareness cue. Consequently, in an open-plan office, occupants are usually aware of each other's presence even when they are not in view. The system is therefore arranged to provide auditory awareness of specific people - for example the members of a specified group, and/or those nearby in the virtual environment. The users can navigate freely about the virtual environment. If the user 1 wishes to discuss matters relating to another interest group, he can navigate to the relevant interest group 301 on the ground plane 400. The user can then arrange to converse with another user 5 in that interest group, for example using electronic mail, by identifying the relevant avatar and transmitting the required text. The user 1 may wish to enter a different mode of communication, e.g. an audio link, with the other user 5. As shown in Figure 5, the user 1 selects the two avatars 1 ,5 (for example by using a "mouse" 1 2 to "click" on the respective avatars, and entering a command to cause the server 9 to set up a telephone connection between the user's respective telephones 1 5, 55, or the audio input/output devices 1 3, 1 7, 53, 57. The other people 6,7 nearby in the virtual environment receive a visual indication that the two users 1 ,5 are in audio communication, by the representation of an "audio zone" 101 surrounding their avatars 1 ,5. The users 1 ,5 may select the discussion to be private, or alternatively may allow the server 9 to arrange for part of the conversation may be fed (at low bit rate and amplitude) into the audio outputs 27, 37, 47 etc of the other users, so that another user e.g. user 6, can "overhear", the conversation and join in if he wishes to contribute. To do so, the user 6 moves into the 'audio zone' 101 , for example by moving his avatar under control of his mouse 62. The server 9 responds to this movement by establishing a three-way conference call, connecting user 6, by means of his telephone 65 or other audio devices 63, 67, to enter the conference at full bandwidth.
Any of the users 1 , 5, 6 can, if desired, cause the discussion to continue in a more formal "virtual meeting space" 102, for example of the type described in International Patent Specification PCT/GB98/03555 filed on November 27th 1 998, and illustrated in Figure 6. This system has facilities such as a shared whiteboard, document distribution, document sharing, etc. Any of the users 1 , 5, or 6 can create the "meeting space" 102 by entering an appropriate command using an input device (e.g. the respective "mouse" 1 2, 52, 62) to instruct the server 9 (or one of the individual user equipments 1 ,5,6) to activate the meeting space application, and to instruct the individual user equipments 1 ,5,6 to interact with the meeting space application to display to each user, using their respective display devices 1 6, 1 7, 56, 57, 66, 67, the activities of the users 1 ,5,6 within the meeting space 1 02, as indicated by the individual users' use of their respective input devices ( 1 1 , 1 2, 1 3, 51 , 52, 53, 61 , 62, 63) .
As the meeting space application 102 is activated, various conferencing tools appear such as a table 1 31 , whiteboard 1 32 and so on and the avatars 1 ,5,6 start to 'act out' the interactions between each other. As shown in Figure 6, the meeting space 102 is displayed to each user 1 , 5, 6 from the viewpoint of that respective user, so the user to whom it is displayed (user 1 in this example) does not appear. Other people 2,3,4 in the social space, not involved in the meeting, are given an indication that the users 1 ,5,6 are in a private meeting and cannot overhear any of the discussion (Figure 7). These other users 2,3,4 are not represented in the meeting space 102, although the overall virtual environment 100 may be represented in a "window".
Any user 5 in the meeting space 102 can show a document to the other users 1 ,6, in the space by entering the details of the document, and a representation 133, 1 33a of the document then appears on the table in front of each avatar 1 ,5,6 A user can select the document (by "clicking" on its representation 1 33 and the document appears in a two dimensional "window" in the respective display 1 6, 56, 66 for him to read.
A user 6 may take notes by "clicking" on a representation of a whiteboard 1 32, causing the respective avatar 6 to move to the whiteboard in the viewpoint of the other users 1 ,5, whilst the user 6 is presented with a full-screen image of the whiteboard on which text, etc, can be entered
Users may leave the meeting space 102, or request admission, at any time - their departure from and entry to the meeting space 102 being represented by appropriate movement of the avatars in the representations of the overall virtual environment 1 00 or meeting space 1 02 displayed to each user 1 ,2,3,4,5,6,7,8, in the same way as the audio zone 101 can be entered and left, except that access to the meeting zone 102 may be refused by the current occupants of the meeting zone. If the meeting is closed, the meeting space 102 is replaced by an informal audio zone 1 01 similar to that of Figure 5, surrounding any participants 1 ,5 present in the meeting zone 102 at the time of closure
If a first user 1 wishes to contact another specific user 2, but does not know when the other user 2 is available, the first user 1 selects the avatar of the other user 2, and the server 9 generates a sound output representative of the current activity of the other user 2. This sound output changes when the other user 2 becomes available. For example when the other user is available a 'background noise' is generated, representative of the other user 2. This background noise could be a sample of speech, or a low bit-rate audio feed could be used. The server 9 infers the availability of the user 2 based on the applications running on the respective computer, the frequency of key presses on the keyboard 21 , whether the telephone 25 is in use, the level 400, 401 , 402 etc the user 2 is currently assigned to etc. The degree of availability may be used to vary audio amplitude or other characteristics, to indicate how likely the user 2 is to welcome an interruption. If the user 2 is not currently logged on, the sound level is zero.
The system may also be used as an "intelligent" screen saver. A computer terminal 3 is allocated to a respective user. When the user is working elsewhere the system is configured such that it reflects his availability. So for example, if the user is currently working at another terminal 4, for example at home, but is willing to be contacted, he can select an 'at work' option, causing the terminal 4 to log on to the server 9 in the conventional way. The server 9 transmits an instruction to the terminal 3 to display a graphic or avatar of the user at a computer. The server 9, by monitoring the operation of the terminal 4, can cause the graphic on the display 37 to indicate the applications being run. Thus people in the office would be able to see from the display 37 on his office terminal 3 that the user is at home and available to be contacted, should he choose to make this known. Similarly, when the user picks up the telephone 45 (as sensed by the telephone 45 itself or at the exchange) the avatar on the display 37 changes so that people know that the user is unavailable to be telephoned, but perhaps still available to be emailed. If both computers 3, 4 have video conferencing capabilities, the system may be arranged such that a user in the office can establish a videoconference with the user 3 by entering a simple command. Another use for the system of the invention is to determine the form in which information should be sent to a user. The attributes of the user identify the type of terminal in use, and the activities currently engaged in. These can then be used to determine in what form (text, audio, video etc) to transmit the information, and whether to interrupt the user's current activity with the new information, merely alert the user to its arrival, or deliver it without announcement. This decision would be based on the complexity of the task the user is currently engaged on (and hence the degree of inconvenience caused by interrupting it), the relevance to the current task of the information being delivered, and its degree of urgency. The information may be abridged according to the "ProSumm" protocol described in the present applicant's International Patent Specification WO98/47083.
Spatialised audio, of the type which uses a plurality of audio outputs 17, in connection with a movable seat and screen, whose position is monitored by the computer equipment 1 , (as described for example in United States Patent 551 5078 (Greschler) may be used to track the user 2 in the virtual environment, and to make audio connections based on which sample or low bit rate audio awareness channel the user 1 is currently facing. For example, the user 1 , hearing the voice of user 2, can turn in his seat towards the sound, and hit a button on the keyboard 1 1 . The current position of the seat is used by the computer 1 to identify the required user 2, causing the telephone 1 5 to ring that user's telephone 25.
The audible output may be conventional audible icons - individual "beeps" to indicate events such as electronic mail arriving, or continuous sound representative of individual users, such as the background sound representative of each user currently logged on. Auditory icons are ideal for alerting users to particular events, but they are less good at conveying continuous states. While it is useful to have a sound to indicate that some new email has arrived in a virtual office environment, it may become distracting if that sound was played continuously until the email is read. If sounds were used to convey other events and states in the office environment such as forthcoming meetings, changes in important documents, the availability of colleagues, etc., then it may become difficult to extract any information at all from the mixture of noises.
For example the presence of mail in a mailbox, the approach of a scheduled appointment, and the whereabouts of nearby users in the virtual environment who are outside the user's field of vision, or too distant to identify readily given the picture resolution of a typical computer monitor, are all characteristics which may require low-level continuous indications, rather than a single sound to indicate an event. The sound may be modified according to the significance of the event - for example the number or urgency of messages, the proximity of the appointment, etc, may be indicated by the amplitude, pitch, or repetition rate of a sound.
This requires a structured approach to delivering auditory information that allows the sound to be organised. A particular example of a continuous indication is music, which can be considered as a system for organising sounds so that the combination of sounds is capable of conveying meaning, and its use in a virtual environment is comparable to the way music is used to support the action in theatre, opera and particularly cinema films. Music as a display device is ideally suited to a virtual environment as, unlike a purely graphical system, it can subtly convey non-specific information without requiring the listener's full attention. This means that as long as the main focus of the work is visually based (e.g. word processing, programming etc.) the music is able to exist alongside the mam work activity without causing interruption.
The use of music in theatre and opera is many centuries old, and is also well-established in films. There can be little doubt that music can change the sense of environment. For example, an apparently innocuous beach scene in the film "Jaws" ( 1 975) is completely altered by the accompanying music, which alone hints at the sinister approach of the unseen shark The information in the virtual environment can also be represented using music; each different user of the environment having a particular theme associated with him/her, the rhythm of which is altered to indicate if that person is busy or not. The music provides a continuous indication of the user's level of activity. Clearly it would be distracting were users required to repeatedly check the environment visually to assess the activity of the other users, so the music is provided to provide an ambient awareness of what's going on, to enhance the environment rather than to replace it, allowing users to gain visual confirmation of events detected through the music.
However, although there are many similarities between virtual environments and films, there is clearly one major difference. All of the action that takes place in a film is scripted and pre-planned whereas in an interactive shared virtual environment much of the content and activity cannot be foreseen Thus the score accompanying a dynamic virtual environment must itself be dynamic and change according to the action taking place It is possible to provide each event, user, etc, with a musical segment or "leitmotif which is played in given circumstances However, this is difficult to arrange and likely to lead to a confused sound, difficult to understand, especially when a lot of activity is taking place
In a preferred arrangement according to the invention, music is used that does not rely on segments being prepared in advance, but can generate music in real time based on the states and events taking place "Generative music" is a term used to describe music that is composed by a computer in real time based on a set of static or changing rules and parameters. The music is defined (written) by setting parameters and rules that affect the system's random choices of note pitch, length, rhythm, instrument etc. One particular preferred implementation uses computer-generated music, such as "Koan", discussed on the SSEYO website at http://www.sseyo.com. This application runs on a PC and uses a midi soundcard to produce the music Koan has built into it many of the rules and organising structures that allow recognisable musical styles to be generated easily, including common western scales as well as Indian and Asian forms. This means that pieces of music can be created that will always conform to these styles but which can have any variety of component parts The Koan system allows the composer to select from over 200 different musical parameters ranging from note length/pitch through to advanced features of midi sound cards, which can be altered to produce a huge variety of different sounds and musical styles. The system to be described here replaces the human composer by input from the virtual environment 100, which selects the parameters to be input to the system.
In a shared environment where music is used to identify people and their level of activity, the use of a generative music system allows musical parts to be defined for each level of activity for each person. In this method however the parts are not explicitly created, but parameters are set so that the system produces music for each person that will sound similar each time it is played. The advantage of the generative method is that it is possible to associate the different levels of activity directly with the parameter that controls the pitch of the notes. This means that the number of levels of activity does not need to be known in advance, and separate pieces don't need to be created for each possible state. The generative system is able to cope with any number of different states and translate this into a continuous variation in the pitch (or other parameter) of the parts.
Clearly it is feasible to represent information from many sources as changes in the musical composition, but the Koan system is a more scaleable technique that relies on less musical knowledge. The problem becomes merely one of collecting the information and deciding which musical feature to associate it with, the generative music application organises the sounds into a coherent output.
In a practical implementation, shown in Figures 8 and 9, a JAVA applet
901 is used to set up communication between the virtual environment 902 and the music application 903 Events in the virtual environment 902 are passed to the applet 901 , which then sends the appropriate messages to alter the behaviour of the music application 903. Likewise, changes occurring in the music application 903 may be passed back to the virtual environment 902 using the applet 901 , for example to control the direction of a sound source An event occurring in the virtual environment system causes a program to be run (either in the server 9 or in an individual terminal 1 ) that consults a look-up table 910 to determine which musical segment 91 1 , 91 2, 91 N should be used to represent this event. The selected segment 91 N is then added to any other segments (92N, 93N, etc) already selected to be part of the music to be played by the music application 903. The music application 903 combines the selected segments to generate the musical output. The interactions between the segments may be quite complex - the addition of one segment may modify the way in which others are presented. The selected segments 91 , 92N, 93N continue to play, continuously or periodically, until the system determines that another event has occurred requiring a segment to be added or deleted. Certain events may require a segment to exhibit a temporal characteristic, for example getting louder or softer or changing in pitch, according to how recently the associated event occurred.

Claims

1 . A computer network ( 100) comprising a plurality of user terminals (1 ,2,3,4,5,6,7,8), each having means for indicating a plurality of attributes relating to the terminal or a user of that terminal, means for monitoring one or more of the terminals (7,8), and means for identifying the attributes of the terminals so monitored, means for displaying on at least one displaying terminal (1 ) information relating to the attributes of the terminals (7,8) monitored thereby, wherein the displaying terminal ( 1 ) is arranged to display an indication of the relatedness of the attributes of the respective monitored terminals (7,8) to those currently indicated by the displaying terminal ( 1 ) .
2. A computer network according to Claim 1 , wherein the indicating means, monitoring means and identification means are agent-based processing operators.
3. A network according to claim 1 or 2, wherein each of a plurality of terminals ( 1 ,2,3,4,5,6,7,8) is capable of monitoring and displaying the attributes of the other members of that plurality.
4. A network according to claim 1 , 2 or 3, wherein the attributes include the activities currently performed by the respective terminals.
5. A network according to claim 4, wherein the display is visual.
6. A network according to claim 4 or 5, wherein the display is audible.
7. A network according to claim 6, wherein the display comprises a generative music system.
8. A network according to any preceding claim, wherein at least a first terminal (3) is configured as a tracking terminal to track the attributes of a specified user currently working elsewhere.
9. A network according to claim 8 wherein the tracking terminal (3) is arranged to monitor the network ( 100) and identify whether the specified user's identity is being used on another terminal (4)
10. A network according to claim 9, wherein the tracking terminal (3) is arranged to identify the activities currently in use at the said other terminal (4).
1 1 A network according to any preceding claim, wherein the terminals
( 1 ,2,3,4,5,6,7,8) interact through the operation of a server terminal (9), the attributes of the, or each, monitored terminal ( 1 ,2,3,4,5,6,7,8) being monitored by the server (9) and reported to each displaying terminal (1 )
1 2. A method of operating a computer network (100) comprising a plurality of user terminals ( 1 ,2,3,4,5,6,7,8), each having means for indicating a plurality of attributes relating to the terminal or a user of that terminal, the method comprising the steps of monitoring one or more terminals (7,8), identifying the attributes of the terminals so monitored, displaying on at least one displaying terminal (1 ) information relating to the attributes of the terminals (7,8) monitored thereby, wherein at least one displaying terminal (1 ) displays an indication of the relatedness of the attributes of the respective monitored terminals (7,8) to those currently indicated by the displaying terminal.
13. A method according to Claim 1 1 , wherein the indicating, identification and monitoring are carried out by the interaction of agent-based processing operators.
14. A method according to claim 1 2 or 1 3, wherein each of a plurality of terminals ( 1 ,2,3,4,5,6,7,8) monitors and displays the attributes of the other members of that plurality
1 5. A method according to claim 1 3 or claim 1 4, wherein the attributes include the activities currently performed by the respective terminals.
1 6 A method according to claim 1 5, wherein the display is visual.
1 7. A method according to claim 1 5 or 1 6, wherein the display is audible.
1 8. A method according to claim 1 7, wherein the display comprises a generative music system.
1 9. A method according to any of claims 1 2 to 1 8, wherein at least one tracking terminal (3) tracks the attributes of a specified user currently working elsewhere.
20. A method according to claim 1 9 wherein the tracking terminal (3) monitors the network and identifies whether the specified user's identity is being used on another terminal (4)
21 . A method according to claim 20, wherein the tracking terminal (3) identifies the activities currently in use at the said other terminal (4).
22. A method according to any of claims 1 2 to 21 , wherein the terminals ( 1 ,2,3,4,5,6,7,8) interact through the operation of a server terminal (9), the attributes of the, or each, monitored terminal (8) being monitored by the server (9) and reported to each displaying terminal ( 1 )
PCT/GB1999/001149 1998-04-17 1999-04-15 Computer network WO1999054822A1 (en)

Priority Applications (4)

Application Number Priority Date Filing Date Title
US09/646,384 US6934747B1 (en) 1998-04-17 1999-04-15 Computer network indicating relatedness of attributes of monitored terminals
EP99915949A EP1071995B1 (en) 1998-04-17 1999-04-15 Computer network
DE69903933T DE69903933T2 (en) 1998-04-17 1999-04-15 computer network
JP2000545102A JP4391691B2 (en) 1998-04-17 1999-04-15 Computer network

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
EP98302981A EP0950956A1 (en) 1998-04-17 1998-04-17 Computer network
EP98302981.0 1998-04-17

Publications (2)

Publication Number Publication Date
WO1999054822A1 true WO1999054822A1 (en) 1999-10-28
WO1999054822B1 WO1999054822B1 (en) 1999-12-23

Family

ID=8234777

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/GB1999/001149 WO1999054822A1 (en) 1998-04-17 1999-04-15 Computer network

Country Status (5)

Country Link
US (1) US6934747B1 (en)
EP (2) EP0950956A1 (en)
JP (1) JP4391691B2 (en)
DE (1) DE69903933T2 (en)
WO (1) WO1999054822A1 (en)

Families Citing this family (28)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
TW401548B (en) * 1996-12-20 2000-08-11 Sony Corp Method and apparatus for sending E-mail, method and apparatus for receiving E-mail, sending program supplying medium, receiving program supplying medium
US6772195B1 (en) * 1999-10-29 2004-08-03 Electronic Arts, Inc. Chat clusters for a virtual world application
US7401131B2 (en) 2000-05-22 2008-07-15 Verizon Business Global Llc Method and system for implementing improved containers in a global ecosystem of interrelated services
US6922685B2 (en) * 2000-05-22 2005-07-26 Mci, Inc. Method and system for managing partitioned data resources
DE10026489A1 (en) * 2000-05-27 2001-12-06 Bosch Gmbh Robert Video conference operating method has synchronised representation of video conference scene provided at each video conference terminal
GB0126908D0 (en) * 2001-11-09 2002-01-02 Ibm Method and system for display of activity of users
US7013006B1 (en) 2002-01-18 2006-03-14 Bellsouth Intellectual Property Corporation Programmable audio alert system and method
US7792279B2 (en) * 2002-01-18 2010-09-07 At&T Intellectual Property I, L.P. Distinguishing audio alerts
WO2003094436A2 (en) * 2002-05-03 2003-11-13 Eyebees B.V. Methods and apparatus for monitoring a plurality of individuals
US7305552B2 (en) 2003-11-26 2007-12-04 Siemens Communications, Inc. Screen saver displaying identity content
US7895223B2 (en) 2005-11-29 2011-02-22 Cisco Technology, Inc. Generating search results based on determined relationships between data objects and user connections to identified destinations
US7769806B2 (en) * 2007-10-24 2010-08-03 Social Communications Company Automated real-time data stream switching in a shared virtual area communication environment
US9009603B2 (en) * 2007-10-24 2015-04-14 Social Communications Company Web browser interface for spatial communication environments
US9357025B2 (en) 2007-10-24 2016-05-31 Social Communications Company Virtual area based telephony communications
US8397168B2 (en) 2008-04-05 2013-03-12 Social Communications Company Interfacing with a spatial virtual communication environment
US9381438B2 (en) * 2007-11-07 2016-07-05 International Business Machines Corporation Dynamically displaying personalized content in an immersive environment
US20090248852A1 (en) * 2008-03-30 2009-10-01 Oded Fuhrmann Evaluating Entities Associations with their Respective Environments
US8042051B2 (en) * 2008-05-30 2011-10-18 International Business Machines Corporation Apparatus for navigation and interaction in a virtual meeting place
US8024662B2 (en) * 2008-05-30 2011-09-20 International Business Machines Corporation Apparatus for navigation and interaction in a virtual meeting place
US8285786B2 (en) * 2008-08-29 2012-10-09 International Business Machines Corporation Confidential presentations in virtual world infrastructure
US8988421B2 (en) * 2008-12-02 2015-03-24 International Business Machines Corporation Rendering avatar details
US8219616B2 (en) * 2008-12-15 2012-07-10 International Business Machines Corporation Use of information channels to provide communications in a virtual environment
US9853922B2 (en) 2012-02-24 2017-12-26 Sococo, Inc. Virtual area communications
FR2942091A1 (en) * 2009-02-10 2010-08-13 Alcatel Lucent MULTIMEDIA COMMUNICATION IN A VIRTUAL ENVIRONMENT
US20120001856A1 (en) * 2010-07-02 2012-01-05 Nokia Corporation Responding to tactile inputs
US8855285B1 (en) * 2013-03-14 2014-10-07 Speek Inc. Systems and methods for conference calling using personal URL
ITRM20130220A1 (en) * 2013-04-12 2014-10-13 Marco Antonio Ferrone SYSTEM E.R.T. (ELECTRONIC TRAVELING IN THE MOST SPORTING SPORTS)
JP2022042227A (en) * 2020-09-02 2022-03-14 富士フイルムビジネスイノベーション株式会社 Information processing apparatus and program

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4974173A (en) * 1987-12-02 1990-11-27 Xerox Corporation Small-scale workspace representations indicating activities by other users
US5349662A (en) * 1992-05-21 1994-09-20 International Business Machines Corporation Method of and apparatus for providing automatic detection of user activity
US5511187A (en) * 1992-12-22 1996-04-23 International Business Machines Corporation Method and system for nonvisual groupware participant status determination in a data processing system
EP0784263A1 (en) * 1995-12-22 1997-07-16 Sun Microsystems, Inc. System and method enabling awareness of others working on similar tasks in a computer work environment
EP0874323A1 (en) * 1997-04-23 1998-10-28 International Business Machines Corporation Web-browser-based conference system

Family Cites Families (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5172415A (en) 1990-06-08 1992-12-15 Fosgate James W Surround processor
US5515078A (en) 1992-06-12 1996-05-07 The Computer Museum, Inc. Virtual-reality positional input and display system
US5533129A (en) 1994-08-24 1996-07-02 Gefvert; Herbert I. Multi-dimensional sound reproduction system
US6029195A (en) * 1994-11-29 2000-02-22 Herz; Frederick S. M. System for customized electronic identification of desirable objects
US5796951A (en) * 1995-12-22 1998-08-18 Intel Corporation System for displaying information relating to a computer network including association devices with tasks performable on those devices
WO1998047083A1 (en) 1997-04-16 1998-10-22 British Telecommunications Public Limited Company Data summariser
WO1999030494A1 (en) 1997-12-09 1999-06-17 British Telecommunications Public Limited Company Conference facility

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4974173A (en) * 1987-12-02 1990-11-27 Xerox Corporation Small-scale workspace representations indicating activities by other users
US5349662A (en) * 1992-05-21 1994-09-20 International Business Machines Corporation Method of and apparatus for providing automatic detection of user activity
US5511187A (en) * 1992-12-22 1996-04-23 International Business Machines Corporation Method and system for nonvisual groupware participant status determination in a data processing system
EP0784263A1 (en) * 1995-12-22 1997-07-16 Sun Microsystems, Inc. System and method enabling awareness of others working on similar tasks in a computer work environment
EP0874323A1 (en) * 1997-04-23 1998-10-28 International Business Machines Corporation Web-browser-based conference system

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
HONDA S ET AL: "A VIRTUAL OFFICE ENVIRONMENT BASED ON A SHARED ROOM REALIZING AWARENESS SPACE AND TRANSMITTING AWARENESS INFORMATION", PROCEEDINGS OF THE INTERNATIONAL SYMPOSIUM ON HUMAN FACTORS IN TELECOMMUNICATIONS, OSLO, NORWAY, 12 May 1997 (1997-05-12) - 16 May 1997 (1997-05-16), pages 287 - 295, XP002074473 *

Also Published As

Publication number Publication date
EP1071995B1 (en) 2002-11-13
JP2002512401A (en) 2002-04-23
EP1071995A1 (en) 2001-01-31
DE69903933D1 (en) 2002-12-19
JP4391691B2 (en) 2009-12-24
WO1999054822B1 (en) 1999-12-23
US6934747B1 (en) 2005-08-23
EP0950956A1 (en) 1999-10-20
DE69903933T2 (en) 2003-08-28

Similar Documents

Publication Publication Date Title
US6934747B1 (en) Computer network indicating relatedness of attributes of monitored terminals
US11380020B2 (en) Promoting communicant interactions in a network communications environment
CN102138156B (en) Persisting a group in an instant messaging application
US7124167B1 (en) Computer based system for directing communications over electronic networks
US9137134B2 (en) Recursive identification of individuals for casual collaborative conferencing
US6148328A (en) Method and system for signaling presence of users in a networked environment
US20090287783A1 (en) Method of social interaction between communication device users
US20100017483A1 (en) Multi-topic instant messaging chat session
CN108028763A (en) The telecommunication platform of cloud computing
KR20120118019A (en) Web browser interface for spatial communication environments
AU2005200442A1 (en) Command based group SMS with mobile message receiver and server
JP2003526292A (en) Communication system with media tool and method
JP2002522998A (en) Computer architecture and processes for audio conferencing over local and global networks, including the Internet and intranets
US20240087180A1 (en) Promoting Communicant Interactions in a Network Communications Environment
Bowskill et al. Wearable location mediated telecommunications; a first step towards contextual communication
CN113300934A (en) Communication method, device, equipment and storage medium
Boyer et al. Virtual community prescence awareness
RU2218593C2 (en) Method for telecommunications in computer networks
Gross Towards ubiquitous awareness: the PRAVTA prototype
WO2002035782A2 (en) Method and device for transmitting streaming multimedia messages
KR20020028438A (en) Method for chatting service with integrated voice and character data and computer-readable medium thereof
WO2001063514A1 (en) Applying dynamic user interfaces to multimedia communication via a computer network
KR20050077907A (en) Apparatus for providing multimedia data by using flash and method thereof
Mutemwa et al. Cell phone notification via Bluetooth for Web 2.0 applications
Kaltenbrunner et al. Marvin: supporting awareness through audio in collaborative virtual environments

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A1

Designated state(s): JP US

AL Designated countries for regional patents

Kind code of ref document: A1

Designated state(s): AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE

121 Ep: the epo has been informed by wipo that ep was designated in this application
AK Designated states

Kind code of ref document: B1

Designated state(s): JP US

AL Designated countries for regional patents

Kind code of ref document: B1

Designated state(s): AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE

DFPE Request for preliminary examination filed prior to expiration of 19th month from priority date (pct application filed before 20040101)
WWE Wipo information: entry into national phase

Ref document number: 09646384

Country of ref document: US

WWE Wipo information: entry into national phase

Ref document number: 1999915949

Country of ref document: EP

WWP Wipo information: published in national office

Ref document number: 1999915949

Country of ref document: EP

WWG Wipo information: grant in national office

Ref document number: 1999915949

Country of ref document: EP