US20060064461A1 - Using login-based addressing to communicate with listed users - Google Patents

Using login-based addressing to communicate with listed users Download PDF

Info

Publication number
US20060064461A1
US20060064461A1 US11/265,059 US26505905A US2006064461A1 US 20060064461 A1 US20060064461 A1 US 20060064461A1 US 26505905 A US26505905 A US 26505905A US 2006064461 A1 US2006064461 A1 US 2006064461A1
Authority
US
United States
Prior art keywords
user
video
audio
users
communication
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US11/265,059
Inventor
Lester Ludwig
J. Lauwers
Keith Lantz
Gerald Burnett
Emmett Burns
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Pragmatus AV LLC
Original Assignee
Collaboration Properties Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Family has litigation
First worldwide family litigation filed litigation Critical https://patents.darts-ip.com/?family=46257787&utm_source=google_patent&utm_medium=platform_link&utm_campaign=public_patent_search&patent=US20060064461(A1) "Global patent litigation dataset” by Darts-ip is licensed under a Creative Commons Attribution 4.0 International License.
Priority claimed from US08/131,523 external-priority patent/US5689641A/en
Priority claimed from US09/702,737 external-priority patent/US7185054B1/en
Application filed by Collaboration Properties Inc filed Critical Collaboration Properties Inc
Priority to US11/265,059 priority Critical patent/US20060064461A1/en
Publication of US20060064461A1 publication Critical patent/US20060064461A1/en
Priority to US11/622,583 priority patent/US7487210B2/en
Priority to US11/622,627 priority patent/US7412482B2/en
Priority to US11/623,177 priority patent/US7437411B2/en
Priority to US11/623,387 priority patent/US7398296B2/en
Priority to US11/623,630 priority patent/US7444373B2/en
Priority to US11/624,351 priority patent/US7441001B2/en
Priority to US11/624,860 priority patent/US7437412B2/en
Priority to US11/625,330 priority patent/US20070083596A1/en
Priority to US11/668,625 priority patent/US7822813B2/en
Priority to US11/669,003 priority patent/US7831663B2/en
Priority to US11/737,723 priority patent/US7730132B2/en
Assigned to AVISTAR COMMUNICATIONS CORPORATION reassignment AVISTAR COMMUNICATIONS CORPORATION MERGER (SEE DOCUMENT FOR DETAILS). Assignors: COLLABORATION PROPERTIES, INC.
Assigned to BALDWIN ENTERPRISES, INC., AS COLLATERAL AGENT reassignment BALDWIN ENTERPRISES, INC., AS COLLATERAL AGENT SECURITY AGREEMENT Assignors: AVISTAR COMMUNICATIONS CORPORATION
Assigned to VICOR, INC. reassignment VICOR, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: BURNS, EMMETT R., BURNETT, GERALD J., LANTZ, KEITH A., LAUWERS, J. CHRIS, LUDWIG, LESTER F.
Assigned to COLLABORATION PROPERTIES, INC. reassignment COLLABORATION PROPERTIES, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: VICOR, INC.
Assigned to AVISTAR COMMUNICATIONS CORPORATION reassignment AVISTAR COMMUNICATIONS CORPORATION RELEASE OF SECURITY INTEREST IN INTELLECTUAL PROPERTY Assignors: BALDWIN ENTERPRISES, INC., AS COLLATERAL AGENT
Assigned to AVISTAR COMMUNICATIONS CORPORATION reassignment AVISTAR COMMUNICATIONS CORPORATION RELEASE BY SECURED PARTY (SEE DOCUMENT FOR DETAILS). Assignors: BALDWIN ENTERPRISES, INC.
Assigned to INTELLECTUAL VENTURES FUND 61 LLC reassignment INTELLECTUAL VENTURES FUND 61 LLC ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: AVISTAR COMMUNICATIONS CORPORATION
Assigned to PRAGMATUS AV LLC reassignment PRAGMATUS AV LLC ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: INTELLECTUAL VENTURES FUND 61 LLC
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q10/00Administration; Management
    • G06Q10/10Office automation; Time management
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F15/00Digital computers in general; Data processing equipment in general
    • G06F15/16Combinations of two or more digital computers each having at least an arithmetic unit, a program unit and a register, e.g. for a simultaneous processing of several programs
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L12/00Data switching networks
    • H04L12/02Details
    • H04L12/16Arrangements for providing special services to substations
    • H04L12/18Arrangements for providing special services to substations for broadcast or conference, e.g. multicast
    • H04L12/1813Arrangements for providing special services to substations for broadcast or conference, e.g. multicast for computer conferences, e.g. chat rooms
    • H04L12/1818Conference organisation arrangements, e.g. handling schedules, setting up parameters needed by nodes to attend a conference, booking network resources, notifying involved parties
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10TECHNICAL SUBJECTS COVERED BY FORMER USPC
    • Y10STECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10S707/00Data processing: database and file management or data structures
    • Y10S707/99931Database or file accessing
    • Y10S707/99933Query processing, i.e. searching
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10TECHNICAL SUBJECTS COVERED BY FORMER USPC
    • Y10STECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10S707/00Data processing: database and file management or data structures
    • Y10S707/99951File or database maintenance
    • Y10S707/99952Coherency, e.g. same view to multiple users
    • Y10S707/99953Recoverability

Definitions

  • the present invention relates to computer-based systems for enhancing collaboration between and among individuals who are separated by distance and/or time (referred to herein as “distributed collaboration”). Principal among the invention's goals is to replicate in a desktop environment, to the maximum extent possible, the full range, level and intensity of interpersonal communication and information sharing which would occur if all the participants were together in the same room at the same time (referred to herein as “face-to-face collaboration”).
  • collaboration often involves the sharing of visual information—e.g., printed material such as articles, drawings, photographs, charts and graphs, as well as videotapes and computer-based animations, visualizations and other displays—in such a way that the participants can collectively and interactively examine, discuss, annotate and revise the information.
  • visual information e.g., printed material such as articles, drawings, photographs, charts and graphs, as well as videotapes and computer-based animations, visualizations and other displays—in such a way that the participants can collectively and interactively examine, discuss, annotate and revise the information.
  • This combination of spoken words, gestures, visual cues and interactive data sharing significantly enhances the effectiveness of collaboration in a variety of contexts, such as “brainstorming” sessions among professionals in a particular field, consultations between one or more experts and one or more clients, sensitive business or political negotiations, and the like.
  • beneficial effects of face-to-face collaboration will be realized only to the extent that each of the remotely located participants can be “recreate
  • prior art desktop videoconferencing systems provide, at best, only a partial solution to the challenges of distributed collaboration in real time, primarily because of their lack of high-quality video (which is necessary for capturing the visual cues discussed above) and their limited data sharing capabilities.
  • telephone answering machines, voice mail, fax machines and conventional electronic mail systems provide incomplete solutions to the problems presented by deferred (asynchronous) collaboration because they are totally incapable of communicating visual cues, gestures, etc and, like conventional videoconferencing systems, are generally limited in the richness of the data that can be exchanged.
  • workstations audio and video capture and processing capabilities have recently been integrated into desktop and portable personal computers and workstations (hereinafter generically referred to as “workstations”). These capabilities have been used primarily in desktop multimedia authoring systems for producing CD-ROM-based works. While such systems are capable of processing, combining, and recording audio, video and data locally (i.e., at the desktop), they do not adequately support networked collaborative environments, principally due to the substantial bandwidth requirements for real-time transmission of high-quality, digitized audio and full-motion video which preclude conventional LANs from supporting more than a few workstations.
  • FIG. 1 is an enterprise view of a desk-top collaboration system embodiment of the present invention.
  • FIGS. 2A and 2B are photographs which attempt to illustrate, to the extent possible in a still image, the high-quality of the full-motion video and related user interface displays that appear on typical CMW screens which may be generated during operation of a preferred embodiment of the invention.
  • FIG. 3 is a block and schematic diagram of a preferred embodiment of a “multimedia local area network” (MLAN) in accordance with a desktop collaboration system embodiment of the present invention.
  • MLAN multimedia local area network
  • FIG. 4 is a block and schematic diagram illustrating how a plurality of geographically dispersed MLANs of the type shown in FIG. 3 can be connected via a wide area network in accordance with the present invention.
  • FIG. 5 is a schematic diagram illustrating how collaboration sites at distant locations L 1 -L 8 are conventionally interconnected over a wide area network by individually connecting each site to every other site.
  • FIG. 6 is a schematic diagram illustrating how collaboration sites at distant locations L 1 -L 8 are interconnected over a wide area network in a preferred embodiment of the invention using a multi-hopping approach.
  • FIG. 7 is a block diagram illustrating a preferred embodiment of video mosaicing circuitry provided in the MLAN of FIG. 3 .
  • FIGS. 8A, 8B and 8 C illustrate the video window on a typical CMW screen which may be generated during operation of a preferred embodiment of the present invention, and which contains only the callee for two-party calls ( 8 A) and a video mosaic of all participants, e.g., for four-party ( 8 B) or eight-party ( 8 C) conference calls.
  • FIG. 9 is a block diagram illustrating a preferred embodiment of audio mixing circuitry provided in the MLAN of FIG. 3 .
  • FIG. 10 is a block diagram illustrating video cut-and-paste circuitry provided in the MLAN of FIG. 3 .
  • FIG. 11 is a schematic diagram illustrating typical operation of the video cut-and-paste circuitry in FIG. 10 .
  • FIGS. 12-17 (consisting of FIGS. 12A, 12B , 13 A, 13 B, 14 A, 14 B, 15 A, 15 B, 16 , 17 A and 17 B) illustrate various examples of how a preferred embodiment of the present invention provides video mosaicing, video cut-and-pasting, and audio mixing at a plurality of distant sites for transmission over a wide area network in order to provide, at the CMW of each conference participant, video images and audio captured from the other conference participants.
  • FIGS. 18A and 18B illustrate various preferred embodiments of a CMW which may be employed in accordance with the present invention.
  • FIG. 19 is a schematic diagram of a preferred embodiment of a CMW add-on box containing integrated audio and video L/O circuitry in accordance with the present invention.
  • FIG. 20 illustrates CMW software in accordance with a preferred embodiment of the present invention, integrated with standard multitasking operating system and applications software.
  • FIG. 21 illustrates software modules which may be provided for running on the MLAN Server in the MLAN of FIG. 3 for controlling operation of the AV and Data Networks.
  • FIG. 22 illustrates an enlarged example of “speed-dial” face icons of certain collaboration participants in a Collaboration Initiator window on a typical CMW screen which may be generated during operation of a preferred embodiment of the present invention.
  • FIG. 23 is a diagrammatic representation of the basic operating events occurring in a preferred embodiment of the present invention during initiation of a two-party call.
  • FIG. 24 is a block and schematic diagram illustrating how physical connections are established in the MLAN of FIG. 3 for physically connecting first and second workstations for a two-party videoconference call.
  • FIG. 25 is a block and schematic diagram illustrating how physical connections are preferably established in MLANs such as illustrated in FIG. 3 , for a two-party call between a first CMW located at one site and a second CMW located at a remote site.
  • FIGS. 26 and 27 are block and schematic diagrams illustrating how conference bridging is preferably provided in the MLAN of FIG. 3 .
  • FIG. 28 diagrammatically illustrates how a snapshot with annotations may be stored in a plurality of bitmaps during data sharing.
  • FIG. 29 is a schematic and diagrammatic illustration of the interaction among multimedia mail (MMM), multimedia call/conference recording (MMCR) and multimedia document management (MMDM) facilities.
  • MMM multimedia mail
  • MMCR multimedia call/conference recording
  • MMDM multimedia document management
  • FIG. 30 is a schematic and diagrammatic illustration of the multimedia document architecture employed in a preferred embodiment of the invention.
  • FIG. 31A illustrates a centralized Audio/Video Storage Server.
  • FIG. 31B is a schematic and diagrammatic illustration of the interactions between the Audio/Video Storage Server and the remainder of the CMW System.
  • FIG. 31C illustrates an alternative embodiment of the interactions illustrated in FIG. 31B .
  • FIG. 31D is a schematic and diagrammatic illustration of the integration of MMM MMCR and MMDM facilities in a preferred embodiment of the invention.
  • FIG. 32 illustrates a generalized hardware implementation of a scalable Audio/Video Storage Server.
  • FIG. 33 illustrates a higher throughput version of the server illustrated in FIG. 32 , using SCSI-based crosspoint switching to increase the number of possible simultaneous file transfers.
  • FIG. 34 illustrates the resulting multimedia collaboration environment achieved by the integration of audio/video/data teleconferencing and MMCR, MMM and MMDM.
  • FIGS. 35-42 illustrate a series of CMW screens which may be generated during operation of a preferred embodiment of the present invention for a typical scenario involving a remote expert who takes advantage of many of the features provided by the present invention.
  • desk-top teleconferencing includes real-time audio and/or video teleconferencing, as well as data conferencing
  • Data conferencing includes snapshot sharing (sharing of “snapshots” of selected regions of the user's screen), application sharing (shared control of running applications), shared whiteboard (equivalent to sharing a “blank” window), and associated telepointing and annotation capabilities. Teleconferences may be recorded and stored for later playback, including both audio/video and all data interactions.
  • desktop teleconferencing supports real-time interactions
  • multimedia mail permits the asynchronous exchange of arbitrary multimedia documents, including previously recorded teleconferences.
  • multimedia capabilities underlying desktop teleconferencing and multimedia mail also greatly facilitate the creation, viewing, and manipulation of high-quality multimedia documents in general, including animations and visualizations that might be developed, for example, in the course of information analysis and modeling. Further, these animations and visualizations may be generated for individual rather than collaborative use, such that the present invention has utility beyond a collaboration context.
  • the preferred embodiment of the invention is a collaborative multimedia workstation (CMW) system wherein very high-quality audio and video capabilities can be readily superimposed onto an enterprise's existing computing and network infrastructure, including workstations, LANs, WANs, and building wiring.
  • CMS collaborative multimedia workstation
  • the system architecture employs separate real-time and asynchronous networks—the former for real-time audio and video, and the latter for non-real-time audio and video, text, graphics and other data, as well as control signals.
  • These networks are interoperable across different computers (e.g., Macintosh, Intel-based PCs, and Sun workstations), operating systems (e.g., Apple System 7, DOS/Windows, and UNIX) and network operating systems (e.g., Novell Netware and Sun ONC+). In many cases, both networks can actually share the same cabling and wall jack connector.
  • the system architecture also accommodates the situation in which the user's desktop computing and/or communications equipment provides varying levels of media-handling capability.
  • a collaboration session may include participants whose equipment provides capabilities ranging from audio only (a telephone) or data only (a personal computer with a modem) to a full complement of real-time, high-fidelity audio and full-motion video, and high-speed data network facilities.
  • the CMW system architecture is readily, scalable to very large enterprise-wide network environments accommodating thousands of users. Further, it is an open architecture that can accommodate appropriate standards. Finally, the CMW system incorporates an intuitive, yet powerful, user interface, making the system easy to learn and use.
  • the present invention thus provides a distributed multimedia collaboration environment that achieves the benefits of face-to-face collaboration as nearly as possible, leverages (“snaps on to”) existing computing and network infrastructure to the maximum extent possible, scales to very large networks consisting of thousand of workstations, accommodates emerging standards, and is easy to learn and use.
  • each of a plurality of “multimedia local area networks” (MLANs) 10 connects, via lines 13 , a plurality of CMWs 12 - 1 to 12 - 10 and provides audio/video/data networking for supporting collaboration among CMW users WAN 15 in turn connects multiple MLANs 10 , and typically includes appropriate combinations of common carrier analog and digital transmission networks
  • MLANs 10 on the same physical premises may be connected via bridges/routes 11 , as shown, to WANs and one another.
  • the system of FIG. 1 accommodates both “real time” delay and jitter-sensitive signals (e.g., real-time audio and video teleconferencing) and classical asynchronous data (e.g., data control signals as well as shared textual, graphics and other media) communication among multiple CMWs 12 regardless of their location.
  • jitter-sensitive signals e.g., real-time audio and video teleconferencing
  • classical asynchronous data e.g., data control signals as well as shared textual, graphics and other media
  • shared textual, graphics and other media shared textual, graphics and other media
  • CMW 12 in FIG. 1 may use any of a variety of types of operating systems, such as Apple System 7, UNIX, DOS/Windows and OS/2.
  • the CMWs can also have different types of window systems. Specific preferred embodiments of a CMW 12 are described hereinafter in connection with FIGS. 18A and 18B . Note that this invention allows for a mix of operating systems and window systems across individual CMWs.
  • CMW 12 in FIG. 1 provides real-time audio/video/data capabilities along with the usual data processing capabilities provided by its operating system
  • CMW 12 also provides for bidirectional communication, via lines 13 , within MLAN 10 , for audio/video signals as well as data signals
  • Audio/video signals transmitted from a CMW 12 typically comprise a high-quality live video image and audio of the CMW operator. These signals are obtained from a video camera and microphone provided at the CMW (via an add-on unit or partially or totally integrated into the CMW), processed, and then made available to low-cost network transmission subsystems.
  • Audio/video signals received by a CMW 12 from MLAN 10 may typically include: video images of one or more conference participants and associated audio, video and data from multimedia mail, previously recorded audio/video from previous calls and conferences, and standard broadcast television (e.g. CNN). Received video signals are displayed on the CMW screen or on an adjacent monitor, and the accompanying audio is reproduced by a speaker provided in or near the CMW.
  • the required transducers and signal processing hardware could be integrated into the CMW, or be provided via a CMW add-on unit, as appropriate.
  • FIG. 2A illustrates a CMW screen containing live, full-motion video of three conference participants
  • FIG. 2B illustrates data shared and annotated by those conferees (lower left window).
  • MLAN 10 having ten CMWs ( 12 - 1 , 12 - 2 ,- 12 - 10 ), coupled therein via lines 13 a and 13 b
  • MLAN 10 typically extends over a distance from a few hundred feet to a few miles, and is usually located within a building or a group of proximate buildings.
  • the current preferred embodiment uses analog networking for audio and video, it also physically separates the real-time and asynchronous switching vehicles and, in particular, assumes an analog audio/video switch.
  • a common switching vehicle e.g., ATM
  • the MLAN 10 thus can be implemented in the preferred embodiment using conventional technology, such as typical Data LAN hubs 25 and A/V Switching Circuitry 30 (as used in television studios and other closed-circuit television networks), linked to the CMWs 12 via appropriate transceivers and unshielded twisted pair (UTP) wiring.
  • lines 13 which interconnect each CMW 12 within its respective MLAN 10 , comprise two sets of lines 13 a and 13 b Lines 13 a provide bidirectional communication of audio/video within MLAN 10 , while lines 13 b provide for the bidirectional communication of data.
  • This separation permits conventional LANs to be used for data communications and a supplemental network to be used for audio/video communications.
  • audio/video/data networking can also be implemented using a single pair of lines for both audio/video and data communications via a very wide variety of analog and digital multiplexing schemes.
  • lines 13 a and 13 b may be implemented in various ways, it is currently preferred to use commonly installed 4-pair UTP telephone wires, wherein one pair is used for incoming video with accompanying audio (mono or stereo) multiplexed in, wherein another pair is used for outgoing multiplexed audio/video, and wherein the remaining two pairs are used for carrying incoming and outgoing data in ways consistent with existing LANs.
  • 10BaseT Ethernet uses RJ-45 pins 1, 2, 4, and 6, leaving pins 3, 5, 7, and 8 available for the two A/V twisted pairs.
  • the resulting system is compatible with standard (AT&T 258A, EIA/TIA 568, 8P8C, 10BaseT, ISDN, 6P6C, etc.) telephone wiring found commonly throughout telephone and LAN cable plants in most office buildings throughout the world.
  • These UTP wires are used in a hierarchy or peer arrangements of star topologies to create MLAN 10 , described below. Note that the distance range of the data wires often must match that of the video and audio Various UTP-compatible data.
  • LAN networks may be used, such as Ethernet, token ring, FDDI, ATM etc. For distances longer than the maximum distance specified by the data LAN protocol, data signals can be additionally processed for proper UTP operations.
  • lines 13 a from each CMW 12 are coupled to a conventional Data LAN hub 25 , which facilitates the communication of data (including control signals) among such CMWs
  • Lines 13 b in FIG. 3 are connected to A/V Switching Circuitry 30 .
  • One or more conference bridges 35 are coupled to A/V Switching Circuitry 30 and possibly (if needed) the Data LAN hub 25 , via lines 35 b and 35 a , respectively, for providing multi-party Conferencing in a particularly advantageous manner, as will hereinafter be described in detail
  • a WAN gateway 40 provides for bidirectional communication between MLAN 10 and WAN 15 in FIG. 1 .
  • Data LAN hub 25 and A/V Switching Circuitry 30 are coupled to WAN gateway 40 via outputs 25 a and 30 a , respectively.
  • Other devices connect to the A/V Switching Circuitry 30 and Data LAN hub 25 to add additional features, such as multimedia mail, conference recording, etc.) as discussed below.
  • MLAN Server 60 supports the TCP/IP network protocol suite. Accordingly, software processes on CMWs 12 communicate with one another and MLAN Server 60 via MLAN 10 using these protocols. Other network protocols could also be used, such as IPX. The manner in which software running on MLAN Server 60 controls the operation of MLAN 10 will be described in detail hereinafter.
  • Data LAN hub 25 , A/V Switching Circuitry 30 and MLAN Server 60 also provide respective lines 25 b , 30 b , and 60 e for coupling to additional multimedia resources 16 ( FIG. 1 ), such as multimedia document management, multimedia databases, radio/TV channel setc Data LAN hub 25 (via bridges/routers 11 in FIG.
  • A/V Switching Circuitry 30 additionally provide lines 25 c and 30 c for coupling to one or more other MLANs 10 which may be in the same locality (i.e., not far enough away to require use of WAN technology)
  • WAN gateways 40 are used to provide highest quality compression methods and standards in a shared resource fashion, thus minimizing costs at the workstation for a given WAN quality level, as discussed below.
  • FIG. 2B illustrates a CMW screen showing a multimedia EMAIL mailbox (top left window) containing references to a number of received messages along with a video enclosure (top right window) to the selected message.
  • A/V Switching Circuitry 30 (whether digital or analog as in the preferred embodiment) provides common audio/video switching for CMWs 12 , conference bridges 35 , WAN gateway 40 and multimedia resources 16 , as determined by MLAN Server 60 , which in turn controls conference bridges 35 and WAN gateway 44 ).
  • asynchronous data is communicated within MLAN 10 utilizing common data communications formats where possible (e.g., for snapshot sharing) so that the system can handle such data in a common manner, regardless of origin, thereby facilitating multimedia mail and data sharing as well as audio/video communications.
  • an initiating CMW 12 signals MLAN Server 60 via Data LAN hub 25 identifying the desired conference participants. After determining which of these conferees will accept the call, MLAN Server 60 controls A/V Switching Circuitry 30 (and CMW software via the data network) to set up the required audio/video and data paths to conferees at the same location as the initiating CMW.
  • the respective MLAN Servers 60 of the involved MLANs 10 control their respective A/V Switching Circuitry 30 , conference bridges 35 , and WAN gateways 40 to set up appropriate communication paths (via WAN 15 in FIG. 1 ) as required for interconnecting the conferees MLAN Servers 60 also communicate with one another via data paths so that each MLAN 10 contains updated information as to the capabilities of all of the system CMWs 12 , and also the current locations of all parties available for teleconferencing.
  • the data conferencing component of the above-described system supports the sharing of visual information at one or more CMWs (as described in greater detail below). This encompasses both “snapshot sharing” (sharing “snapshots” of complete or partial screens, or of one or more selected windows) and “application sharing” (sharing both the control and display of running applications) When transferring images, lossless or slightly lossy image compression can be used to reduce network bandwidth requirements and user-perceived delay while maintaining high image quality.
  • any participant can point at or annotate the shared data.
  • These associated telepointers and annotations appear on every participant's CMW screen as they are drawn (i.e., effectively in real time).
  • FIG. 2B which illustrates a typical CMW screen during a multi-party teleconferencing session, wherein the screen contains annotated shared data as well as video images of the conferees.
  • all or portions of the audio/video and data of the teleconference can be recorded at a CMW (or within MLAN 10 ), complete with all the data interactions.
  • audio/video file services can be implemented either at the individual CMWs 12 or by employing a centralized audio/video storage server.
  • This is one example of the many types of additional servers that can be added to the basic system of MLANs 10 .
  • a similar approach is used for incorporating other multimedia services, such as commercial TV channels, multimedia mail, multimedia document management, multimedia conference recording, visualization servers, etc. (as described in greater detail below)
  • multimedia services such as commercial TV channels, multimedia mail, multimedia document management, multimedia conference recording, visualization servers, etc.
  • GUI and software used for snapshot sharing can also be used as an input/output interface for multimedia mail and more general forms of multimedia documents. This can be accomplished by structuring the interprocess communication protocols to be uniform across all these applications. More complicated examples—specifically multimedia conference recording, multimedia mail and multimedia document management—will be presented in detail below.
  • FIG. 4 Next to be described in connection with FIG. 4 is the advantageous manner in which the present invention provides for real-time audio/video/data communication among geographically dispersed MLANs 10 via WAN 15 ( FIG. 1 ), whereby communication delays, cost and degradation of video quality are significantly minimized from what would otherwise be expected.
  • MLANs 10 are illustrated at locations A, B, C and D CMWs 12 - 1 to 12 - 10 , A/V Switching Circuitry 30 , Data LAN hub 25 , and WAN gateway 40 at each location correspond to those shown in FIGS. 1 and 3 WAN gateway 40 in FIG. 4 will be seen to comprise a router/codec (R&C) bank 42 coupled to WAN 15 via WAN switching multiplexer 44 .
  • the router is used for data interconnection and the codec is used for audio/video interconnection (for multimedia mail and document transmission, as well as videoconferencing). Codecs from multiple vendors, or supporting various compression algorithms may be employed.
  • the router and codec are combined with the switching multiplexer to form a single integrated unit.
  • WAN 15 is comprised of T1 or ISDN common-carrier-provided digital links (switched or dedicated), in which case WAN switching multiplexers 44 are of the appropriate type (T1, ISDN, fractional T1, T3, switched 56 Kbps, etc.). Note that the WAN switching multiplexer 44 typically creates subchannels whose bandwidth is a multiple of 64 Kbps (i.e., 256 Kbps, 384, 768, etc.) among the T1, T3 or ISDN carriers. Inverse multiplexers may be required when using 56 Kbps dedicated or switched services from these carriers.
  • router/codec bank 42 in FIG. 4 provides conventional analog-to-digital conversion and compression of audio/video signals received from A/V Switching Circuitry 30 for transmission to WAN 15 via WAN switching multiplexer 44 , along with transmission and routing of data signals received from Data LAN hub 25 .
  • each router/codec bank 42 in FIG. 4 provides digital-to-analog conversion and decompression of audio/video digital signals received from WAN 15 via WAN switching multiplexer 44 for transmission to A/V Switching Circuitry 30 , along with the transmission to Data LAN hub 25 of data signals received from WAN 15 .
  • the system also provides optimal routes for audio/video signals through the WAN.
  • location. A can take either a direct route to location D via path 47 , or a two-hop route through location C via paths 48 and 49 . If the direct path 47 linking location A and location D is unavailable, the multipath route via location C and paths 48 and 49 could be used.
  • MLAN 10 to MLAN 10 i.e., site-to-site
  • FIG. 5 shows that using the conventional “fully connected mesh” location-to-location approach, thirty-six WAN links are required for interconnecting the nine locations L 1 to L 8 .
  • FIG. 6 shows that using the above multi-hop capabilities, only nine WAN links are required, as shown in FIG. 6 .
  • the number of locations increase, the difference in cost becomes even greater, growing as the square of the number of sites. For example, for 100 locations, the conventional approach would require about 5,000 WAN links, while the multi-hop approach of the present invention would typically require 300 or fewer (possibly considerably fewer) WAN links.
  • specific WAN links for the multi-hop approach of the invention would require higher bandwidth to carry the additional traffic, the cost involved is very much smaller as compared to the cost for the very much larger number of WAN links required by the conventional approach.
  • the WAN switching multiplexer routes audio/video signals directly from the WAN network interface through an available codec to MLAN 10 and vice versa.
  • video signals are routed from one network interface on the WAN switching multiplexer to another network interface.
  • A/V Switching Circuitry 30 could be used for this purpose, the preferred embodiment provides switching functionality inside the WAN switching multiplexer. By doing so, it avoids having to route audio/video signals through codecs to the analog switching circuitry, thereby avoiding additional codec delays at the intermediate locations.
  • a product capable of performing the basic switching functions described above for WAN switching multiplexer 44 is available from Teleos Corporation, Eatontown, N.J. This product is not known to have been used for providing audio/video multi-hopping and dynamic switching among various WAN links as described above.
  • the preferred embodiment of the present invention provides a particularly advantageous way of minimizing delay, cost and degradation of video quality in a multi-party video teleconference involving geographically dispersed sites, while still delivering full conference views of all participants.
  • each site has to allocate (in router/codec bank 42 in FIG. 4 ) a separate codec for each participant, as well as a like number of WAN trunks (via WAN switching multiplexer 44 in FIG. 4 ).
  • the preferred embodiment of the invention advantageously permits each wide area audio/video teleconference to use only one codec at each site, and a minimum number of WAN digital trunks.
  • the preferred embodiment achieves this most important result by employing “distributed” video mosaicing via a video “cut-and-paste” technology along with distributed audio mixing.
  • FIG. 7 illustrates a preferred way of providing video mosaicing in the MLAN of FIG. 3 —i.e., by combining the individual analog video pictures from the individuals participating in a teleconference into a single analog mosaic picture.
  • analog video signals 112 - 1 to 112 - n from the participants of a teleconference are applied to video mosaicing circuitry 36 , which in the preferred embodiment is provided as part of conference bridge 35 in FIG. 3 .
  • These analog video inputs 112 - 1 to 112 - n are obtained from the A/V Switching Circuitry 30 ( FIG. 3 ) and may include video signals from CMWs at one or more distant sites (received via WAN gateway 40 ) as well as from other CMWs at the local site.
  • video mosaicing circuitry 36 is capable of receiving N individual analog video picture signals (where N is a squared integer, i.e., 4, 9, 16, etc.). Circuitry 36 first reduces the size of the N input video signals by reducing the resolutions of each by a factor of M (where M is the square root of N (i.e., 2, 3, 4, etc.), and then arranging them in an M-by-M mosaic of N images. The resulting single analog mosaic 36 a obtained from video mosaicing circuitry 36 is then transmitted to the individual CMWs for display on the screens thereof.
  • N is a squared integer, i.e., 4, 9, 16, etc.
  • a mosaic containing four participants is shown in FIG. 8B .
  • FIG. 9 illustrates preferred audio mixing circuitry 38 for use in conjunction with the video mosaicing circuitry 36 in FIG. 7 , both of which may be part of conference bridges 35 in FIG. 3 .
  • audio signals 114 - 1 to 114 - n are applied to audio summing circuitry 38 for combination.
  • These input audio signals 114 - 1 to 114 - n may include audio signals from local participants as well as audio sums from participants at distant sites.
  • Audio mixing circuitry 38 provides a respective “minus-1” sum output 38 a - 1 , 38 a - 2 , etc., for each participant. Thus, each participant hears every conference participant's audio except his/her own.
  • sums are decomposed and formed in a distributed fashion, creating partial sums at one site which are completed at other sites by appropriate signal insertion. Accordingly, audio mixing circuitry 38 is able to provide one or more additional sums, such as indicated by output 38 , for sending to other sites having conference participants.
  • video cut-and-paste circuitry 39 is provided for this purpose, and may also be incorporated in the conference bridges 35 in FIG. 3 .
  • video cut-and-paste circuitry receives analog video inputs 116 , which may be comprised of one or more mosaics or single video pictures received from one or more distant sites and a mosaic or single video picture produced by the local site. It is assumed that the local video mosaicing circuitry 36 ( FIG. 7 ) and the video cut-and-paste circuitry 39 have the capability of handling all of the applied individual video pictures, or at least are able to choose which ones are to be displayed based on existing available signals.
  • the video cut-and-paste circuitry 39 digitizes the incoming analog video inputs 116 , selectively rearranges the digital signals on a region-by-region basis to produce a single digital M-by-M mosaic, having individual pictures in selected regions, and then converts the resulting digital mosaic back to analog form to provide a single analog mosaic picture 39 a for sending to local participants (and other sites where required) having the individual input video pictures in appropriate regions.
  • This resulting cut-and-paste analog mosaic 39 a will provide the same type of display as illustrated in FIG. 8B .
  • video cut-and-paste circuitry 39 will provide additional cut-and-paste mosaics 39 b - 1 , 39 b - 2 , etc. for this purpose.
  • FIG. 11 diagrammatically illustrates an example of how video cut-and-paste circuitry may operate to provide the cut-and-paste analog mosaic 39 a .
  • four digitized individual signals 116 a , 116 b , 116 c and 116 d derived from the input video signals are “pasted” into selected regions of a digital frame buffer 17 to form a digital 2 ⁇ 2 mosaic, which is converted into an output analog video mosaic 39 a or 39 b in FIG. 10 .
  • the required audio partial sums may be provided by audio mixing circuitry 39 in FIG. 9 in the same manner, replacing each cut-and-paste video operation with a partial sum operation.
  • FIGS. 12-17 will illustrate how these capabilities may advantageously be used in combination in the context of wide-area videoconferencing.
  • the teleconference is assumed to have four participants, designated as A, B, C and D, in which case 2 ⁇ 2 (quad) mosaics are employed. It is to be understood that greater numbers of participants could be provided.
  • FIGS. 12A and 12B illustrate a 2-site example having two participants A and B at Site # 1 and two participants C and D at Site # 2 Note that this example requires mosaicing and cut-and-paste at both sites.
  • FIGS. 13A and 13B illustrate another 2-site example, but having three participants A, B and C at Site # 1 and one participant D at Site # 2 . Note that this example requires mosaicing at both sites, but cut-and-paste only at Site # 2 .
  • FIGS. 14A and 14B illustrate a 3-site example having participants A and B at Site # 1 , participant C at Site # 2 , and participant D at Site # 3 .
  • the two local videos A and B are put into a mosaic which is sent to both Site # 2 and Site # 3 .
  • cut-and-paste is used to insert the single video (C or D) at that site into the empty region in the imported A, B, and D or C mosaic, as shown. Accordingly, mosaicing is required at all three sites, and cut-and-paste is required for only Site # 2 and Site # 3 .
  • FIGS. 15A and 15B illustrate another 3-site example having participant A at Site # 1 , participant B at Site # 2 , and participants C and D at Site # 3 .
  • mosaicing and cut-and-paste are required at all sites Site # 2 additionally has the capability to send different cut-and-paste mosaics to Sites # 1 and Site # 3 .
  • Site # 2 creates minus-1 audio mixes for Site # 1 and Site # 2 , but only provides a partial audio mix (A&B) for Site # 3 .
  • These partial mixes are completed at Site # 3 by mixing in C's signal to complete D's mix (A+B+C) and D's signal to complete C's mix (A+B+D).
  • FIG. 16 illustrates a 4-site example employing a star topology, having one participant at each site; that is, participant A is at Site # 1 , participant B is at Site # 2 , participant C is at Site # 3 , and participant D is at Site # 4 .
  • An audio implementation is not illustrated for this example, since standard minus-1 mixing can be performed at Site # 1 , and the appropriate sums transmitted to the other sites.
  • FIGS. 17A and 17B illustrate a 4-site example that also has only one participant at each site, but uses a line topology rather than a star topology as in the example of FIG. 16 . Note that this example requires mosaicing and cut-and-paste at all sites. Also note that Site # 2 and Site # 3 are each required to transmit two different types of cut-and-paste mosaics.
  • the preferred embodiment also provides the capability of allowing a conference participant to select a close-up of a participant displayed on a mosaic. This capability is provided whenever a full individual video picture is available at that user's site. In such case, the A/V Switching Circuitry 30 ( FIG. 3 ) switches the selected full video picture (whether obtained locally or from another site) to the CMW that requests the close-up.
  • FIGS. 18A, 18B , 19 and 20 are various preferred embodiments of a CMW in accordance with the invention.
  • FIG. 18A One embodiment of a CMW 12 of the present invention is illustrated in FIG. 18A
  • personal computers e.g., an Apple Macintosh or an IBM-compatible PC, desktop or laptop
  • workstations e.g., a Sun SPARC station
  • the currently available personal computers and workstations serve as a base workstation platform.
  • the addition of certain audio and video I/O devices to the standard components of the base platform 100 (where standard components include the display monitor 200 , keyboard 300 and mouse or tablet (or other pointing device) 400 ), all of which connect with the base platform box through standard peripheral ports 101 , 102 and 103 , enables the CMW to generate and receive real-time audio and video signals.
  • These devices include a video camera 500 for capturing the user's image, gestures and surroundings (particularly the user's face and upper body), a microphone 600 for capturing the user's spoken words (and any other sounds generated at the CMW), a speaker 700 for presenting incoming audio signals (such as the spoken words of another participant to a videoconference or audio annotations to a document), a video input card 130 in the base platform 100 for capturing incoming video signals (e.g., the image of another participant to a videoconference, or videomail), and a video display card 120 for displaying video and graphical output on monitor 200 (where video is typically displayed in a separate window).
  • a video camera 500 for capturing the user's image, gestures and surroundings (particularly the user's face and upper body)
  • a microphone 600 for capturing the user's spoken words (and any other sounds generated at the CMW)
  • a speaker 700 for presenting incoming audio signals (such as the spoken words of another participant to a videoconference or audio annotations to a document)
  • peripheral audio and video I/O devices are readily available from a variety of vendors and are just beginning to become standard features in (and often physically integrated into the monitor and/or base platform of) certain personal computers and workstations. See, e.g., the aforementioned BYTE article (“Video Conquers the Desktop”), which describes current models of Apple's Macintosh AV series personal computers and Silicon Graphics' Indy workstations.
  • Add-on box 800 integrates these audio and video I/O devices with additional functions (such as adaptive echo canceling and signal switching) and interfaces with AV Network 901
  • Network 901 is the part of the MLAN 10 which carries bidirectional audio and video signals among the CMWs and A/V Switching Circuitry 30 —e.g., utilizing existing UTP wiring to carry audio and video signals (digital or analog, as in the present embodiment).
  • the AV network 901 is separate and distract from the Data Network 902 portion of the MLAN 10 , which carries bidirectional data signals among the CMWs and the Data LAN hub (e.g., an Ethernet network that also utilizes UTP wiring in the present embodiment with a network interface card 110 in each CMW).
  • the Data LAN hub e.g., an Ethernet network that also utilizes UTP wiring in the present embodiment with a network interface card 110 in each CMW.
  • each CMW will typically be a node on both the AV and the Data Networks.
  • Add-on box 800 There are several approaches to implementing Add-on box 800 .
  • video camera 500 and microphone 600 capture and transmit outgoing video and audio signals into ports 801 and 802 , respectively, of Add-on box 800 .
  • These signals are transmitted via Audio/Video I/O port 805 across AV Network 901 .
  • Incoming video and audio signals (from another videoconference participant) are received across AV network 901 through Audio/Video I/O port 805 .
  • the video signals are sent out of V-OUT port 803 of CMW add-on box 800 to video input card 130 of base platform 100 , where they are displayed (typically in a separate video window) on monitor 200 utilizing the standard base platform video display card 120 .
  • the audio signals are sent out of A-OUT port 804 of CMW add-on box 800 and played through speaker 700 while the video signals are displayed on monitor 200 .
  • the same signal flow occurs for other non-teleconferencing applications of audio and video.
  • Add-on box 800 can be controlled by CMW software (illustrated in FIG. 20 ) executed by base platform 100 Control signals can be communicated between base platform port 104 and Add-on box Control port 806 (e.g., an RS-232, Centronics, SCSI or other standard communications port).
  • CMW software illustrated in FIG. 20
  • Control signals can be communicated between base platform port 104 and Add-on box Control port 806 (e.g., an RS-232, Centronics, SCSI or other standard communications port).
  • Add-on box 800 itself can be implemented as an add-in card to the base platform 100 Connections to the audio and video I/O devices need not change, though the connection for base platform control can be implemented internally (e.g., via the system bus) rather than through an external RS-232 or SCSI peripheral port.
  • Various additional levels of integration can also be achieved as will be evident to those skilled in the art.
  • microphones, speakers, video cameras and UTP transceivers can be integrated into the base platform 100 itself, and all media handling technology and communications can be integrated onto a single card.
  • a handset/headset jack enables the use of an integrated audio I/O device as an alternate to the separate microphone and speaker.
  • a telephone interface could be integrated into add-on box 800 as a local implementation of computer-integrated telephony.
  • a “hold” (i.e., audio and video mute) switch and/or a separate audio mute switch could be added to Add-on box 800 if such an implementation were deemed preferable to a software-based interface.
  • Video signals generated at the CMW are sent to CMW add-on box 800 via V-IN port 801 . They then typically pass unaffected through Loopback/AV Mute circuitry 830 via video ports 833 (input) and 834 (output) and into A/V Transceivers 840 (via Video In port 842 ) where they are transformed from standard video cable signals to UTP signals and sent out via port 845 and Audio/Video I/O port 805 onto AV Network 901 .
  • the Loopback/AV Mute circuitry 830 can, however, be placed in various modes under software control via Control port 806 (implemented, for example, as a standard UART). If in loopback mode (e.g., for testing incoming and outgoing signals at the CMW), the video signals would be routed back out V-OUT port 803 via video port 831 . If in a mute mode (e.g., muting audio, video or both), video signals might, for example, be disconnected and no video signal would be sent out video port 834 Loop back and muting switching functionality is also provided for audio in a similar way. Note: that computer control of loopback is very useful for remote testing and diagnostics while manual override of computer control on mute is effective for assured privacy from use of the workstation for electronic spying.
  • Video input (e.g., captured by the video camera at the CMW of another videoconference participant) is handled in a similar fashion. It is received along AV Network 901 through Audio/Video I/O port 805 and port 845 of A/V Transceivers 840 , where it is sent out Video Out port 841 to video port 832 of Loopback/AV Mute circuitry 830 , which typically passes such signals out video port 831 to V-OUT port 803 (for receipt by a video input card or other display mechanism, such as LCD display 810 of CMW Side Mount unit 850 in FIG. 18B , to be discussed).
  • a video input card or other display mechanism such as LCD display 810 of CMW Side Mount unit 850 in FIG. 18B , to be discussed.
  • Audio input and output passes through A/V transceivers 840 (via Audio In port 844 and Audio Out port 843 ) and Loopback/AV Mute circuitry 830 (through audio ports 837 / 838 and 836 / 835 ) in a similar manner.
  • the audio input and output ports of Add-on box 800 interface with standard amplifier and equalization circuitry, as well as an adaptive room echo canceler 814 to eliminate echo, minimize feedback and provide enhanced audio performance when using a separate microphone and speaker.
  • use of adaptive room echo cancelers provides high-quality audio interactions in wide area conferences.
  • adaptive room echo canceling requires training periods (typically involving an objectionable blast of high-amplitude white noise or tone sequences) for alignment with each acoustic environment, it is preferred that separate echo canceling be dedicated to each workstation rather than sharing a smaller group of echo cancelers across a larger group of workstations.
  • Audio inputs passing through audio port 835 of Loopback/AV Mute circuitry 830 provide audio signals to a speaker (via standard Echo Canceler circuitry 814 and A-OUT port 804 and/or to a handset or headset (via I/O ports 807 and 808 , respectively, under volume control circuitry 815 controlled by software through Control port 806 ). In all cases, incoming audio signals pass through power amplifier circuitry 812 before being sent out of Add-on box 80 ) to the appropriate audio-emitting transducer.
  • Outgoing audio signals generated at the CMW enter Add-on box 800 via A-IN port 802 (for a microphone) or Handset or Headset I/O ports 807 and 808 , respectively.
  • outgoing audio signals pass through standard preamplifier ( 811 ) and equalization ( 813 ) circuitry, whereupon the desired signal is selected by standard “Select” switching circuitry 816 (under software control through Control port 806 ) and passed to audio port 837 of Loopback/AV Mute circuitry, 830 .
  • A/V Transceivers 840 may include muxing/demuxing facilities so as to enable the transmission of audio/video signals on a single pair of wires, e.g., by encoding audio signals digitally in the vertical retrace interval of the analog video signal.
  • implementation of other audio and video enhancements such as stereo audio and external audio/video I/O ports (e.g., for recording signals generated at the CMW), are also well within the capabilities of one skilled in the art. If stereo audio is used in teleconferencing (i.e., to create useful spatial metaphors for users), a second echo canceller may be recommended.
  • FIG. 18B Another embodiment of the CMW of this invention, illustrated in FIG. 18B , utilizes a separate (fully self-contained) “Side Mount” approach which includes its own dedicated video display.
  • This embodiment is advantageous in a variety of situations, such as instances in which additional screen display area is desired (e.g., in a laptop computer or desktop system with a small monitor) or where it is impossible or undesirable to retrofit older, existing or specialized desktop computers for audio/video support.
  • video camera 500 , microphone 600 and speaker 700 of FIG. 18A are integrated together with the functionality of Add-on box 80 ).
  • Side Mount 850 eliminates the necessity of external connections to these integrated audio and video I/O devices, and includes an LCD display 810 for displaying the incoming video signal (which thus eliminates the need for a base platform video input card 130 ).
  • Side Mount unit 850 can become virtually a standalone device that does not require a separate computer for services using only audio and video. This also provides a way of supplementing a network of full-feature workstations with a few low-cost additional “audio video intercoms” for certain sectors of an enterprise (such as clerical, reception, factory floor, etc.).
  • a portable laptop implementation can be made to deliver multimedia mail with video, audio and synchronized annotations via CD-ROM or an add-on videotape unit with separate video, audio and time code tracks (a stereo videotape player can use the second audio channel for time code signals).
  • Videotapes or CD-ROMs can be created in main offices and express mailed, thus avoiding the need for high-bandwidth networking when on the road.
  • Cellular phone links can be used to obtain both voice and data communications (via modems) Modem-based data communications are sufficient to support remote control of mail or presentation playback, annotation, file transfer and fax features.
  • the laptop can then be brought into the office and attached to a docking station where the available MLAN 10 and additional functions adapted from Add-on box 800 can be supplied, providing full CMW capability.
  • CMW software modules 160 are illustrated generally in FIG. 20 and discussed in greater detail below in conjunction with the software running on MLAN Server 60 of FIG. 3
  • Software 160 allows the user to initiate and manage (in conjunction with the server software) videoconferencing, data conferencing, multimedia mail and other collaborative sessions with other users across the network.
  • standard multitasking operating system/GUI software 180 e.g., Apple Macintosh System 7, Microsoft Windows 3.1, or UNIX with the “X Window System” and Motif or other GUI “window manager” software
  • other applications 170 such as word processing and spreadsheet programs
  • Software modules 161 - 168 communicate with operating system/GUI software 180 and other applications 170 utilizing standard function calls and interapplication protocols.
  • the central component of the Collaborative Multimedia Workstation software is the Collaboration Initiator 161 . All collaborative functions can be accessed through this module
  • the Collaboration Initiator When the Collaboration Initiator is started, it exchanges initial configuration information with the Audio Video Network Manager (AVNM) 60 (shown in FIG. 3 ) through Data Network 902 .
  • AVNM Audio Video Network Manager
  • Information is also sent from the Collaboration Initiator to the AVNM indicating the location of the user, the types of services available on that workstation (e.g., videoconferencing, data conferencing, telephony, etc.) and other relevant initialization information.
  • the Collaboration Initiator presents a user interface that allows the user to initiate collaborative sessions (both real-time and asynchronous).
  • session participants can be selected from a graphical rolodex 163 that contains a scrollable list of user names or from a list of quick-dial buttons 162
  • Quick-dial buttons show the face icons for the users they represent.
  • the icon representing the user is retrieved by the Collaboration Initiator from the Directory Server 66 on MLAN Server 60 when it starts up Users can dynamically add new quick-dial buttons by dragging the corresponding entries from the graphical rolodex onto the quick-dial panel.
  • the user elects to initiate a collaborative session, he or she selects one or more desired participants by, for example, clicking on that name to select the desired participant from the system rolodex or a personal rolodex, or by clicking on the quick-dial button for that participant (see, e.g., FIG. 2A ).
  • the user selects the desired session type—e.g., by clicking on a CALL button to initiate a videoconference call, a SHARE button to initiate the sharing of a snapshot image or blank whiteboard, or a MAIL button to send mail.
  • the user can double-click on the rolodex name or a face icon to initiate the default session type—e.g., an audio/video conference call.
  • the system also allows sessions to be invoked from the keyboard. It provides a graphical editor to bind combinations of participants and session types to certain hot keys. Pressing this hot key (possibly in conjunction with a modifier key, e.g., ⁇ Shift> or ⁇ Ctrl>) will cause the Collaboration Initiator to start a session of the specified type with the given participants.
  • a modifier key e.g., ⁇ Shift> or ⁇ Ctrl>
  • Collaboration Initiator module 161 retrieves necessary addressing information from Directory Service 66 (see FIG. 21 ). In the case of a videoconference call, the Collaboration Initiator then communicates with the AVNM (as described in greater detail below) to set up the necessary data structures and manage the various states of that call, and to control A/V Switching Circuitry 30 , which selects the appropriate audio and video signals to be transmitted to/from each participant's CMW.
  • AVNM as described in greater detail below
  • the Collaboration Initiator locates, via the AVNM, the Collaboration Initiator modules at the CMWs of the chosen recipients, and sends a message causing the Collaboration Initiator modules to invoke the Snapshot Sharing modules 164 at each participant's CMW. Subsequent videoconferencing and data conferencing functionality is discussed in greater detail below in the context of particular usage scenarios.
  • Additional collaborative services such as Mail 165 , Application Sharing 166 , Computer-Integrated Telephony 167 and Computer Integrated Fax 168 —are also available from the CMW by utilizing Collaboration Initiator module 161 to initiate the session (i.e., to contact the participants) and to invoke the appropriate application necessary to manage the collaborative session
  • Collaboration Initiator contacts Directory Service 66 for address information (e.g., EMAIL address, fax number, etc.) for the selected participants and invokes the appropriate collaboration tools with the obtained address information.
  • the Collaboration Initiator queries the Service Server module 69 inside AVNM 63 for the current location of the specified participants.
  • this location information uses this location information to coordinate session setup.
  • the various Collaboration Initiators will invoke modules 166 , 167 or 168 (including activating any necessary devices such as the connection between the telephone and the CMW's audio I/O port). Further details on multimedia mail are provided below.
  • FIG. 21 diagrammatically illustrates software 62 comprised of various modules (as discussed above) provided for running on MLAN Server 60 ( FIG. 3 ) in the preferred embodiment. It is to be understood that additional software modules could also be provided. It is also to be understood that, although the software illustrated in FIG. 21 offers various significant advantages, as will become evident hereinafter, different forms and arrangements of software may also be employed within the scope of the invention. The software can also be implemented in various sub-parts running as separate processes.
  • clients e.g., software-controlling workstations, VCRs, laserdisks, multimedia resources, etc.
  • the MLAN Server Software Modules 62 communicate with the MLAN Server Software Modules 62 using the TCP/IP network protocols.
  • the AVNM 63 cooperates with the Service Server 69 , Conference Bridge Manager (CBM 64 in FIG. 21 ) and the WAN Network Manager (WNM 65 in FIG. 21 ) to manage communications within and among both MLANs 10 and WANs 15 ( FIGS. 1 and 3 ).
  • the AVNM additionally cooperates with Audio/Video Storage Server 67 and other multimedia services 68 in FIG. 21 to support various types of collaborative interactions as described herein
  • CBM 64 in FIG. 21 operates as a client of the AVNM 63 to manage conferencing by controlling the operation of conference bridges 35 .
  • WNM 65 manages the allocation of paths (codecs and trunks) provided by WAN gateway 40 for accomplishing the communications to other sites called for by the AVNM.
  • the AVNM 63 manages A/V Switching Circuitry 30 in FIG. 3 for selectively routing audio/video signals to and from CMWs 12 , and also to and from WAN gateway 40 , as called for by clients
  • Audio/video devices e.g., CMWs 12 , conference bridges 35 , multimedia resources 16 and WAN gateway 40 in FIG. 3
  • the AVNM combines these four connections into a port abstraction, wherein each port represents an addressable bidirectional audio/video channel.
  • Each device connected to the network has at least one port Different ports may share the same physical connections on the switch.
  • a conference bridge may typically have four ports (for 2 ⁇ 2 mosaicing) that share the same video-out connection. Not all devices need both video and audio connections at a port.
  • a TV tuner port needs only incoming audio/video connections.
  • the AVNM provides connectivity between audio/video devices by connecting their ports Connecting ports is achieved by switching one port's physical input connections to the other port's physical output connections (for both audio and video) and vice-versa
  • Client programs can specify which of the 4 physical connections on its ports should be switched. This allows client programs to establish unidirectional calls (e.g., by specifying that only the port's input connections should be switched and not the port's output connections) and audio-only or video-only calls (by specifying audio connections only or video connections only).
  • Service Server 69 Before client programs can access audio/video resources through the AVNM, they must register the collaborative services they provide with the Service Server 69 . Examples of these services indicate “video call”, “snapshot sharing”, “conference” and “video file sharing.” These service records are entered into the Service Server's service database. The service database thus keeps track of the location of client programs and the types of collaborative sessions in which they can participate. This allows the Collaboration Initiator to find collaboration participants no matter where they are located. The service database is replicated by all Service Servers: Service Servers communicate with other Service Servers in other MLANs throughout the system to exchange their service records.
  • Clients may create a plurality of services, depending on the collaborative capabilities desired.
  • a client can specify the network resources (e.g ports) that will be used by this service.
  • service information is used to associate a user with the audio/video ports physically connected to the particular CMW into which the user is logged in Clients that want to receive requests do so by putting their services in listening mode. If clients want to accept incoming data shares, but want to block incoming video calls, they must create different services.
  • a client can create an exclusive service on a set of ports to prevent other clients from creating services on these ports. This is useful, for example, to prevent multiple conference bridges from managing the same set of conference bridge ports.
  • the AVNM 63 ( FIG. 21 ), in cooperation with the Service Server 69 , CBM 64 and participating CMWs provide for managing.
  • the participating CMWs may include workstations located at both local and remote sites.
  • a CMW includes a Collaboration Initiator software module 161 (see FIG. 20 ) which is used to establish person-to-person and multiparty calls.
  • the corresponding collaboration initiator window advantageously provides quick-dial face icons of frequently dialed persons, as illustrated, for example, in FIG. 22 , which is an enlarged view of typical face icons along with various initiating buttons (described in greater detail below in connection with FIGS. 35-42 ).
  • Videoconference calls can be initiated, for example, merely by double-clicking on these icons.
  • the CMW typically provides a screen display that includes a live video picture of the remote conference participant, as illustrated for example in FIG. 8A .
  • this display also includes control buttons/menu items that can be used to place the remote participant on hold, to resume a call on hold, to add one or more participants to the call, to initiate data sharing and to hang up the call.
  • FIG. 23 The basic underlying software-controlled operations occurring for a two-party call are diagrammatically illustrated in FIG. 23
  • a caller initiates a call (e.g., by selecting a user from the graphical rolodex and clicking the call button or by double-clicking the face icon of the callee on the quick-dial panel)
  • the caller's Collaboration Initiator responds by identifying the selected user and requesting that user's address from Directory Service 66 , as indicated by ( 2 ) in FIG. 23
  • Directory Service 66 looks up the callee's address in the directory database, as indicated by ( 3 ) in FIG. 23 , and then returns it to the caller's Collaboration Initiator, as illustrated by ( 4 ) in FIG. 23 .
  • the caller's Collaboration Initiator sends a request to the AVNM to place a video call to caller with the specified address, as indicated by ( 5 ) in FIG. 23 .
  • the AVNM queries the Service Server to find the service instance of type “video call” whose name corresponds to the callee's address. This service record identifies the location of the callee's Collaboration Initiator as well as the network ports that the callee is connected to. If no service instance is found for the callee, the AVNM notifies the caller that the callee is not logged in. If the callee is local, the AVNM sends a call event to the callee's Collaboration Initiator, as indicated by ( 6 ) in FIG. 23 .
  • the AVNM forwards the call request ( 5 ) through the WAN gateway 40 for transmission, via WAN 15 ( FIG. 1 ) to the Collaboration Initiator of the callee's CMW at the remote site.
  • the callee's Collaboration Initiator can respond to the call event in a variety of ways.
  • a user-selectable sound is generated to announce the incoming call.
  • the Collaboration Initiator can then act in one of two modes.
  • “Telephone Mode” the Collaboration Initiator displays an invitation message on the CMW screen that contains the name of the caller and buttons to accept or refuse the call.
  • the Collaboration Initiator will then accept or refuse the call, depending on which button is pressed by the callee
  • “Intercom Mode” the Collaboration Initiator accepts all incoming calls automatically, unless there is already another call active on the callee's CMW, in which case behavior reverts to Telephone Mode.
  • the callee's Collaboration Initiator then notifies the AVNM as to whether the call will be accepted or refused. If the call is accepted, ( 7 ), the AVNM sets up the necessary communication paths between the caller and the callee required to establish the call. The AVNM then notifies the caller's Collaboration Initiator that the call has been established by sending it an accept event ( 8 ). If the caller and callee are at different sites, their AVNMs will coordinate in setting up the communication paths at both sites, as required by the call.
  • the AVNM may provide for managing connections among CMWs and other multimedia resources for audio/video/data communications in various ways. The manner employed in the preferred embodiment will next be described.
  • the AVNM manages the switches in the A/V Switching Circuitry 30 in FIG. 3 to provide port-to-port connections in response to connection requests from clients.
  • the primary data structure used by the AVNM for managing these connections will be referred to as a callhandle, which is comprised of a plurality of bits, including state bits.
  • Each port-to-port connection managed by the AVNM comprises two callhandles, one associated with each end of the connection.
  • the callhandle at the client port of the connection permits the client to manage the client's end of the connection.
  • the callhandle mode bits determine the current state of the callhandle and which of a port's four switch connections (video in, video out, audio in, audio out) are involved in a call.
  • AVNM clients send call requests to the AVNM whenever they want to initiate a call.
  • the client specifies the local service in which the call will be involved, the name of the specific port to use for the call, identifying information as to the callee, and the call mode.
  • the AVNM creates a callhandle on the caller's port.
  • All callhandles are created in the “idle” state.
  • the AVNM then puts the caller's callhandle in the “active” state.
  • the AVNM next creates a callhandle for the callee and sends it a call event, which places the callee's callhandle in the “ringing” state.
  • the callee accepts the call, its callhandle is placed in the “active” state, which results in a physical connection between the caller and the callee.
  • Each port can have an arbitrary number of callhandles bound to it, but typically only one of these callhandles can be active at the same time.
  • AVNM clients can send requests to the AVNM to change the state of the call, which can advantageously be accomplished by controlling the callhandle states. For example, during a call, a call request from another party could arrive. This arrival could be signaled to the user by providing an alert indication in a dialogue box on the user's CMW screen. The user could refuse the call by clicking on a refuse button in the dialogue box, or by clicking on a “hold” button on the active call window to put the current call on hold and allow the incoming call to be accepted.
  • the placing of the currently active call on hold can advantageously be accomplished by changing the caller's callhandle from the active state to a “hold” state, which permits the caller to answer incoming calls or initiate new calls, without releasing the previous call. Since the connection set-up to the callee will be retained, a call on hold can conveniently be resumed by the caller clicking on a resume button on the active call window, which returns the corresponding callhandle back to the active state. Typically, multiple calls can be put on hold in this manner.
  • the CMW advantageously provides a hold list display, identifying these on-hold calls and (optionally) the length of time that each party is on hold. A corresponding face icon could be used to identify each on-hold call.
  • buttons could be provided in this hold display which would allow the user to send a preprogrammed message to a party on hold. For example, this message could advise the callee when the call will be resumed, or could state that the call is being terminated and will be reinitiated at a later time.
  • FIG. 24 diagrammatically illustrates how two-party calls are connected for CMWs WS- 1 and WS- 2 , located at the same MLAN 10 .
  • CMWs WS- 1 and WS- 2 are coupled to the local A/V Switching Circuitry 30 via ports 81 and 82 , respectively.
  • a callhandle is created for each port. If CMW WS- 2 accepts the call, these two callhandles become active and in response thereto, the AVNM causes the A/V Switching Circuitry 30 to set up the appropriate connections between ports 81 and 82 , as indicated by the dashed line 83 .
  • FIG. 25 diagrammatically illustrates how two-party calls are connected for CMWs WS- 1 and WS- 2 when located in different MLANs 10 a and 10 b .
  • CMW WS- 1 of MLAN 10 a is connected to a port 91 a of A/V Switching Circuitry 30 a of MLAN 10 a
  • CMW WS- 2 is connected to a port 91 b of the A/V Switching Circuitry 30 b of MLAN 10 b
  • MLANs 10 a and 10 b can communicate with each other via ports 92 a and 92 b (through respective WAN gateways 40 a and 40 b and WAN 15 ).
  • a call between CMWs WS- 1 and WS- 2 can then be established by AVNM of MLAN 10 a in response to the creation of callhandles at ports 91 a and 92 a , setting up appropriate connections between these ports as indicated by dashed line 93 a , and by AVNM of. MLAN 10 b , in response to callhandles created at ports 91 b and 92 b , setting up appropriate connections between these ports as indicated by dashed line 93 b .
  • Appropriate paths 94 a and 94 b in WAN gateways 40 a and 40 b are set up by the WAN network manager 65 ( FIG. 21 ) in each network.
  • the CMW provides a screen that is similar to the screen for two-party calls, which displays a live video picture of the callee's image in a video window.
  • the screen includes a video mosaic containing a live video picture of each of the conference participants (including the CMW user's own picture), as shown, for example, in FIG. 8B .
  • FIG. 8B Of course, other embodiments could show only the remote conference participants (and not the local CMW user) in the conference mosaic (or show a mosaic containing both participants in a two-party call).
  • FIG. 8B the controls shown in FIG.
  • the multi-party conference screen also includes buttons/menu items that can be used to place individual conference participants on hold, to remove individual participants from the conference, to adjourn the entire conference, or to provide a “close-up” image of a single individual (in place of the video mosaic).
  • Multi-party conferencing requires all the mechanisms employed for 2-party calls. In addition, it requires the conference bridge manager CBM 64 ( FIG. 21 ) and the conference bridges 36 ( FIG. 3 ).
  • the CBM acts as a client of the AVNM in managing the operation of the conference bridges 36 .
  • the CBM also acts a server to other clients on the network.
  • the CBM makes conferencing services available by creating service records of type “conference” in the AVNM service database and associating these services with the ports on A/V Switching Circuitry 30 for connection to conference bridges 36 .
  • the preferred embodiment provides two ways for initiating a conference call.
  • the first way is to add one or more parties to an existing two-party call.
  • an ADD button is provided by both the Collaboration Initiator and the Rolodex, as illustrated in FIGS. 2A and 22 .
  • a user selects the party to be added (by clicking on the user's rolodex name or face icon as described above) and clicks on the ADD button to invite that new party. Additional parties can be invited in a similar manner.
  • the second way to initiate a conference call is to select the parties in a similar manner and then click on the CALL button (also provided in the Collaboration Initiator and Rolodex windows on the user's CMW screen).
  • Another alternative embodiment is to initiate a conference call from the beginning by clicking on a CONFERENCE/MOSAIC icon/button/menu item on the CMW screen.
  • This could initiate a conference call with the call initiator as the sole participant (i.e., causing a conference bridge to be allocated such that the caller's image also appears on his/her own screen in a video mosaic, which will also include images of subsequently added participants).
  • New participants could be invited, for example, by selecting each new party's face icon and then clicking on the ADD button.
  • Each conference uses four bridge ports 136 - 1 , 136 - 2 , 136 - 3 and 136 - 4 provided on A/V Switching Circuitry 30 a , which are respectively coupled to bidirectional audio/video lines 36 - 1 , 36 - 2 , 36 - 3 and 36 - 4 connected to conference bridge 36 .
  • A/V Switching Circuitry 30 a which are respectively coupled to bidirectional audio/video lines 36 - 1 , 36 - 2 , 36 - 3 and 36 - 4 connected to conference bridge 36 .
  • a conference call may be provided for additional parties, as well as simultaneously occurring conference calls.
  • the Collaboration Initiator determines that a conference is to be initiated, it queries the AVNM for a conference service. If such a service is available, the Collaboration Initiator requests the associated CBM to allocate a conference bridge. The Collaboration Initiator then places an audio/video call to the CBM to initiate the conference When the CBM accepts the call, the AVNM couples port 101 of CMW WS- 1 to lines 36 - 1 of conference bridge 36 by a connection 137 produced in response to callhandles created for port 101 of WS- 1 and bridge port 136 - 1 .
  • the Collaboration initiator on WS- 1 sends an add request to the CBM.
  • the CBM calls WS- 3 via WS- 3 port 103
  • the AVNM creates callhandles for WS- 3 port 103 and bridge port 136 - 2
  • WS- 3 accepts the call, its callhandle is made “active,” resulting in connection 138 being provided to connect WS- 3 and lines 136 - 2 of conference bridge 36 .
  • CMW WS- 1 next adds CMW WS- 5 and then CMW WS- 8 , callhandles for their respective ports and bridge ports 136 - 3 and 136 - 4 are created, in turn, as described above for WS- 1 and WS- 3 , resulting in connections 139 and 140 being provided to connect WS- 5 and WS- 9 to conference bridge lines 36 - 3 and 36 - 4 , respectively.
  • the conferees WS- 1 , WS- 3 , WS- 5 and WS- 8 are thus coupled to conference bridge lines 136 - 1 , 136 - 2 , 136 - 3 and 136 - 4 , respectively as shown in FIG. 26 .
  • the video mosaicing circuitry 36 and audio mixing circuitry 38 incorporated in conference bridge 36 operate as previously described, to form a resulting four-picture mosaic ( FIG. 8B ) that is sent to all of the conference participants, which in this example are CMWs WS- 1 , WS- 2 , WS- 5 and WS- 8 Users may leave a conference by just hanging up, which causes the AVNM to delete the associated callhandles and to send a hangup notification to CBM When CBM receives the notification, it notifies all other conference participants that the participant has exited. In the preferred embodiment, this results in a blackened portion of that participant's video mosaic image being displayed on the screen of all remaining participants.
  • each incoming single video picture or mosaic from another site is connected to a respective one of the conference bridge lines 36 - 1 to 36 - 4 via WAN gateway 40 .
  • Additional conference participants can then be added as described above for the situations described above in which the conference is initiated by the user of WS- 1 either selecting multiple participants initially or merely selecting a “conference” and then adding subsequent participants.
  • Data conferencing is implemented in the preferred embodiment by certain Snapshot Sharing software provided at the CMW (see FIG. 20 ).
  • This software permits a “snapshot” of a selected portion of a participant's CMW screen (such as a window) to be displayed on the CMW screens of other selected participants (whether or not those participants are also involved in a videoconference). Any number of snapshots may be shared simultaneously. Once displayed, any participant can then telepoint on or annotate the snapshot, which animated actions and results will appear (virtually simultaneously) on the screens of all other participants.
  • the annotation capabilities provided include lines of several different widths and text of several different sizes. Also, to facilitate participant identification, these annotations may be provided in a different color for each participant. Any annotation may also be erased by any participant FIG. 2B (lower left window) illustrates a CMW screen having a shared graph on which participants have drawn and typed to call attention to or supplement specific portions of the shared image.
  • a participant may initiate data conferencing with selected participants (selected and added as described above for videoconference calls) by clicking on a SHARE button on the screen (available in the Rolodex or Collaboration Initiator windows, shown in FIG. 2A , as are CALL and ADD buttons), followed by selection of the window to be shared
  • a participant clicks on his SHARE button his Collaboration Initiator module 161 ( FIG. 20 ) queries the AVNM to locate the Collaboration Initiators of the selected participants, resulting in invocation of their respective Snapshot Sharing modules 164 .
  • the Snapshot Sharing software modules at the CMWs of each of the selected participants query their local operating system 180 to determine available graphic formats, and then send this information to the initiating Snapshot Sharing module, which determines the format that will produce the most advantageous display quality and performance for each selected participant
  • each participant may telepoint on or annotate the snapshot, which actions and results are displayed on the CMW screens of all participants. This is preferably accomplished by monitoring the actions made at the CMW (e.g., by tracking mouse movements) and sending these “operating system commands” to the CMWs of the other participants, rather than continuously exchanging bitmaps, as would be the case with traditional “remote control” products.
  • the original unchanged snapshot is stored in a first bitmap 210 a .
  • a second bitmap 210 b stores the combination of the original snapshot and any annotations.
  • the original unchanged snapshot can be restored (i.e., erasing all annotations) using bitmap 210 a
  • Selective erasures can be accomplished by copying into (i.e., restoring) the desired erased area of bitmap 210 b with the corresponding portion from bitmap 210 a.
  • the user might display the first page of a spreadsheet on his screen, use the SHARE button to share that page, discuss and perhaps annotate it, then return to the spreadsheet application to position to the next page, use the REGRAB button to share the new page, and so on.
  • This mechanism represents a simple, effective step toward application sharing.
  • a user may instead choose to share a snapshot that had previously been saved as a file. This is achieved via the LOAD button, which causes a dialogue box to appear, prompting the user to select a file Conversely, via the SAVE button, any snapshot may be saved, with all current annotations.
  • FIG. 29 is a schematic and diagrammatic view illustrating how multimedia calls/conferences, MMCR, MMM and MMDM work together to provide the above-described features.
  • MM Editing Utilities shown supplementing MMM and MMDM may be identical.
  • the creation, storage, retrieval and editing of multimedia documents serve as the basic element common to MMCR, MMM and MMDM. Accordingly, the preferred embodiment advantageously provides a universal format for multimedia documents.
  • This format defines multimedia documents as a collection of individual components in multiple media combined with an overall structure and timing component that captures the identities, detailed dependencies, references to, and relationships among the various other components.
  • the information provided by this structuring component forms the basis for spatial layout, order of presentation, hyperlinks, temporal synchronization, etc., with respect to the composition of a multimedia document
  • FIG. 30 shows the structure of such documents as well as their relationship with editing and storage facilities.
  • multimedia documents are advantageously structured for authoring, storage, playback and editing by storing some data under conventional file systems and some data in special-purpose storage servers as will be discussed later.
  • the Conventional File System 504 can be used to store all non-time-sensitive portions of a multimedia document.
  • non-time-sensitive data that can be stored in a conventional type of computer file system:
  • time-sensitive media and media that have synchronization requirements with other media components.
  • Some of these time-sensitive media can be stored on conventional file systems while others may require special-purpose storage facilities.
  • time-sensitive media examples are small audio files and short or low-quality video clips (e.g as might be produced using QuickTime or Video for Windows).
  • Other examples include window event lists as supported by the Window-Event Record and Play system 512 shown in FIG. 30 .
  • This component allows for storing and replaying a user's interactions with application programs by capturing the requests and events exchanged between the client program and the window system in a time-stamped sequence. After this “record” phase, the resulting information is stored in a conventional file that can later be retrieved and “played” back. During playback the same sequence of window system requests and events reoccurs with the same relative timing as when they were recorded. In prior-art systems, this capability has been used for creating automated demonstrations. In the present invention it can be used, for example, to reproduce annotated snapshots as they occurred at recording
  • Snapshot Share 514 shown in FIG. 30 is a utility used in multimedia calls and conferencing for capturing window or screen snapshots, sharing with one or more call or conference participants, and permitting group annotation, telepointing, and re-grabs.
  • this utility is adapted so that its captured images and window events can be recorded by the Window-Event Record and Play system 512 while being used by only one person.
  • By synchronizing events associated with a video or audio stream to specific frame numbers or time codes, a multimedia call or conference can be recorded and reproduced in its entirety.
  • the same functionality is preferably used to create multimedia mail whose authoring steps are virtually identical to participating in a multimedia call or conference (though other forms of MMM are not precluded).
  • Some time-sensitive media require dedicated storage servers in order to satisfy real-time requirements.
  • High-quality audio/video segments for example, require dedicated real-time audio/video storage servers.
  • a preferred embodiment of such a server will be described later. Next to be considered is how the current invention guarantees synchronization between different media components.
  • Audio or video segments can exist without being accompanied by the other. If audio and video are recorded simultaneously (“co-recorded”), the preferred embodiment allows the case where their streams are recorded and played back with automatic synchronization—as would result from conventional VCRs, laserdisks, or time-division multiplexed (“interleaved”) audio/video streams. This excludes the need to tightly synchronize (i.e., “lip-sync”) separate audio and video sequences. Rather, reliance is on the co-recording capability of the Real-Time Audio/Video Storage Server 502 to deliver all closely synchronized audio and video directly at its signal outputs.
  • Each recorded video sequence is tagged with time codes (e.g. SMPTE at 1/30 second intervals) or video frame numbers.
  • Each recorded audio sequence is tagged with time codes (e.g., SMPTE or MIDI) or, if co-recorded with video, video frame numbers.
  • the preferred embodiment also provides synchronization between window events and audio and/or video streams. The following functions are supported:
  • Media-time-driven Synchronization synchronization of window events to an audio, video, or audio/video stream, using the real-time media as the timing source.
  • machine-time-driven synchronization is used throughout the document Whenever audio and/or video is playing, media-time-synchronization is used.
  • the system supports transition between machine-time and media-time synchronization whenever an audio/video segment is started or stopped.
  • viewing a multimedia document might proceed as follows:
  • the present invention can include many special-purpose servers that provide storage of time-sensitive media (e.g. audio/video streams) and support coordination with other media.
  • time-sensitive media e.g. audio/video streams
  • This section describes the preferred embodiment for audio/video storage and recording services.
  • a centralized server 502 coupled to MLAN 10 , as illustrated in FIG. 31 .
  • a centralized server 502 provides the following advantages:
  • the total amount of storage hardware required can be far less (due to better utilization resulting from statistical averaging).
  • File system administration e.g. backups and file system replication, etc.
  • File system replication are far less costly and higher performance.
  • the Real-Time Audio/Video Storage Server 502 shown in FIG. 31A structures and manages the audio/video files recorded and stored on its storage devices
  • Storage devices may typically include computer-controlled VCRs, as well as rewritable magnetic or optical disks.
  • server 502 in FIG. 31 A includes disks 60 e for recording and playback. Analog information is transferred between disks 60 e and the.
  • A/V Switching Circuitry 30 via analog I/O 62 Control is provided by control 64 coupled to Data LAN hub 25 .
  • the centralized audio/video storage and playback server 502 in FIG. 31A performs the following functions:
  • the storage server needs to provide a real-time transmission path between the storage medium and the appropriate audio/video network port for each simultaneous client accessing the server. For example, if one user is viewing a video file at the same time several other people are creating and storing new video files on the same disk, multiple simultaneous paths to the storage media are required. Similarly, video mail sent to large distribution groups, video databases, and similar functions may also require simultaneous access to the same video files, again imposing multiple access requirements on the video storage capabilities.
  • VCRs For storage servers that are based on computer-controlled VCRs or rewritable laserdisks, a real-time transmission path is readily available through the direct analog connection between the disk or tape and the network port.
  • each VCR or laserdisk can only be accessed by one client program at the same time (multi-head laserdisks are an exception). Therefore, storage servers based on VCRs and laserdisks are difficult to scale for multiple access usage.
  • multiple access to the same material is provided by fie replication and staging, which greatly increases storage requirements and the need for moving information quickly among storage media units serving different users.
  • Video systems based on magnetic disks are more readily scalable for simultaneous use by multiple people.
  • a generalized hardware implementation of such a scalable storage and playback system- 502 is illustrated in FIG. 32 .
  • Individual I/O cards 530 supporting digital and analog I/O are linked by intra-chassis digital networking (e.g. buses) for fie transfer within chassis 532 holding some number of these cards
  • Multiple chassis 532 are linked by inter-chassis networking.
  • the Digital Video Storage System available from Parallax Graphics is an example of such a system implementation.
  • the bandwidth available for the transfer of files among disks is ultimately limited by the bandwidth of these intra-chassis and inter-chassis networking. or systems that use sufficiently powerful video compression schemes, real-time delivery requirements for a small number of users can be met by existing file system software (such as the Unix file system), provided that the block-size of the storage system is optimized for video storage and that sufficient buffering is provided by the operating system software to guarantee continuous flow of the audio/video data.
  • existing file system software such as the Unix file system
  • FIG. 33 which uses crosspoint switching, such as provided by SCSI Crossbar 540 , which increases the total bandwidth of the inter-chassis and intra-chassis network, thereby increasing the number of possible simultaneous file transfers.
  • the audio/video storage system can leverage the previously described network facilities: the MLANs 10 can be used to establish a multimedia network connection between client workstations and the audio/video storage servers Audio/Video editors and viewers running on the client workstation use the same software interfaces as the multimedia teleconferencing system to establish these network connections.
  • Client workstations use the existing audio/video network to connect to the storage server's network ports. These network ports are connected to compression/decompression engines that plug into the server bus. These engines compress the audio/video streams that come in over the network and store them on the local disk. Similarly, for playback, the server reads stored video segments from its local disk and routes them through the decompression engines back to client workstations for local display.
  • the present invention allows for alternative delivery strategies. or example, some compression algorithms are asymmetric, meaning that decompression requires much less compute power than compression. In some cases, real-time decompression can even be done in software, without requiring any special-purpose decompression hardware. As a result, there is no need to decompress stored audio and video on the storage server and play it back in realtime over the network. Instead, it can be more efficient to transfer an entire audio/video file from the storage server to the client workstation, cache it on the workstation's disk, and play it back locally.
  • Multimedia conference recording will next be considered.
  • recording (storage) capabilities are preferably provided for audio and video of all parties, and also for all shared windows, including any telepointing and annotations provided during the teleconference.
  • these capabilities are provided in a way such that they can be replayed with accurate correspondence in time to the recorded audio and video, such as by synchronizing to frame numbers or time code events.
  • a preferred way of capturing audio and video from calls would be to record all calls and conferences as if they were multi-party conferences (even for two-party calls), using video mosaicing, audio mixing and cut-and-pasting, as previously described in connection with FIGS. 7-11 .
  • MMCR as described will advantageously permit users at their desktop to review real-time collaboration as it previously occurred, including during a later teleconference.
  • the output of a MMCR session is a multimedia document that can be stored, viewed, and edited using the multimedia document facilities described earlier.
  • FIG. 31D shows how conference recording relates to the various system components described earlier.
  • the Multimedia Conference Record/Play system 522 provides the user with the additional GUIs (graphical user interfaces) and other functions required to provide the previously described MMCR functionality.
  • the Conference Invoker 518 shown in FIG. 31D is a utility that coordinates the audio/video calls that must be made to connect the audio/video storage server 502 with special recording outputs on conference bridge hardware ( 35 in FIG. 3 ). The resulting recording is linked to information identifying the conference, a function also performed by this utility.
  • MMM multimedia mail
  • the captured information is preferably a superset of that captured during MMCR, except that no other user is involved and the user is given a chance to review and edit before sending the message.
  • the Multimedia Mail system 524 in FIG. 31D provides the user with the additional GUIs and other functions required to provide the previously described MMM functionality
  • Multimedia Mail relies on a conventional Email system 506 shown in FIG. 31D for creating, transporting, and browsing messages.
  • multimedia document editors and viewers are used for creating and viewing message bodies.
  • Multimedia documents (as described above) consist of time-insensitive components and time sensitive components.
  • the Conventional Email system 506 relies on the Conventional File system 504 and Real-Time Audio/Video Storage Server 502 for storage support.
  • the time insensitive components are transported within the Conventional Email system 506 , while the real-time components may be separately transported through the audio/video network using file transfer utilities associated with the Real-Time Audio/Video Storage Server 502 .
  • MMDM Multimedia document management
  • Multimedia documents can be authored as mail in the MMM system or as call/conference recordings in the MMCR system and then passed on to the MMDM system.
  • multimedia documents can also be authored by means other than MMM and MMCR.
  • Multimedia documents stored within the MMDM system can be reviewed and searched.
  • Multimedia documents stored within the MMDM system can be used as material in the creation of subsequent MMM.
  • Multimedia documents stored within the MMDM system can be edited to create other multimedia documents.
  • the Multimedia Document Management system 526 in FIG. 31D provides the user with the additional GUIs and other functions required to provide the previously described MMDM functionality.
  • the MMDM includes sophisticated searching and editing capabilities in connection with the MMDM multimedia document such that a user can rapidly access desired selected portions of a stored multimedia document.
  • the Specialized Search system 520 in FIG. 31D comprises utilities that allow users to do more sophisticated searches across and within multimedia documents. This includes context-based and content-based searches (employing operations such as speech and image recognition, information filters, etc.), time-based searches, and event-based searches (window events, call management events, speech/audio events, etc.).
  • FIG. 34 The resulting multimedia collaboration environment achieved by the above-described integration of audio/video/data teleconferencing, MMCR, MMM and MMDM is illustrated in FIG. 34 . It will be evident that each user can collaborate with other users in real-time despite separations in space and time. In addition, collaborating users can access information already available within their computing and information systems, including information captured from previous collaborations. Note in FIG. 34 that space and time separations are supported in the following ways:
  • the present invention smoothly spans these three venus.
  • the focus of the scenario revolves around the activities of a trader who is a specialist in securities.
  • the setting is the start of his day at his desk in a major financial center (NYC) at a major U.S. investment bank.
  • NYC major financial center
  • the Expert has been actively watching a particular security over the past week and upon his arrival into the office, he notices it is on the rise. Before going home last night, he previously set up his system to filter overnight news on a particular family of securities and a security within that family. He scans the filtered news and sees a story that may have a long-term impact on this security in question. He believes he needs to act now in order to get a good price on the security. Also, through filtered mail, he sees that his counterpart in London, who has also been watching this security, is interested in getting our Expert's opinion once he arrives at work.
  • the Expert issues a multimedia mail message on the security to the head of sales worldwide for use in working with their client base. Also among the recipients is an analyst in the research department and his counterpart in London.
  • the Expert in preparation for his previously established “on-call” office hours, consults with others within the corporation (using the videoconferencing and other collaborative techniques described above), accesses company records from his CMW, and analyzes such information, employing software-assisted analytic techniques. His office hours are now at hand, so he enters “intercom” mode, which enables incoming calls to appear automatically (without requiring the Expert to “answer his phone” and elect to accept or reject the call).
  • the Expert's computer beeps, indicating an incoming call, and the image of a field representative 201 and his client 202 who are located at a bank branch somewhere in the U.S. appears in video window 203 of the Expert's screen (shown in FIG. 35 ). Note that, unless the call is convened to a “conference” call (whether explicitly via a menu selection or implicitly by calling two or more other participants or adding a third participant to a call), the callers will see only each other in the video window and will not see themselves as part of a video mosaic.
  • the Collaboration Initiator window 204 from which the Expert can (utilizing Collaboration Initiator software module 161 shown in FIG. 20 ) initiate and control various collaborative sessions.
  • the user can initiate with a selected participant a video call (CALL button) or the addition of that selected participant to an existing video call (ADD button), as well as a share session (SHARE button) using a selected window or region on the screen (or a blank region via the WHITEBOARD button for subsequent annotation).
  • the user can also invoke his MAIL software (MAIL button) and prepare outgoing or check incoming Email messages (the presence of which is indicated by a picture of an envelope in the dog's mouth in In Box icon 205 ), as well as check for “I called” messages from other callers (MESSAGES button) left via the LEAVE WORD button in video window 203
  • Video window 203 also contains buttons from which many of these and certain additional features can be invoked, such as hanging up a video call (HANGUP button), putting a call on hold (HOLD button), resuming a call previously put on hold (RESUME button) or muting the audio portion of a call (MUTE button).
  • HANGUP button hanging up a video call
  • HOLD button putting a call on hold
  • MUTE button muting the audio portion of a call
  • the user can invoke the recording of a conference by the conference RECORD button.
  • a standard desktop window 206 containing icons from which other programs (
  • the Expert is now engaged in a videoconference with field representative 201 and his client 202 .
  • the field representative shares with the Expert a graphical image 210 (pie chart of client portfolio holdings) of his client's portfolio holdings (by clicking on his SHARE button, corresponding to the SHARE button in video window 203 of the Expert's screen, and selecting that image from his screen, resulting in the shared image appearing in the Share window 211 of the screen of all participants to the share) and begins to discuss the client's investment dilemma.
  • the field representative also invokes a command to secretly bring up the client profile on the Expert's screen.
  • the Expert illustrates his advice by creating (using his own modeling software) and sharing a new graphical image 220 ( FIG. 37 ) with the field representative and his client.
  • Either party to the share can annotate that image using the drawing tools 221 (and the TEXT button, which permits typed characters to be displayed) provided within Share window 211 , or “regrab” a modified version of the original image (by using the REGRAB button), or remove all such annotations (by using the CLEAR button of Share window 211 ), or “grab” a new image to share (by clicking on the GRAB button of Share window 211 and selecting that new image from the screen).
  • any participant to a shared session can add a new participant by selecting that participant from the rolodex or quick-dial list (as described above for video calls and for data conferencing) and clicking the ADD button of Share window 211 .
  • field representative 201 makes annotations 222 to image 220 in order to illustrate his concerns. While responding to the concerns of field representative 201 , the Expert hears a beep and receives a visual notice (New Call window 223 ) on his screen (not visible to the field representative and his client), indicating the existence of a new incoming call and identifying the caller. At this point, the Expert can accept the new call (ACCEPT button), refuse the new call (REFUSE button, which will result in a message being displayed on the caller's screen indicating that the Expert is unavailable) or add the new caller to the Expert's existing call (ADD button).
  • ACCEPT button refuse the new call
  • REFUSE button which will result in a message being displayed on the caller's screen indicating that the Expert is unavailable
  • ADD button add the new caller to the Expert's existing call
  • the Expert elects yet another option (not shown)—to defer the call and leave the caller a standard message that the Expert will call back in X minutes (in this case 1 minute). The Expert then elects also to defer his existing call, telling the field representative and his client that he will call them back in 5 minutes, and then elects to return the initial deferred call.
  • the Expert's act of deferring a call results not only in a message being sent to the caller, but also in the caller's name (and perhaps other information associated with the call, such as the time the call was deferred or is to be resumed) being displayed in a list 230 (see FIG. 38 ) on the Expert's screen from which the call can be reinitiated.
  • the “state” of the call e.g., the information being shared
  • deferring a call actually breaks the logical and physical connections, requiring that the entire call be reinitiated by the Collaboration initiator and the AVNM as described above.
  • video window 203 now displays, as illustrated in FIG. 38 , a video mosaic consisting of three smaller images (instead of a single large image displaying only caller 231 ) of the Expert 233 , caller 231 and London expert 232 .
  • an urgent PRIORITY request (New Call window 234 ) is received from the Expert's boss (who is engaged in a three-party videoconference call with two members of the bank's operations department and is attempting to add the Expert to that call to answer a quick question).
  • the Expert puts his three-party videoconference on hold (merely by clicking the HOLD button in video window 203 ) and accepts (via the ACCEPT button of New Call window 234 ) the urgent call from his boss, which results in the Expert being added to the boss' three-party videoconference call.
  • video window 203 is now replaced with a four-person video mosaic representing a four-party conference call consisting of the Expert 233 , his boss 241 and the two members 242 and 243 of the bank's operations department.
  • the Expert quickly answers the boss' question and, by clicking on the RESUME button (of video window 203 ) adjacent to the names of the other participants to the call on hold, simultaneously hangs up on the conference call with his boss and resumes his three-party conference call involving the securities issue, as illustrated in video window 203 of FIG. 40 .
  • the Expert receives notification of an incoming multimedia mail message—e.g., a beep accompanied by the appearance of an envelope 252 in the dog's mouth in In Box icon 205 shown in FIG. 40 .
  • an incoming multimedia mail message e.g., a beep accompanied by the appearance of an envelope 252 in the dog's mouth in In Box icon 205 shown in FIG. 40 .
  • he quickly scans his incoming multimedia mail message by clicking on In Box icon 205 , which invokes his mail software, and then selecting the incoming message for a quick scan, as generally illustrated in the top two windows of FIG. 2B . He decides it can wait for further review as the sender is an analyst other than the one helping on his security question.
  • Call 261 is the Head of Sales for the Chicago office Working at home
  • Call 262 is an urgent international call
  • call indicator 262 The Expert notes from call indicator 262 that this call is not only an international call (shown in the top portion of the New Call window), but he realizes it is from a laptop user in the field in Central Mexico.
  • the Expert elects to prioritize his calls in the following manner: 262 , 261 and 263 . He therefore quickly answers call 261 (by clicking on its ACCEPT button) and puts that call on hold while deferring call 263 in the manner discussed above. He then proceeds to accept the call identified by international call indicator 262 .
  • FIG. 42 deferred call indicator 271 and the indicator for the call placed on hold (next to the highlighted RESUME button in video window 203 ), as well as the image of caller 272 from the laptop in the field in Central Mexico.
  • Mexican caller 272 is outdoors and has no direct access to any wired telephone connection, his laptop has two wireless modems permitting dial-up access to two data connections in the nearest field office (through which his calls were routed).
  • the system automatically (based upon the laptop's registered service capabilities) allocated one connection for an analog telephone voice call (using his laptop's built-in microphone and speaker and the Expert's computer-integrated telephony capabilities) to provide audio teleconferencing.
  • the other connection provides control, data conferencing and one-way digital video (i.e., the laptop user cannot see the image of the Expert) from the laptop's built-in camera, albeit at a very slow frame rate (e.g., 3-10 small frames per second) due to the relatively slow dial-up phone connection.
  • a very slow frame rate e.g., 3-10 small frames per second
  • the present invention accommodates such capabilities, supplementing an audio telephone connection with limited (i.e., relatively slow) one-way video and data conferencing functionality.
  • limited i.e., relatively slow
  • the present invention will accommodate such improvements automatically.
  • other participants need not be reduced to this “lowest common denominator.” For example, additional participants could be added to the call illustrated in FIG. 42 as described above, and such participants could have full videoconferencing, data conferencing and other collaborative functionality vis-a-vis one another, while having limited functionality only with caller 272 .
  • the salesperson 272 also needs the Expert s help on another issue. He has hard copy only of a client's portfolio and needs some advice on its composition before he meets with the client tomorrow. He says he will fax it to the Expert for analysis Upon receiving the fax—on his CMW, via computer-integrated fax—the Expert asks if he should either send the Mexican caller a “QuickTime” movie (a lower quality compressed video standard from Apple Computer) on his laptop tonight or send a higher-quality CD via FedX tomorrow—the notion being that the Expert can produce an actual video presentation with models and annotations in video form. The salesperson can then play it to his client tomorrow afternoon and it will be as if the Expert is in the room. The Mexican caller decides he would prefer the CD.
  • the Expert learns, in the course of his call with remote laptop caller 272 , that he missed an important issue during his previous quick scan of his incoming multimedia mail message.
  • the Expert is upset that the sender of the message did not utilize the “video highlight” feature to highlight this aspect of the message.
  • This feature permits the composer of the message to define “tags” (e.g., by clicking a TAG button, not shown) during record time which are stored with the message along with a “time stamp,” and which cause a predefined or selectable audio and/or visual indicator to be played/displayed at that precise point in the message during playback.
  • the Expert decides to merge the two calls together by adding the call on hold to his existing call.
  • both the Expert and the previously held caller will have full video capabilities vis-a-vis one another and will see a three-way mosaic image (with the image of caller 272 at a slower frame rate), whereas caller 272 will have access only to the audio portion of this three-way conference call, though he will have data conferencing functionality with both of the other participants.
  • the Expert forwards the multimedia mail message to both caller 272 and the other participant, and all three of them review the video enclosure in greater detail and discuss the concern raised by caller 272 . They share certain relevant data as described above and realize that they need to ask a quick question of another remote expert. They add that expert to the call (resulting in the addition of a fourth image to the video mosaic, also not shown) for less than a minute while they obtain a quick answer to their question. They then continue their three-way call until the Expert provides his advice and then adjourns the call.
  • the Expert composes a new multimedia mail message, recording his image and audio synchronized (as described above) to the screen displays resulting from his simultaneous interaction with his CMW (e.g., running a program that performs certain calculations and displays a graph while the Expert illustrates certain points by telepointing on the screen, during which time his image and spoken words are also captured). He sends this message to a number of salesforce recipients whose identities are determined automatically by an outgoing mail filter that utilizes a database of information on each potential recipient (e.g., selecting only those whose clients have investment policies which allow this type of investment).
  • CMW e.g., running a program that performs certain calculations and displays a graph while the Expert illustrates certain points by telepointing on the screen, during which time his image and spoken words are also captured. He sends this message to a number of salesforce recipients whose identities are determined automatically by an outgoing mail filter that utilizes a database of information on each potential recipient (e.g., selecting only those whose clients have investment policies which allow this type of investment
  • the Expert then receives an audio and visual reminder (not shown) that a particular video feed (e.g., a short segment of a financial cable television show featuring new financial instruments) will be triggered automatically in a few minutes. He uses this time to search his local securities database, which is dynamically updated from financial information feeds (e.g., prepared from a broadcast textual stream of current financial events with indexed headers that automatically applies data filters to select incoming events relating to certain securities). The video feed is then displayed on the Expert's screen and he watches this short video segment.
  • a particular video feed e.g., a short segment of a financial cable television show featuring new financial instruments
  • financial information feeds e.g., prepared from a broadcast textual stream of current financial events with indexed headers that automatically applies data filters to select incoming events relating to certain securities.
  • the Expert After analyzing this extremely up-to-date information, the Expert then reinitiates his previously deferred call, from indicator 271 shown in FIG. 42 , which he knows is from the Head of Sales in Los Angeles, who is seeking to provide his prime clients with securities advice on another securities transaction based upon the most recent available information.
  • the Expert's call is not answered directly, though he receives a short prerecorded video message (left by the caller who had to leave his home for a meeting across town soon after his priority message was deferred) asking that the Expert leave him a multimedia mail reply message with advice for a particular client, and explaining that he will access this message remotely from his laptop as soon as his meeting is concluded.
  • the Expert complies with this request and composes and sends this mail message.
  • the Expert then receives an audio and visual reminder on his screen indicating that his office hours will end in two minutes. He switches from “intercom” mode to “telephone” mode so that he will no longer be disturbed without an opportunity to reject incoming calls via the New Call window described above. He then receives and accepts a final call concerning an issue from an electronic meeting several months ago, which was recorded in its entirety.
  • the Expert accesses this recorded meeting from his “corporate memory.” He searches the recorded meeting (which appears in a second video window on his screen as would a live meeting, along with standard controls for stop/play/rewind/fast forward/etc.) for an event that will trigger his memory using his fast forward controls, but cannot locate the desired portion of the meeting. He then elects to search the ASCII text log (which was automatically extracted in the background after the meeting had been recorded, using the latest voice recognition techniques), but still cannot locate the desired portion of the meeting. Finally, he applies an information filter to perform a content-oriented (rather than literal) search and finds the portion of the meeting he was seeking. After quickly reviewing this short portion of the previously recorded meeting, the Expert responds to the caller's question, adjourns the call and concludes his office hours.
  • the above scenario involves many state-of-the-art desktop tools (e.g., video and information feeds, information filtering and voice recognition) that can be leveraged by our Expert during videoconferencing, data conferencing and other collaborative activities provided by the present invention—because this invention, instead of providing a dedicated videoconferencing system, provides a desktop multimedia collaboration system that integrates into the Expert's existing workstation/LAN/WAN environment.
  • desktop tools e.g., video and information feeds, information filtering and voice recognition

Landscapes

  • Engineering & Computer Science (AREA)
  • Business, Economics & Management (AREA)
  • Entrepreneurship & Innovation (AREA)
  • Strategic Management (AREA)
  • Theoretical Computer Science (AREA)
  • Human Resources & Organizations (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Economics (AREA)
  • Tourism & Hospitality (AREA)
  • Signal Processing (AREA)
  • Computer Hardware Design (AREA)
  • Marketing (AREA)
  • Operations Research (AREA)
  • Quality & Reliability (AREA)
  • Data Mining & Analysis (AREA)
  • General Business, Economics & Management (AREA)
  • Multimedia (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Software Systems (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
  • Telephonic Communication Services (AREA)
  • Data Exchanges In Wide-Area Networks (AREA)

Abstract

A multimedia collaboration system that integrates separate real-time and asynchronous networks—the former for real-time audio and video, and the latter for control signals and textual, graphical and other data—in a manner that is interoperable across different computer and network operating system platforms and which closely approximates the experience of face-to-face collaboration, while liberating the participants from the limitations of time and distance. These capabilities are achieved by exploiting a variety of hardware, software and networking technologies in a manner that preserves the quality and integrity of audio/video/data and other multimedia information, even after wide area transmission, and at a significantly reduced networking cost as compared to what would be required by presently known approaches. The system architecture is readily scalable to the largest enterprise network environments. It accommodates differing levels of collaborative capabilities available to individual users and permits high-quality audio and video capabilities to be readily superimposed onto existing personal computers and workstations and their interconnecting LANs and WANs. In a particular preferred embodiment, a plurality of geographically dispersed multimedia LANs are interconnected by a WAN. The demands made on the WAN are significantly reduced by employing multi-hopping techniques, including dynamically avoiding the unnecessary decompression of data at intermediate hops, and exploiting video mosaicing, cut-and-paste and audio mixing technologies so that significantly fewer wide area transmission paths are required while maintaining the high quality of the transmitted audio/video.

Description

    CROSS-REFERENCE TO RELATED APPLICATIONS
  • The present application is a Continuation of U.S. patent application Ser. No. 09/702,737 filed Nov. 1, 2000, which is a Continuation of U.S. patent application Ser. No. 08/994,848, filed Dec. 19, 1997, now U.S. Pat. No. 6,237,025, which is a continuation of U.S. patent application Ser. No. 08/660,461, filed Jun. 7, 1996, now U.S. Pat. No. 5,802,294, which is a continuation of U.S. patent application Ser. No. 08/131,523, filed Oct. 1, 1993, now U.S. Pat. No. 5,689,641, the disclosures of which are incorporated herein by reference.
  • This application is also related to U.S. patent application Ser. No. ______ filed Oct. 28, 2005 (Attorney Docket No. CA1420); U.S. patent application Ser. No. ______ filed Oct. 28, 2005 (Attorney Docket No. CA1422); U.S. patent application Ser. No. ______, filed Oct. 28, 2005 (Attorney Docket No. CA1423); and U.S. patent application Ser. No. ______ filed Oct. 28, 2005 (Attorney Docket No. CA 1424).
  • This application is also related to U.S. patent application Ser. No. 10/721,343, filed Nov. 26, 2003; U.S. patent application Ser. No. 10/722,051 filed Nov. 26, 2003; U.S. patent application Ser. No. 10/721,905 filed Nov. 26, 2003; U.S. patent application Ser. No. 10/721,385 filed Nov. 26, 2003.
  • BACKGROUND OF THE INVENTION
  • The present invention relates to computer-based systems for enhancing collaboration between and among individuals who are separated by distance and/or time (referred to herein as “distributed collaboration”). Principal among the invention's goals is to replicate in a desktop environment, to the maximum extent possible, the full range, level and intensity of interpersonal communication and information sharing which would occur if all the participants were together in the same room at the same time (referred to herein as “face-to-face collaboration”).
  • It is well known to behavioral scientists that interpersonal communication involves a large number of subtle and complex visual cues, referred to by names like “eye contact” and “body language,” which provide additional information over and above the spoken words and explicit gestures. These cues are, for the most part, processed subconsciously by the participants, and often control the course of a meeting.
  • In addition to spoken words, demonstrative gestures and behavioral cues, collaboration often involves the sharing of visual information—e.g., printed material such as articles, drawings, photographs, charts and graphs, as well as videotapes and computer-based animations, visualizations and other displays—in such a way that the participants can collectively and interactively examine, discuss, annotate and revise the information. This combination of spoken words, gestures, visual cues and interactive data sharing significantly enhances the effectiveness of collaboration in a variety of contexts, such as “brainstorming” sessions among professionals in a particular field, consultations between one or more experts and one or more clients, sensitive business or political negotiations, and the like. In distributed collaboration settings, then, where the participants cannot be in the same place at the same time, the beneficial effects of face-to-face collaboration will be realized only to the extent that each of the remotely located participants can be “recreated” at each site.
  • To illustrate the difficulties inherent in reproducing the beneficial effects of face-to-face collaboration in a distributed collaboration environment, consider the case of decision-making in the fast-moving commodities trading markets, where many thousands of dollars of profit (or loss) may depend on an expert trader making the fight decision within hours, or even minutes, of receiving a request from a distant client. The expert requires immediate access to a wide range of potentially relevant information such as financial data, historical pricing information, current price quotes, newswire services, government policies and programs, economic forecasts, weather reports, etc. Much of this information can be processed by the expert in isolation. However, before making a decision to buy or sell, he or she will frequently need to discuss the information with other experts, who may be geographically dispersed, and with the client. One or more of these other experts may be in a meeting, on another call, or otherwise temporarily unavailable. In this event, the expert must communicate “asynchronously”—to bridge time as well as distance.
  • As discussed below, prior art desktop videoconferencing systems provide, at best, only a partial solution to the challenges of distributed collaboration in real time, primarily because of their lack of high-quality video (which is necessary for capturing the visual cues discussed above) and their limited data sharing capabilities. Similarly, telephone answering machines, voice mail, fax machines and conventional electronic mail systems provide incomplete solutions to the problems presented by deferred (asynchronous) collaboration because they are totally incapable of communicating visual cues, gestures, etc and, like conventional videoconferencing systems, are generally limited in the richness of the data that can be exchanged.
  • It has been proposed to extend traditional videoconferencing capabilities from conference centers, where groups of participants must assemble in the same room, to the desktop, where individual participants may remain in their office or home. Such a system is disclosed in U.S. Pat. No. 4,710,917 to Tompkins et al for Video Conferencing Network issued on Dec. 1, 1987. It has also been proposed to augment such video conferencing systems with limited “video mail” facilities. However, such dedicated videoconferencing systems (and extensions thereof) do not effectively leverage the investment in existing embedded information infrastructures—such as desktop personal computers and workstations, local area network (LAN) and wide area network (WAN) environments, building wiring, etc.—to facilitate interactive sharing of data in the form of text, images, charts, graphs, recorded video, screen displays and the like. That is, they attempt to add computing capabilities to a videoconferencing system, rather than adding multimedia and collaborative capabilities to the user's existing computer system. Thus, while such systems may be useful in limited contexts, they do not provide the capabilities required for maximally effective collaboration, and are not cost-effective.
  • Conversely, audio and video capture and processing capabilities have recently been integrated into desktop and portable personal computers and workstations (hereinafter generically referred to as “workstations”). These capabilities have been used primarily in desktop multimedia authoring systems for producing CD-ROM-based works. While such systems are capable of processing, combining, and recording audio, video and data locally (i.e., at the desktop), they do not adequately support networked collaborative environments, principally due to the substantial bandwidth requirements for real-time transmission of high-quality, digitized audio and full-motion video which preclude conventional LANs from supporting more than a few workstations. Thus, although currently available desktop multimedia computers frequently include videoconferencing and other multimedia or collaborative capabilities within their advertised feature set (see, e.g., A Reinhardt, “Video Conquers the Desktop,” BYTE, September 1993, pp. 64-90), such systems have not yet solved the many problems inherent in any practical implementation of a scalable collaboration system.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • FIG. 1 is an enterprise view of a desk-top collaboration system embodiment of the present invention.
  • FIGS. 2A and 2B are photographs which attempt to illustrate, to the extent possible in a still image, the high-quality of the full-motion video and related user interface displays that appear on typical CMW screens which may be generated during operation of a preferred embodiment of the invention.
  • FIG. 3 is a block and schematic diagram of a preferred embodiment of a “multimedia local area network” (MLAN) in accordance with a desktop collaboration system embodiment of the present invention.
  • FIG. 4 is a block and schematic diagram illustrating how a plurality of geographically dispersed MLANs of the type shown in FIG. 3 can be connected via a wide area network in accordance with the present invention.
  • FIG. 5 is a schematic diagram illustrating how collaboration sites at distant locations L1-L8 are conventionally interconnected over a wide area network by individually connecting each site to every other site.
  • FIG. 6 is a schematic diagram illustrating how collaboration sites at distant locations L1-L8 are interconnected over a wide area network in a preferred embodiment of the invention using a multi-hopping approach.
  • FIG. 7 is a block diagram illustrating a preferred embodiment of video mosaicing circuitry provided in the MLAN of FIG. 3.
  • FIGS. 8A, 8B and 8C illustrate the video window on a typical CMW screen which may be generated during operation of a preferred embodiment of the present invention, and which contains only the callee for two-party calls (8A) and a video mosaic of all participants, e.g., for four-party (8B) or eight-party (8C) conference calls.
  • FIG. 9 is a block diagram illustrating a preferred embodiment of audio mixing circuitry provided in the MLAN of FIG. 3.
  • FIG. 10 is a block diagram illustrating video cut-and-paste circuitry provided in the MLAN of FIG. 3.
  • FIG. 11 is a schematic diagram illustrating typical operation of the video cut-and-paste circuitry in FIG. 10.
  • FIGS. 12-17 (consisting of FIGS. 12A, 12B, 13A, 13B, 14A, 14B, 15A, 15B, 16, 17A and 17B) illustrate various examples of how a preferred embodiment of the present invention provides video mosaicing, video cut-and-pasting, and audio mixing at a plurality of distant sites for transmission over a wide area network in order to provide, at the CMW of each conference participant, video images and audio captured from the other conference participants.
  • FIGS. 18A and 18B illustrate various preferred embodiments of a CMW which may be employed in accordance with the present invention.
  • FIG. 19 is a schematic diagram of a preferred embodiment of a CMW add-on box containing integrated audio and video L/O circuitry in accordance with the present invention.
  • FIG. 20 illustrates CMW software in accordance with a preferred embodiment of the present invention, integrated with standard multitasking operating system and applications software.
  • FIG. 21 illustrates software modules which may be provided for running on the MLAN Server in the MLAN of FIG. 3 for controlling operation of the AV and Data Networks.
  • FIG. 22 illustrates an enlarged example of “speed-dial” face icons of certain collaboration participants in a Collaboration Initiator window on a typical CMW screen which may be generated during operation of a preferred embodiment of the present invention.
  • FIG. 23 is a diagrammatic representation of the basic operating events occurring in a preferred embodiment of the present invention during initiation of a two-party call.
  • FIG. 24 is a block and schematic diagram illustrating how physical connections are established in the MLAN of FIG. 3 for physically connecting first and second workstations for a two-party videoconference call.
  • FIG. 25 is a block and schematic diagram illustrating how physical connections are preferably established in MLANs such as illustrated in FIG. 3, for a two-party call between a first CMW located at one site and a second CMW located at a remote site.
  • FIGS. 26 and 27 are block and schematic diagrams illustrating how conference bridging is preferably provided in the MLAN of FIG. 3.
  • FIG. 28 diagrammatically illustrates how a snapshot with annotations may be stored in a plurality of bitmaps during data sharing.
  • FIG. 29 is a schematic and diagrammatic illustration of the interaction among multimedia mail (MMM), multimedia call/conference recording (MMCR) and multimedia document management (MMDM) facilities.
  • FIG. 30 is a schematic and diagrammatic illustration of the multimedia document architecture employed in a preferred embodiment of the invention.
  • FIG. 31A illustrates a centralized Audio/Video Storage Server.
  • FIG. 31B is a schematic and diagrammatic illustration of the interactions between the Audio/Video Storage Server and the remainder of the CMW System.
  • FIG. 31C illustrates an alternative embodiment of the interactions illustrated in FIG. 31B.
  • FIG. 31D is a schematic and diagrammatic illustration of the integration of MMM MMCR and MMDM facilities in a preferred embodiment of the invention.
  • FIG. 32 illustrates a generalized hardware implementation of a scalable Audio/Video Storage Server.
  • FIG. 33 illustrates a higher throughput version of the server illustrated in FIG. 32, using SCSI-based crosspoint switching to increase the number of possible simultaneous file transfers.
  • FIG. 34 illustrates the resulting multimedia collaboration environment achieved by the integration of audio/video/data teleconferencing and MMCR, MMM and MMDM.
  • FIGS. 35-42 illustrate a series of CMW screens which may be generated during operation of a preferred embodiment of the present invention for a typical scenario involving a remote expert who takes advantage of many of the features provided by the present invention.
  • SUMMARY OF THE INVENTION
  • In accordance with the present invention, computer hardware, software and communications technologies are combined in novel ways to produce a multimedia collaboration system that greatly facilitates distributed collaboration, in part by replicating the benefits of face-to-face collaboration. The system tightly integrates a carefully selected set of multimedia and collaborative capabilities, principal among which are desktop teleconferencing and multimedia mail.
  • As used herein, desk-top teleconferencing includes real-time audio and/or video teleconferencing, as well as data conferencing Data conferencing, in turn, includes snapshot sharing (sharing of “snapshots” of selected regions of the user's screen), application sharing (shared control of running applications), shared whiteboard (equivalent to sharing a “blank” window), and associated telepointing and annotation capabilities. Teleconferences may be recorded and stored for later playback, including both audio/video and all data interactions.
  • While desktop teleconferencing supports real-time interactions, multimedia mail permits the asynchronous exchange of arbitrary multimedia documents, including previously recorded teleconferences. Indeed, it is to be understood that the multimedia capabilities underlying desktop teleconferencing and multimedia mail also greatly facilitate the creation, viewing, and manipulation of high-quality multimedia documents in general, including animations and visualizations that might be developed, for example, in the course of information analysis and modeling. Further, these animations and visualizations may be generated for individual rather than collaborative use, such that the present invention has utility beyond a collaboration context.
  • The preferred embodiment of the invention is a collaborative multimedia workstation (CMW) system wherein very high-quality audio and video capabilities can be readily superimposed onto an enterprise's existing computing and network infrastructure, including workstations, LANs, WANs, and building wiring.
  • In a preferred embodiment, the system architecture employs separate real-time and asynchronous networks—the former for real-time audio and video, and the latter for non-real-time audio and video, text, graphics and other data, as well as control signals. These networks are interoperable across different computers (e.g., Macintosh, Intel-based PCs, and Sun workstations), operating systems (e.g., Apple System 7, DOS/Windows, and UNIX) and network operating systems (e.g., Novell Netware and Sun ONC+). In many cases, both networks can actually share the same cabling and wall jack connector.
  • The system architecture also accommodates the situation in which the user's desktop computing and/or communications equipment provides varying levels of media-handling capability. For example, a collaboration session—whether real-time or asynchronous—may include participants whose equipment provides capabilities ranging from audio only (a telephone) or data only (a personal computer with a modem) to a full complement of real-time, high-fidelity audio and full-motion video, and high-speed data network facilities.
  • The CMW system architecture is readily, scalable to very large enterprise-wide network environments accommodating thousands of users. Further, it is an open architecture that can accommodate appropriate standards. Finally, the CMW system incorporates an intuitive, yet powerful, user interface, making the system easy to learn and use.
  • The present invention thus provides a distributed multimedia collaboration environment that achieves the benefits of face-to-face collaboration as nearly as possible, leverages (“snaps on to”) existing computing and network infrastructure to the maximum extent possible, scales to very large networks consisting of thousand of workstations, accommodates emerging standards, and is easy to learn and use. The specific nature of the invention, as well as its objects, features, advantages and uses, will become more readily apparent from the following detailed description and examples, and from the accompanying drawings.
  • DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENT
  • Overall System Architecture
  • Referring initially to FIG. 1, illustrated therein is an overall diagrammatic view of a multimedia collaboration system in accordance with the present invention. As shown, each of a plurality of “multimedia local area networks” (MLANs) 10 connects, via lines 13, a plurality of CMWs 12-1 to 12-10 and provides audio/video/data networking for supporting collaboration among CMW users WAN 15 in turn connects multiple MLANs 10, and typically includes appropriate combinations of common carrier analog and digital transmission networks Multiple MLANs 10 on the same physical premises may be connected via bridges/routes 11, as shown, to WANs and one another.
  • In accordance with the present invention, the system of FIG. 1 accommodates both “real time” delay and jitter-sensitive signals (e.g., real-time audio and video teleconferencing) and classical asynchronous data (e.g., data control signals as well as shared textual, graphics and other media) communication among multiple CMWs 12 regardless of their location. Although only ten CMWs 12 are illustrated in FIG. 1, it will be understood that many more could be provided. As also indicated in FIG. 1, various other multimedia resources 16 (e.g., VCRs, laserdiscs, TV feeds, etc.) are connected to MLANs 10 and are thereby accessible by individual CMWs 12.
  • CMW 12 in FIG. 1 may use any of a variety of types of operating systems, such as Apple System 7, UNIX, DOS/Windows and OS/2. The CMWs can also have different types of window systems. Specific preferred embodiments of a CMW 12 are described hereinafter in connection with FIGS. 18A and 18B. Note that this invention allows for a mix of operating systems and window systems across individual CMWs.
  • In the preferred embodiment, CMW 12 in FIG. 1 provides real-time audio/video/data capabilities along with the usual data processing capabilities provided by its operating system CMW 12 also provides for bidirectional communication, via lines 13, within MLAN 10, for audio/video signals as well as data signals Audio/video signals transmitted from a CMW 12 typically comprise a high-quality live video image and audio of the CMW operator. These signals are obtained from a video camera and microphone provided at the CMW (via an add-on unit or partially or totally integrated into the CMW), processed, and then made available to low-cost network transmission subsystems.
  • Audio/video signals received by a CMW 12 from MLAN 10 may typically include: video images of one or more conference participants and associated audio, video and data from multimedia mail, previously recorded audio/video from previous calls and conferences, and standard broadcast television (e.g. CNN). Received video signals are displayed on the CMW screen or on an adjacent monitor, and the accompanying audio is reproduced by a speaker provided in or near the CMW. In general, the required transducers and signal processing hardware could be integrated into the CMW, or be provided via a CMW add-on unit, as appropriate.
  • In the preferred embodiment, it has been found particularly advantageous to provide the above-described video at standard NTSC-quality TV performance (i.e., 30 frames per second at 640×480 pixels per frame and the equivalent of 24 bits of color per pixel) with accompanying high-fidelity audio (typically between 7 and 15 KHz). For example, FIG. 2A illustrates a CMW screen containing live, full-motion video of three conference participants, while FIG. 2B illustrates data shared and annotated by those conferees (lower left window).
  • Multimedia Local Area Network
  • Referring next to FIG. 3, illustrated therein is a preferred embodiment of MLAN 10 having ten CMWs (12-1, 12-2,-12-10), coupled therein via lines 13 a and 13 b MLAN 10 typically extends over a distance from a few hundred feet to a few miles, and is usually located within a building or a group of proximate buildings.
  • Given the current state of networking technologies, it is useful (for the sake of maintaining quality and minimizing costs) to provide separate signal paths for real-time audio/video and classical asynchronous data communications (including digitized audio and video enclosures of multimedia mail messages that are free from real-time delivery constraints). At the moment, analog methods for carrying real-time audio/video are preferred. In the future, digital methods may be used. Eventually, digital audio and video signal paths may be multiplexed with the data signal path as a common digital stream. Another alternative is to multiplex real-time and asynchronous data paths together using analog multiplexing methods. For the purposes of the present application, however, we will treat these two signal paths as using physically separate wires. Further, as the current preferred embodiment uses analog networking for audio and video, it also physically separates the real-time and asynchronous switching vehicles and, in particular, assumes an analog audio/video switch. In the future, a common switching vehicle (e.g., ATM) could be used.
  • The MLAN 10 thus can be implemented in the preferred embodiment using conventional technology, such as typical Data LAN hubs 25 and A/V Switching Circuitry 30 (as used in television studios and other closed-circuit television networks), linked to the CMWs 12 via appropriate transceivers and unshielded twisted pair (UTP) wiring. Note in FIG. 1 that lines 13, which interconnect each CMW 12 within its respective MLAN 10, comprise two sets of lines 13 a and 13 b Lines 13 a provide bidirectional communication of audio/video within MLAN 10, while lines 13 b provide for the bidirectional communication of data. This separation permits conventional LANs to be used for data communications and a supplemental network to be used for audio/video communications. Although this separation is advantageous in the preferred embodiment, it is again to be understood that audio/video/data networking can also be implemented using a single pair of lines for both audio/video and data communications via a very wide variety of analog and digital multiplexing schemes.
  • While lines 13 a and 13 b may be implemented in various ways, it is currently preferred to use commonly installed 4-pair UTP telephone wires, wherein one pair is used for incoming video with accompanying audio (mono or stereo) multiplexed in, wherein another pair is used for outgoing multiplexed audio/video, and wherein the remaining two pairs are used for carrying incoming and outgoing data in ways consistent with existing LANs. For example, 10BaseT Ethernet uses RJ-45 pins 1, 2, 4, and 6, leaving pins 3, 5, 7, and 8 available for the two A/V twisted pairs. The resulting system is compatible with standard (AT&T 258A, EIA/TIA 568, 8P8C, 10BaseT, ISDN, 6P6C, etc.) telephone wiring found commonly throughout telephone and LAN cable plants in most office buildings throughout the world. These UTP wires are used in a hierarchy or peer arrangements of star topologies to create MLAN 10, described below. Note that the distance range of the data wires often must match that of the video and audio Various UTP-compatible data. LAN networks may be used, such as Ethernet, token ring, FDDI, ATM etc. For distances longer than the maximum distance specified by the data LAN protocol, data signals can be additionally processed for proper UTP operations.
  • As shown in FIG. 3, lines 13 a from each CMW 12 are coupled to a conventional Data LAN hub 25, which facilitates the communication of data (including control signals) among such CMWs Lines 13 b in FIG. 3 are connected to A/V Switching Circuitry 30. One or more conference bridges 35 are coupled to A/V Switching Circuitry 30 and possibly (if needed) the Data LAN hub 25, via lines 35 b and 35 a, respectively, for providing multi-party Conferencing in a particularly advantageous manner, as will hereinafter be described in detail A WAN gateway 40 provides for bidirectional communication between MLAN 10 and WAN 15 in FIG. 1. For this purpose, Data LAN hub 25 and A/V Switching Circuitry 30 are coupled to WAN gateway 40 via outputs 25 a and 30 a, respectively. Other devices connect to the A/V Switching Circuitry 30 and Data LAN hub 25 to add additional features, such as multimedia mail, conference recording, etc.) as discussed below.
  • Control of A/V Switching Circuitry 30, conference bridges 35 and WAN gateway 40 in FIG. 3 is provided by MLAN Server 60 via lines 60 b, 60 c, and 60 d, respectively. In a preferred embodiment, MLAN Server 60 supports the TCP/IP network protocol suite. Accordingly, software processes on CMWs 12 communicate with one another and MLAN Server 60 via MLAN 10 using these protocols. Other network protocols could also be used, such as IPX. The manner in which software running on MLAN Server 60 controls the operation of MLAN 10 will be described in detail hereinafter.
  • Note in FIG. 3 that Data LAN hub 25, A/V Switching Circuitry 30 and MLAN Server 60 also provide respective lines 25 b, 30 b, and 60 e for coupling to additional multimedia resources 16 (FIG. 1), such as multimedia document management, multimedia databases, radio/TV channel setc Data LAN hub 25 (via bridges/routers 11 in FIG. 1) and A/V Switching Circuitry 30 additionally provide lines 25 c and 30 c for coupling to one or more other MLANs 10 which may be in the same locality (i.e., not far enough away to require use of WAN technology) Where WANs are required, WAN gateways 40 are used to provide highest quality compression methods and standards in a shared resource fashion, thus minimizing costs at the workstation for a given WAN quality level, as discussed below.
  • The basic operation of the preferred embodiment of the resulting collaboration system shown in FIGS. 1 and 3 will next be considered. Important features of the present invention reside in providing not only multi-party real-time desktop audio/video/data teleconferencing among geographically distributed CMWs, but also in providing from the same desktop audio/video/data/text/graphics mail capabilities, as well as access to other resources, such as databases, audio and video files, overview cameras, standard TV channels, etc FIG. 2B illustrates a CMW screen showing a multimedia EMAIL mailbox (top left window) containing references to a number of received messages along with a video enclosure (top right window) to the selected message.
  • A/V Switching Circuitry 30 (whether digital or analog as in the preferred embodiment) provides common audio/video switching for CMWs 12, conference bridges 35, WAN gateway 40 and multimedia resources 16, as determined by MLAN Server 60, which in turn controls conference bridges 35 and WAN gateway 44). Similarly, asynchronous data is communicated within MLAN 10 utilizing common data communications formats where possible (e.g., for snapshot sharing) so that the system can handle such data in a common manner, regardless of origin, thereby facilitating multimedia mail and data sharing as well as audio/video communications.
  • For example, to provide multi-party teleconferencing, an initiating CMW 12 signals MLAN Server 60 via Data LAN hub 25 identifying the desired conference participants. After determining which of these conferees will accept the call, MLAN Server 60 controls A/V Switching Circuitry 30 (and CMW software via the data network) to set up the required audio/video and data paths to conferees at the same location as the initiating CMW.
  • When one or more conferees are at distant locations, the respective MLAN Servers 60 of the involved MLANs 10, on a peer-to-peer basis, control their respective A/V Switching Circuitry 30, conference bridges 35, and WAN gateways 40 to set up appropriate communication paths (via WAN 15 in FIG. 1) as required for interconnecting the conferees MLAN Servers 60 also communicate with one another via data paths so that each MLAN 10 contains updated information as to the capabilities of all of the system CMWs 12, and also the current locations of all parties available for teleconferencing.
  • The data conferencing component of the above-described system supports the sharing of visual information at one or more CMWs (as described in greater detail below). This encompasses both “snapshot sharing” (sharing “snapshots” of complete or partial screens, or of one or more selected windows) and “application sharing” (sharing both the control and display of running applications) When transferring images, lossless or slightly lossy image compression can be used to reduce network bandwidth requirements and user-perceived delay while maintaining high image quality.
  • In all cases, any participant can point at or annotate the shared data. These associated telepointers and annotations appear on every participant's CMW screen as they are drawn (i.e., effectively in real time). For example, note FIG. 2B which illustrates a typical CMW screen during a multi-party teleconferencing session, wherein the screen contains annotated shared data as well as video images of the conferees. As described in greater detail below, all or portions of the audio/video and data of the teleconference can be recorded at a CMW (or within MLAN 10), complete with all the data interactions.
  • In the above-described preferred embodiment, audio/video file services can be implemented either at the individual CMWs 12 or by employing a centralized audio/video storage server. This is one example of the many types of additional servers that can be added to the basic system of MLANs 10. A similar approach is used for incorporating other multimedia services, such as commercial TV channels, multimedia mail, multimedia document management, multimedia conference recording, visualization servers, etc. (as described in greater detail below) Certainly, applications that run self-contained on a CMW can be readily added, but the invention extends this capability greatly in the way that MLAN 10, storage and other functions are implemented and leveraged.
  • In particular, standard signal formats, network interfaces, user interface messages, and call models can allow virtually any multimedia resource to be smoothly integrated into the system. Factors facilitating such smooth integration include: (i) a common mechanism for user access across the network; (ii) a common metaphor (e.g., placing a call) for the user to initiate use of such resource; (iii) the ability for one function (e.g., a multimedia conference or multimedia database) to access and exchange information with another function (e.g., multimedia mail); and (iv) the ability to extend such access of one networked function by another networked function to relatively complex nestings of simpler functions (for example, record a multimedia conference in which a group of users has accessed multimedia mail messages and transferred them to a multimedia database, and then send part of the conference recording just created as a new multimedia mail message, utilizing a multimedia mail editor if necessary).
  • A simple example of the smooth integration of functions made possible by the above-described approach is that the GUI and software used for snapshot sharing (described below) can also be used as an input/output interface for multimedia mail and more general forms of multimedia documents. This can be accomplished by structuring the interprocess communication protocols to be uniform across all these applications. More complicated examples—specifically multimedia conference recording, multimedia mail and multimedia document management—will be presented in detail below.
  • Wide Area Network
  • Next to be described in connection with FIG. 4 is the advantageous manner in which the present invention provides for real-time audio/video/data communication among geographically dispersed MLANs 10 via WAN 15 (FIG. 1), whereby communication delays, cost and degradation of video quality are significantly minimized from what would otherwise be expected.
  • Four MLANs 10 are illustrated at locations A, B, C and D CMWs 12-1 to 12-10, A/V Switching Circuitry 30, Data LAN hub 25, and WAN gateway 40 at each location correspond to those shown in FIGS. 1 and 3 WAN gateway 40 in FIG. 4 will be seen to comprise a router/codec (R&C) bank 42 coupled to WAN 15 via WAN switching multiplexer 44. The router is used for data interconnection and the codec is used for audio/video interconnection (for multimedia mail and document transmission, as well as videoconferencing). Codecs from multiple vendors, or supporting various compression algorithms may be employed. In the preferred embodiment, the router and codec are combined with the switching multiplexer to form a single integrated unit.
  • Typically, WAN 15 is comprised of T1 or ISDN common-carrier-provided digital links (switched or dedicated), in which case WAN switching multiplexers 44 are of the appropriate type (T1, ISDN, fractional T1, T3, switched 56 Kbps, etc.). Note that the WAN switching multiplexer 44 typically creates subchannels whose bandwidth is a multiple of 64 Kbps (i.e., 256 Kbps, 384, 768, etc.) among the T1, T3 or ISDN carriers. Inverse multiplexers may be required when using 56 Kbps dedicated or switched services from these carriers.
  • In the MLAN 10 to WAN 15 direction, router/codec bank 42 in FIG. 4 provides conventional analog-to-digital conversion and compression of audio/video signals received from A/V Switching Circuitry 30 for transmission to WAN 15 via WAN switching multiplexer 44, along with transmission and routing of data signals received from Data LAN hub 25. In the WAN 15 to MLAN 10 direction, each router/codec bank 42 in FIG. 4 provides digital-to-analog conversion and decompression of audio/video digital signals received from WAN 15 via WAN switching multiplexer 44 for transmission to A/V Switching Circuitry 30, along with the transmission to Data LAN hub 25 of data signals received from WAN 15.
  • The system also provides optimal routes for audio/video signals through the WAN. For example, in FIG. 4, location. A can take either a direct route to location D via path 47, or a two-hop route through location C via paths 48 and 49. If the direct path 47 linking location A and location D is unavailable, the multipath route via location C and paths 48 and 49 could be used.
  • In a more complex network, several multi-hop routes are typically available, in which case the routing system handles the decision making, which for example can be based on network loading considerations. Note the resulting two-level network hierarchy: a MLAN 10 to MLAN 10 (i.e., site-to-site) service connecting codecs with one another only at connection endpoints.
  • The cost savings made possible by providing the above-described multi-hop capability (with intermediate codec bypassing) are very significant as will become evident by noting the examples of FIGS. 5 and 6. FIG. 5 shows that using the conventional “fully connected mesh” location-to-location approach, thirty-six WAN links are required for interconnecting the nine locations L1 to L8. On the other hand, using the above multi-hop capabilities, only nine WAN links are required, as shown in FIG. 6. As the number of locations increase, the difference in cost becomes even greater, growing as the square of the number of sites. For example, for 100 locations, the conventional approach would require about 5,000 WAN links, while the multi-hop approach of the present invention would typically require 300 or fewer (possibly considerably fewer) WAN links. Although specific WAN links for the multi-hop approach of the invention would require higher bandwidth to carry the additional traffic, the cost involved is very much smaller as compared to the cost for the very much larger number of WAN links required by the conventional approach.
  • At the endpoints of a wide-area call, the WAN switching multiplexer routes audio/video signals directly from the WAN network interface through an available codec to MLAN 10 and vice versa. At intermediate hops in the network, however, video signals are routed from one network interface on the WAN switching multiplexer to another network interface. Although A/V Switching Circuitry 30 could be used for this purpose, the preferred embodiment provides switching functionality inside the WAN switching multiplexer. By doing so, it avoids having to route audio/video signals through codecs to the analog switching circuitry, thereby avoiding additional codec delays at the intermediate locations.
  • A product capable of performing the basic switching functions described above for WAN switching multiplexer 44 is available from Teleos Corporation, Eatontown, N.J. This product is not known to have been used for providing audio/video multi-hopping and dynamic switching among various WAN links as described above.
  • In addition to the above-described multiple-hop approach, the preferred embodiment of the present invention provides a particularly advantageous way of minimizing delay, cost and degradation of video quality in a multi-party video teleconference involving geographically dispersed sites, while still delivering full conference views of all participants. Normally, in order for the CMWs at all sites to be provided with live audio/video of every participant in a teleconference simultaneously, each site has to allocate (in router/codec bank 42 in FIG. 4) a separate codec for each participant, as well as a like number of WAN trunks (via WAN switching multiplexer 44 in FIG. 4).
  • As will next be described, however, the preferred embodiment of the invention advantageously permits each wide area audio/video teleconference to use only one codec at each site, and a minimum number of WAN digital trunks. Basically, the preferred embodiment achieves this most important result by employing “distributed” video mosaicing via a video “cut-and-paste” technology along with distributed audio mixing.
  • Distributed Video Mosaicing
  • FIG. 7 illustrates a preferred way of providing video mosaicing in the MLAN of FIG. 3—i.e., by combining the individual analog video pictures from the individuals participating in a teleconference into a single analog mosaic picture. As shown in FIG. 7, analog video signals 112-1 to 112-n from the participants of a teleconference are applied to video mosaicing circuitry 36, which in the preferred embodiment is provided as part of conference bridge 35 in FIG. 3. These analog video inputs 112-1 to 112-n are obtained from the A/V Switching Circuitry 30 (FIG. 3) and may include video signals from CMWs at one or more distant sites (received via WAN gateway 40) as well as from other CMWs at the local site.
  • In the preferred embodiment, video mosaicing circuitry 36 is capable of receiving N individual analog video picture signals (where N is a squared integer, i.e., 4, 9, 16, etc.). Circuitry 36 first reduces the size of the N input video signals by reducing the resolutions of each by a factor of M (where M is the square root of N (i.e., 2, 3, 4, etc.), and then arranging them in an M-by-M mosaic of N images. The resulting single analog mosaic 36 a obtained from video mosaicing circuitry 36 is then transmitted to the individual CMWs for display on the screens thereof.
  • As will become evident hereinafter, it may be preferable to send a different mosaic to distant sites, in which case video mosaicing circuitry 36 would provide an additional mosaic 36 b for this purpose. A typical displayed mosaic picture (N=4, M=2) showing three participants is illustrated in FIG. 2A. A mosaic containing four participants is shown in FIG. 8B. It will be appreciated that, since a mosaic (36 a or 36 b) can be transmitted as a single video picture to another site, via WAN 15 (FIGS. 1 and 4), only one codec and digital trunk are required. Of course, if only a single individual video picture is required to be sent from a site, it may be sent directly without being included in a mosaic. Note that for large conferences it is possible to employ multiple video mosaics, one for each video window supported by the CMWs (see, e.g., FIG. 8C). In very large conferences, it is also possible to display video only from a select focus group whose members are selected by a dynamic “floor control” mechanism. Also note that, with additional mosaic hardware, it is possible to give each CMW its own mosaic. This can be used in small conferences to raise the maximum number of participants (from M sup 2 to M sup 2+1—i.e., 5, 10, 17, etc.) or to give everyone in a large conference their own “focus group” view.
  • Also note that the entire video mosaicing approach described thus far and continued below applies should digital video transmission be used in lieu of analog transmission, particularly since both mosaic and video window implementations use digital formats internally and in current products are transformed to and from analog for external interfacing. In particular, note that mosaicing can be done digitally without decompression with many existing compression schemes. Further, with an all-digital approach, mosaicing can be done as needed directly on the CMW.
  • FIG. 9 illustrates preferred audio mixing circuitry 38 for use in conjunction with the video mosaicing circuitry 36 in FIG. 7, both of which may be part of conference bridges 35 in FIG. 3. As shown in FIG. 9, audio signals 114-1 to 114-n are applied to audio summing circuitry 38 for combination. These input audio signals 114-1 to 114-n may include audio signals from local participants as well as audio sums from participants at distant sites. Audio mixing circuitry 38 provides a respective “minus-1” sum output 38 a-1, 38 a-2, etc., for each participant. Thus, each participant hears every conference participant's audio except his/her own.
  • In the preferred embodiment, sums are decomposed and formed in a distributed fashion, creating partial sums at one site which are completed at other sites by appropriate signal insertion. Accordingly, audio mixing circuitry 38 is able to provide one or more additional sums, such as indicated by output 38, for sending to other sites having conference participants.
  • Next to be considered is the manner in which video cut-and-paste techniques are advantageously employed in the preferred embodiment. It will be understood that, since video mosaics and/or individual video pictures may be sent from one or more other sites, the problem arises as to how these situations are handled. In the preferred embodiment, video cut-and-paste circuitry 39, as illustrated in FIG. 10, is provided for this purpose, and may also be incorporated in the conference bridges 35 in FIG. 3.
  • Referring to FIG. 10, video cut-and-paste circuitry receives analog video inputs 116, which may be comprised of one or more mosaics or single video pictures received from one or more distant sites and a mosaic or single video picture produced by the local site. It is assumed that the local video mosaicing circuitry 36 (FIG. 7) and the video cut-and-paste circuitry 39 have the capability of handling all of the applied individual video pictures, or at least are able to choose which ones are to be displayed based on existing available signals.
  • The video cut-and-paste circuitry 39 digitizes the incoming analog video inputs 116, selectively rearranges the digital signals on a region-by-region basis to produce a single digital M-by-M mosaic, having individual pictures in selected regions, and then converts the resulting digital mosaic back to analog form to provide a single analog mosaic picture 39 a for sending to local participants (and other sites where required) having the individual input video pictures in appropriate regions. This resulting cut-and-paste analog mosaic 39 a will provide the same type of display as illustrated in FIG. 8B. As will become evident hereinafter, it is sometimes beneficial to send different cut-and-paste mosaics to different sites, in which case video cut-and-paste circuitry 39 will provide additional cut-and-paste mosaics 39 b-1, 39 b-2, etc. for this purpose.
  • FIG. 11 diagrammatically illustrates an example of how video cut-and-paste circuitry may operate to provide the cut-and-paste analog mosaic 39 a. As shown in FIG. 11; four digitized individual signals 116 a, 116 b, 116 c and 116 d derived from the input video signals are “pasted” into selected regions of a digital frame buffer 17 to form a digital 2×2 mosaic, which is converted into an output analog video mosaic 39 a or 39 b in FIG. 10. The required audio partial sums may be provided by audio mixing circuitry 39 in FIG. 9 in the same manner, replacing each cut-and-paste video operation with a partial sum operation.
  • Having described in connection with FIGS. 7-11 how video mosaicing, audio mixing, video cut-and-pasting, and distributed audio mixing may be performed, the following description of FIGS. 12-17 will illustrate how these capabilities may advantageously be used in combination in the context of wide-area videoconferencing. For these examples, the teleconference is assumed to have four participants, designated as A, B, C and D, in which case 2×2 (quad) mosaics are employed. It is to be understood that greater numbers of participants could be provided. Also, two or more simultaneously occurring teleconferences could also be handled, in which case additional mosaicing, cut-and-paste and audio mixing circuitry would be provided at the various sites along with additional WAN paths For each example, the “A” figure illustrates the video mosaicing and cut-and-pasting provided, and the corresponding “B” figure (having the same figure number) illustrates the associated audio mixing provided. Note that these figures indicate typical delays that might be encountered for each example (with a single “UNIT” delay ranging from 0-450 milliseconds, depending upon available compression technology).
  • FIGS. 12A and 12B illustrate a 2-site example having two participants A and B at Site # 1 and two participants C and D at Site # 2 Note that this example requires mosaicing and cut-and-paste at both sites.
  • FIGS. 13A and 13B illustrate another 2-site example, but having three participants A, B and C at Site # 1 and one participant D at Site # 2. Note that this example requires mosaicing at both sites, but cut-and-paste only at Site # 2.
  • FIGS. 14A and 14B illustrate a 3-site example having participants A and B at Site # 1, participant C at Site # 2, and participant D at Site # 3. At Site # 1, the two local videos A and B are put into a mosaic which is sent to both Site # 2 and Site # 3. At Site # 2 and Site # 3, cut-and-paste is used to insert the single video (C or D) at that site into the empty region in the imported A, B, and D or C mosaic, as shown. Accordingly, mosaicing is required at all three sites, and cut-and-paste is required for only Site # 2 and Site # 3.
  • FIGS. 15A and 15B illustrate another 3-site example having participant A at Site # 1, participant B at Site # 2, and participants C and D at Site # 3. Note that mosaicing and cut-and-paste are required at all sites Site # 2 additionally has the capability to send different cut-and-paste mosaics to Sites # 1 and Site # 3. Further note with respect to FIG. 15B that Site # 2 creates minus-1 audio mixes for Site # 1 and Site # 2, but only provides a partial audio mix (A&B) for Site # 3. These partial mixes are completed at Site # 3 by mixing in C's signal to complete D's mix (A+B+C) and D's signal to complete C's mix (A+B+D).
  • FIG. 16 illustrates a 4-site example employing a star topology, having one participant at each site; that is, participant A is at Site # 1, participant B is at Site # 2, participant C is at Site # 3, and participant D is at Site # 4. An audio implementation is not illustrated for this example, since standard minus-1 mixing can be performed at Site # 1, and the appropriate sums transmitted to the other sites.
  • FIGS. 17A and 17B illustrate a 4-site example that also has only one participant at each site, but uses a line topology rather than a star topology as in the example of FIG. 16. Note that this example requires mosaicing and cut-and-paste at all sites. Also note that Site # 2 and Site # 3 are each required to transmit two different types of cut-and-paste mosaics.
  • The preferred embodiment also provides the capability of allowing a conference participant to select a close-up of a participant displayed on a mosaic. This capability is provided whenever a full individual video picture is available at that user's site. In such case, the A/V Switching Circuitry 30 (FIG. 3) switches the selected full video picture (whether obtained locally or from another site) to the CMW that requests the close-up.
  • Next to be described in connection with FIGS. 18A, 18B, 19 and 20 are various preferred embodiments of a CMW in accordance with the invention. Collaborative Multimedia Workstation Hardware
  • One embodiment of a CMW 12 of the present invention is illustrated in FIG. 18A Currently available personal computers (e.g., an Apple Macintosh or an IBM-compatible PC, desktop or laptop) and workstations (e,g., a Sun SPARC station) can be adapted to work with the present invention to provide such features as real-time videoconferencing, data conferencing, multimedia mail, etc. In business situations, it can be advantageous to set up a laptop to operate with reduced functionality via cellular telephone links and removable storage media (e.g., CD-ROM, video tape with timecode support etc.), but take on full capability back in the office via a docking station connected to the MLAN 10. This requires a voice and data modem as yet another function server attached to the MLAN.
  • The currently available personal computers and workstations serve as a base workstation platform. The addition of certain audio and video I/O devices to the standard components of the base platform 100 (where standard components include the display monitor 200, keyboard 300 and mouse or tablet (or other pointing device) 400), all of which connect with the base platform box through standard peripheral ports 101, 102 and 103, enables the CMW to generate and receive real-time audio and video signals. These devices include a video camera 500 for capturing the user's image, gestures and surroundings (particularly the user's face and upper body), a microphone 600 for capturing the user's spoken words (and any other sounds generated at the CMW), a speaker 700 for presenting incoming audio signals (such as the spoken words of another participant to a videoconference or audio annotations to a document), a video input card 130 in the base platform 100 for capturing incoming video signals (e.g., the image of another participant to a videoconference, or videomail), and a video display card 120 for displaying video and graphical output on monitor 200 (where video is typically displayed in a separate window).
  • These peripheral audio and video I/O devices are readily available from a variety of vendors and are just beginning to become standard features in (and often physically integrated into the monitor and/or base platform of) certain personal computers and workstations. See, e.g., the aforementioned BYTE article (“Video Conquers the Desktop”), which describes current models of Apple's Macintosh AV series personal computers and Silicon Graphics' Indy workstations.
  • Add-on box 800 (shown in FIG. 18A and illustrated in greater detail in FIG. 19) integrates these audio and video I/O devices with additional functions (such as adaptive echo canceling and signal switching) and interfaces with AV Network 901 Network 901 is the part of the MLAN 10 which carries bidirectional audio and video signals among the CMWs and A/V Switching Circuitry 30—e.g., utilizing existing UTP wiring to carry audio and video signals (digital or analog, as in the present embodiment).
  • In the present embodiment, the AV network 901 is separate and distract from the Data Network 902 portion of the MLAN 10, which carries bidirectional data signals among the CMWs and the Data LAN hub (e.g., an Ethernet network that also utilizes UTP wiring in the present embodiment with a network interface card 110 in each CMW). Note that each CMW will typically be a node on both the AV and the Data Networks.
  • There are several approaches to implementing Add-on box 800. In a typical videoconference, video camera 500 and microphone 600 capture and transmit outgoing video and audio signals into ports 801 and 802, respectively, of Add-on box 800. These signals are transmitted via Audio/Video I/O port 805 across AV Network 901. Incoming video and audio signals (from another videoconference participant) are received across AV network 901 through Audio/Video I/O port 805. The video signals are sent out of V-OUT port 803 of CMW add-on box 800 to video input card 130 of base platform 100, where they are displayed (typically in a separate video window) on monitor 200 utilizing the standard base platform video display card 120. The audio signals are sent out of A-OUT port 804 of CMW add-on box 800 and played through speaker 700 while the video signals are displayed on monitor 200. The same signal flow occurs for other non-teleconferencing applications of audio and video.
  • Add-on box 800 can be controlled by CMW software (illustrated in FIG. 20) executed by base platform 100 Control signals can be communicated between base platform port 104 and Add-on box Control port 806 (e.g., an RS-232, Centronics, SCSI or other standard communications port).
  • Many other embodiments of the CMW illustrated in FIG. 18A will work in accordance with the present invention. For example, Add-on box 800 itself can be implemented as an add-in card to the base platform 100 Connections to the audio and video I/O devices need not change, though the connection for base platform control can be implemented internally (e.g., via the system bus) rather than through an external RS-232 or SCSI peripheral port. Various additional levels of integration can also be achieved as will be evident to those skilled in the art. For example, microphones, speakers, video cameras and UTP transceivers can be integrated into the base platform 100 itself, and all media handling technology and communications can be integrated onto a single card.
  • A handset/headset jack enables the use of an integrated audio I/O device as an alternate to the separate microphone and speaker. A telephone interface could be integrated into add-on box 800 as a local implementation of computer-integrated telephony. A “hold” (i.e., audio and video mute) switch and/or a separate audio mute switch could be added to Add-on box 800 if such an implementation were deemed preferable to a software-based interface.
  • The internals of Add-on box 800 of FIG. 18A are illustrated in FIG. 19. Video signals generated at the CMW (e.g., captured by camera 500 of FIG. 18A) are sent to CMW add-on box 800 via V-IN port 801. They then typically pass unaffected through Loopback/AV Mute circuitry 830 via video ports 833 (input) and 834 (output) and into A/V Transceivers 840 (via Video In port 842) where they are transformed from standard video cable signals to UTP signals and sent out via port 845 and Audio/Video I/O port 805 onto AV Network 901.
  • The Loopback/AV Mute circuitry 830 can, however, be placed in various modes under software control via Control port 806 (implemented, for example, as a standard UART). If in loopback mode (e.g., for testing incoming and outgoing signals at the CMW), the video signals would be routed back out V-OUT port 803 via video port 831. If in a mute mode (e.g., muting audio, video or both), video signals might, for example, be disconnected and no video signal would be sent out video port 834 Loop back and muting switching functionality is also provided for audio in a similar way. Note: that computer control of loopback is very useful for remote testing and diagnostics while manual override of computer control on mute is effective for assured privacy from use of the workstation for electronic spying.
  • Video input (e.g., captured by the video camera at the CMW of another videoconference participant) is handled in a similar fashion. It is received along AV Network 901 through Audio/Video I/O port 805 and port 845 of A/V Transceivers 840, where it is sent out Video Out port 841 to video port 832 of Loopback/AV Mute circuitry 830, which typically passes such signals out video port 831 to V-OUT port 803 (for receipt by a video input card or other display mechanism, such as LCD display 810 of CMW Side Mount unit 850 in FIG. 18B, to be discussed).
  • Audio input and output (e.g., for playback through speaker 700 and capture by microphone 600 of FIG. 18A) passes through A/V transceivers 840 (via Audio In port 844 and Audio Out port 843) and Loopback/AV Mute circuitry 830 (through audio ports 837/838 and 836/835) in a similar manner. The audio input and output ports of Add-on box 800 interface with standard amplifier and equalization circuitry, as well as an adaptive room echo canceler 814 to eliminate echo, minimize feedback and provide enhanced audio performance when using a separate microphone and speaker. In particular, use of adaptive room echo cancelers provides high-quality audio interactions in wide area conferences. Because adaptive room echo canceling requires training periods (typically involving an objectionable blast of high-amplitude white noise or tone sequences) for alignment with each acoustic environment, it is preferred that separate echo canceling be dedicated to each workstation rather than sharing a smaller group of echo cancelers across a larger group of workstations.
  • Audio inputs passing through audio port 835 of Loopback/AV Mute circuitry 830 provide audio signals to a speaker (via standard Echo Canceler circuitry 814 and A-OUT port 804 and/or to a handset or headset (via I/ O ports 807 and 808, respectively, under volume control circuitry 815 controlled by software through Control port 806). In all cases, incoming audio signals pass through power amplifier circuitry 812 before being sent out of Add-on box 80) to the appropriate audio-emitting transducer.
  • Outgoing audio signals generated at the CMW (e.g., by microphone 600 of FIG. 18A or the mouthpiece of a handset or headset) enter Add-on box 800 via A-IN port 802 (for a microphone) or Handset or Headset I/ O ports 807 and 808, respectively. In all cases, outgoing audio signals pass through standard preamplifier (811) and equalization (813) circuitry, whereupon the desired signal is selected by standard “Select” switching circuitry 816 (under software control through Control port 806) and passed to audio port 837 of Loopback/AV Mute circuitry, 830.
  • It is to be understood that A/V Transceivers 840 may include muxing/demuxing facilities so as to enable the transmission of audio/video signals on a single pair of wires, e.g., by encoding audio signals digitally in the vertical retrace interval of the analog video signal. Implementation of other audio and video enhancements, such as stereo audio and external audio/video I/O ports (e.g., for recording signals generated at the CMW), are also well within the capabilities of one skilled in the art. If stereo audio is used in teleconferencing (i.e., to create useful spatial metaphors for users), a second echo canceller may be recommended.
  • Another embodiment of the CMW of this invention, illustrated in FIG. 18B, utilizes a separate (fully self-contained) “Side Mount” approach which includes its own dedicated video display. This embodiment is advantageous in a variety of situations, such as instances in which additional screen display area is desired (e.g., in a laptop computer or desktop system with a small monitor) or where it is impossible or undesirable to retrofit older, existing or specialized desktop computers for audio/video support. In this embodiment, video camera 500, microphone 600 and speaker 700 of FIG. 18A are integrated together with the functionality of Add-on box 80). Side Mount 850 eliminates the necessity of external connections to these integrated audio and video I/O devices, and includes an LCD display 810 for displaying the incoming video signal (which thus eliminates the need for a base platform video input card 130).
  • Given the proximity of Side Mount device 850 to the user, and the direct access to audio/video I/O within that device, various additional controls 820 can be provided at the user's touch (all well within the capabilities of those skilled in the art). Note that, with enough additions, Side Mount unit 850 can become virtually a standalone device that does not require a separate computer for services using only audio and video. This also provides a way of supplementing a network of full-feature workstations with a few low-cost additional “audio video intercoms” for certain sectors of an enterprise (such as clerical, reception, factory floor, etc.).
  • A portable laptop implementation can be made to deliver multimedia mail with video, audio and synchronized annotations via CD-ROM or an add-on videotape unit with separate video, audio and time code tracks (a stereo videotape player can use the second audio channel for time code signals). Videotapes or CD-ROMs can be created in main offices and express mailed, thus avoiding the need for high-bandwidth networking when on the road. Cellular phone links can be used to obtain both voice and data communications (via modems) Modem-based data communications are sufficient to support remote control of mail or presentation playback, annotation, file transfer and fax features. The laptop can then be brought into the office and attached to a docking station where the available MLAN 10 and additional functions adapted from Add-on box 800 can be supplied, providing full CMW capability.
  • Collaborative Multimedia Workstation Software
  • CMW software modules 160 are illustrated generally in FIG. 20 and discussed in greater detail below in conjunction with the software running on MLAN Server 60 of FIG. 3 Software 160 allows the user to initiate and manage (in conjunction with the server software) videoconferencing, data conferencing, multimedia mail and other collaborative sessions with other users across the network.
  • Also present on the CMW in this embodiment are standard multitasking operating system/GUI software 180 (e.g., Apple Macintosh System 7, Microsoft Windows 3.1, or UNIX with the “X Window System” and Motif or other GUI “window manager” software) as well as other applications 170, such as word processing and spreadsheet programs Software modules 161-168 communicate with operating system/GUI software 180 and other applications 170 utilizing standard function calls and interapplication protocols.
  • The central component of the Collaborative Multimedia Workstation software is the Collaboration Initiator 161. All collaborative functions can be accessed through this module When the Collaboration Initiator is started, it exchanges initial configuration information with the Audio Video Network Manager (AVNM) 60 (shown in FIG. 3) through Data Network 902. Information is also sent from the Collaboration Initiator to the AVNM indicating the location of the user, the types of services available on that workstation (e.g., videoconferencing, data conferencing, telephony, etc.) and other relevant initialization information.
  • The Collaboration Initiator presents a user interface that allows the user to initiate collaborative sessions (both real-time and asynchronous). In the preferred embodiment, session participants can be selected from a graphical rolodex 163 that contains a scrollable list of user names or from a list of quick-dial buttons 162 Quick-dial buttons show the face icons for the users they represent. In the preferred embodiment, the icon representing the user is retrieved by the Collaboration Initiator from the Directory Server 66 on MLAN Server 60 when it starts up Users can dynamically add new quick-dial buttons by dragging the corresponding entries from the graphical rolodex onto the quick-dial panel.
  • Once the user elects to initiate a collaborative session, he or she selects one or more desired participants by, for example, clicking on that name to select the desired participant from the system rolodex or a personal rolodex, or by clicking on the quick-dial button for that participant (see, e.g., FIG. 2A). In either case, the user then selects the desired session type—e.g., by clicking on a CALL button to initiate a videoconference call, a SHARE button to initiate the sharing of a snapshot image or blank whiteboard, or a MAIL button to send mail. Alternatively, the user can double-click on the rolodex name or a face icon to initiate the default session type—e.g., an audio/video conference call.
  • The system also allows sessions to be invoked from the keyboard. It provides a graphical editor to bind combinations of participants and session types to certain hot keys. Pressing this hot key (possibly in conjunction with a modifier key, e.g., <Shift> or <Ctrl>) will cause the Collaboration Initiator to start a session of the specified type with the given participants.
  • Once the user selects the desired participant and session type, Collaboration Initiator module 161 retrieves necessary addressing information from Directory Service 66 (see FIG. 21). In the case of a videoconference call, the Collaboration Initiator then communicates with the AVNM (as described in greater detail below) to set up the necessary data structures and manage the various states of that call, and to control A/V Switching Circuitry 30, which selects the appropriate audio and video signals to be transmitted to/from each participant's CMW. In the case of a data conferencing session, the Collaboration Initiator locates, via the AVNM, the Collaboration Initiator modules at the CMWs of the chosen recipients, and sends a message causing the Collaboration Initiator modules to invoke the Snapshot Sharing modules 164 at each participant's CMW. Subsequent videoconferencing and data conferencing functionality is discussed in greater detail below in the context of particular usage scenarios.
  • As indicated previously, additional collaborative services—such as Mail 165, Application Sharing 166, Computer-Integrated Telephony 167 and Computer Integrated Fax 168—are also available from the CMW by utilizing Collaboration Initiator module 161 to initiate the session (i.e., to contact the participants) and to invoke the appropriate application necessary to manage the collaborative session When initiating asynchronous collaboration (e.g., mail, fax, etc.), the Collaboration Initiator contacts Directory Service 66 for address information (e.g., EMAIL address, fax number, etc.) for the selected participants and invokes the appropriate collaboration tools with the obtained address information. For real-time sessions, the Collaboration Initiator queries the Service Server module 69 inside AVNM 63 for the current location of the specified participants. Using this location information, it communicates (via the AVNM) with the Collaboration Initiators of the other session participants to coordinate session setup. As a result the various Collaboration Initiators will invoke modules 166, 167 or 168 (including activating any necessary devices such as the connection between the telephone and the CMW's audio I/O port). Further details on multimedia mail are provided below.
  • MLAN Server Software
  • FIG. 21 diagrammatically illustrates software 62 comprised of various modules (as discussed above) provided for running on MLAN Server 60 (FIG. 3) in the preferred embodiment. It is to be understood that additional software modules could also be provided. It is also to be understood that, although the software illustrated in FIG. 21 offers various significant advantages, as will become evident hereinafter, different forms and arrangements of software may also be employed within the scope of the invention. The software can also be implemented in various sub-parts running as separate processes.
  • In the preferred embodiment, clients (e.g., software-controlling workstations, VCRs, laserdisks, multimedia resources, etc.) communicate with the MLAN Server Software Modules 62 using the TCP/IP network protocols. Generally, the AVNM 63 cooperates with the Service Server 69, Conference Bridge Manager (CBM 64 in FIG. 21) and the WAN Network Manager (WNM 65 in FIG. 21) to manage communications within and among both MLANs 10 and WANs 15 (FIGS. 1 and 3).
  • The AVNM additionally cooperates with Audio/Video Storage Server 67 and other multimedia services 68 in FIG. 21 to support various types of collaborative interactions as described herein CBM 64 in FIG. 21 operates as a client of the AVNM 63 to manage conferencing by controlling the operation of conference bridges 35. This includes management of the video mosaicing circuitry 37, audio mixing circuitry 38 and cut-and-paste circuitry 39 preferably incorporated therein WNM 65 manages the allocation of paths (codecs and trunks) provided by WAN gateway 40 for accomplishing the communications to other sites called for by the AVNM.
  • Audio Video Network Manager
  • The AVNM 63 manages A/V Switching Circuitry 30 in FIG. 3 for selectively routing audio/video signals to and from CMWs 12, and also to and from WAN gateway 40, as called for by clients Audio/video devices (e.g., CMWs 12, conference bridges 35, multimedia resources 16 and WAN gateway 40 in FIG. 3) connected to A/V Switching Circuitry 30 in FIG. 3, have physical connections for audio in, audio out, video in and video out. For each device on the network, the AVNM combines these four connections into a port abstraction, wherein each port represents an addressable bidirectional audio/video channel. Each device connected to the network has at least one port Different ports may share the same physical connections on the switch. For example, a conference bridge may typically have four ports (for 2×2 mosaicing) that share the same video-out connection. Not all devices need both video and audio connections at a port. For example, a TV tuner port needs only incoming audio/video connections.
  • In response to client program requests, the AVNM provides connectivity between audio/video devices by connecting their ports Connecting ports is achieved by switching one port's physical input connections to the other port's physical output connections (for both audio and video) and vice-versa Client programs can specify which of the 4 physical connections on its ports should be switched. This allows client programs to establish unidirectional calls (e.g., by specifying that only the port's input connections should be switched and not the port's output connections) and audio-only or video-only calls (by specifying audio connections only or video connections only).
  • Service Server
  • Before client programs can access audio/video resources through the AVNM, they must register the collaborative services they provide with the Service Server 69. Examples of these services indicate “video call”, “snapshot sharing”, “conference” and “video file sharing.” These service records are entered into the Service Server's service database. The service database thus keeps track of the location of client programs and the types of collaborative sessions in which they can participate. This allows the Collaboration Initiator to find collaboration participants no matter where they are located. The service database is replicated by all Service Servers: Service Servers communicate with other Service Servers in other MLANs throughout the system to exchange their service records.
  • Clients may create a plurality of services, depending on the collaborative capabilities desired. When creating a service, a client can specify the network resources (e.g ports) that will be used by this service. In particular, service information is used to associate a user with the audio/video ports physically connected to the particular CMW into which the user is logged in Clients that want to receive requests do so by putting their services in listening mode. If clients want to accept incoming data shares, but want to block incoming video calls, they must create different services.
  • A client can create an exclusive service on a set of ports to prevent other clients from creating services on these ports. This is useful, for example, to prevent multiple conference bridges from managing the same set of conference bridge ports.
  • Next to be considered is the preferred manner in which the AVNM 63 (FIG. 21), in cooperation with the Service Server 69, CBM 64 and participating CMWs provide for managing. A/V Switching Circuitry 30 and conference bridges 35 in FIG. 3 during audio/video/data teleconferencing. The participating CMWs may include workstations located at both local and remote sites.
  • Basic Two-Party Videoconferencing
  • As previously described, a CMW includes a Collaboration Initiator software module 161 (see FIG. 20) which is used to establish person-to-person and multiparty calls. The corresponding collaboration initiator window advantageously provides quick-dial face icons of frequently dialed persons, as illustrated, for example, in FIG. 22, which is an enlarged view of typical face icons along with various initiating buttons (described in greater detail below in connection with FIGS. 35-42).
  • Videoconference calls can be initiated, for example, merely by double-clicking on these icons. When a call is initiated, the CMW typically provides a screen display that includes a live video picture of the remote conference participant, as illustrated for example in FIG. 8A. In the preferred embodiment, this display also includes control buttons/menu items that can be used to place the remote participant on hold, to resume a call on hold, to add one or more participants to the call, to initiate data sharing and to hang up the call.
  • The basic underlying software-controlled operations occurring for a two-party call are diagrammatically illustrated in FIG. 23 When a caller initiates a call (e.g., by selecting a user from the graphical rolodex and clicking the call button or by double-clicking the face icon of the callee on the quick-dial panel), the caller's Collaboration Initiator responds by identifying the selected user and requesting that user's address from Directory Service 66, as indicated by (2) in FIG. 23 Directory Service 66 looks up the callee's address in the directory database, as indicated by (3) in FIG. 23, and then returns it to the caller's Collaboration Initiator, as illustrated by (4) in FIG. 23.
  • The caller's Collaboration Initiator sends a request to the AVNM to place a video call to caller with the specified address, as indicated by (5) in FIG. 23. The AVNM queries the Service Server to find the service instance of type “video call” whose name corresponds to the callee's address. This service record identifies the location of the callee's Collaboration Initiator as well as the network ports that the callee is connected to. If no service instance is found for the callee, the AVNM notifies the caller that the callee is not logged in. If the callee is local, the AVNM sends a call event to the callee's Collaboration Initiator, as indicated by (6) in FIG. 23. If the callee is at a remote site, the AVNM forwards the call request (5) through the WAN gateway 40 for transmission, via WAN 15 (FIG. 1) to the Collaboration Initiator of the callee's CMW at the remote site.
  • The callee's Collaboration Initiator can respond to the call event in a variety of ways. In the preferred embodiment, a user-selectable sound is generated to announce the incoming call. The Collaboration Initiator can then act in one of two modes. In “Telephone Mode,” the Collaboration Initiator displays an invitation message on the CMW screen that contains the name of the caller and buttons to accept or refuse the call. The Collaboration Initiator will then accept or refuse the call, depending on which button is pressed by the callee In “Intercom Mode,” the Collaboration Initiator accepts all incoming calls automatically, unless there is already another call active on the callee's CMW, in which case behavior reverts to Telephone Mode.
  • The callee's Collaboration Initiator then notifies the AVNM as to whether the call will be accepted or refused. If the call is accepted, (7), the AVNM sets up the necessary communication paths between the caller and the callee required to establish the call. The AVNM then notifies the caller's Collaboration Initiator that the call has been established by sending it an accept event (8). If the caller and callee are at different sites, their AVNMs will coordinate in setting up the communication paths at both sites, as required by the call.
  • The AVNM may provide for managing connections among CMWs and other multimedia resources for audio/video/data communications in various ways. The manner employed in the preferred embodiment will next be described.
  • As has been described previously, the AVNM manages the switches in the A/V Switching Circuitry 30 in FIG. 3 to provide port-to-port connections in response to connection requests from clients. The primary data structure used by the AVNM for managing these connections will be referred to as a callhandle, which is comprised of a plurality of bits, including state bits.
  • Each port-to-port connection managed by the AVNM comprises two callhandles, one associated with each end of the connection. The callhandle at the client port of the connection permits the client to manage the client's end of the connection. The callhandle mode bits determine the current state of the callhandle and which of a port's four switch connections (video in, video out, audio in, audio out) are involved in a call.
  • AVNM clients send call requests to the AVNM whenever they want to initiate a call. As part of a call request, the client specifies the local service in which the call will be involved, the name of the specific port to use for the call, identifying information as to the callee, and the call mode. In response, the AVNM creates a callhandle on the caller's port.
  • All callhandles are created in the “idle” state. The AVNM then puts the caller's callhandle in the “active” state. The AVNM next creates a callhandle for the callee and sends it a call event, which places the callee's callhandle in the “ringing” state. When the callee accepts the call, its callhandle is placed in the “active” state, which results in a physical connection between the caller and the callee. Each port can have an arbitrary number of callhandles bound to it, but typically only one of these callhandles can be active at the same time.
  • After a call has been set up, AVNM clients can send requests to the AVNM to change the state of the call, which can advantageously be accomplished by controlling the callhandle states. For example, during a call, a call request from another party could arrive. This arrival could be signaled to the user by providing an alert indication in a dialogue box on the user's CMW screen. The user could refuse the call by clicking on a refuse button in the dialogue box, or by clicking on a “hold” button on the active call window to put the current call on hold and allow the incoming call to be accepted.
  • The placing of the currently active call on hold can advantageously be accomplished by changing the caller's callhandle from the active state to a “hold” state, which permits the caller to answer incoming calls or initiate new calls, without releasing the previous call. Since the connection set-up to the callee will be retained, a call on hold can conveniently be resumed by the caller clicking on a resume button on the active call window, which returns the corresponding callhandle back to the active state. Typically, multiple calls can be put on hold in this manner. As an aid in managing calls that are on hold, the CMW advantageously provides a hold list display, identifying these on-hold calls and (optionally) the length of time that each party is on hold. A corresponding face icon could be used to identify each on-hold call. In addition, buttons could be provided in this hold display which would allow the user to send a preprogrammed message to a party on hold. For example, this message could advise the callee when the call will be resumed, or could state that the call is being terminated and will be reinitiated at a later time.
  • Reference is now directed to FIG. 24 which diagrammatically illustrates how two-party calls are connected for CMWs WS-1 and WS-2, located at the same MLAN 10. As shown in FIG. 24, CMWs WS-1 and WS-2 are coupled to the local A/V Switching Circuitry 30 via ports 81 and 82, respectively. As previously described, when CMW WS-1 calls CMW WS-2, a callhandle is created for each port. If CMW WS-2 accepts the call, these two callhandles become active and in response thereto, the AVNM causes the A/V Switching Circuitry 30 to set up the appropriate connections between ports 81 and 82, as indicated by the dashed line 83.
  • FIG. 25 diagrammatically illustrates how two-party calls are connected for CMWs WS-1 and WS-2 when located in different MLANs 10 a and 10 b. As illustrated in FIG. 25, CMW WS-1 of MLAN 10 a is connected to a port 91 a of A/V Switching Circuitry 30 a of MLAN 10 a, while CMW WS-2 is connected to a port 91 b of the A/V Switching Circuitry 30 b of MLAN 10 b. It will be assumed that MLANs 10 a and 10 b can communicate with each other via ports 92 a and 92 b (through respective WAN gateways 40 a and 40 b and WAN 15). A call between CMWs WS-1 and WS-2 can then be established by AVNM of MLAN 10 a in response to the creation of callhandles at ports 91 a and 92 a, setting up appropriate connections between these ports as indicated by dashed line 93 a, and by AVNM of. MLAN 10 b, in response to callhandles created at ports 91 b and 92 b, setting up appropriate connections between these ports as indicated by dashed line 93 b. Appropriate paths 94 a and 94 b in WAN gateways 40 a and 40 b, respectively, are set up by the WAN network manager 65 (FIG. 21) in each network.
  • Conference Calls
  • Next to be described is the specific manner in which the preferred embodiment provides for multi-party conference calls (involving more than two participants) When a multi-party conference call is initiated, the CMW provides a screen that is similar to the screen for two-party calls, which displays a live video picture of the callee's image in a video window. However, for multi-party calls, the screen includes a video mosaic containing a live video picture of each of the conference participants (including the CMW user's own picture), as shown, for example, in FIG. 8B. Of course, other embodiments could show only the remote conference participants (and not the local CMW user) in the conference mosaic (or show a mosaic containing both participants in a two-party call). In addition to the controls shown in FIG. 8B, the multi-party conference screen also includes buttons/menu items that can be used to place individual conference participants on hold, to remove individual participants from the conference, to adjourn the entire conference, or to provide a “close-up” image of a single individual (in place of the video mosaic).
  • Multi-party conferencing requires all the mechanisms employed for 2-party calls. In addition, it requires the conference bridge manager CBM 64 (FIG. 21) and the conference bridges 36 (FIG. 3). The CBM acts as a client of the AVNM in managing the operation of the conference bridges 36. The CBM also acts a server to other clients on the network. The CBM makes conferencing services available by creating service records of type “conference” in the AVNM service database and associating these services with the ports on A/V Switching Circuitry 30 for connection to conference bridges 36.
  • The preferred embodiment provides two ways for initiating a conference call. The first way is to add one or more parties to an existing two-party call. For this purpose, an ADD button is provided by both the Collaboration Initiator and the Rolodex, as illustrated in FIGS. 2A and 22. To add a new party, a user selects the party to be added (by clicking on the user's rolodex name or face icon as described above) and clicks on the ADD button to invite that new party. Additional parties can be invited in a similar manner. The second way to initiate a conference call is to select the parties in a similar manner and then click on the CALL button (also provided in the Collaboration Initiator and Rolodex windows on the user's CMW screen).
  • Another alternative embodiment is to initiate a conference call from the beginning by clicking on a CONFERENCE/MOSAIC icon/button/menu item on the CMW screen. This could initiate a conference call with the call initiator as the sole participant (i.e., causing a conference bridge to be allocated such that the caller's image also appears on his/her own screen in a video mosaic, which will also include images of subsequently added participants). New participants could be invited, for example, by selecting each new party's face icon and then clicking on the ADD button.
  • Next to be considered with reference to FIGS. 26 and 27 is the manner in which conference calls are handled in the preferred embodiment. For the purposes of this description it will be assumed that up to four parties may participate in a conference call. Each conference uses four bridge ports 136-1, 136-2, 136-3 and 136-4 provided on A/V Switching Circuitry 30 a, which are respectively coupled to bidirectional audio/video lines 36-1, 36-2, 36-3 and 36-4 connected to conference bridge 36. However, from this description it will be apparent how a conference call may be provided for additional parties, as well as simultaneously occurring conference calls.
  • Once the Collaboration Initiator determines that a conference is to be initiated, it queries the AVNM for a conference service. If such a service is available, the Collaboration Initiator requests the associated CBM to allocate a conference bridge. The Collaboration Initiator then places an audio/video call to the CBM to initiate the conference When the CBM accepts the call, the AVNM couples port 101 of CMW WS-1 to lines 36-1 of conference bridge 36 by a connection 137 produced in response to callhandles created for port 101 of WS-1 and bridge port 136-1.
  • When the user of WS-1 selects the appropriate face icon and clicks the ADD button to invite a new participant to the conference, which will be assumed to be CMW WS-3, the Collaboration initiator on WS-1 sends an add request to the CBM. In response, the CBM calls WS-3 via WS-3 port 103 When CBM initiates the call, the AVNM creates callhandles for WS-3 port 103 and bridge port 136-2 When WS-3 accepts the call, its callhandle is made “active,” resulting in connection 138 being provided to connect WS-3 and lines 136-2 of conference bridge 36. Assuming CMW WS-1 next adds CMW WS-5 and then CMW WS-8, callhandles for their respective ports and bridge ports 136-3 and 136-4 are created, in turn, as described above for WS-1 and WS-3, resulting in connections 139 and 140 being provided to connect WS-5 and WS-9 to conference bridge lines 36-3 and 36-4, respectively. The conferees WS-1, WS-3, WS-5 and WS-8 are thus coupled to conference bridge lines 136-1,136-2, 136-3 and 136-4, respectively as shown in FIG. 26.
  • It will be understood that the video mosaicing circuitry 36 and audio mixing circuitry 38 incorporated in conference bridge 36 operate as previously described, to form a resulting four-picture mosaic (FIG. 8B) that is sent to all of the conference participants, which in this example are CMWs WS-1, WS-2, WS-5 and WS-8 Users may leave a conference by just hanging up, which causes the AVNM to delete the associated callhandles and to send a hangup notification to CBM When CBM receives the notification, it notifies all other conference participants that the participant has exited. In the preferred embodiment, this results in a blackened portion of that participant's video mosaic image being displayed on the screen of all remaining participants.
  • The manner in which the CBM and the conference bridge 36 operate when conference participants are located at different sites will be evident from the previously described operation of the cut-and-paste circuitry 39 (FIG. 10) with the video mosaicing circuitry 36 (FIG. 7) and audio mixing circuitry 38 (FIG. 9). In such case, each incoming single video picture or mosaic from another site is connected to a respective one of the conference bridge lines 36-1 to 36-4 via WAN gateway 40.
  • The situation in which a two-party call is converted to a conference call will next be considered in connection with FIG. 27 and the previously considered 2-party call illustrated in FIG. 24 Converting this 2-party call to a conference requires that this two-party call (such as illustrated between WS-1 and WS-2 in FIG. 24) be rerouted dynamically so as to be coupled through conference bridge 36 When the user of WS-1 clicks on the ADD button to add a new party (for example WS-5), the Collaboration Initiator of WS-1 sends a redirect request to the AVNM, which cooperates with the CBM to break the two-party connection 83 in FIG. 24, and then redirect the callhandles created for ports 81 and 83 to callhandles created for bridge ports 136-1 and 136-2, respectively.
  • As shown in FIG. 27, this results in producing a connection 86 between WS-1 and bridge port 136-1, and a connection 87 between WS-2 and bridge port 136-2, thereby creating a conference set-up between WS-1 and WS-2. Additional conference participants can then be added as described above for the situations described above in which the conference is initiated by the user of WS-1 either selecting multiple participants initially or merely selecting a “conference” and then adding subsequent participants.
  • Having described the preferred manner in which two-party calls and conference calls are set up in the preferred embodiment, the preferred manner in which data conferencing is provided between CMWs will next be described.
  • Data Conferencing
  • Data conferencing is implemented in the preferred embodiment by certain Snapshot Sharing software provided at the CMW (see FIG. 20). This software permits a “snapshot” of a selected portion of a participant's CMW screen (such as a window) to be displayed on the CMW screens of other selected participants (whether or not those participants are also involved in a videoconference). Any number of snapshots may be shared simultaneously. Once displayed, any participant can then telepoint on or annotate the snapshot, which animated actions and results will appear (virtually simultaneously) on the screens of all other participants. The annotation capabilities provided include lines of several different widths and text of several different sizes. Also, to facilitate participant identification, these annotations may be provided in a different color for each participant. Any annotation may also be erased by any participant FIG. 2B (lower left window) illustrates a CMW screen having a shared graph on which participants have drawn and typed to call attention to or supplement specific portions of the shared image.
  • A participant may initiate data conferencing with selected participants (selected and added as described above for videoconference calls) by clicking on a SHARE button on the screen (available in the Rolodex or Collaboration Initiator windows, shown in FIG. 2A, as are CALL and ADD buttons), followed by selection of the window to be shared When a participant clicks on his SHARE button, his Collaboration Initiator module 161 (FIG. 20) queries the AVNM to locate the Collaboration Initiators of the selected participants, resulting in invocation of their respective Snapshot Sharing modules 164. The Snapshot Sharing software modules at the CMWs of each of the selected participants query their local operating system 180 to determine available graphic formats, and then send this information to the initiating Snapshot Sharing module, which determines the format that will produce the most advantageous display quality and performance for each selected participant
  • After the snapshot to be shared is displayed on all CMWs, each participant may telepoint on or annotate the snapshot, which actions and results are displayed on the CMW screens of all participants. This is preferably accomplished by monitoring the actions made at the CMW (e.g., by tracking mouse movements) and sending these “operating system commands” to the CMWs of the other participants, rather than continuously exchanging bitmaps, as would be the case with traditional “remote control” products.
  • As illustrated in FIG. 28, the original unchanged snapshot is stored in a first bitmap 210 a. A second bitmap 210 b stores the combination of the original snapshot and any annotations. Thus, when desired (e.g., by clicking on a CLEAR button located in each participant's Share window, as illustrated in FIG. 2B), the original unchanged snapshot can be restored (i.e., erasing all annotations) using bitmap 210 a Selective erasures can be accomplished by copying into (i.e., restoring) the desired erased area of bitmap 210 b with the corresponding portion from bitmap 210 a.
  • Rather than causing a new Share window to be created whenever a snapshot is shared, it is possible to replace the contents of an existing Share window with a new image. This can be achieved in either of two ways. First, the user can click on the GRAB button and then select a new window whose contents should replace the contents of the existing Share window. Second, the user can click on the REGRAB button to cause a (presumably modified) version of the original source window to replace the contents of the existing Share window. This is particularly useful when one participant desires to share a long document that cannot be displayed on the screen in its entirety. or example, the user might display the first page of a spreadsheet on his screen, use the SHARE button to share that page, discuss and perhaps annotate it, then return to the spreadsheet application to position to the next page, use the REGRAB button to share the new page, and so on. This mechanism represents a simple, effective step toward application sharing.
  • Further, instead of sharing a snapshot of data on his current screen, a user may instead choose to share a snapshot that had previously been saved as a file. This is achieved via the LOAD button, which causes a dialogue box to appear, prompting the user to select a file Conversely, via the SAVE button, any snapshot may be saved, with all current annotations.
  • The capabilities described above were carefully selected to be particularly effective in environments where the principal goal is to share existing information, rather than to create new information. In particular, user interfaces are designed to make snapshot capture, telepointing and annotation extremely easy to use. Nevertheless, it is also to be understood that, instead of sharing snapshots, a blank “whiteboard” can also be shared (via the WHITEBOARD button provided by the Rolodex, Collaboration Initiator, and active call windows), and that more complex paintbox capabilities could easily be added for application areas that require such capabilities.
  • As pointed out previously herein, important features of the present invention reside in the manner in which the capabilities and advantages of multimedia mail (MMM), multimedia conference recording (MMCR), and multimedia document management (MMDM) are tightly integrated with audio/video/data teleconferencing to provide a multimedia collaboration system that facilitates an unusually higher level of communication and collaboration between geographically dispersed users than has heretofore been achievable by known prior an systems FIG. 29 is a schematic and diagrammatic view illustrating how multimedia calls/conferences, MMCR, MMM and MMDM work together to provide the above-described features. In the preferred embodiment MM Editing Utilities shown supplementing MMM and MMDM may be identical.
  • Having already described various preferred embodiments and examples of audio/video/data teleconferencing, next to be considered are various preferred ways of integrating MMCR, MMM and MMDM with audio/video/data teleconferencing in accordance with the invention, or this purpose, basic preferred approaches and features of each will be considered along with preferred associated hardware and software.
  • Multimedia Documents
  • In the preferred embodiment, the creation, storage, retrieval and editing of multimedia documents serve as the basic element common to MMCR, MMM and MMDM. Accordingly, the preferred embodiment advantageously provides a universal format for multimedia documents. This format defines multimedia documents as a collection of individual components in multiple media combined with an overall structure and timing component that captures the identities, detailed dependencies, references to, and relationships among the various other components. The information provided by this structuring component forms the basis for spatial layout, order of presentation, hyperlinks, temporal synchronization, etc., with respect to the composition of a multimedia document FIG. 30 shows the structure of such documents as well as their relationship with editing and storage facilities.
  • Each of the components of a multimedia document uses its own editors for creating, editing, and viewing. In addition, each component may use dedicated storage facilities. In the preferred embodiment, multimedia documents are advantageously structured for authoring, storage, playback and editing by storing some data under conventional file systems and some data in special-purpose storage servers as will be discussed later. The Conventional File System 504 can be used to store all non-time-sensitive portions of a multimedia document. In particular, the following are examples of non-time-sensitive data that can be stored in a conventional type of computer file system:
  • 1. structured and unstructured text
  • 2. raster images
  • 3. structured graphics and vector graphics (e.g., PostScript)
  • 4. references to files in other file systems (video, hi-fidelity audio, etc.) via pointers
  • 5. restricted forms of executables
  • 6. structure and timing information for all of the above (spatial layout, order of presentation, hyperlinks, temporal synchronization, etc.)
  • Of particular importance in multimedia documents is support for time-sensitive media and media that have synchronization requirements with other media components. Some of these time-sensitive media can be stored on conventional file systems while others may require special-purpose storage facilities.
  • Examples of time-sensitive media that can be stored on conventional file systems are small audio files and short or low-quality video clips (e.g as might be produced using QuickTime or Video for Windows). Other examples include window event lists as supported by the Window-Event Record and Play system 512 shown in FIG. 30. This component allows for storing and replaying a user's interactions with application programs by capturing the requests and events exchanged between the client program and the window system in a time-stamped sequence. After this “record” phase, the resulting information is stored in a conventional file that can later be retrieved and “played” back. During playback the same sequence of window system requests and events reoccurs with the same relative timing as when they were recorded. In prior-art systems, this capability has been used for creating automated demonstrations. In the present invention it can be used, for example, to reproduce annotated snapshots as they occurred at recording
  • As described above in connection with collaborative workstation software, Snapshot Share 514 shown in FIG. 30 is a utility used in multimedia calls and conferencing for capturing window or screen snapshots, sharing with one or more call or conference participants, and permitting group annotation, telepointing, and re-grabs. Here, this utility is adapted so that its captured images and window events can be recorded by the Window-Event Record and Play system 512 while being used by only one person. By synchronizing events associated with a video or audio stream to specific frame numbers or time codes, a multimedia call or conference can be recorded and reproduced in its entirety. Similarly, the same functionality is preferably used to create multimedia mail whose authoring steps are virtually identical to participating in a multimedia call or conference (though other forms of MMM are not precluded).
  • Some time-sensitive media require dedicated storage servers in order to satisfy real-time requirements. High-quality audio/video segments, for example, require dedicated real-time audio/video storage servers. A preferred embodiment of such a server will be described later. Next to be considered is how the current invention guarantees synchronization between different media components.
  • Media Synchronization
  • A preferred manner for providing multimedia synchronization in the preferred embodiment will next be considered. Only multimedia documents with real-time material need include synchronization functions and information Synchronization for such situations may be provided as described below.
  • Audio or video segments can exist without being accompanied by the other. If audio and video are recorded simultaneously (“co-recorded”), the preferred embodiment allows the case where their streams are recorded and played back with automatic synchronization—as would result from conventional VCRs, laserdisks, or time-division multiplexed (“interleaved”) audio/video streams. This excludes the need to tightly synchronize (i.e., “lip-sync”) separate audio and video sequences. Rather, reliance is on the co-recording capability of the Real-Time Audio/Video Storage Server 502 to deliver all closely synchronized audio and video directly at its signal outputs.
  • Each recorded video sequence is tagged with time codes (e.g. SMPTE at 1/30 second intervals) or video frame numbers. Each recorded audio sequence is tagged with time codes (e.g., SMPTE or MIDI) or, if co-recorded with video, video frame numbers. The preferred embodiment also provides synchronization between window events and audio and/or video streams. The following functions are supported:
  • 1. Media-time-driven Synchronization: synchronization of window events to an audio, video, or audio/video stream, using the real-time media as the timing source.
  • 2. Machine-time-driven-Synchronization:
      • a. synchronization of window events to the system clock
      • b. synchronization of the start of an audio, video, or audio/video segment to the system clock
  • If no audio or video is involved, machine-time-driven synchronization is used throughout the document Whenever audio and/or video is playing, media-time-synchronization is used. The system supports transition between machine-time and media-time synchronization whenever an audio/video segment is started or stopped.
  • As an example, viewing a multimedia document might proceed as follows:
      • Document starts with an annotated share (machine-time-driven synchronization).
      • Next, start audio only (a “voice annotation”) as text and graphical annotations on the share continue (audio is timing source for window events).
      • Audio ends, but annotations continue (machine-time-driven synchronization).
      • Next, start co-recorded audio/video continuing with further annotations on same share (audio is timing source for window events).
      • Next, start a new share during the continuing audio/video recording; annotations happen on both shares (audio is timing source for window events).
      • Audio/video stops, annotations on both shares continue (machine-time-driven synchronization).
      • Document ends.
    Audio/Video Storage
  • As described above, the present invention can include many special-purpose servers that provide storage of time-sensitive media (e.g. audio/video streams) and support coordination with other media. This section describes the preferred embodiment for audio/video storage and recording services.
  • Although storage and recording services could be provided at each CMW, it is preferable to employ a centralized server 502 coupled to MLAN 10, as illustrated in FIG. 31. A centralized server 502, as shown in FIG. 31, provides the following advantages:
  • 1. The total amount of storage hardware required can be far less (due to better utilization resulting from statistical averaging).
  • 2. Bulky and expensive compression/decompression hardware can be pooled on the storage servers and shared by multiple clients. As a result, fewer compression/decompression engines of higher performance are required than if each workstation were equipped with its own compression/decompression hardware.
  • 3. Also, more costly centralized codecs can be used to transfer mail wide area among campuses at far lower costs than attempting to use data WAN technologies.
  • 4. File system administration (e.g. backups and file system replication, etc.) are far less costly and higher performance.
  • The Real-Time Audio/Video Storage Server 502 shown in FIG. 31A structures and manages the audio/video files recorded and stored on its storage devices Storage devices may typically include computer-controlled VCRs, as well as rewritable magnetic or optical disks. For example, server 502 in FIG. 31 A includes disks 60 e for recording and playback. Analog information is transferred between disks 60 e and the. A/V Switching Circuitry 30 via analog I/O 62 Control is provided by control 64 coupled to Data LAN hub 25.
  • At a high level, the centralized audio/video storage and playback server 502 in FIG. 31A performs the following functions:
      • 1. File Management
        • It provides mechanisms for creating, naming, time-stamping, storing, retrieving copying, deleting, and playing back some or all portions of an audio/video file.
      • 2. File Transfer and Replication
        • The audio/video file server supports replication of files on different disks managed by the same file server to facilitate simultaneous access to the same files. Moreover, file transfer facilities are provided to support transmission of audio/video files between itself and other audio/video storage and playback engines. File transfer can also be achieved by using the underlying audio/video network facilities: servers establish a real-time audio/video network connection between themselves so one server can “play back” a file while the second server simultaneously records it.
      • 3. Disk Management
        • The storage facilities support specific disk allocation, garbage collection and defragmentation facilities. They also support mapping disks with other disks (for replication and staging modes, as appropriate) and mapping disks, via I/O equipment, with the appropriate Video/Audio network port.
      • 4. Synchronization support
        • Synchronization between audio and video is ensured by the multiplexing scheme used by the storage media, typically by interleaving the audio and video streams in a time-division-multiplexed fashion. Further, if synchronization is required with other stored media (such as window system graphics), then frame numbers, time codes, or other timing events are generated by the storage server. An advantageous way of providing this synchronization in the preferred embodiment is to synchronize record and playback to received frame number or time code events.
      • 5. Searching
        • To support intra-file searching, at least start, stop, pause, fast forward, reverse, and fast reverse operations are provided. To support inter-file searching, audio/video tagging, or more generalized “go-to” operations and mechanisms, such as frame numbers or time code, are supported at a search-function level.
      • 6. Connection Management
        • The server handles requests for audio/video network connections from client programs (such as video viewers and editors running on client workstations) for real-time recording and real-time playback of audio/video files.
  • Next to be considered is how centralized audio/video storage servers provide for real-time recording and playback of video streams.
  • Real-Time Disk Delivery
  • To support real-time audio/video recording and playback, the storage server needs to provide a real-time transmission path between the storage medium and the appropriate audio/video network port for each simultaneous client accessing the server. For example, if one user is viewing a video file at the same time several other people are creating and storing new video files on the same disk, multiple simultaneous paths to the storage media are required. Similarly, video mail sent to large distribution groups, video databases, and similar functions may also require simultaneous access to the same video files, again imposing multiple access requirements on the video storage capabilities.
  • For storage servers that are based on computer-controlled VCRs or rewritable laserdisks, a real-time transmission path is readily available through the direct analog connection between the disk or tape and the network port. However, because of this single direct connection, each VCR or laserdisk can only be accessed by one client program at the same time (multi-head laserdisks are an exception). Therefore, storage servers based on VCRs and laserdisks are difficult to scale for multiple access usage. In the preferred embodiment, multiple access to the same material is provided by fie replication and staging, which greatly increases storage requirements and the need for moving information quickly among storage media units serving different users.
  • Video systems based on magnetic disks are more readily scalable for simultaneous use by multiple people. A generalized hardware implementation of such a scalable storage and playback system-502 is illustrated in FIG. 32. Individual I/O cards 530 supporting digital and analog I/O are linked by intra-chassis digital networking (e.g. buses) for fie transfer within chassis 532 holding some number of these cards Multiple chassis 532 are linked by inter-chassis networking. The Digital Video Storage System available from Parallax Graphics is an example of such a system implementation.
  • The bandwidth available for the transfer of files among disks is ultimately limited by the bandwidth of these intra-chassis and inter-chassis networking. or systems that use sufficiently powerful video compression schemes, real-time delivery requirements for a small number of users can be met by existing file system software (such as the Unix file system), provided that the block-size of the storage system is optimized for video storage and that sufficient buffering is provided by the operating system software to guarantee continuous flow of the audio/video data.
  • Special-purpose software/hardware solutions can be provided to guarantee higher performance under heavier usage or higher bandwidth conditions, for example, a higher throughput version of FIG. 32 is illustrated in FIG. 33, which uses crosspoint switching, such as provided by SCSI Crossbar 540, which increases the total bandwidth of the inter-chassis and intra-chassis network, thereby increasing the number of possible simultaneous file transfers.
  • Real-Time Network Delivery
  • By using the same audio/video format as used for audio/video teleconferencing, the audio/video storage system can leverage the previously described network facilities: the MLANs 10 can be used to establish a multimedia network connection between client workstations and the audio/video storage servers Audio/Video editors and viewers running on the client workstation use the same software interfaces as the multimedia teleconferencing system to establish these network connections.
  • The resulting architecture is shown in FIG. 31B Client workstations use the existing audio/video network to connect to the storage server's network ports. These network ports are connected to compression/decompression engines that plug into the server bus. These engines compress the audio/video streams that come in over the network and store them on the local disk. Similarly, for playback, the server reads stored video segments from its local disk and routes them through the decompression engines back to client workstations for local display.
  • The present invention allows for alternative delivery strategies. or example, some compression algorithms are asymmetric, meaning that decompression requires much less compute power than compression. In some cases, real-time decompression can even be done in software, without requiring any special-purpose decompression hardware. As a result, there is no need to decompress stored audio and video on the storage server and play it back in realtime over the network. Instead, it can be more efficient to transfer an entire audio/video file from the storage server to the client workstation, cache it on the workstation's disk, and play it back locally. These observations lead to a modified architecture as presented in FIG. 31C. In this architecture, clients interact with the storage server as follows:
      • To record video, clients set up real-time audio/video network connections to the storage server as before (this connection could make use of an analog line).
      • In response to a connection request, the storage server allocates a compression module to the new client.
      • As soon as the client starts recording, the storage server routes the output from the compression hardware to an audio/video file allocated on its local storage devices.
      • For playback, this audio/video file gets transferred over the data network to the client workstation and pre-staged on the workstation's local disk.
      • The client uses local decompression software and/or hardware to play back the audio/video on its local audio and video hardware.
  • This approach frees up audio/video network ports and compression/decompression engines on the server. As a result, the server is scaled to support a higher number of simultaneous recording sessions, thereby further reducing the cost of the system. Note that such an architecture can be part of a preferred embodiment for reasons other than compression/decompression asymmetry (such as the economics of the technology of the day, existing embedded base in the enterprise, etc.).
  • Multimedia Conference Recording
  • Multimedia conference recording (MMCR) will next be considered. For full-feature multimedia desktop calls and conferencing (e.g. audio/video calls or conferences with snapshot share), recording (storage) capabilities are preferably provided for audio and video of all parties, and also for all shared windows, including any telepointing and annotations provided during the teleconference. Using the multimedia synchronization facilities described above, these capabilities are provided in a way such that they can be replayed with accurate correspondence in time to the recorded audio and video, such as by synchronizing to frame numbers or time code events.
  • A preferred way of capturing audio and video from calls would be to record all calls and conferences as if they were multi-party conferences (even for two-party calls), using video mosaicing, audio mixing and cut-and-pasting, as previously described in connection with FIGS. 7-11. It will be appreciated that MMCR as described will advantageously permit users at their desktop to review real-time collaboration as it previously occurred, including during a later teleconference. The output of a MMCR session is a multimedia document that can be stored, viewed, and edited using the multimedia document facilities described earlier.
  • FIG. 31D shows how conference recording relates to the various system components described earlier. The Multimedia Conference Record/Play system 522 provides the user with the additional GUIs (graphical user interfaces) and other functions required to provide the previously described MMCR functionality.
  • The Conference Invoker 518 shown in FIG. 31D is a utility that coordinates the audio/video calls that must be made to connect the audio/video storage server 502 with special recording outputs on conference bridge hardware (35 in FIG. 3). The resulting recording is linked to information identifying the conference, a function also performed by this utility.
  • Multimedia Mail
  • Now considering multimedia mail (MMM), it will be understood that MMM adds to the above-described MMCR the capability of delivering delayed collaboration, as well as the additional ability to review the information multiple times and, as described hereinafter, to edit, re-send, and archive it. The captured information is preferably a superset of that captured during MMCR, except that no other user is involved and the user is given a chance to review and edit before sending the message.
  • The Multimedia Mail system 524 in FIG. 31D provides the user with the additional GUIs and other functions required to provide the previously described MMM functionality Multimedia Mail relies on a conventional Email system 506 shown in FIG. 31D for creating, transporting, and browsing messages. However, multimedia document editors and viewers are used for creating and viewing message bodies. Multimedia documents (as described above) consist of time-insensitive components and time sensitive components. The Conventional Email system 506 relies on the Conventional File system 504 and Real-Time Audio/Video Storage Server 502 for storage support. The time insensitive components are transported within the Conventional Email system 506, while the real-time components may be separately transported through the audio/video network using file transfer utilities associated with the Real-Time Audio/Video Storage Server 502.
  • Multimedia Document Management
  • Multimedia document management (MMDM) provides long-term, high-volume storage for MMCR and MMM. The MMDM system assists in providing the following capabilities to a CMW user:
  • 1. Multimedia documents can be authored as mail in the MMM system or as call/conference recordings in the MMCR system and then passed on to the MMDM system.
  • 2. To the degree supported by external compatible multimedia editing and authoring systems, multimedia documents can also be authored by means other than MMM and MMCR.
  • 3. Multimedia documents stored within the MMDM system can be reviewed and searched.
  • 4. Multimedia documents stored within the MMDM system can be used as material in the creation of subsequent MMM.
  • 5. Multimedia documents stored within the MMDM system can be edited to create other multimedia documents.
  • The Multimedia Document Management system 526 in FIG. 31D provides the user with the additional GUIs and other functions required to provide the previously described MMDM functionality. The MMDM includes sophisticated searching and editing capabilities in connection with the MMDM multimedia document such that a user can rapidly access desired selected portions of a stored multimedia document. The Specialized Search system 520 in FIG. 31D comprises utilities that allow users to do more sophisticated searches across and within multimedia documents. This includes context-based and content-based searches (employing operations such as speech and image recognition, information filters, etc.), time-based searches, and event-based searches (window events, call management events, speech/audio events, etc.).
  • Classes of Collaboration
  • The resulting multimedia collaboration environment achieved by the above-described integration of audio/video/data teleconferencing, MMCR, MMM and MMDM is illustrated in FIG. 34. It will be evident that each user can collaborate with other users in real-time despite separations in space and time. In addition, collaborating users can access information already available within their computing and information systems, including information captured from previous collaborations. Note in FIG. 34 that space and time separations are supported in the following ways:
      • 1. Same time, different place
        • Multimedia calls and conferences
      • 2. Different time, same place
        • MMDM access to stored MMCR and MMM information, or use of MMM directly (i.e., copying mail to oneself)
      • 3. Different time, different place
        • MMM
      • 4. Same time, same place
        • Collaborative, face-to-face, multimedia document creation
  • By use of the same user interfaces and network functions, the present invention smoothly spans these three venus.
  • Remote Access to Expertise
  • In order to illustrate how the present invention may be implemented and operated, an exemplary preferred embodiment will be described having features applicable to the aforementioned scenario involving remote access to expertise. It is to be understood that this exemplary embodiment is merely illustrative, and is not to be considered as limiting the scope of the invention, since the invention may be adapted for other applications (such as in engineering and manufacturing) or uses having more or less hardware, software and operating features and combined in various ways.
  • Consider the following scenario involving access from remote sites to an in-house corporate “expert” in the trading of financial instruments such as in the securities market:
  • The focus of the scenario revolves around the activities of a trader who is a specialist in securities. The setting is the start of his day at his desk in a major financial center (NYC) at a major U.S. investment bank.
  • The Expert has been actively watching a particular security over the past week and upon his arrival into the office, he notices it is on the rise. Before going home last night, he previously set up his system to filter overnight news on a particular family of securities and a security within that family. He scans the filtered news and sees a story that may have a long-term impact on this security in question. He believes he needs to act now in order to get a good price on the security. Also, through filtered mail, he sees that his counterpart in London, who has also been watching this security, is interested in getting our Expert's opinion once he arrives at work.
  • The Expert issues a multimedia mail message on the security to the head of sales worldwide for use in working with their client base. Also among the recipients is an analyst in the research department and his counterpart in London. The Expert, in preparation for his previously established “on-call” office hours, consults with others within the corporation (using the videoconferencing and other collaborative techniques described above), accesses company records from his CMW, and analyzes such information, employing software-assisted analytic techniques. His office hours are now at hand, so he enters “intercom” mode, which enables incoming calls to appear automatically (without requiring the Expert to “answer his phone” and elect to accept or reject the call).
  • The Expert's computer beeps, indicating an incoming call, and the image of a field representative 201 and his client 202 who are located at a bank branch somewhere in the U.S. appears in video window 203 of the Expert's screen (shown in FIG. 35). Note that, unless the call is convened to a “conference” call (whether explicitly via a menu selection or implicitly by calling two or more other participants or adding a third participant to a call), the callers will see only each other in the video window and will not see themselves as part of a video mosaic.
  • Also illustrated on the Expert's screen in FIG. 35 is the Collaboration Initiator window 204 from which the Expert can (utilizing Collaboration Initiator software module 161 shown in FIG. 20) initiate and control various collaborative sessions. For example, the user can initiate with a selected participant a video call (CALL button) or the addition of that selected participant to an existing video call (ADD button), as well as a share session (SHARE button) using a selected window or region on the screen (or a blank region via the WHITEBOARD button for subsequent annotation). The user can also invoke his MAIL software (MAIL button) and prepare outgoing or check incoming Email messages (the presence of which is indicated by a picture of an envelope in the dog's mouth in In Box icon 205), as well as check for “I called” messages from other callers (MESSAGES button) left via the LEAVE WORD button in video window 203 Video window 203 also contains buttons from which many of these and certain additional features can be invoked, such as hanging up a video call (HANGUP button), putting a call on hold (HOLD button), resuming a call previously put on hold (RESUME button) or muting the audio portion of a call (MUTE button). In addition, the user can invoke the recording of a conference by the conference RECORD button. Also present on the Expert's screen is a standard desktop window 206 containing icons from which other programs (whether or not part of this invention) can be launched.
  • Returning to the example, the Expert is now engaged in a videoconference with field representative 201 and his client 202. In the course of this videoconference, as illustrated in FIG. 36, the field representative shares with the Expert a graphical image 210 (pie chart of client portfolio holdings) of his client's portfolio holdings (by clicking on his SHARE button, corresponding to the SHARE button in video window 203 of the Expert's screen, and selecting that image from his screen, resulting in the shared image appearing in the Share window 211 of the screen of all participants to the share) and begins to discuss the client's investment dilemma. The field representative also invokes a command to secretly bring up the client profile on the Expert's screen.
  • After considering this information, reviewing the shared portfolio and asking clarifying questions, the Expert illustrates his advice by creating (using his own modeling software) and sharing a new graphical image 220 (FIG. 37) with the field representative and his client. Either party to the share can annotate that image using the drawing tools 221 (and the TEXT button, which permits typed characters to be displayed) provided within Share window 211, or “regrab” a modified version of the original image (by using the REGRAB button), or remove all such annotations (by using the CLEAR button of Share window 211), or “grab” a new image to share (by clicking on the GRAB button of Share window 211 and selecting that new image from the screen). In addition, any participant to a shared session can add a new participant by selecting that participant from the rolodex or quick-dial list (as described above for video calls and for data conferencing) and clicking the ADD button of Share window 211. One can also save the shared image (SAVE button), load a previously saved image to be shared (LOAD button), or print an image (PRINT button).
  • While discussing the Expert's advice, field representative 201 makes annotations 222 to image 220 in order to illustrate his concerns. While responding to the concerns of field representative 201, the Expert hears a beep and receives a visual notice (New Call window 223) on his screen (not visible to the field representative and his client), indicating the existence of a new incoming call and identifying the caller. At this point, the Expert can accept the new call (ACCEPT button), refuse the new call (REFUSE button, which will result in a message being displayed on the caller's screen indicating that the Expert is unavailable) or add the new caller to the Expert's existing call (ADD button). In this case, the Expert elects yet another option (not shown)—to defer the call and leave the caller a standard message that the Expert will call back in X minutes (in this case 1 minute). The Expert then elects also to defer his existing call, telling the field representative and his client that he will call them back in 5 minutes, and then elects to return the initial deferred call.
  • It should be noted that the Expert's act of deferring a call results not only in a message being sent to the caller, but also in the caller's name (and perhaps other information associated with the call, such as the time the call was deferred or is to be resumed) being displayed in a list 230 (see FIG. 38) on the Expert's screen from which the call can be reinitiated. Moreover, the “state” of the call (e.g., the information being shared) is retained so that it can be recreated when the call is reinitiated Unlike a “hold” (described above), deferring a call actually breaks the logical and physical connections, requiring that the entire call be reinitiated by the Collaboration initiator and the AVNM as described above.
  • Upon returning to the initial deferred call, the Expert engages in a videoconference with caller 231, a research analyst who is located 10 floors up from the Expert with a complex question regarding a particular security Caller 231 decides to add London expert 232 to the videoconference (via the ADD button in Collaboration Initiator window 204) to provide additional information regarding the factual history of the security Upon selecting the ADD button, video window 203 now displays, as illustrated in FIG. 38, a video mosaic consisting of three smaller images (instead of a single large image displaying only caller 231) of the Expert 233, caller 231 and London expert 232.
  • During this videoconference, an urgent PRIORITY request (New Call window 234) is received from the Expert's boss (who is engaged in a three-party videoconference call with two members of the bank's operations department and is attempting to add the Expert to that call to answer a quick question). The Expert puts his three-party videoconference on hold (merely by clicking the HOLD button in video window 203) and accepts (via the ACCEPT button of New Call window 234) the urgent call from his boss, which results in the Expert being added to the boss' three-party videoconference call.
  • As illustrated in FIG. 39, video window 203 is now replaced with a four-person video mosaic representing a four-party conference call consisting of the Expert 233, his boss 241 and the two members 242 and 243 of the bank's operations department. The Expert quickly answers the boss' question and, by clicking on the RESUME button (of video window 203) adjacent to the names of the other participants to the call on hold, simultaneously hangs up on the conference call with his boss and resumes his three-party conference call involving the securities issue, as illustrated in video window 203 of FIG. 40.
  • While that call was on hold, however, analyst 231 and London expert 232 were still engaged in a two-way videoconference (with a blackened portion of the video mosaic on their screens indicating that the Expert was on hold) and had shared and annotated a graphical image 250 (see annotations 251 to image 250 of FIG. 40) illustrating certain financial concerns. Once the Expert resumed the call, analyst 231 added the Expert to the share session, causing Share window 211 containing annotated image 250 to appear on the Expert's screen Optionally, snapshot sharing could progress while the video was on hold.
  • Before concluding his conference regarding the securities, the Expert receives notification of an incoming multimedia mail message—e.g., a beep accompanied by the appearance of an envelope 252 in the dog's mouth in In Box icon 205 shown in FIG. 40. Once he concludes his call, he quickly scans his incoming multimedia mail message by clicking on In Box icon 205, which invokes his mail software, and then selecting the incoming message for a quick scan, as generally illustrated in the top two windows of FIG. 2B. He decides it can wait for further review as the sender is an analyst other than the one helping on his security question.
  • He then reinitiates (by selecting deferred call indicator 230, shown in FIG. 40) his deferred call with field representative 201 and his client 202, as shown in FIG. 41. Note that the full state of the call is also recreated, including restoration of previously shared image 220 with annotations 222 as they existed when the call was deferred (see FIG. 37). Note also in FIG. 41 that, having reviewed his only unread incoming multimedia mail message, In Box icon 205 no longer shows an envelope in the dog's mouth, indicating that the Expert currently has no unread incoming messages.
  • As the Expert continues to provide advice and pricing information to field representative 201, he receives notification of three priority calls 261-263 in short succession Call 261 is the Head of Sales for the Chicago office Working at home, she had instructed her CMW to alert her of all urgent news or messages, and was subsequently alerted to the arrival of the Expert's earlier multimedia mail message Call 262 is an urgent international call Call 263 is from the Head of Sales in Los Angeles. The Expert quickly winds down and then concludes his call with field representative 201.
  • The Expert notes from call indicator 262 that this call is not only an international call (shown in the top portion of the New Call window), but he realizes it is from a laptop user in the field in Central Mexico. The Expert elects to prioritize his calls in the following manner: 262, 261 and 263. He therefore quickly answers call 261 (by clicking on its ACCEPT button) and puts that call on hold while deferring call 263 in the manner discussed above. He then proceeds to accept the call identified by international call indicator 262.
  • Note in FIG. 42 deferred call indicator 271 and the indicator for the call placed on hold (next to the highlighted RESUME button in video window 203), as well as the image of caller 272 from the laptop in the field in Central Mexico. Although Mexican caller 272 is outdoors and has no direct access to any wired telephone connection, his laptop has two wireless modems permitting dial-up access to two data connections in the nearest field office (through which his calls were routed). The system automatically (based upon the laptop's registered service capabilities) allocated one connection for an analog telephone voice call (using his laptop's built-in microphone and speaker and the Expert's computer-integrated telephony capabilities) to provide audio teleconferencing. The other connection provides control, data conferencing and one-way digital video (i.e., the laptop user cannot see the image of the Expert) from the laptop's built-in camera, albeit at a very slow frame rate (e.g., 3-10 small frames per second) due to the relatively slow dial-up phone connection.
  • It is important to note that, despite the limited capabilities of the wireless laptop equipment, the present invention accommodates such capabilities, supplementing an audio telephone connection with limited (i.e., relatively slow) one-way video and data conferencing functionality. As telephony and video compression technologies improve, the present invention will accommodate such improvements automatically. Moreover, even with one participant to a teleconference having limited capabilities, other participants need not be reduced to this “lowest common denominator.” For example, additional participants could be added to the call illustrated in FIG. 42 as described above, and such participants could have full videoconferencing, data conferencing and other collaborative functionality vis-a-vis one another, while having limited functionality only with caller 272.
  • As his day evolved, the off-site salesperson 272 in Mexico was notified by his manager through the laptop about a new security and became convinced that his client would have particular interest in this issue. The salesperson therefore decided to contact the Expert as shown in FIG. 42. While discussing the security issues, the Expert again shares all captured graphs, charts, etc.
  • The salesperson 272 also needs the Expert s help on another issue. He has hard copy only of a client's portfolio and needs some advice on its composition before he meets with the client tomorrow. He says he will fax it to the Expert for analysis Upon receiving the fax—on his CMW, via computer-integrated fax—the Expert asks if he should either send the Mexican caller a “QuickTime” movie (a lower quality compressed video standard from Apple Computer) on his laptop tonight or send a higher-quality CD via FedX tomorrow—the notion being that the Expert can produce an actual video presentation with models and annotations in video form. The salesperson can then play it to his client tomorrow afternoon and it will be as if the Expert is in the room. The Mexican caller decides he would prefer the CD.
  • Continuing with this scenario, the Expert learns, in the course of his call with remote laptop caller 272, that he missed an important issue during his previous quick scan of his incoming multimedia mail message. The Expert is upset that the sender of the message did not utilize the “video highlight” feature to highlight this aspect of the message. This feature permits the composer of the message to define “tags” (e.g., by clicking a TAG button, not shown) during record time which are stored with the message along with a “time stamp,” and which cause a predefined or selectable audio and/or visual indicator to be played/displayed at that precise point in the message during playback.
  • Because this issue relates to the caller that the Expert has on hold, the Expert decides to merge the two calls together by adding the call on hold to his existing call. As noted above, both the Expert and the previously held caller will have full video capabilities vis-a-vis one another and will see a three-way mosaic image (with the image of caller 272 at a slower frame rate), whereas caller 272 will have access only to the audio portion of this three-way conference call, though he will have data conferencing functionality with both of the other participants.
  • The Expert forwards the multimedia mail message to both caller 272 and the other participant, and all three of them review the video enclosure in greater detail and discuss the concern raised by caller 272. They share certain relevant data as described above and realize that they need to ask a quick question of another remote expert. They add that expert to the call (resulting in the addition of a fourth image to the video mosaic, also not shown) for less than a minute while they obtain a quick answer to their question. They then continue their three-way call until the Expert provides his advice and then adjourns the call.
  • The Expert composes a new multimedia mail message, recording his image and audio synchronized (as described above) to the screen displays resulting from his simultaneous interaction with his CMW (e.g., running a program that performs certain calculations and displays a graph while the Expert illustrates certain points by telepointing on the screen, during which time his image and spoken words are also captured). He sends this message to a number of salesforce recipients whose identities are determined automatically by an outgoing mail filter that utilizes a database of information on each potential recipient (e.g., selecting only those whose clients have investment policies which allow this type of investment).
  • The Expert then receives an audio and visual reminder (not shown) that a particular video feed (e.g., a short segment of a financial cable television show featuring new financial instruments) will be triggered automatically in a few minutes. He uses this time to search his local securities database, which is dynamically updated from financial information feeds (e.g., prepared from a broadcast textual stream of current financial events with indexed headers that automatically applies data filters to select incoming events relating to certain securities). The video feed is then displayed on the Expert's screen and he watches this short video segment.
  • After analyzing this extremely up-to-date information, the Expert then reinitiates his previously deferred call, from indicator 271 shown in FIG. 42, which he knows is from the Head of Sales in Los Angeles, who is seeking to provide his prime clients with securities advice on another securities transaction based upon the most recent available information. The Expert's call is not answered directly, though he receives a short prerecorded video message (left by the caller who had to leave his home for a meeting across town soon after his priority message was deferred) asking that the Expert leave him a multimedia mail reply message with advice for a particular client, and explaining that he will access this message remotely from his laptop as soon as his meeting is concluded. The Expert complies with this request and composes and sends this mail message.
  • The Expert then receives an audio and visual reminder on his screen indicating that his office hours will end in two minutes. He switches from “intercom” mode to “telephone” mode so that he will no longer be disturbed without an opportunity to reject incoming calls via the New Call window described above. He then receives and accepts a final call concerning an issue from an electronic meeting several months ago, which was recorded in its entirety.
  • The Expert accesses this recorded meeting from his “corporate memory.” He searches the recorded meeting (which appears in a second video window on his screen as would a live meeting, along with standard controls for stop/play/rewind/fast forward/etc.) for an event that will trigger his memory using his fast forward controls, but cannot locate the desired portion of the meeting. He then elects to search the ASCII text log (which was automatically extracted in the background after the meeting had been recorded, using the latest voice recognition techniques), but still cannot locate the desired portion of the meeting. Finally, he applies an information filter to perform a content-oriented (rather than literal) search and finds the portion of the meeting he was seeking. After quickly reviewing this short portion of the previously recorded meeting, the Expert responds to the caller's question, adjourns the call and concludes his office hours.
  • It should be noted that the above scenario involves many state-of-the-art desktop tools (e.g., video and information feeds, information filtering and voice recognition) that can be leveraged by our Expert during videoconferencing, data conferencing and other collaborative activities provided by the present invention—because this invention, instead of providing a dedicated videoconferencing system, provides a desktop multimedia collaboration system that integrates into the Expert's existing workstation/LAN/WAN environment.
  • It should also be noted that all of the preceding collaborative activities in this scenario took place during a relatively short portion of the expert's day (e.g., less than an hour of cumulative time) while the Expert remained in his office and continued to utilize the tools and information available from his desktop. Prior to this invention, such a scenario would not have been possible because many of these activities could have taken place only with face-to-face collaboration, which in many circumstances is not feasible or economical and which thus may well have resulted in a loss of the associated business opportunities.
  • Although the present invention has been described in connection with particular preferred embodiments and examples, it is to be understood that many modifications and variations can be made in hardware, software, operation, uses, protocols and data formats without departing from the scope to which the inventions disclosed herein are entitled. For example, for certain applications, it will be useful to provide some or all of the audio/video signals in digital form. Accordingly, the present invention is to be considered as including all apparatus and methods encompassed by the appended claims.

Claims (44)

1. A method of real-time communication between a plurality of users each with respective communication devices having associated displays, the method comprising:
providing each of the plurality of users with collaboration initiation software at their communication devices;
allowing at least first and second users to log in to their respective communication devices;
registering the communication devices of at least the first and second users with at least one communication network;
in response to the users logging in, enabling the system to maintain service records for at least the first and second users, the service records including user identification information and an associated location information identifying the user's location no matter where that user is located;
causing display of a user identifier for at least the second user in a list of users;
allowing the first user to request communications with the second user by selecting the displayed second user's identifier;
in response to the selection, retrieving necessary location information of the second user from at least one service record;
using the retrieved information to establish a connection between the first and second users, thereby enabling real-time communication, including at least audio, between the first and second users.
2. The method of claim 1, wherein at least one user identifier comprises an icon representing a user.
3. The method of claim 2, further comprising indicating to the first user if another user is not logged in.
4. The method of claim 3, wherein indicating to the first user whether another user is not logged in occurs if no service record is found for the other user.
5. The method of claim 3, wherein the service record further includes information as to the capabilities of at least one of the communication devices.
6. The method of claim 4, wherein the location information includes address information.
7. The method of claim 7, wherein information in at least one service record is usable to associate a user with ports associated with the communication device into which that user is logged in.
8. The method of claim 6, further comprising creating at least first and second call handles, respectively associated with the first and second users' communication devices, each call handle defining at least a call state for its respective device.
9. The method of claim 2, further comprising allowing at least one of the first and second users to send messages to the other.
10. The method of claim 4, wherein the communication includes video.
11. The method of claim 4, wherein the first and second users establish communications using respective first and second communications networks.
12. The method of claim 11, wherein at least one of the networks is a wireless network.
13. The method of claim 11, wherein at least one of the first and second communications networks is a wide area network.
14. The method of claim 4, during the audio communication between the first and second users, further comprising:
notifying a third user of a request by the first user to initiate an audio communication with the third user;
allowing the third user to decide whether to accept the request; and
notifying the first user of the decision by the third user.
15. The method of claim 14, upon third user's acceptance of the request, further comprising:
allowing the first user to add the third user into the audio communication involving the first and second users.
16. The method of claim 4, during the audio communication between the first and second users, further comprising:
notifying the first or second user of a request by a third user to initiate an audio communication with the respective first or second user;
allowing the respective first or second user to decide whether to accept the request; and
allowing the respective first or second user to add the third user into the audio communication involving the first and respective first and second users upon the second user's acceptance of the request.
17. The method of claim 1, wherein the retrieved location information is provided to the communications device of the first user.
18. The method of claim 1, wherein the service record is maintained by at least one communications network.
19. The method of claim 11, wherein the service record for the second user is maintained by an entity other than the network the second user uses to establish the communications.
20. The method of claim 4, wherein the audio communication includes digital audio.
21. The method of claim 2, wherein at least one of the service records is entered when the associated user logs in.
22. The method of claim 4, further comprising:
detecting an attempt by a first user to initiate a communication with a second user;
causing an indication of the attempt to initiate communications to appear automatically on second user's display; and
allowing the second user to accept or reject the communication.
23. A system for real-time, communication between a plurality of separated users, comprising:
(a) at least one communication device for use by each of the plurality of users and having an associated display;
(b) communication initiation software associated with the communication devices;
(c) at least one communication network to which at least first and second users can connect by using the collaboration initiation software at their respective communication devices;
(d) at least one service record for each of the first and second logged in users, the at least one service record including user identification information and an associated location where each user is logged in no matter where that user is located and being maintained in response to the respective user logging in; and
(e) computer software for
(i) causing display of a user identifier for at least the second user in a list of users and;
(ii) generating a signal in response to a first user selecting the displayed second user's identifier; and
wherein the system is configured to
cause the retrieving of necessary location information of the second user;
use the retrieval location information to cause the establishing of a connection between the first and second users; and
thereby enable real-time communication, including at least audio, between the first and second users.
24. The system of claim 23, wherein at least one user identifier comprises an icon representing a user.
25. The system of claim 24, wherein the system is configured to indicate to the first user if another user is not logged in.
26. The method of claim 25, wherein indicating to the first user whether another user is not logged in occurs if no service record is found for the other user.
27. The system of claim 25, wherein the service record further includes information as to the capabilities of at least one of the communication devices.
28. The system of claim 26, wherein the location information includes addressing information.
29. The system of claim 23, wherein information in at least one service record is used to associate a user with the ports associated with the communication device into which that user is logged in.
30. The system of claim 29, further configured to create at least first and second call handles, respectively associated with the first and second users' communication devices, each call handle defining at least a call sate for its respective device.
31. The system of claim 24, further configured to allow at least one of the first and second users to send messages to the other.
32. The system of claim 25, wherein the communication includes video.
33. The system of claim 25, wherein the first and second users establish communications using respective first and second communications networks.
34. The system of claim 25, wherein at least one of the networks is a wireless network.
35. The system of claim 25, wherein at least one of the first and second communications networks is a wide area network.
36. The system of claim 25, during the audio communication between the first and second users, further configured to:
notify a third user of a request by the first user to initiate an audio communication with the third user;
allow the third user to decide whether to accept the request; and
notify the first user of the decision by the third user.
37. The system of claim 36, upon third user's acceptance of the request, further configured to allow the first user to add the third user into the audio communication involving the first and second users.
38. The system of claim 25, during the audio communication between the first and second users, further configured to:
notify the respective first or second user of a request by a third user to initiate an audio communication with the first user;
allow the respective first or second user to decide whether to accept the request; and
allow the respective first or second user to add the third user into the audio communication involving the first and second users upon the second user's acceptance of the request.
39. The system of claim 23, wherein the retrieved location information is provided to the communications device of the first user.
40. The system of claim 23, wherein the service record is maintained by at least one communications network.
41. The system of claim 33, wherein the service record for the second user is maintained by an entity other than the network the second user uses to establish the communications.
42. The system of claim 24, wherein the audio communication includes digital audio.
43. The system of claim 30, wherein at least one of the service records is entered when the associated user logs in to the network.
44. The system of claim 24, wherein the system is configured to:
(a) detect an attempt by a second user to initiate a communication with the first user;
(b) cause an indication of the attempt to initiate communications to appear automatically on the first user's display; and
(c) allow the first user to accept or reject the communication.
US11/265,059 1993-10-01 2005-11-01 Using login-based addressing to communicate with listed users Abandoned US20060064461A1 (en)

Priority Applications (12)

Application Number Priority Date Filing Date Title
US11/265,059 US20060064461A1 (en) 1993-10-01 2005-11-01 Using login-based addressing to communicate with listed users
US11/622,583 US7487210B2 (en) 1993-10-01 2007-01-12 Method for managing real-time communications
US11/622,627 US7412482B2 (en) 1993-10-01 2007-01-12 System for managing real-time communications
US11/623,177 US7437411B2 (en) 1993-10-01 2007-01-15 Communication of a selected type over a wide area network
US11/623,630 US7444373B2 (en) 1993-10-01 2007-01-16 Wireless real-time communication
US11/623,387 US7398296B2 (en) 1993-10-01 2007-01-16 Networked audio communication over two networks
US11/624,351 US7441001B2 (en) 1993-10-01 2007-01-18 Real-time wide-area communications between ports
US11/624,860 US7437412B2 (en) 1993-10-01 2007-01-19 Real-time communication of a selected type
US11/625,330 US20070083596A1 (en) 1993-10-01 2007-01-21 Storing and Accessing Media Files
US11/668,625 US7822813B2 (en) 1993-10-01 2007-01-30 Storing and accessing media files
US11/669,003 US7831663B2 (en) 1993-10-01 2007-01-30 Storage and playback of media files
US11/737,723 US7730132B2 (en) 1993-10-01 2007-04-19 Storing and accessing media files

Applications Claiming Priority (5)

Application Number Priority Date Filing Date Title
US08/131,523 US5689641A (en) 1993-10-01 1993-10-01 Multimedia collaboration system arrangement for routing compressed AV signal through a participant site without decompressing the AV signal
US08/660,461 US5802294A (en) 1993-10-01 1996-06-07 Teleconferencing system in which location video mosaic generator sends combined local participants images to second location video mosaic generator for displaying combined images
US08/994,848 US6237025B1 (en) 1993-10-01 1997-12-19 Multimedia collaboration system
US09/702,737 US7185054B1 (en) 1993-10-01 2000-11-01 Participant display and selection in video conference calls
US11/265,059 US20060064461A1 (en) 1993-10-01 2005-11-01 Using login-based addressing to communicate with listed users

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US09/702,737 Continuation US7185054B1 (en) 1993-10-01 2000-11-01 Participant display and selection in video conference calls

Related Child Applications (11)

Application Number Title Priority Date Filing Date
US11/622,583 Continuation US7487210B2 (en) 1993-10-01 2007-01-12 Method for managing real-time communications
US11/622,627 Continuation US7412482B2 (en) 1993-10-01 2007-01-12 System for managing real-time communications
US11/623,177 Continuation US7437411B2 (en) 1993-10-01 2007-01-15 Communication of a selected type over a wide area network
US11/623,630 Continuation US7444373B2 (en) 1993-10-01 2007-01-16 Wireless real-time communication
US11/623,387 Continuation US7398296B2 (en) 1993-10-01 2007-01-16 Networked audio communication over two networks
US11/624,351 Continuation US7441001B2 (en) 1993-10-01 2007-01-18 Real-time wide-area communications between ports
US11/624,860 Continuation US7437412B2 (en) 1993-10-01 2007-01-19 Real-time communication of a selected type
US11/625,330 Continuation US20070083596A1 (en) 1993-10-01 2007-01-21 Storing and Accessing Media Files
US11/669,003 Continuation US7831663B2 (en) 1993-10-01 2007-01-30 Storage and playback of media files
US11/668,625 Continuation US7822813B2 (en) 1993-10-01 2007-01-30 Storing and accessing media files
US11/737,723 Continuation US7730132B2 (en) 1993-10-01 2007-04-19 Storing and accessing media files

Publications (1)

Publication Number Publication Date
US20060064461A1 true US20060064461A1 (en) 2006-03-23

Family

ID=46257787

Family Applications (25)

Application Number Title Priority Date Filing Date
US09/879,460 Expired - Fee Related US6594688B2 (en) 1993-10-01 2001-06-11 Dedicated echo canceler for a workstation
US10/120,559 Expired - Fee Related US6789105B2 (en) 1993-10-01 2002-04-09 Multiple-editor authoring of multimedia documents including real-time video and time-insensitive media
US10/120,307 Expired - Fee Related US7054904B2 (en) 1993-10-01 2002-04-09 Marking and searching capabilities in multimedia documents within multimedia collaboration networks
US10/382,554 Abandoned US20030187940A1 (en) 1993-10-01 2003-03-04 Teleconferencing employing multiplexing of video and data conferencing signals
US10/382,553 Expired - Fee Related US6959322B2 (en) 1993-10-01 2003-03-04 UTP based video conferencing
US10/721,343 Expired - Fee Related US7206809B2 (en) 1993-10-01 2003-11-26 Method for real-time communication between plural users
US10/721,905 Expired - Fee Related US7421470B2 (en) 1993-10-01 2003-11-26 Method for real-time communication between plural users
US10/722,051 Expired - Fee Related US7433921B2 (en) 1993-10-01 2003-11-26 System for real-time communication between plural users
US10/721,385 Expired - Fee Related US7152093B2 (en) 1993-10-01 2003-11-26 System for real-time communication between plural users
US11/265,394 Abandoned US20060075121A1 (en) 1993-10-01 2005-11-01 Registration based addressing and call handles to establish communication
US11/265,059 Abandoned US20060064461A1 (en) 1993-10-01 2005-11-01 Using login-based addressing to communicate with listed users
US11/264,926 Expired - Fee Related US7908320B2 (en) 1993-10-01 2005-11-01 Tracking user locations over multiple networks to enable real time communications
US11/265,256 Abandoned US20060059266A1 (en) 1993-10-01 2005-11-01 Registration based addressing over multiple networks with digital audio communication
US11/264,936 Abandoned US20060041617A1 (en) 1993-10-01 2005-11-01 Log-in based communications plus two data types
US11/622,583 Expired - Fee Related US7487210B2 (en) 1993-10-01 2007-01-12 Method for managing real-time communications
US11/622,627 Expired - Fee Related US7412482B2 (en) 1993-10-01 2007-01-12 System for managing real-time communications
US11/623,177 Expired - Fee Related US7437411B2 (en) 1993-10-01 2007-01-15 Communication of a selected type over a wide area network
US11/623,630 Expired - Fee Related US7444373B2 (en) 1993-10-01 2007-01-16 Wireless real-time communication
US11/623,387 Expired - Fee Related US7398296B2 (en) 1993-10-01 2007-01-16 Networked audio communication over two networks
US11/624,351 Expired - Fee Related US7441001B2 (en) 1993-10-01 2007-01-18 Real-time wide-area communications between ports
US11/624,860 Expired - Fee Related US7437412B2 (en) 1993-10-01 2007-01-19 Real-time communication of a selected type
US11/625,330 Abandoned US20070083596A1 (en) 1993-10-01 2007-01-21 Storing and Accessing Media Files
US11/669,003 Expired - Fee Related US7831663B2 (en) 1993-10-01 2007-01-30 Storage and playback of media files
US11/668,625 Expired - Fee Related US7822813B2 (en) 1993-10-01 2007-01-30 Storing and accessing media files
US11/737,723 Expired - Fee Related US7730132B2 (en) 1993-10-01 2007-04-19 Storing and accessing media files

Family Applications Before (10)

Application Number Title Priority Date Filing Date
US09/879,460 Expired - Fee Related US6594688B2 (en) 1993-10-01 2001-06-11 Dedicated echo canceler for a workstation
US10/120,559 Expired - Fee Related US6789105B2 (en) 1993-10-01 2002-04-09 Multiple-editor authoring of multimedia documents including real-time video and time-insensitive media
US10/120,307 Expired - Fee Related US7054904B2 (en) 1993-10-01 2002-04-09 Marking and searching capabilities in multimedia documents within multimedia collaboration networks
US10/382,554 Abandoned US20030187940A1 (en) 1993-10-01 2003-03-04 Teleconferencing employing multiplexing of video and data conferencing signals
US10/382,553 Expired - Fee Related US6959322B2 (en) 1993-10-01 2003-03-04 UTP based video conferencing
US10/721,343 Expired - Fee Related US7206809B2 (en) 1993-10-01 2003-11-26 Method for real-time communication between plural users
US10/721,905 Expired - Fee Related US7421470B2 (en) 1993-10-01 2003-11-26 Method for real-time communication between plural users
US10/722,051 Expired - Fee Related US7433921B2 (en) 1993-10-01 2003-11-26 System for real-time communication between plural users
US10/721,385 Expired - Fee Related US7152093B2 (en) 1993-10-01 2003-11-26 System for real-time communication between plural users
US11/265,394 Abandoned US20060075121A1 (en) 1993-10-01 2005-11-01 Registration based addressing and call handles to establish communication

Family Applications After (14)

Application Number Title Priority Date Filing Date
US11/264,926 Expired - Fee Related US7908320B2 (en) 1993-10-01 2005-11-01 Tracking user locations over multiple networks to enable real time communications
US11/265,256 Abandoned US20060059266A1 (en) 1993-10-01 2005-11-01 Registration based addressing over multiple networks with digital audio communication
US11/264,936 Abandoned US20060041617A1 (en) 1993-10-01 2005-11-01 Log-in based communications plus two data types
US11/622,583 Expired - Fee Related US7487210B2 (en) 1993-10-01 2007-01-12 Method for managing real-time communications
US11/622,627 Expired - Fee Related US7412482B2 (en) 1993-10-01 2007-01-12 System for managing real-time communications
US11/623,177 Expired - Fee Related US7437411B2 (en) 1993-10-01 2007-01-15 Communication of a selected type over a wide area network
US11/623,630 Expired - Fee Related US7444373B2 (en) 1993-10-01 2007-01-16 Wireless real-time communication
US11/623,387 Expired - Fee Related US7398296B2 (en) 1993-10-01 2007-01-16 Networked audio communication over two networks
US11/624,351 Expired - Fee Related US7441001B2 (en) 1993-10-01 2007-01-18 Real-time wide-area communications between ports
US11/624,860 Expired - Fee Related US7437412B2 (en) 1993-10-01 2007-01-19 Real-time communication of a selected type
US11/625,330 Abandoned US20070083596A1 (en) 1993-10-01 2007-01-21 Storing and Accessing Media Files
US11/669,003 Expired - Fee Related US7831663B2 (en) 1993-10-01 2007-01-30 Storage and playback of media files
US11/668,625 Expired - Fee Related US7822813B2 (en) 1993-10-01 2007-01-30 Storing and accessing media files
US11/737,723 Expired - Fee Related US7730132B2 (en) 1993-10-01 2007-04-19 Storing and accessing media files

Country Status (1)

Country Link
US (25) US6594688B2 (en)

Cited By (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20080181218A1 (en) * 2007-01-31 2008-07-31 Gorzynski Mark E Coordinated media control system
US20080229405A1 (en) * 2007-03-16 2008-09-18 Yoshihiro Fujii Communication System, Communication System Management Apparatus, Terminal Connection Control Method, and Program
US20080228876A1 (en) * 2007-03-13 2008-09-18 Byron Johnson System and method for online collaboration
US20080243994A1 (en) * 2007-03-30 2008-10-02 Alexander Kropivny Method, Apparatus, System, and Medium for Supporting Multiple-Party Communications
US20080242422A1 (en) * 2007-03-30 2008-10-02 Uranus International Limited Method, Apparatus, System, Medium, and Signals for Supporting Game Piece Movement in a Multiple-Party Communication
US20080244702A1 (en) * 2007-03-30 2008-10-02 Uranus International Limited Method, Apparatus, System, Medium, and Signals for Intercepting a Multiple-Party Communication
US20080244615A1 (en) * 2007-03-30 2008-10-02 Uranus International Limited Method, Apparatus, System, Medium, and Signals for Supporting a Multiple-Party Communication on a Plurality of Computer Servers
US20080244013A1 (en) * 2007-03-30 2008-10-02 Alexander Kropivny Method, Apparatus, System, Medium, and Signals for Publishing Content Created During a Communication
US20080244461A1 (en) * 2007-03-30 2008-10-02 Alexander Kropivny Method, Apparatus, System, Medium, and Signals For Supporting Pointer Display In A Multiple-Party Communication
US20090070420A1 (en) * 2006-05-01 2009-03-12 Schuyler Quackenbush System and method for processing data signals
US20110093599A1 (en) * 2009-10-20 2011-04-21 Avaya Inc. Hierarchal structuring of nodes in a peer-to-peer network

Families Citing this family (830)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8352400B2 (en) 1991-12-23 2013-01-08 Hoffberg Steven M Adaptive pattern recognition based controller apparatus and method and human-factored interface therefore
US7006881B1 (en) * 1991-12-23 2006-02-28 Steven Hoffberg Media recording device with remote graphic user interface
US6594688B2 (en) 1993-10-01 2003-07-15 Collaboration Properties, Inc. Dedicated echo canceler for a workstation
US6343313B1 (en) * 1996-03-26 2002-01-29 Pixion, Inc. Computer conferencing system with real-time multipoint, multi-speed, multi-stream scalability
IL119364A (en) 1996-10-06 2000-11-21 Icq Inc Communications system
US7490169B1 (en) 1997-03-31 2009-02-10 West Corporation Providing a presentation on a network having a plurality of synchronized media types
EP1021917A4 (en) * 1997-03-31 2002-05-15 Broadband Associates Method and system for providing a presentation on a network
US7412533B1 (en) 1997-03-31 2008-08-12 West Corporation Providing a presentation on a network having a plurality of synchronized media types
US6298356B1 (en) * 1998-01-16 2001-10-02 Aspect Communications Corp. Methods and apparatus for enabling dynamic resource collaboration
US6256620B1 (en) * 1998-01-16 2001-07-03 Aspect Communications Method and apparatus for monitoring information access
US6380950B1 (en) * 1998-01-20 2002-04-30 Globalstreams, Inc. Low bandwidth television
US9924234B2 (en) 1998-07-23 2018-03-20 Comcast Ip Holdings I, Llc Data structure and methods for providing an interactive program
US6425131B2 (en) * 1998-12-30 2002-07-23 At&T Corp. Method and apparatus for internet co-browsing over cable television and controlled through computer telephony
US7966078B2 (en) * 1999-02-01 2011-06-21 Steven Hoffberg Network media appliance system and method
US6529949B1 (en) * 2000-02-07 2003-03-04 Interactual Technologies, Inc. System, method and article of manufacture for remote unlocking of local content located on a client device
US7346920B2 (en) * 2000-07-07 2008-03-18 Sonic Solutions, A California Corporation System, method and article of manufacture for a common cross platform framework for development of DVD-Video content integrated with ROM content
US20050166232A1 (en) * 1999-04-21 2005-07-28 Lamkin Allan B... Presentation of media content from multiple media sources
US20050182828A1 (en) * 1999-04-21 2005-08-18 Interactual Technologies, Inc. Platform specific execution
US20060041639A1 (en) * 1999-04-21 2006-02-23 Interactual Technologies, Inc. Platform detection
US6671718B1 (en) * 1999-06-28 2003-12-30 Mark Meister Email client application incorporating an active transmit authorization request
US6721410B1 (en) * 1999-08-10 2004-04-13 Nortel Networks Limited Recursive identification of individuals for casual collaborative conferencing
US8145776B1 (en) * 1999-10-15 2012-03-27 Sony Corporation Service providing apparatus and method, and information processing apparatus and method as well as program storage medium
US7392481B2 (en) * 2001-07-02 2008-06-24 Sonic Solutions, A California Corporation Method and apparatus for providing content-owner control in a networked device
US20040220926A1 (en) * 2000-01-03 2004-11-04 Interactual Technologies, Inc., A California Cpr[P Personalization services for entities from multiple sources
US7328239B1 (en) 2000-03-01 2008-02-05 Intercall, Inc. Method and apparatus for automatically data streaming a multiparty conference session
US8645137B2 (en) 2000-03-16 2014-02-04 Apple Inc. Fast, language-independent method for user authentication by voice
US7096185B2 (en) * 2000-03-31 2006-08-22 United Video Properties, Inc. User speech interfaces for interactive media guidance applications
US7844670B2 (en) * 2000-04-03 2010-11-30 Paltalk Holdings, Inc. Method and computer program product for establishing real-time communications between networked computers
US20020023134A1 (en) * 2000-04-03 2002-02-21 Roskowski Steven G. Method and computer program product for establishing real-time communications between networked computers
US20040027375A1 (en) * 2000-06-12 2004-02-12 Ricus Ellis System for controlling a display of the user interface of a software application
US20020019845A1 (en) * 2000-06-16 2002-02-14 Hariton Nicholas T. Method and system for distributed scripting of presentations
US20060259926A1 (en) 2000-07-20 2006-11-16 Digital Deck, Inc. Adaptable programming guide for networked devices
US20020053081A1 (en) * 2000-10-31 2002-05-02 Digitaldeck, Inc. Adaptable programming guide for networked devices
US20020029384A1 (en) * 2000-07-20 2002-03-07 Griggs Theodore L. Mechanism for distributing content data
US20050022241A1 (en) * 2000-10-31 2005-01-27 Griggs Theodore L. Adaptable programming guide for networked devices
US6636259B1 (en) * 2000-07-26 2003-10-21 Ipac Acquisition Subsidiary I, Llc Automatically configuring a web-enabled digital camera to access the internet
US8224776B1 (en) 2000-07-26 2012-07-17 Kdl Scan Designs Llc Method and system for hosting entity-specific photo-sharing websites for entity-specific digital cameras
US7689510B2 (en) * 2000-09-07 2010-03-30 Sonic Solutions Methods and system for use in network management of content
US9292516B2 (en) * 2005-02-16 2016-03-22 Sonic Solutions Llc Generation, organization and/or playing back of content based on incorporated parameter identifiers
CA2422519A1 (en) * 2000-09-28 2002-04-04 Vigilos, Inc. System and method for dynamic interaction with remote devices
US8392552B2 (en) * 2000-09-28 2013-03-05 Vig Acquisitions Ltd., L.L.C. System and method for providing configurable security monitoring utilizing an integrated information system
AU2001296925A1 (en) 2000-09-28 2002-04-08 Vigilos, Inc. Method and process for configuring a premises for monitoring
US7287088B1 (en) 2000-10-06 2007-10-23 Fotomedia Technologies, Llc Transmission bandwidth and memory requirements reduction in a portable image capture device by eliminating duplicate image transmissions
US8223975B2 (en) 2008-06-19 2012-07-17 Xm Satellite Radio Inc. Method and apparatus for multiplexing audio program channels from one or more received broadcast streams to provide a playlist style listening experience to users
US9008812B2 (en) 2008-06-19 2015-04-14 Sirius Xm Radio Inc. Method and apparatus for using selected content tracks from two or more program channels to automatically generate a blended mix channel for playback to a user upon selection of a corresponding preset button on a user interface
US20020059377A1 (en) * 2000-11-14 2002-05-16 Jagadish Bandhole Collaborative computing systems using dynamic computing environments
US20020129359A1 (en) * 2000-11-27 2002-09-12 Lichner Randall Manton Cross platform system and method for the management and distribution of pay per view video on demand
US7957514B2 (en) 2000-12-18 2011-06-07 Paltalk Holdings, Inc. System, method and computer program product for conveying presence information via voice mail
US8977683B2 (en) * 2000-12-26 2015-03-10 Polycom, Inc. Speakerphone transmitting password information to a remote device
WO2002093397A1 (en) * 2000-12-26 2002-11-21 Polycom, Inc. System and method for coordinating a conference using a dedicated server
US20020107910A1 (en) * 2001-02-02 2002-08-08 Yan Zhao Client/server two-way communication system framework under HTTP protocol
US8761363B2 (en) * 2001-02-27 2014-06-24 Verizon Data Services Llc Methods and systems for automatic forwarding of communications to a preferred device
US8873730B2 (en) 2001-02-27 2014-10-28 Verizon Patent And Licensing Inc. Method and apparatus for calendared communications flow control
US8472428B2 (en) * 2001-02-27 2013-06-25 Verizon Data Services Llc Methods and systems for line management
US8494135B2 (en) * 2001-02-27 2013-07-23 Verizon Data Services Llc Methods and systems for contact management
US8503639B2 (en) * 2001-02-27 2013-08-06 Verizon Data Services Llc Method and apparatus for adaptive message and call notification
US7903796B1 (en) 2001-02-27 2011-03-08 Verizon Data Services Llc Method and apparatus for unified communication management via instant messaging
US7912193B2 (en) 2001-02-27 2011-03-22 Verizon Data Services Llc Methods and systems for call management with user intervention
US8467502B2 (en) 2001-02-27 2013-06-18 Verizon Data Services Llc Interactive assistant for managing telephone communications
US8750482B2 (en) * 2001-02-27 2014-06-10 Verizon Data Services Llc Methods and systems for preemptive rejection of calls
US8761355B2 (en) * 2002-11-25 2014-06-24 Telesector Resources Group, Inc. Methods and systems for notification of call to device
US8472606B2 (en) * 2001-02-27 2013-06-25 Verizon Data Services Llc Methods and systems for directory information lookup
US8488766B2 (en) * 2001-02-27 2013-07-16 Verizon Data Services Llc Methods and systems for multiuser selective notification
US6976017B1 (en) * 2001-02-27 2005-12-13 Verizon Data Services Inc. Method and apparatus for context based querying
US8798251B2 (en) * 2001-02-27 2014-08-05 Verizon Data Services Llc Methods and systems for computer enhanced conference calling
US8774380B2 (en) 2001-02-27 2014-07-08 Verizon Patent And Licensing Inc. Methods and systems for call management with user intervention
US8751571B2 (en) * 2001-02-27 2014-06-10 Verizon Data Services Llc Methods and systems for CPN triggered collaboration
US8488761B2 (en) * 2001-02-27 2013-07-16 Verizon Data Services Llc Methods and systems for a call log
US8503650B2 (en) 2001-02-27 2013-08-06 Verizon Data Services Llc Methods and systems for configuring and providing conference calls
US20080032738A1 (en) * 2001-03-07 2008-02-07 Palm, Inc. Portable wireless network
US20020183038A1 (en) * 2001-05-31 2002-12-05 Palm, Inc. System and method for crediting an account associated with a network access node
RU2272316C2 (en) * 2001-03-16 2006-03-20 Нетомат, Инк. Joint use, control and transmission of information along computer network
JP3679731B2 (en) * 2001-05-24 2005-08-03 キヤノン株式会社 Image communication system
US7237243B2 (en) * 2001-06-11 2007-06-26 Microsoft Corporation Multiple device management method and system
US7594218B1 (en) * 2001-07-24 2009-09-22 Adobe Systems Incorporated System and method for providing audio in a media file
US6750897B1 (en) 2001-08-16 2004-06-15 Verizon Data Services Inc. Systems and methods for implementing internet video conferencing using standard phone calls
JP2003076452A (en) * 2001-09-05 2003-03-14 Nagano Fujitsu Component Kk Computer switching unit and computer system, program for exchanging data, storage medium storing it and computer body incorporating it
US7058689B2 (en) * 2001-10-16 2006-06-06 Sprint Communications Company L.P. Sharing of still images within a video telephony call
JP2003150527A (en) * 2001-11-05 2003-05-23 Internatl Business Mach Corp <Ibm> Chat system, terminal unit therefor, chat server and program
US7266594B2 (en) * 2001-11-07 2007-09-04 Microsoft Corporation Method and system for configuring a computer for real-time communication
US7861169B2 (en) 2001-11-19 2010-12-28 Ricoh Co. Ltd. Multimedia print driver dialog interfaces
US7747655B2 (en) 2001-11-19 2010-06-29 Ricoh Co. Ltd. Printable representations for time-based media
US7539504B2 (en) * 2001-12-05 2009-05-26 Espre Solutions, Inc. Wireless telepresence collaboration system
US8144854B2 (en) 2001-12-31 2012-03-27 Polycom Inc. Conference bridge which detects control information embedded in audio information to prioritize operations
US8102984B2 (en) * 2001-12-31 2012-01-24 Polycom Inc. Speakerphone and conference bridge which receive and provide participant monitoring information
US8705719B2 (en) 2001-12-31 2014-04-22 Polycom, Inc. Speakerphone and conference bridge which receive and provide participant monitoring information
US20040205482A1 (en) * 2002-01-24 2004-10-14 International Business Machines Corporation Method and apparatus for active annotation of multimedia content
US7480715B1 (en) 2002-01-25 2009-01-20 Vig Acquisitions Ltd., L.L.C. System and method for performing a predictive threat assessment based on risk factors
US20030142471A1 (en) * 2002-01-29 2003-07-31 Palm, Inc. Replaceable cover for handheld computer
US7693484B2 (en) * 2002-01-29 2010-04-06 Palm, Inc. Dynamic networking modes method and apparatus
US7096037B2 (en) * 2002-01-29 2006-08-22 Palm, Inc. Videoconferencing bandwidth management for a handheld computer system and method
US20030218638A1 (en) * 2002-02-06 2003-11-27 Stuart Goose Mobile multimodal user interface combining 3D graphics, location-sensitive speech interaction and tracking technologies
US6937712B2 (en) * 2002-02-22 2005-08-30 Marc S. Lemchen Network-based intercom system and method for simulating a hardware based dedicated intercom system
US6792091B2 (en) * 2002-02-22 2004-09-14 Marc S. Lemchen Network-based intercom system and method for simulating a hardware based dedicated intercom system
US9392120B2 (en) 2002-02-27 2016-07-12 Verizon Patent And Licensing Inc. Methods and systems for call management with user intervention
US20030167335A1 (en) * 2002-03-04 2003-09-04 Vigilos, Inc. System and method for network-based communication
US20030206172A1 (en) * 2002-03-05 2003-11-06 Vigilos, Inc. System and method for the asynchronous collection and management of video data
ATE296009T1 (en) * 2002-03-11 2005-06-15 Siemens Ag METHOD AND DEVICE FOR SETTING UP A CONNECTION BETWEEN A FIRST AND A SECOND SUBSCRIBER IN A TELECOMMUNICATIONS NETWORK
US7454760B2 (en) 2002-04-22 2008-11-18 Rosebud Lms, Inc. Method and software for enabling n-way collaborative work over a network of computers
US20080040675A1 (en) * 2002-04-30 2008-02-14 Aol Llc Instant messaging interface having a tear-off element
US7421661B1 (en) 2002-04-30 2008-09-02 Aol Llc Instant messaging interface having an informational tool tip
US20030202110A1 (en) * 2002-04-30 2003-10-30 Owens James W. Arrangement of images
US8751247B1 (en) * 2002-05-23 2014-06-10 At&T Intellectual Property Ii, L.P. Network-based collaborative control of distributed multimedia content
US20070086724A1 (en) * 2002-07-17 2007-04-19 Jeff Grady Interface systems for portable digital media storage and playback devices
US20040019640A1 (en) * 2002-07-25 2004-01-29 Bartram Linda Ruth System and method for distributing shared storage for collaboration across multiple devices
US20040162637A1 (en) 2002-07-25 2004-08-19 Yulun Wang Medical tele-robotic system with a master remote station with an arbitrator
US7275215B2 (en) 2002-07-29 2007-09-25 Cerulean Studios, Llc System and method for managing contacts in an instant messaging environment
US20040033810A1 (en) * 2002-08-12 2004-02-19 Griffin Patricia Nichole Audio visual communications system
US7185098B2 (en) * 2002-09-30 2007-02-27 Microsoft Corporation Conferencing using flexible connection ports
US20040107270A1 (en) * 2002-10-30 2004-06-03 Jamie Stephens Method and system for collaboration recording
US20040090452A1 (en) * 2002-11-13 2004-05-13 Schmidt Mark H. One-click recording
US7668763B1 (en) 2002-11-25 2010-02-23 Xcm Development, Llc Tax return outsourcing and systems for protecting data
CA2515492A1 (en) * 2003-02-10 2004-08-26 Raindance Communications, Inc. Methods and apparatus for automatically adding a media component to an established multimedia collaboration session
US7164423B1 (en) 2003-04-30 2007-01-16 Apple Computer, Inc. Method and apparatus for providing an animated representation of a reorder operation
US7295852B1 (en) 2003-05-01 2007-11-13 Palm, Inc. Automated telephone conferencing method and system
US7669134B1 (en) 2003-05-02 2010-02-23 Apple Inc. Method and apparatus for displaying information during an instant messaging session
JP2004336343A (en) * 2003-05-07 2004-11-25 Canon Inc Image processing system
WO2004104745A2 (en) * 2003-05-14 2004-12-02 Digital Deck, Inc. Infrared signal distribution and management system and method
US7242316B2 (en) * 2003-05-14 2007-07-10 Digital Deck, Inc. Device control apparatus and method
US8022837B2 (en) * 2003-05-14 2011-09-20 Resource Consortium Limited, Inc. Infrared signal distribution and management system and method
TW200501668A (en) * 2003-05-14 2005-01-01 Digitaldeck Inc Distributed media management apparatus and method
US20040249889A1 (en) * 2003-06-03 2004-12-09 Drommond David E. Method and system for processing interactive communications
US20040255327A1 (en) * 2003-06-12 2004-12-16 Digital Deck, Inc. Media content distribution system and method
US20040255308A1 (en) * 2003-06-16 2004-12-16 Microsoft Corporation Method and system for activity-based user interfaces
US7865180B2 (en) * 2003-06-23 2011-01-04 Palm, Inc. Automated telephone conferencing method and system
US7421469B1 (en) * 2003-06-26 2008-09-02 Cisco Technology, Inc. Initiating a collaborative computing session from an advanced capability telephone
US8001187B2 (en) * 2003-07-01 2011-08-16 Apple Inc. Peer-to-peer active content sharing
US20050010874A1 (en) * 2003-07-07 2005-01-13 Steven Moder Virtual collaborative editing room
US20050015293A1 (en) * 2003-07-16 2005-01-20 International Business Machines Corporation Collaboration enhanced workflow system
US8239233B1 (en) 2003-07-17 2012-08-07 Xcm Development, Llc Work flow systems and processes for outsourced financial services
GB0319360D0 (en) * 2003-08-18 2003-09-17 Nokia Corp Setting up communication sessions
CA2442297A1 (en) * 2003-09-24 2005-03-24 Sylvie Martin Colour coded system for writing
US8077341B2 (en) 2003-09-25 2011-12-13 Ricoh Co., Ltd. Printer with audio or video receiver, recorder, and real-time content-based processing logic
JP2005108230A (en) 2003-09-25 2005-04-21 Ricoh Co Ltd Printing system with embedded audio/video content recognition and processing function
US7864352B2 (en) 2003-09-25 2011-01-04 Ricoh Co. Ltd. Printer with multimedia server
US7944861B2 (en) * 2003-10-14 2011-05-17 Tele-Town Hall, Llc System and process for mass telephony conference call
US8385526B2 (en) * 2003-10-14 2013-02-26 Tele-Town Hall, LLC. System and process for mass telephony conference call
US7852998B1 (en) * 2003-10-14 2010-12-14 Tele-Town Hall, Llc System and process for mass telephony conference call
US20050125541A1 (en) * 2003-12-04 2005-06-09 Randall Frank Integrating multiple communication modes
US8515023B2 (en) * 2003-12-05 2013-08-20 Stanley F Schoenbach Real-time interpreting systems and methods
US7813836B2 (en) 2003-12-09 2010-10-12 Intouch Technologies, Inc. Protocol for a remotely controlled videoconferencing robot
US8020105B1 (en) 2003-12-19 2011-09-13 Apple Inc. Unified user interface for instant messaging and email
FR2864412B1 (en) * 2003-12-19 2006-02-10 Gemplus Card Int METHOD FOR SAVING PERSONAL DATA FROM A SUBSCRIBER TO A TELECOMMUNICATIONS NETWORK, SERVER AND DEVICE THEREFOR
US7765263B1 (en) * 2003-12-19 2010-07-27 Apple Inc. Method and apparatus for processing electronic messages
US7343561B1 (en) 2003-12-19 2008-03-11 Apple Inc. Method and apparatus for message display
US20050138115A1 (en) * 2003-12-23 2005-06-23 Luis Llamas Real time mobile skill set information system
US7525584B2 (en) * 2004-01-05 2009-04-28 Lifesize Communications, Inc. Fast edge directed demosaicing
IL159838A0 (en) * 2004-01-13 2004-06-20 Yehuda Binder Information device
US20060294187A1 (en) * 2004-01-20 2006-12-28 Thomas Hakel Integrated web communications system
US20050182942A1 (en) * 2004-02-17 2005-08-18 Doru Calin Methods and devices for obtaining domain access rights
US8255835B2 (en) * 2004-02-24 2012-08-28 Research In Motion Limited Method and system for managing unread electronic messages
US20050204438A1 (en) * 2004-02-26 2005-09-15 Yulun Wang Graphical interface for a remote presence system
US20050193104A1 (en) * 2004-02-27 2005-09-01 Wyse Technology Inc. User interface for remote computing devices
JP4295644B2 (en) * 2004-03-08 2009-07-15 京セラ株式会社 Mobile terminal, broadcast recording / playback method for mobile terminal, and broadcast recording / playback program
US20050210394A1 (en) * 2004-03-16 2005-09-22 Crandall Evan S Method for providing concurrent audio-video and audio instant messaging sessions
JP2005277445A (en) * 2004-03-22 2005-10-06 Fuji Xerox Co Ltd Conference video image processing apparatus, and conference video image processing method and program
US9161068B2 (en) * 2004-03-24 2015-10-13 Onstream Media Corporation Remotely accessed virtual recording room
US20050229228A1 (en) * 2004-04-07 2005-10-13 Sandeep Relan Unicast cable content delivery
US8274666B2 (en) 2004-03-30 2012-09-25 Ricoh Co., Ltd. Projector/printer for displaying or printing of documents
US7603615B2 (en) * 2004-03-30 2009-10-13 Ricoh Co., Ltd. Multimedia projector-printer
US20060036466A1 (en) * 2004-03-31 2006-02-16 Tarr Christopher A System and method for providing custom stock images
US7085558B2 (en) * 2004-04-15 2006-08-01 International Business Machines Corporation Conference call reconnect system
US8738614B2 (en) * 2004-04-23 2014-05-27 Qualcomm Incorporated Methods and apparatus for providing hierarchical content flow in a data network
US9185147B1 (en) * 2004-04-29 2015-11-10 Paul Erich Keel System and methods for remote collaborative intelligence analysis
US8209376B1 (en) * 2004-05-06 2012-06-26 Apple Inc. Application-specific group listing
US7890604B2 (en) 2004-05-07 2011-02-15 Microsoft Corproation Client-side callbacks to server events
US7324069B2 (en) * 2004-05-14 2008-01-29 Pixar Animation review methods and apparatus
US9026578B2 (en) 2004-05-14 2015-05-05 Microsoft Corporation Systems and methods for persisting data between web pages
US8065600B2 (en) * 2004-05-14 2011-11-22 Microsoft Corporation Systems and methods for defining web content navigation
US8091029B1 (en) * 2004-05-28 2012-01-03 Adobe Systems Incorporated Presenter-only meeting canvas for collaborative electronic meeting space
US7734802B1 (en) 2004-05-28 2010-06-08 Adobe Systems Incorporated Dynamically adaptable collaborative electronic meeting space
US7379968B2 (en) * 2004-06-03 2008-05-27 International Business Machines Corporation Multiple moderation for networked conferences
US7881235B1 (en) 2004-06-25 2011-02-01 Apple Inc. Mixed media conferencing
US7865834B1 (en) 2004-06-25 2011-01-04 Apple Inc. Multi-way video conferencing user interface
US8234335B1 (en) * 2004-06-29 2012-07-31 Sprint Spectrum L.P. Customized grouping of participants in real-time conference set-up
US20060010197A1 (en) * 2004-07-06 2006-01-12 Francis Ovenden Multimedia collaboration and communications
US8077963B2 (en) 2004-07-13 2011-12-13 Yulun Wang Mobile robot with a head-based movement mapping scheme
US7684374B2 (en) * 2004-07-28 2010-03-23 Broadcom Corporation Handling of multimedia call sessions and attachments using multi-network simulcasting
US20070011234A1 (en) * 2004-07-29 2007-01-11 Xcm Development, Llc Computer conferencing system and features
US7634539B2 (en) * 2004-08-02 2009-12-15 Sap Ag Relations between collaboration workspaces
US20060064643A1 (en) 2004-09-14 2006-03-23 Hariton Nicholas T Distributed scripting for presentations with touch screen displays
US7624149B2 (en) * 2004-09-15 2009-11-24 Microsoft Corporation Instant messaging with audio
US9189756B2 (en) * 2004-09-21 2015-11-17 International Business Machines Corporation Case management system and method for collaborative project teaming
US7552801B2 (en) 2004-09-27 2009-06-30 Electronics Micro Systems, Inc. Elevator systems having display systems with pluralities of sequentially connected monitor units
US7640502B2 (en) * 2004-10-01 2009-12-29 Microsoft Corporation Presentation facilitation
US20060085515A1 (en) * 2004-10-14 2006-04-20 Kevin Kurtz Advanced text analysis and supplemental content processing in an instant messaging environment
US7903137B2 (en) * 2004-10-15 2011-03-08 Lifesize Communications, Inc. Videoconferencing echo cancellers
US7545435B2 (en) * 2004-10-15 2009-06-09 Lifesize Communications, Inc. Automatic backlight compensation and exposure control
US7720232B2 (en) * 2004-10-15 2010-05-18 Lifesize Communications, Inc. Speakerphone
US7692683B2 (en) * 2004-10-15 2010-04-06 Lifesize Communications, Inc. Video conferencing system transcoder
US7688345B2 (en) * 2004-10-15 2010-03-30 Lifesize Communications, Inc. Audio output in video conferencing and speakerphone based on call type
US7602141B2 (en) * 2004-10-15 2009-10-13 Lifesize Communications, Inc. Battery operated speakerphone and charging stand
US7864221B2 (en) * 2004-10-15 2011-01-04 Lifesize Communications, Inc. White balance for video applications
US8237770B2 (en) 2004-10-15 2012-08-07 Lifesize Communications, Inc. Audio based on speaker position and/or conference location
US20060106929A1 (en) * 2004-10-15 2006-05-18 Kenoyer Michael L Network conference communications
US8116500B2 (en) * 2004-10-15 2012-02-14 Lifesize Communications, Inc. Microphone orientation and size in a speakerphone
US7720236B2 (en) * 2004-10-15 2010-05-18 Lifesize Communications, Inc. Updating modeling information based on offline calibration experiments
US7572073B2 (en) * 2004-10-15 2009-08-11 Lifesize Communications, Inc. Camera support mechanism
US7760887B2 (en) * 2004-10-15 2010-07-20 Lifesize Communications, Inc. Updating modeling information based on online data gathering
US8149739B2 (en) * 2004-10-15 2012-04-03 Lifesize Communications, Inc. Background call validation
US20060132595A1 (en) * 2004-10-15 2006-06-22 Kenoyer Michael L Speakerphone supporting video and audio features
US7970151B2 (en) * 2004-10-15 2011-06-28 Lifesize Communications, Inc. Hybrid beamforming
US8054336B2 (en) * 2004-10-15 2011-11-08 Lifesize Communications, Inc. High definition pan tilt zoom camera with embedded microphones and thin cable for data and power
US7667728B2 (en) * 2004-10-15 2010-02-23 Lifesize Communications, Inc. Video and audio conferencing system with spatial audio
US7717629B2 (en) * 2004-10-15 2010-05-18 Lifesize Communications, Inc. Coordinated camera pan tilt mechanism
US7826624B2 (en) * 2004-10-15 2010-11-02 Lifesize Communications, Inc. Speakerphone self calibration and beam forming
US7473040B2 (en) * 2004-10-15 2009-01-06 Lifesize Communications, Inc. High definition camera pan tilt mechanism
US8477173B2 (en) * 2004-10-15 2013-07-02 Lifesize Communications, Inc. High definition videoconferencing system
US7730129B2 (en) 2004-10-20 2010-06-01 Inbit, Inc. Collaborative communication platforms
US20060101022A1 (en) * 2004-10-25 2006-05-11 Microsoft Corporation System and process for providing an interactive, computer network-based, virtual team worksite
US8150920B2 (en) * 2004-11-01 2012-04-03 Nokia Corporation System, method and computer program product for providing content to a terminal and directing the rendering of such content at the terminal
JP4599135B2 (en) * 2004-11-02 2010-12-15 キヤノン株式会社 Information processing apparatus and information processing method
US7711681B2 (en) 2004-11-05 2010-05-04 Accenture Global Services Gmbh System for distributed information presentation and interaction
US8826328B2 (en) * 2004-11-12 2014-09-02 Opentv, Inc. Communicating primary content streams and secondary content streams including targeted advertising to a remote unit
US7870192B2 (en) * 2004-12-16 2011-01-11 International Business Machines Corporation Integrated voice and video conferencing management
WO2006066052A2 (en) * 2004-12-16 2006-06-22 Sonic Solutions Methods and systems for use in network management of content
WO2006070343A2 (en) * 2004-12-29 2006-07-06 Nds Limited Freeze and record
US20060150109A1 (en) * 2004-12-30 2006-07-06 Motorola, Inc. Shared user interface
US20060167996A1 (en) * 2005-01-13 2006-07-27 Orsolini Garry S System and method for enabling electronic presentations
US7747685B2 (en) * 2005-01-20 2010-06-29 International Business Machines Corporation Method for automatic detection of display sharing and alert generation in instant messaging
WO2006078094A1 (en) * 2005-01-22 2006-07-27 Nhn Corporation System and method for sharing knowledge through on-line human network
US7533182B2 (en) * 2005-01-24 2009-05-12 Starz Media, Llc Portable screening room
US7524119B2 (en) * 2005-02-03 2009-04-28 Paul Sandhu Apparatus and method for viewing radiographs
US20060190960A1 (en) * 2005-02-14 2006-08-24 Barker Geoffrey T System and method for incorporating video analytics in a monitoring network
US20060210023A1 (en) * 2005-02-24 2006-09-21 Larry Spruce Voice messenger LX100
US7511767B2 (en) * 2005-03-29 2009-03-31 Snell & Wilcox Limited Video storage device, and method of controlling a video storage device and video mixer
US8468125B2 (en) * 2005-04-12 2013-06-18 International Business Machines Corporation Automatically moving multidimensional data between live datacubes of enterprise software systems
US20060232550A1 (en) * 2005-04-15 2006-10-19 Buckner Nathan C Integrated mouse in remote control
US8831647B2 (en) * 2005-04-19 2014-09-09 Devereux Research Ab Llc Presence-enabled mobile access
US20070064851A1 (en) * 2005-09-02 2007-03-22 Sbc Knowledge Ventures Lp Method for synchronizing a customer edge router or customer premise equipment associated therewith
US7991167B2 (en) * 2005-04-29 2011-08-02 Lifesize Communications, Inc. Forming beams with nulls directed at noise sources
US7593539B2 (en) * 2005-04-29 2009-09-22 Lifesize Communications, Inc. Microphone and speaker arrangement in speakerphone
US7970150B2 (en) * 2005-04-29 2011-06-28 Lifesize Communications, Inc. Tracking talkers using virtual broadside scan and directed beams
US7907164B2 (en) * 2005-05-02 2011-03-15 Lifesize Communications, Inc. Integrated videoconferencing system
US20060248210A1 (en) * 2005-05-02 2006-11-02 Lifesize Communications, Inc. Controlling video display mode in a video conferencing system
US20070165106A1 (en) * 2005-05-02 2007-07-19 Groves Randall D Distributed Videoconferencing Processing
US20060257827A1 (en) * 2005-05-12 2006-11-16 Blinktwice, Llc Method and apparatus to individualize content in an augmentative and alternative communication device
US8386628B1 (en) 2005-05-23 2013-02-26 Glance Networks, Inc. Method and apparatus for reducing the amount of information that must be transmitted to slower viewers over a remote viewing session
US8185583B2 (en) * 2005-06-03 2012-05-22 Siemens Enterprise Communications, Inc. Visualization enhanced presence system
US9165281B2 (en) * 2005-06-07 2015-10-20 Hewlett-Packard Development Company, L.P. System and method for enabling electronic presentations
US8199791B2 (en) 2005-06-08 2012-06-12 Polycom, Inc. Mixed voice and spread spectrum data signaling with enhanced concealment of data
JP4745726B2 (en) * 2005-06-13 2011-08-10 キヤノン株式会社 File management apparatus, control method therefor, computer program, and computer-readable storage medium
US8085318B2 (en) * 2005-10-11 2011-12-27 Apple Inc. Real-time image capture and manipulation based on streaming data
US7663691B2 (en) 2005-10-11 2010-02-16 Apple Inc. Image capture using display device as light source
US20060284895A1 (en) * 2005-06-15 2006-12-21 Marcu Gabriel G Dynamic gamma correction
US9514441B2 (en) * 2005-06-17 2016-12-06 Blackberry Limited Method and apparatus for switching between concurrent messaging sessions
US8826136B2 (en) * 2005-06-27 2014-09-02 Core Wireless Licensing S.A.R.L. System and method for enabling collaborative media stream editing
US20070266325A1 (en) * 2005-07-15 2007-11-15 Dittoware Inc. System and method for delivering presentations
US8613620B2 (en) * 2005-07-26 2013-12-24 Interactive Sports Direct Incorporated Method and system for providing web based interactive lessons with improved session playback
US8977965B1 (en) 2005-08-19 2015-03-10 At&T Intellectual Property Ii, L.P. System and method for controlling presentations using a multimodal interface
US9116989B1 (en) 2005-08-19 2015-08-25 At&T Intellectual Property Ii, L.P. System and method for using speech for data searching during presentations
US8677377B2 (en) 2005-09-08 2014-03-18 Apple Inc. Method and apparatus for building an intelligent automated assistant
CN1794722B (en) * 2005-09-19 2010-05-05 华为技术有限公司 Method of transfering off-line message and instant message server
US8130934B1 (en) * 2005-09-26 2012-03-06 At&T Intellectual Property Ii, L.P. Method and apparatus for providing network based muting of call legs
US9198728B2 (en) 2005-09-30 2015-12-01 Intouch Technologies, Inc. Multi-camera mobile teleconferencing platform
US20070100952A1 (en) * 2005-10-27 2007-05-03 Yen-Fu Chen Systems, methods, and media for playback of instant messaging session histrory
US20070101190A1 (en) * 2005-10-27 2007-05-03 International Business Machines Corporation Systems, methods, and media for sharing input device movement information in an instant messaging system
US9026915B1 (en) 2005-10-31 2015-05-05 At&T Intellectual Property Ii, L.P. System and method for creating a presentation using natural language
US8010728B1 (en) 2005-11-07 2011-08-30 Koninklijke Philips Electronics N.V. Multi-function docking assembly for portable digital media storage and playback device
JP4298736B2 (en) * 2005-11-09 2009-07-22 キヤノン株式会社 Information processing apparatus, electronic document processing method, and program
US8209620B2 (en) * 2006-01-31 2012-06-26 Accenture Global Services Limited System for storage and navigation of application states and interactions
US7599520B2 (en) * 2005-11-18 2009-10-06 Accenture Global Services Gmbh Detection of multiple targets on a plane of interest
US20070124370A1 (en) * 2005-11-29 2007-05-31 Microsoft Corporation Interactive table based platform to facilitate collaborative activities
US20070198744A1 (en) * 2005-11-30 2007-08-23 Ava Mobile, Inc. System, method, and computer program product for concurrent collaboration of media
FI20055644A0 (en) 2005-12-02 2005-12-02 Nokia Corp group Communications
WO2007068090A1 (en) * 2005-12-12 2007-06-21 Audiokinetic Inc. System and method for authoring media content
US20070182822A1 (en) * 2005-12-12 2007-08-09 Microsoft Corporation Media Composer
US9614977B1 (en) 2005-12-16 2017-04-04 At&T Intellectual Property Ii, L.P. Real-time media control for audio and multimedia conferencing services
US8311129B2 (en) * 2005-12-16 2012-11-13 Lifesize Communications, Inc. Temporal video filtering
US20070155395A1 (en) * 2005-12-29 2007-07-05 Nandu Gopalakrishnan Scheduling mobile users based on cell load
US8286089B2 (en) 2005-12-30 2012-10-09 Research In Motion Limited Representing new messages on a communication device
US8125931B2 (en) * 2006-01-10 2012-02-28 Utbk, Inc. Systems and methods to provide availability indication
US7720091B2 (en) * 2006-01-10 2010-05-18 Utbk, Inc. Systems and methods to arrange call back
IL173222A0 (en) * 2006-01-18 2006-06-11 Clip In Touch Internat Ltd Apparatus and method for creating and transmitting unique dynamically personalized multimedia messages
US20070171275A1 (en) * 2006-01-24 2007-07-26 Kenoyer Michael L Three Dimensional Videoconferencing
US8125509B2 (en) * 2006-01-24 2012-02-28 Lifesize Communications, Inc. Facial recognition for a videoconference
US8487976B2 (en) * 2006-01-24 2013-07-16 Lifesize Communications, Inc. Participant authentication for a videoconference
US8125508B2 (en) 2006-01-24 2012-02-28 Lifesize Communications, Inc. Sharing participant information in a videoconference
US8120638B2 (en) * 2006-01-24 2012-02-21 Lifesize Communications, Inc. Speech to text conversion in a videoconference
US8098599B2 (en) * 2006-02-13 2012-01-17 Tp Lab Inc. Method and system for multiple party telephone call
KR20070085125A (en) * 2006-02-22 2007-08-27 삼성전자주식회사 Method for managing an audio-video network and apparatus for the same
US20070208994A1 (en) * 2006-03-03 2007-09-06 Reddel Frederick A V Systems and methods for document annotation
JP4897520B2 (en) * 2006-03-20 2012-03-14 株式会社リコー Information distribution system
US20080013756A1 (en) * 2006-03-28 2008-01-17 Numark Industries, Llc Media storage manager and player
US8442197B1 (en) 2006-03-30 2013-05-14 Avaya Inc. Telephone-based user interface for participating simultaneously in more than one teleconference
US7969461B2 (en) 2006-03-30 2011-06-28 Polycom, Inc. System and method for exchanging connection information for videoconferencing units using instant messaging
US7589757B2 (en) * 2006-03-30 2009-09-15 Polycom, Inc. System and method for exchanging connection information for videoconferencing units using e-mails
EP1843550B1 (en) * 2006-04-05 2011-12-21 Motorola Mobility, Inc. Transport medium selection to optimize device operation
US20070239839A1 (en) * 2006-04-06 2007-10-11 Buday Michael E Method for multimedia review synchronization
US20070244671A1 (en) * 2006-04-13 2007-10-18 Drone Iyangar Systems and methods for storing, retrieving, and sharing design and construction data
KR100810269B1 (en) * 2006-04-18 2008-03-06 삼성전자주식회사 Wireless terminal and a method for performing video communication service using that
US7889851B2 (en) * 2006-04-20 2011-02-15 Cisco Technology, Inc. Accessing a calendar server to facilitate initiation of a scheduled call
US8015247B1 (en) * 2006-05-24 2011-09-06 Aol Inc. Joint communication sessions
US20070279483A1 (en) * 2006-05-31 2007-12-06 Beers Ted W Blended Space For Aligning Video Streams
US8407670B2 (en) * 2006-06-02 2013-03-26 Microsoft Corporation Collaborative code conflict detection, notification and resolution
CA2654580A1 (en) * 2006-06-09 2007-12-21 Aastra Usa, Inc. Automated group communication
US8849679B2 (en) 2006-06-15 2014-09-30 Intouch Technologies, Inc. Remote controlled robot system that provides medical images
US20070293206A1 (en) * 2006-06-20 2007-12-20 Microsoft Corporation Microsoft Patent Group Cell phone communication management
US7676582B2 (en) * 2006-06-30 2010-03-09 Microsoft Corporation Optimized desktop sharing viewer join
US20080005235A1 (en) * 2006-06-30 2008-01-03 Microsoft Corporation Collaborative integrated development environment using presence information
US20080013916A1 (en) * 2006-07-17 2008-01-17 Videothang Llc Systems and methods for encoding, editing and sharing multimedia files
US8296362B2 (en) * 2006-07-18 2012-10-23 Cisco Technology, Inc. Methods and apparatuses for dynamically enforcing privileges for use during a data collaboration session
US20080027726A1 (en) * 2006-07-28 2008-01-31 Eric Louis Hansen Text to audio mapping, and animation of the text
US7667762B2 (en) * 2006-08-01 2010-02-23 Lifesize Communications, Inc. Dual sensor video camera
US8294823B2 (en) * 2006-08-04 2012-10-23 Apple Inc. Video communication systems and methods
US7839434B2 (en) 2006-08-04 2010-11-23 Apple Inc. Video communication systems and methods
US7764632B2 (en) 2006-08-24 2010-07-27 Interwise Ltd. Software bridge for multi-point multi-media teleconferencing and telecollaboration
US9304675B2 (en) 2006-09-06 2016-04-05 Apple Inc. Portable electronic device for instant messaging
US9318108B2 (en) 2010-01-18 2016-04-19 Apple Inc. Intelligent automated assistant
US20080065695A1 (en) * 2006-09-11 2008-03-13 Pivi Unlimited Llc System and method for nondeterministic media playback selected from a plurality of distributed media libraries
US8266535B2 (en) 2006-09-11 2012-09-11 Broadnet Teleservices, Llc Teleforum apparatus and method
US7831727B2 (en) * 2006-09-11 2010-11-09 Apple Computer, Inc. Multi-content presentation of unassociated content types
US9202509B2 (en) 2006-09-12 2015-12-01 Sonos, Inc. Controlling and grouping in a multi-zone media system
US8483853B1 (en) 2006-09-12 2013-07-09 Sonos, Inc. Controlling and manipulating groupings in a multi-zone media system
US8788080B1 (en) 2006-09-12 2014-07-22 Sonos, Inc. Multi-channel pairing in a media system
US20080066001A1 (en) * 2006-09-13 2008-03-13 Majors Kenneth D Conferencing system with linked chat
US7853649B2 (en) * 2006-09-21 2010-12-14 Apple Inc. Audio processing for improved user experience
US8837697B2 (en) * 2006-09-29 2014-09-16 Verint Americas Inc. Call control presence and recording
US8199886B2 (en) 2006-09-29 2012-06-12 Verint Americas, Inc. Call control recording
US8433756B2 (en) * 2006-10-02 2013-04-30 Tp Lab, Inc. Multiple terminal collaboration system
US8271713B2 (en) * 2006-10-13 2012-09-18 Philips Electronics North America Corporation Interface systems for portable digital media storage and playback devices
US10037781B2 (en) * 2006-10-13 2018-07-31 Koninklijke Philips N.V. Interface systems for portable digital media storage and playback devices
US9918039B2 (en) * 2006-10-13 2018-03-13 Koninklijke Philips N.V. Interface systems for portable digital media storage and playback devices
EP2074822A4 (en) * 2006-10-13 2010-07-21 Koninkl Philips Electronics Nv Interface systems for portable digital media storage and playback devices
US20080138028A1 (en) * 2006-10-13 2008-06-12 Jeff Grady Interface systems for portable digital media storage and playback devices
KR101079591B1 (en) * 2006-11-21 2011-11-07 삼성전자주식회사 Display apparatus having the video call function, method thereof and video call system
US8384753B1 (en) 2006-12-15 2013-02-26 At&T Intellectual Property I, L. P. Managing multiple data sources
US20080147430A1 (en) * 2006-12-18 2008-06-19 Hawthorne John D Virtual Event Attendance Systems
US9425973B2 (en) * 2006-12-26 2016-08-23 International Business Machines Corporation Resource-based synchronization between endpoints in a web-based real time collaboration
JP5230096B2 (en) * 2006-12-27 2013-07-10 キヤノン株式会社 VIDEO / AUDIO OUTPUT DEVICE AND VIDEO / AUDIO OUTPUT METHOD
US20080215993A1 (en) * 2006-12-28 2008-09-04 New Tier, Inc. Communicator Program Manager
WO2008085205A2 (en) * 2006-12-29 2008-07-17 Prodea Systems, Inc. System and method for providing network support services and premises gateway support infrastructure
US8661035B2 (en) * 2006-12-29 2014-02-25 International Business Machines Corporation Content management system and method
US8407289B2 (en) * 2007-01-08 2013-03-26 Cisco Technology, Inc. Methods and apparatuses for dynamically suggesting an application based on a collaboration session
US8010711B2 (en) * 2007-01-26 2011-08-30 Digital Video Chip, Llc Universal multimedia
US8040369B1 (en) 2007-02-02 2011-10-18 Hewlett-Packard Development Company, L.P. System and method for transferring objects in a video conferencing session
JP2008245234A (en) * 2007-02-26 2008-10-09 Sony Corp Wireless communication device and wireless communication system
US9405830B2 (en) * 2007-02-28 2016-08-02 Aol Inc. Personalization techniques using image clouds
US8334891B2 (en) * 2007-03-05 2012-12-18 Cisco Technology, Inc. Multipoint conference video switching
RU2008108856A (en) * 2007-03-09 2009-09-20 Май Холливуд Лтд. (Il) DEVICE, SYSTEM AND METHOD OF ELECTRONIC COMMUNICATION USING AUDIO-VISUAL CLIPS
US7984158B2 (en) * 2007-03-20 2011-07-19 Microsoft Corporation Web service for coordinating actions of clients
US8219087B2 (en) * 2007-03-21 2012-07-10 Cellco Partnership Method, apparatus, computer program product and system for providing dynamic assignment of session capabilities
US8874445B2 (en) * 2007-03-30 2014-10-28 Verizon Patent And Licensing Inc. Apparatus and method for controlling output format of information
US7813277B2 (en) * 2007-06-29 2010-10-12 Packeteer, Inc. Lockless bandwidth management for multiprocessor networking devices
US8977255B2 (en) 2007-04-03 2015-03-10 Apple Inc. Method and system for operating a multi-function portable electronic device using voice-activation
US20080255840A1 (en) * 2007-04-16 2008-10-16 Microsoft Corporation Video Nametags
US8031222B2 (en) * 2007-04-25 2011-10-04 Microsoft Corporation Multiple resolution capture in real time communications
US8264521B2 (en) 2007-04-30 2012-09-11 Cisco Technology, Inc. Media detection and packet distribution in a multipoint conference
US8310521B2 (en) * 2007-04-30 2012-11-13 Microsoft Corp. Insertion of virtual video into live video
US9160783B2 (en) 2007-05-09 2015-10-13 Intouch Technologies, Inc. Robot system that operates through a network firewall
US20080300989A1 (en) * 2007-05-31 2008-12-04 Eyewonder, Inc. Systems and methods for generating, reviewing, editing, and transmitting an advertising unit in a single environment
US8122378B2 (en) * 2007-06-08 2012-02-21 Apple Inc. Image capture and manipulation
US20080307324A1 (en) * 2007-06-08 2008-12-11 Apple Inc. Sharing content in a videoconference session
US20090178062A1 (en) * 2007-06-08 2009-07-09 Apple Inc. Framework and Methods for Providing Communication Services to Client Applications
US7865017B2 (en) * 2007-06-08 2011-01-04 Seiko Epson Corporation Selection of regions within an image
US20080303949A1 (en) * 2007-06-08 2008-12-11 Apple Inc. Manipulating video streams
US20080307105A1 (en) * 2007-06-11 2008-12-11 Microsoft Corporation Streaming media archiver for live events
US9674675B2 (en) * 2007-06-20 2017-06-06 Qualcomm Incorporated Synchronizing floor control and media sharing in a half-duplex PTT system
US9210202B2 (en) 2007-06-20 2015-12-08 Qualcomm Incorporated System and method for sharing media in a group communication among wireless communication devices
US8237765B2 (en) * 2007-06-22 2012-08-07 Lifesize Communications, Inc. Video conferencing device which performs multi-way conferencing
FR2917931A1 (en) * 2007-06-22 2008-12-26 France Telecom METHOD AND SYSTEM FOR CONNECTING PEOPLE IN A TELECOMMUNICATIONS SYSTEM.
EP2007114B1 (en) * 2007-06-22 2016-08-10 Alcatel Lucent A system for providing information to users sharing a nomadic experience
US8532276B2 (en) 2007-06-26 2013-09-10 Ingenio Llc Systems and methods to provide telephonic connections via concurrent calls
US8782527B2 (en) * 2007-06-27 2014-07-15 Microsoft Corp. Collaborative phone-based file exchange
US8266534B2 (en) * 2007-06-28 2012-09-11 Microsoft Corporation Collaborative generation of meeting minutes and agenda confirmation
US9954996B2 (en) 2007-06-28 2018-04-24 Apple Inc. Portable electronic device with conversation management for incoming instant messages
US8526632B2 (en) * 2007-06-28 2013-09-03 Microsoft Corporation Microphone array for a camera speakerphone
US20090013264A1 (en) * 2007-06-28 2009-01-08 Anand Ganesh Basawapatna Enhanced interactive electronic meeting system
US20090006608A1 (en) * 2007-06-28 2009-01-01 Microsoft Corporation Dynamically enhancing meeting participation through compilation of data
US8165416B2 (en) * 2007-06-29 2012-04-24 Microsoft Corporation Automatic gain and exposure control using region of interest detection
US8330787B2 (en) * 2007-06-29 2012-12-11 Microsoft Corporation Capture device movement compensation for speaker indexing
US8139100B2 (en) * 2007-07-13 2012-03-20 Lifesize Communications, Inc. Virtual multiway scaler compensation
US8347207B2 (en) * 2007-07-16 2013-01-01 International Business Machines Corporation Automatically moving annotations associated with multidimensional data between live datacubes
US7840415B2 (en) 2007-07-30 2010-11-23 Eric Schifone Method and platform for providing an interactive internet computer-driven/IP based streaming video/audio apparatus
CN101365096B (en) * 2007-08-09 2012-05-23 华为技术有限公司 Method for providing video content, related service apparatus and system
KR101403839B1 (en) * 2007-08-16 2014-06-03 엘지전자 주식회사 Mobile communication terminal with touchscreen and display control method thereof
KR101435800B1 (en) * 2007-08-20 2014-08-29 엘지전자 주식회사 Portable terminal, method for transmitting data in the portable terminal and program recording medium
US10091345B2 (en) * 2007-09-04 2018-10-02 Apple Inc. Media out interface
US8060063B1 (en) 2007-09-07 2011-11-15 Sprint Communications Company L.P. Presenting messages on a mobile device that is currently presenting other media content
US8086223B2 (en) * 2007-09-07 2011-12-27 International Business Machines Corporation Using proximity information to dynamically change greeting messages on a mobile telephony device
US20090077473A1 (en) * 2007-09-13 2009-03-19 James Williams System and method of managing workflow of an authoring assignment
US9661267B2 (en) * 2007-09-20 2017-05-23 Lifesize, Inc. Videoconferencing system discovery
US9060094B2 (en) 2007-09-30 2015-06-16 Optical Fusion, Inc. Individual adjustment of audio and video properties in network conferencing
US8004529B2 (en) * 2007-10-01 2011-08-23 Apple Inc. Processing an animation file to provide an animated icon
US8578000B2 (en) * 2008-12-05 2013-11-05 Social Communications Company Realtime kernel
US8732236B2 (en) * 2008-12-05 2014-05-20 Social Communications Company Managing network communications between network nodes and stream transport protocol
US20090112983A1 (en) * 2007-10-29 2009-04-30 Motorola, Inc. Floor control in a communications system
US8661096B2 (en) * 2007-11-05 2014-02-25 Cyberlink Corp. Collaborative editing in a video editing system
US8739214B2 (en) * 2007-11-08 2014-05-27 At&T Intellectual Property I, L.P. Methods, computer program products, and virtual servers for a virtual collaborative environment
US9602295B1 (en) 2007-11-09 2017-03-21 Avaya Inc. Audio conferencing server for the internet
US8032169B2 (en) * 2007-11-28 2011-10-04 Motorola Solutions, Inc. System and method for providing low overhead floor control in a distributed peer-to-peer communications network
US9268837B2 (en) 2007-12-04 2016-02-23 International Business Machines Corporation Data entry commentary and sheet reconstruction for multidimensional enterprise system
US10002189B2 (en) 2007-12-20 2018-06-19 Apple Inc. Method and apparatus for searching using an active ontology
US9886503B2 (en) 2007-12-27 2018-02-06 Sirius Xm Radio Inc. Method and apparatus for multiplexing audio program channels from one or more received broadcast streams to provide a playlist style listening experience to users
US8291308B2 (en) 2007-12-28 2012-10-16 International Business Machines Corporation Annotating collaborative information structures
US9330720B2 (en) 2008-01-03 2016-05-03 Apple Inc. Methods and apparatus for altering audio output signals
US8327272B2 (en) 2008-01-06 2012-12-04 Apple Inc. Portable multifunction device, method, and graphical user interface for viewing and managing electronic calendars
US8015149B1 (en) * 2008-01-15 2011-09-06 Adobe Systems Incorporated Asset repository
KR20090081310A (en) * 2008-01-23 2009-07-28 삼성전자주식회사 Apparatus and method for providing host function in a multi-standby portable terminal
US20090210789A1 (en) * 2008-02-14 2009-08-20 Microsoft Corporation Techniques to generate a visual composition for a multimedia conference event
US8275108B2 (en) 2008-02-26 2012-09-25 International Business Machines Corporation Hierarchal control of teleconferences
US20090228558A1 (en) * 2008-03-05 2009-09-10 Brenner Michael R Time management for outgoing electronic mail
US8869032B2 (en) 2008-03-13 2014-10-21 International Business Machines Corporation Telecom web browsers, and methods for defining a telecom web browser
US8831197B2 (en) * 2008-03-14 2014-09-09 Cisco Technology, Inc. One button conference initiation
US9357164B2 (en) * 2008-03-18 2016-05-31 Cisco Technology, Inc. Establishing a remotely hosted conference initiated with one button push
US8126048B2 (en) * 2008-03-18 2012-02-28 Seiko Epson Corporation Recording streaming delta-encoded data
US8139923B2 (en) * 2008-03-19 2012-03-20 Seiko Epson Corporation Playback of recorded streaming delta-encoded data
US8867711B2 (en) * 2008-03-19 2014-10-21 International Business Machines Corporation Telecom web browsers, and methods for defining a telecom web browser
US10875182B2 (en) 2008-03-20 2020-12-29 Teladoc Health, Inc. Remote presence system mounted to operating room hardware
US10091460B2 (en) * 2008-03-31 2018-10-02 Disney Enterprises, Inc. Asynchronous online viewing party
US8484291B1 (en) 2008-04-02 2013-07-09 Glance Networks, Inc. Method and apparatus for selecting commands for transmission from an updated queue
US8996376B2 (en) 2008-04-05 2015-03-31 Apple Inc. Intelligent text-to-speech conversion
US9514444B2 (en) * 2009-01-15 2016-12-06 Sococo, Inc. Encapsulating virtual area based communicant assemblies
US8896658B2 (en) * 2008-04-10 2014-11-25 Creative Technology Ltd Interface for voice communications
US8179418B2 (en) 2008-04-14 2012-05-15 Intouch Technologies, Inc. Robotic based health care system
US8170241B2 (en) 2008-04-17 2012-05-01 Intouch Technologies, Inc. Mobile tele-presence system with a microphone system
US9310980B2 (en) 2012-08-21 2016-04-12 Beyondvia Technologies Systems and methods for performance comparisons utilizing an infinite cylinder user interface
US10304126B2 (en) 2008-04-30 2019-05-28 Beyondvia Technologies Visual communication systems and methods designing and building entire experiences
US8769423B2 (en) 2008-04-30 2014-07-01 Beyondvia Technologies Content distribution systems and methods using digital images
US9305026B2 (en) 2008-04-30 2016-04-05 Beyondvia Technologies Systems and methods for digital images utilizing an infinite cylinder user interface
US8806353B2 (en) * 2008-04-30 2014-08-12 Beyondvia Technologies Mobile electronic communications dialog using sequenced digital images
US8856003B2 (en) * 2008-04-30 2014-10-07 Motorola Solutions, Inc. Method for dual channel monitoring on a radio device
US8379851B2 (en) * 2008-05-12 2013-02-19 Microsoft Corporation Optimized client side rate control and indexed file layout for streaming media
US10496753B2 (en) 2010-01-18 2019-12-03 Apple Inc. Automatically adapting user interfaces for hands-free interaction
US7529804B1 (en) 2008-05-15 2009-05-05 International Business Machines Corporation System and method for comprehensive automatic color customization in an email message based on cultural perspective
US20090307610A1 (en) * 2008-06-10 2009-12-10 Melonie Elizabeth Ryan Method for a plurality of users to be simultaneously matched to interact one on one in a live controlled environment
US8260289B2 (en) * 2008-06-11 2012-09-04 International Business Machines Corporation Dynamic user-to-user address book sharing in a mobile station
ITMC20080120A1 (en) * 2008-06-20 2009-12-21 Nazario Luzi VIDEO RECOVERY SYSTEM IN PARTICULAR FOR FISHERIES.
US20090319916A1 (en) * 2008-06-24 2009-12-24 Microsoft Corporation Techniques to auto-attend multimedia conference events
US20090327907A1 (en) * 2008-06-30 2009-12-31 Microsoft Corporation Integrating character-based profiles within a social network
US20090327906A1 (en) * 2008-06-30 2009-12-31 Microsoft Corporation Supporting brand assets in a social networking service
US9193065B2 (en) 2008-07-10 2015-11-24 Intouch Technologies, Inc. Docking system for a tele-presence robot
US9842192B2 (en) 2008-07-11 2017-12-12 Intouch Technologies, Inc. Tele-presence robot system with multi-cast features
US8271887B2 (en) * 2008-07-17 2012-09-18 The Boeing Company Systems and methods for whiteboard collaboration and annotation
US10127231B2 (en) * 2008-07-22 2018-11-13 At&T Intellectual Property I, L.P. System and method for rich media annotation
US20100030549A1 (en) 2008-07-31 2010-02-04 Lee Michael M Mobile device having human language translation capability with positional feedback
US8909810B2 (en) * 2008-08-05 2014-12-09 Isabella Products, Inc. Systems and methods for multimedia content sharing
US8407609B2 (en) 2008-08-21 2013-03-26 Linqware Inc. System and method for providing and tracking the provision of audio and visual presentations via a computer network
WO2010025343A1 (en) * 2008-08-28 2010-03-04 Jonas Lawrence A System for integrating multiple im networks and social networking websites
US8489999B2 (en) * 2008-09-02 2013-07-16 Accenture Global Services Limited Shared user interface surface system
GB2463109B (en) * 2008-09-05 2013-03-13 Skype Communication system and method
GB2463108B (en) * 2008-09-05 2012-08-29 Skype Communication system and method
GB2463104A (en) 2008-09-05 2010-03-10 Skype Ltd Thumbnail selection of telephone contact using zooming
GB2463105A (en) 2008-09-05 2010-03-10 Skype Ltd Viewer activity dependent video telephone call ringing
GB2463110B (en) * 2008-09-05 2013-01-16 Skype Communication system and method
GB2463124B (en) * 2008-09-05 2012-06-20 Skype Ltd A peripheral device for communication over a communications sytem
GB2463103A (en) * 2008-09-05 2010-03-10 Skype Ltd Video telephone call using a television receiver
GB2463107A (en) * 2008-09-05 2010-03-10 Skype Ltd A remote control unit of a media device for placing/receiving calls, comprising activating one of the two wireless transceivers when needed.
US20100064033A1 (en) * 2008-09-08 2010-03-11 Franco Travostino Integration of an internal cloud infrastructure with existing enterprise services and systems
US8731519B2 (en) 2008-09-08 2014-05-20 At&T Mobility Ii Llc Mobile handset extension to a device
US20100070881A1 (en) * 2008-09-12 2010-03-18 At&T Intellectual Property I, L.P. Project facilitation and collaboration application
US8340819B2 (en) 2008-09-18 2012-12-25 Intouch Technologies, Inc. Mobile videoconferencing robot system with network adaptive driving
US8856641B2 (en) * 2008-09-24 2014-10-07 Yahoo! Inc. Time-tagged metainformation and content display method and system
US20100080361A1 (en) * 2008-09-29 2010-04-01 Conrad Edward Houghton Method for Sharing Audio-only content, Audio-Visual content, and Visual-only content between Subscribers on a Telephone call
JP5369587B2 (en) * 2008-09-30 2013-12-18 富士通株式会社 Information processing apparatus, information processing system, method, and program
US8676904B2 (en) 2008-10-02 2014-03-18 Apple Inc. Electronic devices with voice command and contextual data processing capabilities
US8514265B2 (en) * 2008-10-02 2013-08-20 Lifesize Communications, Inc. Systems and methods for selecting videoconferencing endpoints for display in a composite video image
US8996165B2 (en) 2008-10-21 2015-03-31 Intouch Technologies, Inc. Telepresence robot with a camera boom
KR101531508B1 (en) * 2008-10-27 2015-06-26 엘지전자 주식회사 A method and an apparatus for data transmitting and receiving using an mobile terminal
US20100110160A1 (en) * 2008-10-30 2010-05-06 Brandt Matthew K Videoconferencing Community with Live Images
US8154580B2 (en) * 2008-11-21 2012-04-10 Creative Technology Ltd System and method for facilitating user communication from a location
US9401937B1 (en) * 2008-11-24 2016-07-26 Shindig, Inc. Systems and methods for facilitating communications amongst multiple users
US8405702B1 (en) 2008-11-24 2013-03-26 Shindig, Inc. Multiparty communications systems and methods that utilize multiple modes of communication
US9138891B2 (en) 2008-11-25 2015-09-22 Intouch Technologies, Inc. Server connectivity control for tele-presence robot
US8463435B2 (en) 2008-11-25 2013-06-11 Intouch Technologies, Inc. Server connectivity control for tele-presence robot
KR101050642B1 (en) 2008-12-04 2011-07-19 삼성전자주식회사 Watch phone and method of conducting call in watch phone
WO2010067118A1 (en) 2008-12-11 2010-06-17 Novauris Technologies Limited Speech recognition involving a mobile device
US20100153112A1 (en) * 2008-12-16 2010-06-17 Motorola, Inc. Progressively refining a speech-based search
US8230352B2 (en) * 2008-12-18 2012-07-24 Sap Ag Collaboration interface for a multi-channel collaboration window with context support
JP5200928B2 (en) * 2008-12-29 2013-06-05 ブラザー工業株式会社 Video conference system, bandwidth control method, conference control device, video conference terminal device, and program
US8112480B2 (en) * 2009-01-16 2012-02-07 Microsoft Corporation Signaling support for sharer switching in application sharing
US8849680B2 (en) 2009-01-29 2014-09-30 Intouch Technologies, Inc. Documentation through a remote presence robot
KR101525760B1 (en) * 2009-02-26 2015-06-04 삼성전자주식회사 User Interface for supporting call function and Portable Device using the same
US8456510B2 (en) * 2009-03-04 2013-06-04 Lifesize Communications, Inc. Virtual distributed multipoint control unit
US8643695B2 (en) * 2009-03-04 2014-02-04 Lifesize Communications, Inc. Videoconferencing endpoint extension
US20100228825A1 (en) * 2009-03-06 2010-09-09 Microsoft Corporation Smart meeting room
CN103415534A (en) * 2009-03-10 2013-11-27 贝勒研究院 Antigen presenting cell targeted cancer vaccines
US8549401B1 (en) * 2009-03-30 2013-10-01 Symantec Corporation Systems and methods for automatically generating computer-assistance videos
US9344745B2 (en) * 2009-04-01 2016-05-17 Shindig, Inc. Group portraits composed using video chat systems
US20100257457A1 (en) * 2009-04-07 2010-10-07 De Goes John A Real-time content collaboration
CN101534411B (en) * 2009-04-08 2012-12-12 华为终端有限公司 Control method for video conference, terminal and system based on image
US8977684B2 (en) 2009-04-14 2015-03-10 Citrix Systems, Inc. Systems and methods for computer and voice conference audio transmission during conference call via VoIP device
US8897920B2 (en) 2009-04-17 2014-11-25 Intouch Technologies, Inc. Tele-presence robot system with software modularity, projector and laser pointer
US8621011B2 (en) 2009-05-12 2013-12-31 Avaya Inc. Treatment of web feeds as work assignment in a contact center
US8332755B2 (en) 2009-05-27 2012-12-11 Microsoft Corporation Force-feedback within telepresence
US20100306670A1 (en) * 2009-05-29 2010-12-02 Microsoft Corporation Gesture-based document sharing manipulation
US8095119B2 (en) * 2009-06-02 2012-01-10 Microsoft Corporation In-call contact information display
US8711198B2 (en) * 2009-06-04 2014-04-29 Hewlett-Packard Development Company, L.P. Video conference
US10706373B2 (en) 2011-06-03 2020-07-07 Apple Inc. Performing actions associated with task items that represent tasks to perform
US10241644B2 (en) 2011-06-03 2019-03-26 Apple Inc. Actionable reminder entries
US10241752B2 (en) 2011-09-30 2019-03-26 Apple Inc. Interface for a virtual digital assistant
US9858925B2 (en) 2009-06-05 2018-01-02 Apple Inc. Using context information to facilitate processing of commands in a virtual assistant
US8305421B2 (en) * 2009-06-29 2012-11-06 Lifesize Communications, Inc. Automatic determination of a configuration for a conference
US9431006B2 (en) 2009-07-02 2016-08-30 Apple Inc. Methods and apparatuses for automatic speech recognition
US9025497B2 (en) 2009-07-10 2015-05-05 Qualcomm Incorporated Media forwarding for a group communication session in a wireless communications system
US9088630B2 (en) * 2009-07-13 2015-07-21 Qualcomm Incorporated Selectively mixing media during a group communication session within a wireless communications system
US20110035683A1 (en) * 2009-08-07 2011-02-10 Larry Stead Method and apparatus for synchronous, collaborative media consumption
US20110037864A1 (en) * 2009-08-17 2011-02-17 Microseven Systems, LLC Method and apparatus for live capture image
US9165073B2 (en) 2009-08-17 2015-10-20 Shoutpoint, Inc. Apparatus, system and method for a web-based interactive video platform
US8384755B2 (en) 2009-08-26 2013-02-26 Intouch Technologies, Inc. Portable remote presence robot
US11399153B2 (en) 2009-08-26 2022-07-26 Teladoc Health, Inc. Portable telepresence apparatus
US9020472B2 (en) * 2009-09-04 2015-04-28 Twalkin, Inc. Social networking-based teleconferencing system and method
US8363810B2 (en) * 2009-09-08 2013-01-29 Avaya Inc. Method and system for aurally positioning voice signals in a contact center environment
US8448073B2 (en) * 2009-09-09 2013-05-21 Viewplicity, Llc Multiple camera group collaboration system and method
US9367876B2 (en) 2009-09-18 2016-06-14 Salesforce.Com, Inc. Systems and methods for multimedia multipoint real-time conferencing allowing real-time bandwidth management and prioritized media distribution
US9158409B2 (en) * 2009-09-29 2015-10-13 Beijing Lenovo Software Ltd Object determining method, object display method, object switching method and electronic device
US8547880B2 (en) * 2009-09-30 2013-10-01 Avaya Inc. Method and system for replaying a portion of a multi-party audio interaction
US8677240B2 (en) * 2009-10-05 2014-03-18 Harris Corporation Video processing system providing association between displayed video and media content and related methods
US8358746B2 (en) * 2009-10-15 2013-01-22 Avaya Inc. Method and apparatus for unified interface for heterogeneous session management
US8363796B2 (en) * 2009-10-15 2013-01-29 Avaya Inc. Selection and initiation of IVR scripts by contact center agents
US8350891B2 (en) * 2009-11-16 2013-01-08 Lifesize Communications, Inc. Determining a videoconference layout based on numbers of participants
US20110115906A1 (en) * 2009-11-17 2011-05-19 Amroad Technology, Inc. Monitoring system and control method thereof
JP5400599B2 (en) * 2009-12-18 2014-01-29 株式会社日立製作所 GUI customization method, system, and program
US8526798B2 (en) * 2009-12-23 2013-09-03 Western Digital Technologies, Inc. Portable content container displaying A/V files in response to a command received from a consumer device
US10553209B2 (en) 2010-01-18 2020-02-04 Apple Inc. Systems and methods for hands-free notification summaries
US10276170B2 (en) 2010-01-18 2019-04-30 Apple Inc. Intelligent automated assistant
US10679605B2 (en) 2010-01-18 2020-06-09 Apple Inc. Hands-free list-reading by intelligent automated assistant
US10705794B2 (en) 2010-01-18 2020-07-07 Apple Inc. Automatically adapting user interfaces for hands-free interaction
DE202011111062U1 (en) 2010-01-25 2019-02-19 Newvaluexchange Ltd. Device and system for a digital conversation management platform
US8572303B2 (en) 2010-02-03 2013-10-29 Tal Lavian Portable universal communication device
US9001819B1 (en) 2010-02-18 2015-04-07 Zvi Or-Bach Systems and methods for visual presentation and selection of IVR menu
US8553859B1 (en) 2010-02-03 2013-10-08 Tal Lavian Device and method for providing enhanced telephony
US8879698B1 (en) 2010-02-03 2014-11-04 Tal Lavian Device and method for providing enhanced telephony
US8681951B1 (en) 2010-02-03 2014-03-25 Tal Lavian Systems and methods for visual presentation and selection of IVR menu
US8625756B1 (en) 2010-02-03 2014-01-07 Tal Lavian Systems and methods for visual presentation and selection of IVR menu
US8903073B2 (en) 2011-07-20 2014-12-02 Zvi Or-Bach Systems and methods for visual presentation and selection of IVR menu
US8548131B1 (en) 2010-02-03 2013-10-01 Tal Lavian Systems and methods for communicating with an interactive voice response system
US8687777B1 (en) 2010-02-03 2014-04-01 Tal Lavian Systems and methods for visual presentation and selection of IVR menu
US8537989B1 (en) 2010-02-03 2013-09-17 Tal Lavian Device and method for providing enhanced telephony
US8548135B1 (en) 2010-02-03 2013-10-01 Tal Lavian Systems and methods for visual presentation and selection of IVR menu
US8406388B2 (en) 2011-07-18 2013-03-26 Zvi Or-Bach Systems and methods for visual presentation and selection of IVR menu
US8594280B1 (en) 2010-02-03 2013-11-26 Zvi Or-Bach Systems and methods for visual presentation and selection of IVR menu
US11154981B2 (en) 2010-02-04 2021-10-26 Teladoc Health, Inc. Robot user interface for telepresence robot system
US8682667B2 (en) 2010-02-25 2014-03-25 Apple Inc. User profiling for selecting user specific voice input processing information
US20110219097A1 (en) * 2010-03-04 2011-09-08 Dolby Laboratories Licensing Corporation Techniques For Client Device Dependent Filtering Of Metadata
US8670017B2 (en) 2010-03-04 2014-03-11 Intouch Technologies, Inc. Remote presence system including a cart that supports a robot face and an overhead camera
US8869141B2 (en) * 2010-03-09 2014-10-21 Avistar Communications Corp. Scalable high-performance interactive real-time media architectures for virtual desktop environments
EP3110227B1 (en) * 2010-03-11 2018-10-03 Nokia Technologies Oy Method and apparatus for device-to-device communication setup
JP4957821B2 (en) * 2010-03-18 2012-06-20 コニカミノルタビジネステクノロジーズ株式会社 CONFERENCE SYSTEM, INFORMATION PROCESSING DEVICE, DISPLAY METHOD, AND DISPLAY PROGRAM
US9203790B2 (en) * 2010-03-26 2015-12-01 Socon Media, Inc. Method, system and computer program product for controlled networked communication
US20130298040A1 (en) * 2010-04-30 2013-11-07 American Teleconferencing Services, Ltd. Systems, Methods, and Computer Programs for Providing Simultaneous Online Conferences
US10343283B2 (en) 2010-05-24 2019-07-09 Intouch Technologies, Inc. Telepresence robot system that can be accessed by a cellular phone
US10808882B2 (en) 2010-05-26 2020-10-20 Intouch Technologies, Inc. Tele-robotic system with a robot face placed on a chair
US9166712B2 (en) 2010-06-22 2015-10-20 Sirius Xm Radio Inc. Method and apparatus for multiplexing audio program channels from one or more received broadcast streams to provide a playlist style listening experience to users
US20120011465A1 (en) * 2010-07-06 2012-01-12 Marcelo Amaral Rezende Digital whiteboard system
US20120030595A1 (en) * 2010-07-29 2012-02-02 Seiko Epson Corporation Information storage medium, terminal apparatus, and image generation method
US10019741B2 (en) * 2010-08-09 2018-07-10 Western Digital Technologies, Inc. Methods and systems for a personal multimedia content archive
US8802957B2 (en) * 2010-08-16 2014-08-12 Boardwalk Technology Group, Llc Mobile replacement-dialogue recording system
JP5672862B2 (en) * 2010-08-27 2015-02-18 ソニー株式会社 Imaging apparatus, imaging system, and imaging method
US8744065B2 (en) 2010-09-22 2014-06-03 Avaya Inc. Method and system for monitoring contact center transactions
KR101824647B1 (en) * 2010-10-12 2018-02-02 삼성전자주식회사 Method for synchronizing between devices, user device, and computer readable storage medium thereof
US9129138B1 (en) 2010-10-29 2015-09-08 Western Digital Technologies, Inc. Methods and systems for a portable data locker
US9736312B2 (en) 2010-11-17 2017-08-15 Avaya Inc. Method and system for controlling audio signals in multiple concurrent conference calls
US8938216B2 (en) 2010-11-24 2015-01-20 Cisco Technology, Inc. Geographical location information/signal quality-context based recording and playback of multimedia data from a conference session
US9264664B2 (en) 2010-12-03 2016-02-16 Intouch Technologies, Inc. Systems and methods for dynamic bandwidth allocation
US10762293B2 (en) 2010-12-22 2020-09-01 Apple Inc. Using parts-of-speech tagging and named entity recognition for spelling correction
US8977310B2 (en) 2010-12-30 2015-03-10 Motorola Solutions, Inc. Methods for coordinating wireless coverage between different wireless networks for members of a communication group
US12093036B2 (en) 2011-01-21 2024-09-17 Teladoc Health, Inc. Telerobotic system with a dual application screen presentation
US11265652B2 (en) 2011-01-25 2022-03-01 Sonos, Inc. Playback device pairing
US11429343B2 (en) 2011-01-25 2022-08-30 Sonos, Inc. Stereo playback configuration and control
US8411132B2 (en) 2011-01-27 2013-04-02 Audio Properties, Inc. System and method for real-time media data review
US9323250B2 (en) 2011-01-28 2016-04-26 Intouch Technologies, Inc. Time-dependent navigation of telepresence robots
EP2668008A4 (en) 2011-01-28 2018-01-24 Intouch Technologies, Inc. Interfacing with a mobile telepresence robot
US9154248B2 (en) 2011-02-14 2015-10-06 Sirius Xm Radio Inc. Method and apparatus for enhanced playback of content while switching among channels of broadcast or streamed content while being received
US9531818B2 (en) * 2011-03-01 2016-12-27 Lenovo (Beijing) Limited Information notification methods, apparatuses and electronic devices
US8698872B2 (en) * 2011-03-02 2014-04-15 At&T Intellectual Property I, Lp System and method for notification of events of interest during a video conference
WO2012118917A2 (en) 2011-03-03 2012-09-07 Social Communications Company Realtime communications and network browsing client
US9262612B2 (en) 2011-03-21 2016-02-16 Apple Inc. Device access using voice authentication
US8719369B1 (en) 2011-04-25 2014-05-06 Brian Brown Customized message generation
US10769739B2 (en) 2011-04-25 2020-09-08 Intouch Technologies, Inc. Systems and methods for management of information among medical providers and facilities
US9098611B2 (en) 2012-11-26 2015-08-04 Intouch Technologies, Inc. Enhanced video interaction for a user interface of a telepresence network
US20140139616A1 (en) 2012-01-27 2014-05-22 Intouch Technologies, Inc. Enhanced Diagnostics for a Telepresence Robot
US20120304062A1 (en) * 2011-05-23 2012-11-29 Speakertext, Inc. Referencing content via text captions
US20120307989A1 (en) * 2011-06-02 2012-12-06 Hawley Michael J Method and apparatus for multimedia calls
US10057736B2 (en) 2011-06-03 2018-08-21 Apple Inc. Active transport based notifications
US8375085B2 (en) * 2011-07-06 2013-02-12 Avaya Inc. System and method of enhanced collaboration through teleportation
US8994660B2 (en) 2011-08-29 2015-03-31 Apple Inc. Text correction processing
CN103001982B (en) * 2011-09-09 2017-04-26 华为技术有限公司 Real-time sharing method, real-time sharing device and real-time sharing system
US9448708B1 (en) * 2011-10-19 2016-09-20 Google Inc. Theming for virtual collaboration
US8836751B2 (en) 2011-11-08 2014-09-16 Intouch Technologies, Inc. Tele-presence system with a user interface that displays different communication links
KR101659674B1 (en) * 2011-11-27 2016-09-30 가부시키가이샤 시너지드라이브 Voice link system
US20130139256A1 (en) 2011-11-30 2013-05-30 Elwha LLC, a limited liability corporation of the State of Delaware Deceptive indicia profile generation from communications interactions
US9601117B1 (en) * 2011-11-30 2017-03-21 West Corporation Method and apparatus of processing user data of a multi-speaker conference call
US10250939B2 (en) * 2011-11-30 2019-04-02 Elwha Llc Masking of deceptive indicia in a communications interaction
US9378366B2 (en) 2011-11-30 2016-06-28 Elwha Llc Deceptive indicia notification in a communications interaction
US20130139259A1 (en) 2011-11-30 2013-05-30 Elwha Llc Deceptive indicia profile generation from communications interactions
US9832510B2 (en) 2011-11-30 2017-11-28 Elwha, Llc Deceptive indicia profile generation from communications interactions
US9245020B2 (en) * 2011-12-14 2016-01-26 Microsoft Technology Licensing, Llc Collaborative media sharing
US20130159173A1 (en) * 2011-12-19 2013-06-20 Sridhar Sivaraman Shared Mobile Payments
CA2862431A1 (en) * 2012-01-11 2013-07-18 Smart Technologies Ulc Method of displaying input during a collaboration session and interactive board employing same
US8731148B1 (en) 2012-03-02 2014-05-20 Tal Lavian Systems and methods for visual presentation and selection of IVR menu
US8867708B1 (en) 2012-03-02 2014-10-21 Tal Lavian Systems and methods for visual presentation and selection of IVR menu
US10134385B2 (en) 2012-03-02 2018-11-20 Apple Inc. Systems and methods for name pronunciation
US9483461B2 (en) 2012-03-06 2016-11-01 Apple Inc. Handling speech synthesis of content for multiple languages
US8406155B1 (en) 2012-03-19 2013-03-26 Google Inc. Cloud based contact center platform powered by individual multi-party conference rooms
US9137029B1 (en) 2012-03-19 2015-09-15 Google Inc. State and availability monitoring for customer support services for multimedia conferences
US9071681B1 (en) 2012-03-19 2015-06-30 Google Inc. Inbound telephony orchestrator for hangout-based contact center platform
US9251313B2 (en) 2012-04-11 2016-02-02 Intouch Technologies, Inc. Systems and methods for visualizing and managing telepresence devices in healthcare networks
US8902278B2 (en) 2012-04-11 2014-12-02 Intouch Technologies, Inc. Systems and methods for visualizing and managing telepresence devices in healthcare networks
US20130283318A1 (en) * 2012-04-23 2013-10-24 Comcast Cable Communications, LLC. Dynamic Mosaic for Creation of Video Rich User Interfaces
US9729115B2 (en) 2012-04-27 2017-08-08 Sonos, Inc. Intelligently increasing the sound level of player
US9280610B2 (en) 2012-05-14 2016-03-08 Apple Inc. Crowd sourcing information to fulfill user requests
WO2013176758A1 (en) 2012-05-22 2013-11-28 Intouch Technologies, Inc. Clinical workflows utilizing autonomous and semi-autonomous telemedicine devices
US9361021B2 (en) 2012-05-22 2016-06-07 Irobot Corporation Graphical user interfaces including touchpad driving interfaces for telemedicine devices
CN104521226A (en) * 2012-06-06 2015-04-15 卡尔加里科学公司 Image viewing architecture having integrated collaboratively-based secure file transfer mechanism
US9721563B2 (en) 2012-06-08 2017-08-01 Apple Inc. Name recognition system
US9495129B2 (en) 2012-06-29 2016-11-15 Apple Inc. Device, method, and user interface for voice-activated navigation and browsing of a document
US9113033B2 (en) 2012-08-28 2015-08-18 Microsoft Technology Licensing, Llc Mobile video conferencing with digital annotation
US9576574B2 (en) 2012-09-10 2017-02-21 Apple Inc. Context-sensitive handling of interruptions by intelligent digital assistant
US20140082049A1 (en) * 2012-09-18 2014-03-20 Kanan Abbas Babayev Method for media-data broadcasting between communication network users
US9547647B2 (en) 2012-09-19 2017-01-17 Apple Inc. Voice-based media searching
US9565314B2 (en) 2012-09-27 2017-02-07 Dolby Laboratories Licensing Corporation Spatial multiplexing in a soundfield teleconferencing system
US9008330B2 (en) 2012-09-28 2015-04-14 Sonos, Inc. Crossover frequency adjustments for audio speakers
US8706155B1 (en) * 2012-10-02 2014-04-22 Google Inc. Transmission protocol modification to maximize mobile device battery life
US20140114664A1 (en) * 2012-10-20 2014-04-24 Microsoft Corporation Active Participant History in a Video Conferencing System
US20140123014A1 (en) * 2012-11-01 2014-05-01 Inxpo, Inc. Method and system for chat and activity stream capture and playback
USD754161S1 (en) 2012-11-26 2016-04-19 Nero Ag Device with a display screen with graphical user interface
US20140149875A1 (en) * 2012-11-26 2014-05-29 Nero Ag System and method for presentation of a tapestry interface
US20140149932A1 (en) * 2012-11-26 2014-05-29 Nero Ag System and method for providing a tapestry presentation
US20140164951A1 (en) * 2012-12-10 2014-06-12 Microsoft Corporation Group nudge using real-time communication system
KR20240132105A (en) 2013-02-07 2024-09-02 애플 인크. Voice trigger for a digital assistant
US9503485B1 (en) * 2013-03-01 2016-11-22 Whatsapp Inc. Connecting communicating devices in a multi-server communication system
US9368114B2 (en) 2013-03-14 2016-06-14 Apple Inc. Context-sensitive handling of interruptions
WO2014144579A1 (en) 2013-03-15 2014-09-18 Apple Inc. System and method for updating an adaptive speech recognition model
AU2014233517B2 (en) 2013-03-15 2017-05-25 Apple Inc. Training an at least partial voice command system
KR20140117060A (en) * 2013-03-26 2014-10-07 한국전자통신연구원 Method and Apparatus OF Controlling MAC-LAYER ProtocoL for Device-to-Device Communications without ID
US9736203B2 (en) * 2013-03-28 2017-08-15 Ittiam Systems Pte. Ltd. System and method for virtual social colocation
ES2594610T3 (en) * 2013-04-26 2016-12-21 Airbus Defence And Space Limited Data routing within a communications network
US9047901B1 (en) 2013-05-28 2015-06-02 Western Digital Technologies, Inc. Disk drive measuring spiral track error by measuring a slope of a spiral track across a disk radius
WO2014197334A2 (en) 2013-06-07 2014-12-11 Apple Inc. System and method for user-specified pronunciation of words for speech synthesis and recognition
US9582608B2 (en) 2013-06-07 2017-02-28 Apple Inc. Unified ranking with entropy-weighted information for phrase-based semantic auto-completion
WO2014197336A1 (en) 2013-06-07 2014-12-11 Apple Inc. System and method for detecting errors in interactions with a voice-based digital assistant
WO2014197335A1 (en) 2013-06-08 2014-12-11 Apple Inc. Interpreting and acting upon commands that involve sharing information with remote devices
KR101772152B1 (en) 2013-06-09 2017-08-28 애플 인크. Device, method, and graphical user interface for enabling conversation persistence across two or more instances of a digital assistant
US10176167B2 (en) 2013-06-09 2019-01-08 Apple Inc. System and method for inferring user intent from speech inputs
EP3008964B1 (en) 2013-06-13 2019-09-25 Apple Inc. System and method for emergency calls initiated by voice command
US20150007031A1 (en) * 2013-06-26 2015-01-01 Lucid Global, Llc. Medical Environment Simulation and Presentation System
US9087521B2 (en) * 2013-07-02 2015-07-21 Family Systems, Ltd. Systems and methods for improving audio conferencing services
DE112014003653B4 (en) 2013-08-06 2024-04-18 Apple Inc. Automatically activate intelligent responses based on activities from remote devices
US9608944B2 (en) * 2013-08-09 2017-03-28 Beijing Lenovo Software Ltd. Information processing apparatus and information processing method
JP6092049B2 (en) * 2013-08-28 2017-03-08 東芝ライフスタイル株式会社 Imaging system and imaging apparatus
US9653117B2 (en) * 2013-08-29 2017-05-16 Avid Technology, Inc. Interconnected multimedia systems with synchronized playback of media streams
US10271010B2 (en) 2013-10-31 2019-04-23 Shindig, Inc. Systems and methods for controlling the display of content
US8938679B1 (en) * 2013-11-18 2015-01-20 Axure Software Solutions, Inc. Comment system for interactive graphical designs
US10296160B2 (en) 2013-12-06 2019-05-21 Apple Inc. Method for extracting salient dialog usage from live data
US9226073B2 (en) 2014-02-06 2015-12-29 Sonos, Inc. Audio output balancing during synchronized playback
US9226087B2 (en) 2014-02-06 2015-12-29 Sonos, Inc. Audio output balancing during synchronized playback
US10068302B2 (en) * 2014-02-10 2018-09-04 Cerner Innovation, Inc. Integrating video into patient workflows
WO2015131126A1 (en) 2014-02-27 2015-09-03 Cinsay, Inc. Apparatus and method for gathering analytics
US9819906B2 (en) * 2014-03-10 2017-11-14 Cisco Technology, Inc. Selective data content sharing
US10728603B2 (en) 2014-03-14 2020-07-28 Aibuy, Inc. Apparatus and method for automatic provisioning of merchandise
US9471957B2 (en) 2014-03-28 2016-10-18 Smart Technologies Ulc Method for partitioning, managing and displaying a collaboration space and interactive input system employing same
US9640219B1 (en) * 2014-03-31 2017-05-02 Google Inc. Systems and methods for modifying a segment of an uploaded media file
US8942727B1 (en) 2014-04-11 2015-01-27 ACR Development, Inc. User Location Tracking
US9413707B2 (en) 2014-04-11 2016-08-09 ACR Development, Inc. Automated user task management
US9620105B2 (en) 2014-05-15 2017-04-11 Apple Inc. Analyzing audio input for efficient speech and music recognition
US10277643B2 (en) * 2014-05-23 2019-04-30 Samsung Electronics Co., Ltd. Server and method of providing collaboration services and user terminal for receiving collaboration services
US10592095B2 (en) 2014-05-23 2020-03-17 Apple Inc. Instantaneous speaking of content on touch devices
US9502031B2 (en) 2014-05-27 2016-11-22 Apple Inc. Method for supporting dynamic grammars in WFST-based ASR
US10078631B2 (en) 2014-05-30 2018-09-18 Apple Inc. Entropy-guided text prediction using combined word and character n-gram language models
CN110797019B (en) 2014-05-30 2023-08-29 苹果公司 Multi-command single speech input method
US9842101B2 (en) 2014-05-30 2017-12-12 Apple Inc. Predictive conversion of language input
US9760559B2 (en) 2014-05-30 2017-09-12 Apple Inc. Predictive text input
US9430463B2 (en) 2014-05-30 2016-08-30 Apple Inc. Exemplar-based natural language processing
US9734193B2 (en) 2014-05-30 2017-08-15 Apple Inc. Determining domain salience ranking from ambiguous words in natural speech
US9715875B2 (en) 2014-05-30 2017-07-25 Apple Inc. Reducing the need for manual start/end-pointing and trigger phrases
US9633004B2 (en) 2014-05-30 2017-04-25 Apple Inc. Better resolution when referencing to concepts
US9785630B2 (en) 2014-05-30 2017-10-10 Apple Inc. Text prediction using combined word N-gram and unigram language models
US10170123B2 (en) 2014-05-30 2019-01-01 Apple Inc. Intelligent assistant for home automation
US10289433B2 (en) 2014-05-30 2019-05-14 Apple Inc. Domain specific language for encoding assistant dialog
US9053727B1 (en) 2014-06-02 2015-06-09 Western Digital Technologies, Inc. Disk drive opening spiral crossing window based on DC and AC spiral track error
US9338493B2 (en) 2014-06-30 2016-05-10 Apple Inc. Intelligent automated assistant for TV user interactions
US10659851B2 (en) 2014-06-30 2020-05-19 Apple Inc. Real-time digital assistant knowledge updates
US10209947B2 (en) 2014-07-23 2019-02-19 Sonos, Inc. Device grouping
US9671997B2 (en) 2014-07-23 2017-06-06 Sonos, Inc. Zone grouping
US10446141B2 (en) 2014-08-28 2019-10-15 Apple Inc. Automatic speech recognition based on user feedback
US9818400B2 (en) 2014-09-11 2017-11-14 Apple Inc. Method and apparatus for discovering trending terms in speech requests
US10789041B2 (en) 2014-09-12 2020-09-29 Apple Inc. Dynamic thresholds for always listening speech trigger
US9998518B2 (en) * 2014-09-18 2018-06-12 Multipop Llc Media platform for adding synchronized content to media with a duration
US9606986B2 (en) 2014-09-29 2017-03-28 Apple Inc. Integrated word N-gram and class M-gram language models
US10074360B2 (en) 2014-09-30 2018-09-11 Apple Inc. Providing an indication of the suitability of speech recognition
US10127911B2 (en) 2014-09-30 2018-11-13 Apple Inc. Speaker identification and unsupervised speaker adaptation techniques
US9668121B2 (en) 2014-09-30 2017-05-30 Apple Inc. Social reminders
US9646609B2 (en) 2014-09-30 2017-05-09 Apple Inc. Caching apparatus for serving phonetic pronunciations
US9886432B2 (en) 2014-09-30 2018-02-06 Apple Inc. Parsimonious handling of word inflection via categorical stem + suffix N-gram language models
US10552013B2 (en) 2014-12-02 2020-02-04 Apple Inc. Data detection
JP2016111472A (en) * 2014-12-04 2016-06-20 株式会社リコー Image forming apparatus, voice recording method, and voice recording program
US9711141B2 (en) 2014-12-09 2017-07-18 Apple Inc. Disambiguating heteronyms in speech synthesis
US20160182579A1 (en) * 2014-12-23 2016-06-23 Smart Technologies Ulc Method of establishing and managing messaging sessions based on user positions in a collaboration space and a collaboration system employing same
US10885570B2 (en) 2014-12-31 2021-01-05 Aibuy, Inc. System and method for managing a product exchange
US9929824B2 (en) 2015-01-26 2018-03-27 Timecode Systems Limited Networked programmable master clock base stations
US9380264B1 (en) 2015-02-16 2016-06-28 Siva Prasad Vakalapudi System and method for video communication
US10152299B2 (en) 2015-03-06 2018-12-11 Apple Inc. Reducing response latency of intelligent automated assistants
US9865280B2 (en) 2015-03-06 2018-01-09 Apple Inc. Structured dictation using intelligent automated assistants
US9721566B2 (en) 2015-03-08 2017-08-01 Apple Inc. Competing devices responding to voice triggers
US10567477B2 (en) 2015-03-08 2020-02-18 Apple Inc. Virtual assistant continuity
US9886953B2 (en) 2015-03-08 2018-02-06 Apple Inc. Virtual assistant activation
US9899019B2 (en) 2015-03-18 2018-02-20 Apple Inc. Systems and methods for structured stem and suffix language models
US9842105B2 (en) 2015-04-16 2017-12-12 Apple Inc. Parsimonious continuous-space phrase representations for natural language processing
US9921796B2 (en) 2015-04-24 2018-03-20 Kyocera Corporation Sharing of input information superimposed on images
US9699411B2 (en) * 2015-05-09 2017-07-04 Ricoh Company, Ltd. Integration of videoconferencing with interactive electronic whiteboard appliances
US10083688B2 (en) 2015-05-27 2018-09-25 Apple Inc. Device voice control for selecting a displayed affordance
US10222933B2 (en) * 2015-05-30 2019-03-05 Genesys Telecommunications Laboratories, Inc. System and method for managing multiple interactions
US10127220B2 (en) 2015-06-04 2018-11-13 Apple Inc. Language identification from short strings
US10101822B2 (en) 2015-06-05 2018-10-16 Apple Inc. Language input correction
US9578173B2 (en) 2015-06-05 2017-02-21 Apple Inc. Virtual assistant aided communication with 3rd party service in a communication session
US11025565B2 (en) 2015-06-07 2021-06-01 Apple Inc. Personalized prediction of responses for instant messaging
US10255907B2 (en) 2015-06-07 2019-04-09 Apple Inc. Automatic accent detection using acoustic models
US10186254B2 (en) 2015-06-07 2019-01-22 Apple Inc. Context-based endpoint detection
US10248376B2 (en) 2015-06-11 2019-04-02 Sonos, Inc. Multiple groupings in a playback system
US10747498B2 (en) 2015-09-08 2020-08-18 Apple Inc. Zero latency digital assistant
US10671428B2 (en) 2015-09-08 2020-06-02 Apple Inc. Distributed personal assistant
US9697820B2 (en) 2015-09-24 2017-07-04 Apple Inc. Unit-selection text-to-speech synthesis using concatenation-sensitive neural networks
US10366158B2 (en) 2015-09-29 2019-07-30 Apple Inc. Efficient word encoding for recurrent neural network language models
US11010550B2 (en) 2015-09-29 2021-05-18 Apple Inc. Unified language modeling framework for word prediction, auto-completion and auto-correction
US11587559B2 (en) 2015-09-30 2023-02-21 Apple Inc. Intelligent device identification
WO2017059329A1 (en) * 2015-10-02 2017-04-06 Polycom, Inc. Digital storyboards using multiple displays for content presentation and collaboration
US10742733B2 (en) 2015-10-12 2020-08-11 Timecode Systems Limited Synchronizing data between media devices
US9866732B2 (en) 2015-10-12 2018-01-09 Timecode Systems Limited Synchronizing data between personal and timecode devices
US10063637B2 (en) 2015-10-12 2018-08-28 Timecode Systems Limited Synchronizing data between personal and timecode devices
EP3157241B1 (en) 2015-10-12 2020-02-12 Timecode Systems Limited Synchronizing data between personal and timecode devices
US10691473B2 (en) 2015-11-06 2020-06-23 Apple Inc. Intelligent automated assistant in a messaging environment
US10049668B2 (en) 2015-12-02 2018-08-14 Apple Inc. Applying neural network language models to weighted finite state transducers for automatic speech recognition
US10116801B1 (en) 2015-12-23 2018-10-30 Shoutpoint, Inc. Conference call platform capable of generating engagement scores
US10223066B2 (en) 2015-12-23 2019-03-05 Apple Inc. Proactive assistance based on dialog communication between devices
US10088981B2 (en) * 2015-12-29 2018-10-02 Sap Se User engagement application across user interface applications
US20170199719A1 (en) * 2016-01-08 2017-07-13 KIDdesigns Inc. Systems and methods for recording and playing audio
US10574712B2 (en) 2016-02-19 2020-02-25 International Business Machines Corporation Provisioning conference rooms
US10446143B2 (en) 2016-03-14 2019-10-15 Apple Inc. Identification of voice inputs providing credentials
US20170272800A1 (en) * 2016-03-21 2017-09-21 Le Holdings (Beijing) Co., Ltd. Method for bullet screen pushing and electronic device
US20170316787A1 (en) * 2016-04-28 2017-11-02 Quentin Wayne Gilkey Audio watermarking system and processes
US9934775B2 (en) 2016-05-26 2018-04-03 Apple Inc. Unit-selection text-to-speech synthesis based on predicted concatenation parameters
US9972304B2 (en) 2016-06-03 2018-05-15 Apple Inc. Privacy preserving distributed evaluation framework for embedded personalized systems
US10249300B2 (en) 2016-06-06 2019-04-02 Apple Inc. Intelligent list reading
US10049663B2 (en) 2016-06-08 2018-08-14 Apple, Inc. Intelligent automated assistant for media exploration
DK179588B1 (en) 2016-06-09 2019-02-22 Apple Inc. Intelligent automated assistant in a home environment
US10067938B2 (en) 2016-06-10 2018-09-04 Apple Inc. Multilingual word prediction
US10490187B2 (en) 2016-06-10 2019-11-26 Apple Inc. Digital assistant providing automated status report
US10586535B2 (en) 2016-06-10 2020-03-10 Apple Inc. Intelligent digital assistant in a multi-tasking environment
US10192552B2 (en) 2016-06-10 2019-01-29 Apple Inc. Digital assistant providing whispered speech
US10509862B2 (en) 2016-06-10 2019-12-17 Apple Inc. Dynamic phrase expansion of language input
DK179343B1 (en) 2016-06-11 2018-05-14 Apple Inc Intelligent task discovery
DK201670540A1 (en) 2016-06-11 2018-01-08 Apple Inc Application integration with a digital assistant
DK179049B1 (en) 2016-06-11 2017-09-18 Apple Inc Data driven natural language event detection and classification
DK179415B1 (en) 2016-06-11 2018-06-14 Apple Inc Intelligent device arbitration and control
US10133916B2 (en) 2016-09-07 2018-11-20 Steven M. Gottlieb Image and identity validation in video chat events
US10474753B2 (en) 2016-09-07 2019-11-12 Apple Inc. Language identification using recurrent neural networks
US10043516B2 (en) 2016-09-23 2018-08-07 Apple Inc. Intelligent automated assistant
US9936162B1 (en) 2016-10-04 2018-04-03 Avaya Inc. System and method for processing digital images during videoconference
US9774823B1 (en) * 2016-10-04 2017-09-26 Avaya Inc. System and method for processing digital images during videoconference
US10712997B2 (en) 2016-10-17 2020-07-14 Sonos, Inc. Room association based on name
US10699217B2 (en) * 2016-10-20 2020-06-30 Diwo, Llc Method and system for reflective learning
USD820303S1 (en) 2016-10-27 2018-06-12 Apple Inc. Display screen or portion thereof with graphical user interface
USD815137S1 (en) 2016-10-27 2018-04-10 Apple Inc. Display screen or portion thereof with graphical user interface
US11281993B2 (en) 2016-12-05 2022-03-22 Apple Inc. Model and ensemble compression for metric learning
US10593346B2 (en) 2016-12-22 2020-03-17 Apple Inc. Rank-reduced token representation for automatic speech recognition
US10579443B2 (en) 2016-12-27 2020-03-03 Dropbox, Inc. Kernel event triggers for content item security
US11204787B2 (en) 2017-01-09 2021-12-21 Apple Inc. Application integration with a digital assistant
KR102136413B1 (en) * 2017-04-06 2020-07-21 주식회사 스무디 Method, system and non-transitory computer-readable recording medium for providing a multilateral communication service
US11862302B2 (en) 2017-04-24 2024-01-02 Teladoc Health, Inc. Automated transcription and documentation of tele-health encounters
US10417266B2 (en) 2017-05-09 2019-09-17 Apple Inc. Context-aware ranking of intelligent response suggestions
DK201770383A1 (en) 2017-05-09 2018-12-14 Apple Inc. User interface for correcting recognition errors
DK201770439A1 (en) 2017-05-11 2018-12-13 Apple Inc. Offline personal assistant
US10395654B2 (en) 2017-05-11 2019-08-27 Apple Inc. Text normalization based on a data-driven learning network
US10726832B2 (en) 2017-05-11 2020-07-28 Apple Inc. Maintaining privacy of personal information
DK179496B1 (en) 2017-05-12 2019-01-15 Apple Inc. USER-SPECIFIC Acoustic Models
DK179745B1 (en) 2017-05-12 2019-05-01 Apple Inc. SYNCHRONIZATION AND TASK DELEGATION OF A DIGITAL ASSISTANT
DK201770428A1 (en) 2017-05-12 2019-02-18 Apple Inc. Low-latency intelligent automated assistant
US11301477B2 (en) 2017-05-12 2022-04-12 Apple Inc. Feedback analysis of a digital assistant
DK201770432A1 (en) 2017-05-15 2018-12-21 Apple Inc. Hierarchical belief states for digital assistants
DK201770431A1 (en) 2017-05-15 2018-12-20 Apple Inc. Optimizing dialogue policy decisions for digital assistants using implicit feedback
US20180336275A1 (en) 2017-05-16 2018-11-22 Apple Inc. Intelligent automated assistant for media exploration
US10403278B2 (en) 2017-05-16 2019-09-03 Apple Inc. Methods and systems for phonetic matching in digital assistant services
US10311144B2 (en) 2017-05-16 2019-06-04 Apple Inc. Emoji word sense disambiguation
DK179549B1 (en) 2017-05-16 2019-02-12 Apple Inc. Far-field extension for digital assistant services
US10657328B2 (en) 2017-06-02 2020-05-19 Apple Inc. Multi-task recurrent neural network architecture for efficient morphology handling in neural language modeling
US10483007B2 (en) 2017-07-25 2019-11-19 Intouch Technologies, Inc. Modular telehealth cart with thermal imaging and touch screen user interface
US11636944B2 (en) 2017-08-25 2023-04-25 Teladoc Health, Inc. Connectivity infrastructure for a telehealth platform
US10129573B1 (en) * 2017-09-20 2018-11-13 Microsoft Technology Licensing, Llc Identifying relevance of a video
US10445429B2 (en) 2017-09-21 2019-10-15 Apple Inc. Natural language understanding using vocabularies with compressed serialized tries
US10755051B2 (en) 2017-09-29 2020-08-25 Apple Inc. Rule-based natural language processing
US10140467B1 (en) 2017-10-16 2018-11-27 Dropbox, Inc. Workflow functions of content management system enforced by client device
US10331623B2 (en) * 2017-10-16 2019-06-25 Dropbox, Inc. Workflow functions of content management system enforced by client device
US10636424B2 (en) 2017-11-30 2020-04-28 Apple Inc. Multi-turn canned dialog
US10733982B2 (en) 2018-01-08 2020-08-04 Apple Inc. Multi-directional dialog
US11019305B2 (en) 2018-01-12 2021-05-25 Facebook, Inc. Coordinated effects in videos
US11102264B2 (en) 2018-01-12 2021-08-24 Facebook, Inc. Methods and systems for initiating a coordinated effect
US11372698B2 (en) 2018-01-12 2022-06-28 Meta Platforms, Inc. Coordinated effects in experiences
US10733375B2 (en) 2018-01-31 2020-08-04 Apple Inc. Knowledge-based framework for improving natural language understanding
US11488602B2 (en) 2018-02-20 2022-11-01 Dropbox, Inc. Meeting transcription using custom lexicons based on document history
US10467335B2 (en) 2018-02-20 2019-11-05 Dropbox, Inc. Automated outline generation of captured meeting audio in a collaborative document context
TW201936768A (en) * 2018-03-01 2019-09-16 日商富士軟片股份有限公司 Multilayer body, water-soluble resin composition, and kit
US10789959B2 (en) 2018-03-02 2020-09-29 Apple Inc. Training speaker recognition models for digital assistants
US10592604B2 (en) 2018-03-12 2020-03-17 Apple Inc. Inverse text normalization for automatic speech recognition
US10818288B2 (en) 2018-03-26 2020-10-27 Apple Inc. Natural assistant interaction
US10909331B2 (en) 2018-03-30 2021-02-02 Apple Inc. Implicit identification of translation payload with neural machine translation
US10617299B2 (en) 2018-04-27 2020-04-14 Intouch Technologies, Inc. Telehealth cart that supports a removable tablet with seamless audio/video switching
US11145294B2 (en) 2018-05-07 2021-10-12 Apple Inc. Intelligent automated assistant for delivering content from user experiences
US10928918B2 (en) 2018-05-07 2021-02-23 Apple Inc. Raise to speak
US10984780B2 (en) 2018-05-21 2021-04-20 Apple Inc. Global semantic word embeddings using bi-directional recurrent neural networks
US10951859B2 (en) 2018-05-30 2021-03-16 Microsoft Technology Licensing, Llc Videoconferencing device and method
DK201870355A1 (en) 2018-06-01 2019-12-16 Apple Inc. Virtual assistant operation in multi-device environments
US10892996B2 (en) 2018-06-01 2021-01-12 Apple Inc. Variable latency device coordination
DK179822B1 (en) 2018-06-01 2019-07-12 Apple Inc. Voice interaction at a primary device to access call functionality of a companion device
US11386266B2 (en) 2018-06-01 2022-07-12 Apple Inc. Text correction
DK180639B1 (en) 2018-06-01 2021-11-04 Apple Inc DISABILITY OF ATTENTION-ATTENTIVE VIRTUAL ASSISTANT
GB2576692A (en) * 2018-06-01 2020-03-04 Stageclip Ltd Media processing
US11076039B2 (en) 2018-06-03 2021-07-27 Apple Inc. Accelerated task performance
US11615208B2 (en) * 2018-07-06 2023-03-28 Capital One Services, Llc Systems and methods for synthetic data generation
EP3912338B1 (en) 2019-01-14 2024-04-10 Dolby Laboratories Licensing Corporation Sharing physical writing surfaces in videoconferencing
US11437072B2 (en) 2019-02-07 2022-09-06 Moxtra, Inc. Recording presentations using layered keyframes
CN109994187B (en) * 2019-02-14 2024-07-02 平安科技(深圳)有限公司 Medical image information cloud storage system based on patient user identity
US11206308B2 (en) * 2019-04-26 2021-12-21 At&T Intellectual Property I, L.P. Facilitating support functionalities through a support appliance device in advanced networks
USD913315S1 (en) * 2019-05-31 2021-03-16 Apple Inc. Electronic device with graphical user interface
US11689379B2 (en) 2019-06-24 2023-06-27 Dropbox, Inc. Generating customized meeting insights based on user interactions and meeting media
US11418551B2 (en) * 2019-10-18 2022-08-16 Samsung Sds Co., Ltd. Method and system for providing virtual whiteboard based collaboration service
US11350185B2 (en) 2019-12-13 2022-05-31 Bank Of America Corporation Text-to-audio for interactive videos using a markup language
US10805665B1 (en) 2019-12-13 2020-10-13 Bank Of America Corporation Synchronizing text-to-audio with interactive videos in the video framework
CN113014854B (en) * 2020-04-30 2022-11-11 北京字节跳动网络技术有限公司 Method, device, equipment and medium for generating interactive record
CA3123547A1 (en) * 2020-08-31 2022-02-28 Evertz Microsystems Ltd. Systems and methods for down-sampled video routing
EP3985669A1 (en) * 2020-10-16 2022-04-20 Moodagent A/S Methods and systems for automatically matching audio content with visual input
US11379424B2 (en) * 2020-10-30 2022-07-05 Docusign, Inc. Edit interface in an online document system
US11837363B2 (en) 2020-11-04 2023-12-05 Hill-Rom Services, Inc. Remote management of patient environment
CN113010698B (en) * 2020-11-18 2023-03-10 北京字跳网络技术有限公司 Multimedia interaction method, information interaction method, device, equipment and medium
US11863594B2 (en) * 2021-01-07 2024-01-02 Samsung Electronics Co., Ltd. Electronic device and method for processing call request in electronic device
US20220301449A1 (en) * 2021-03-16 2022-09-22 Radix Technologies Ltd. System and method for remote classroom management
US11722535B2 (en) * 2021-03-30 2023-08-08 Snap Inc. Communicating with a user external to a virtual conference
JP2022184560A (en) * 2021-06-01 2022-12-13 富士フイルムビジネスイノベーション株式会社 Information processing device and information processing program
US11704626B2 (en) * 2021-07-09 2023-07-18 Prezi, Inc. Relocation of content item to motion picture sequences at multiple devices
WO2023235329A1 (en) * 2022-06-03 2023-12-07 Apple Inc. Framework for simultaneous subject and desk capture during videoconferencing

Citations (99)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US563507A (en) * 1896-07-07 nachmias
US3723653A (en) * 1967-10-24 1973-03-27 Matsushita Electric Ind Co Ltd Television telephone system
US3873771A (en) * 1972-04-11 1975-03-25 Telescan Communications System Simultaneous transmission of a video and an audio signal through an ordinary telephone transmission line
US4005265A (en) * 1970-11-13 1977-01-25 U.S. Philips Corporation Videophone system synchronizer
US4210927A (en) * 1977-05-11 1980-07-01 Hitachi, Ltd. Method for transmitting a color video signal on a narrow-band transmission line
US4430526A (en) * 1982-01-25 1984-02-07 Bell Telephone Laboratories, Incorporated Interactive graphics transmission system employing an adaptive stylus for reduced bandwidth
US4441180A (en) * 1979-06-01 1984-04-03 Licentia Patent-Verwaltungs-Gmbh Service integrated communication transmission and interchange system
US4451705A (en) * 1982-05-28 1984-05-29 Bell Telephone Laboratories, Incorporated Call completion circuit and method
US4500960A (en) * 1982-06-28 1985-02-19 At&T Bell Laboratories Geographically distributed multiprocessor time-shared communication processing system
US4516156A (en) * 1982-03-15 1985-05-07 Satellite Business Systems Teleconferencing method and system
US4529839A (en) * 1983-10-25 1985-07-16 At&T Bell Laboratories Multilocation video conference terminal including an arrangement to reduce disruption in video switching
US4574374A (en) * 1983-10-25 1986-03-04 At&T Bell Laboratories Multilocation video conference terminal including rapid video switching
US4577065A (en) * 1983-11-03 1986-03-18 At&T Bell Laboratories Meet-me conference arrangement
US4645872A (en) * 1982-04-01 1987-02-24 John Hopkins University Videophone network system
US4650929A (en) * 1984-02-29 1987-03-17 Heinrich-Hertz-Institut Fur Nachrichtentechnik Berlin Gmbh Communication system for videoconferencing
US4653090A (en) * 1985-12-16 1987-03-24 American Telephone & Telegraph (At&T) Graphics based call management
US4672442A (en) * 1984-07-17 1987-06-09 Kokusai Denshin Denwa Co., Ltd. Frame rate conversion system in television signal
US4720850A (en) * 1986-03-14 1988-01-19 American Telephone And Telegraph Company At&T Bell Laboratories Communication system control arrangement
US4739509A (en) * 1986-03-21 1988-04-19 American Telephone And Telegraph Company Automated message chronicling system
US4741025A (en) * 1985-03-01 1988-04-26 Nec Corporation Speech additive distribution equipment for conferencing system
US4796293A (en) * 1987-12-18 1989-01-03 Communications Network Enhancement Inc. Enhanced dedicated teleconferencing system
US4800344A (en) * 1985-03-21 1989-01-24 And Yet, Inc. Balun
US4817018A (en) * 1987-01-29 1989-03-28 International Business Machines Corporation Electronic calendaring method which provides for automatic assignment of alternates in requested events
US4827798A (en) * 1987-06-04 1989-05-09 Anchor Tech., Inc. Apparatus and method for exerting a braking torque upon a vehicle
US4893326A (en) * 1987-05-04 1990-01-09 Video Telecom Corp. Video-telephone communications system
US4914586A (en) * 1987-11-06 1990-04-03 Xerox Corporation Garbage collector for hypermedia systems
US4922523A (en) * 1984-12-29 1990-05-01 Hashimoto Corporation Videotex with telephone call screening system
US4931872A (en) * 1986-11-24 1990-06-05 Gte Laboratories Incorporated Methods of and apparatus for the generation of split-screen video displays
US4987492A (en) * 1987-09-28 1991-01-22 Stults Robert A User interface control for communication system
US4995078A (en) * 1988-06-09 1991-02-19 Monslow H Vincent Television broadcast system for selective transmission of viewer-chosen programs at viewer-requested times
US4995071A (en) * 1988-07-08 1991-02-19 Telenorma Telefonbau Und Normalzeit Gmbh Video conference installation
US4998243A (en) * 1989-10-10 1991-03-05 Racal Data Communications Inc. ISDN terminal adapter with teleconference provision
US5003532A (en) * 1989-06-02 1991-03-26 Fujitsu Limited Multi-point conference system
US5010399A (en) * 1989-07-14 1991-04-23 Inline Connection Corporation Video transmission and control system utilizing internal telephone lines
US5012509A (en) * 1988-12-23 1991-04-30 Hitachi, Ltd. Communication conference apparatus and method for remote conferencing
US5014267A (en) * 1989-04-06 1991-05-07 Datapoint Corporation Video conferencing network
US5016976A (en) * 1990-03-30 1991-05-21 The United States Of America As Represented By The Secretary Of The Air Force Modified compact 2f optical correlator
US5027400A (en) * 1988-08-19 1991-06-25 Hitachi Ltd. Multimedia bidirectional broadcast system
US5099510A (en) * 1990-06-11 1992-03-24 Communications Network Enhancement Inc. Teleconferencing with bridge partitioning and other features
US5109515A (en) * 1987-09-28 1992-04-28 At&T Bell Laboratories User and application program transparent resource sharing multiple computer interface architecture with kernel process level transfer of user requested services
US5109517A (en) * 1990-10-09 1992-04-28 Ast Research, Inc. System for selectively controlling slots in an IBM-AT/NEC 9801 dual-compatible computer
US5121426A (en) * 1989-12-22 1992-06-09 At&T Bell Laboratories Loudspeaking telephone station including directional microphone
US5127041A (en) * 1990-06-01 1992-06-30 Spectrum Information Technologies, Inc. System and method for interfacing computers to diverse telephone networks
US5190586A (en) * 1989-11-28 1993-03-02 Nippon Paint Co., Ltd. Pigment composite particles prepared by suspension polymorization
US5191611A (en) * 1989-04-03 1993-03-02 Lang Gerald S Method and apparatus for protecting material on storage media and for transferring material on storage media to various recipients
US5195087A (en) * 1990-08-31 1993-03-16 At&T Bell Laboratories Telephone system with monitor on hold feature
US5195086A (en) * 1990-04-12 1993-03-16 At&T Bell Laboratories Multiple call control method in a multimedia conferencing system
US5200989A (en) * 1988-06-16 1993-04-06 Italtel Societa Italiana Wide band communication system transmitting video and audio signals among a plurality of users
US5202828A (en) * 1991-05-15 1993-04-13 Apple Computer, Inc. User interface system having programmable user interface elements
US5202957A (en) * 1990-08-09 1993-04-13 Future Communications Full motion video telephone system
US5216633A (en) * 1990-11-30 1993-06-01 Samsung Electronics Co., Ltd. Nonvolatile semiconductor memory device including access code circuitry
US5218627A (en) * 1990-12-19 1993-06-08 U S West Advanced Technologies Decentralized video telecommunication system
US5224094A (en) * 1990-04-16 1993-06-29 Motorola, Inc. Communication system network that includes full duplex conference calling
US5283637A (en) * 1990-08-20 1994-02-01 Christine Holland Trustee/Goolcharan Trust Telecommunication system for transmitting full motion video
US5293619A (en) * 1991-05-30 1994-03-08 Sandia Corporation Method and apparatus for collaborative use of application program
US5303343A (en) * 1989-09-27 1994-04-12 Kabushiki Kaisha Toshiba Multi-medium store-and-forward exchange apparatus and method of controlling the apparatus
US5315633A (en) * 1991-12-20 1994-05-24 Unisys Corporation Digital video switch for video teleconferencing
US5319795A (en) * 1991-02-27 1994-06-07 Nec Corporation Cellular mobile communications system using frequencies commonly shared by neighboring cells for handoff operations
US5323445A (en) * 1991-03-07 1994-06-21 Mitsubishi Denki Kabushiki Kaisha Multi-location television conference system
US5323472A (en) * 1990-03-27 1994-06-21 The Boeing Company Optical image analyzer using optical correlation and opto-electronic feedback
US5325423A (en) * 1992-11-13 1994-06-28 Multimedia Systems Corporation Interactive multimedia communication system
US5379374A (en) * 1990-11-21 1995-01-03 Hitachi, Ltd. Collaborative information processing system and workstation
US5382972A (en) * 1988-09-22 1995-01-17 Kannes; Deno Video conferencing system for courtroom and other applications
US5384772A (en) * 1993-09-01 1995-01-24 Intel Corporation Method and apparatus for audio flow control during teleconferencing
US5384598A (en) * 1992-10-20 1995-01-24 International Business Machines Corporation System and method for frame differencing video compression and decompression with frame rate scalability
US5386581A (en) * 1989-03-28 1995-01-31 Matsushita Electric Industrial Co., Ltd. Multimedia data editing apparatus including visual graphic display of time information
US5390316A (en) * 1990-08-31 1995-02-14 International Business Machines Corporation Multicomputer complex having a distributed shared memory system for providing a single system view from multiple consoles
US5392223A (en) * 1992-07-29 1995-02-21 International Business Machines Corp. Audio/video communications processor
US5392400A (en) * 1992-07-02 1995-02-21 International Business Machines Corporation Collaborative computing system using pseudo server process to allow input from different server processes individually and sequence number map for maintaining received data sequence
US5392346A (en) * 1992-02-25 1995-02-21 At&T Corp. Mobile log-in capability featuring fixed physical (terminal-dependent) translations and portable logical (user-dependent) translations
US5392277A (en) * 1993-06-11 1995-02-21 At&T Corp. Routing to intelligence
US5404435A (en) * 1991-07-29 1995-04-04 International Business Machines Corporation Non-text object storage and retrieval
US5404316A (en) * 1992-08-03 1995-04-04 Spectra Group Ltd., Inc. Desktop digital video processing system
US5408662A (en) * 1992-07-17 1995-04-18 Fuji Xerox Co., Ltd. System for performing a cooperative operation on common data
US5408526A (en) * 1992-10-29 1995-04-18 At&T Corp. Conference calling system
US5410595A (en) * 1992-11-12 1995-04-25 Motorola, Inc. Apparatus and method for noise reduction for a full-duplex speakerphone or the like
US5416618A (en) * 1992-11-10 1995-05-16 The United States Of America As Represented By The Administrator Of The National Aeronautics & Space Administration Full complex modulation using two one-parameter spatial light modulators
US5422883A (en) * 1992-10-16 1995-06-06 International Business Machines Corporation Call setup and channel allocation for a multi-media network bus
US5485504A (en) * 1991-08-07 1996-01-16 Alcatel N.V. Hand-held radiotelephone with video transmission and display
US5491797A (en) * 1992-11-30 1996-02-13 Qwest Communications Schedulable automatically configured video conferencing system
US5491695A (en) * 1991-07-17 1996-02-13 Digital Access Corporation Means and method of dial up bridging of network for high bandwidth digital communication
US5506954A (en) * 1993-11-24 1996-04-09 Intel Corporation PC-based conferencing system
US5515491A (en) * 1992-12-31 1996-05-07 International Business Machines Corporation Method and system for managing communications within a collaborative data processing system
US5517652A (en) * 1990-05-30 1996-05-14 Hitachi, Ltd. Multi-media server for treating multi-media information and communication system empolying the multi-media server
US5529024A (en) * 1993-10-29 1996-06-25 Daimler-Benz A.G. Fuel injection system for an internal-combustion engine
US5594495A (en) * 1992-06-03 1997-01-14 Digital Equipment Corporation Video data scaling for video teleconferencing workstations communicating by digital data network
US5602580A (en) * 1993-09-17 1997-02-11 Tseng; Ling-Yuan Video communication controller using FM sideband transmission
US5608653A (en) * 1992-06-03 1997-03-04 Digital Equipment Corporation Video teleconferencing for networked workstations
US5610972A (en) * 1992-03-05 1997-03-11 Bell Atlantic Network Services, Inc. Personal communications service using wireline/wireless integration
US5623690A (en) * 1992-06-03 1997-04-22 Digital Equipment Corporation Audio/video storage and retrieval for multimedia workstations by interleaving audio and video data in data file
US5719786A (en) * 1993-02-03 1998-02-17 Novell, Inc. Digital media data stream network management system
US5751336A (en) * 1995-10-12 1998-05-12 International Business Machines Corporation Permutation based pyramid block transmission scheme for broadcasting in video-on-demand storage systems
US5864844A (en) * 1993-02-18 1999-01-26 Apple Computer, Inc. System and method for enhancing a user interface with a computer based training tool
US5872923A (en) * 1993-03-19 1999-02-16 Ncr Corporation Collaborative video conferencing system
US5881168A (en) * 1992-03-30 1999-03-09 Canon Kabushiki Kaisha Image processing apparatus method and system for converting character codes of character images extracted from an original image and positions of the images for transmission
US6181867B1 (en) * 1995-06-07 2001-01-30 Intervu, Inc. Video storage and retrieval system
US6336143B1 (en) * 1993-09-27 2002-01-01 International Business Machines Corporation Method and apparatus for multimedia data interchange with pacing capability in a distributed data processing system
US6343314B1 (en) * 1993-10-01 2002-01-29 Collaboration Properties, Inc. Remote participant hold and disconnect during videoconferencing
US6577324B1 (en) * 1992-06-03 2003-06-10 Compaq Information Technologies Group, L.P. Video and audio multimedia pop-up documentation by performing selected functions on selected topics

Family Cites Families (125)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US544476A (en) * 1895-08-13 Half to dorr b
US515491A (en) * 1894-02-27 Atomizer
US3974337A (en) 1967-10-24 1976-08-10 Matsushita Electric Industrial Co., Ltd. FM television telephone system
US3873711A (en) 1971-07-06 1975-03-25 Betz Laboratories Synergistic compositions for the control of aerobacter aerogenes
US3721777A (en) 1971-11-26 1973-03-20 Bell Telephone Labor Inc Echo path delay simulator for use with adaptive echo cancellers
GB1458663A (en) 1972-12-15 1976-12-15 Ard Anstalt Microphone circuits
FR2312904A1 (en) 1975-05-27 1976-12-24 Poirier Alain IMPROVEMENTS TO TELECONFERENCE SYSTEMS AND DEVICES VIA VIDEO PHONE
FR2484747A1 (en) 1980-06-11 1981-12-18 Faye Andre METHOD AND DEVICE FOR ESTABLISHING BIDIRECTIONAL AND / OR MULTIDIRECTIONAL COMMUNICATIONS BETWEEN DIFFERENT PEOPLE GROUPED IN DIFFERENT GEOGRAPHICALLY REMOTE STATIONS AND CONNECTED TO THEM THROUGH TELEPHONE LINES
US4400724A (en) * 1981-06-08 1983-08-23 The United States Of America As Represented By The Secretary Of The Army Virtual space teleconference system
US4786741A (en) * 1982-11-15 1988-11-22 Scientific Design Company, Inc. Preparation of alkylene carbonates
US4476349A (en) * 1982-03-30 1984-10-09 At&T Bell Laboratories Call message service
US4599611A (en) 1982-06-02 1986-07-08 Digital Equipment Corporation Interactive computer-based information display system
US4479195A (en) 1982-09-07 1984-10-23 At&T Bell Laboratories Data conference system
US4475193A (en) 1982-09-30 1984-10-02 Astech, Inc. Power line carrier multi telephone extension system for full duplex conferencing between telephones
US4625081A (en) * 1982-11-30 1986-11-25 Lotito Lawrence A Automated telephone voice service system
US4531024A (en) 1983-10-25 1985-07-23 At&T Bell Laboratories Multilocation video conference terminal including video switching contention control
US4529840A (en) 1983-10-26 1985-07-16 At&T Bell Laboratories Multilocation video conference terminal including controllable conference location reconfiguration
DE3507152A1 (en) 1984-02-29 1985-08-29 Heinrich-Hertz-Institut für Nachrichtentechnik Berlin GmbH, 1000 Berlin Communication system for videoconferences
FR2575882B1 (en) 1985-01-08 1988-06-10 Bull Sems METHOD FOR THE SAFE BROADCAST OF MESSAGES IN A RING NETWORK AND DEVICE FOR CARRYING OUT THE METHOD
US4691347A (en) 1985-02-15 1987-09-01 American Telephone And Telegraph Company, At&T Bell Laboratories Method and apparatus for controlling a conference
US4716585A (en) 1985-04-05 1987-12-29 Datapoint Corporation Gain switched audio conferencing network
US4686698A (en) 1985-04-08 1987-08-11 Datapoint Corporation Workstation for interfacing with a video conferencing network
US4847829A (en) * 1985-04-08 1989-07-11 Datapoint Corporation Video conferencing network
US4710917A (en) 1985-04-08 1987-12-01 Datapoint Corporation Video conferencing network
US4713806A (en) * 1986-03-14 1987-12-15 American Telephone And Telegraph Company, At&T Bell Laboratories Communication system control arrangement
US4850009A (en) 1986-05-12 1989-07-18 Clinicom Incorporated Portable handheld terminal including optical bar code reader and electromagnetic transceiver means for interactive wireless communication with a base communications station
US4837798A (en) 1986-06-02 1989-06-06 American Telephone And Telegraph Company Communication system having unified messaging
US4661609A (en) * 1986-07-31 1987-04-28 Texaco Inc. Process for cosynthesis of ethylene glycol and dimethyl carbonate
JPS647791A (en) 1987-06-30 1989-01-11 Nec Corp Multiscreen video conference method and device therefor
US4888795A (en) 1987-06-30 1989-12-19 Nec Corporation Videotelephone apparatus for transmitting high and low resolution video signals over telephone exchange lines
US4977520A (en) 1988-02-03 1990-12-11 Ibm Corp. Method to facilitate a reply to electronic meeting invitation in an interactive multi-terminal system employing electronic calendars
US4885747A (en) 1988-02-17 1989-12-05 International Business Machines Corp. Broadband and baseband LAN
US4924496A (en) 1988-05-12 1990-05-08 Romek Figa D/B/A Abraham & Sons Automatic incoming telephone call originating number and party display system
US4890320A (en) 1988-06-09 1989-12-26 Monslow H Vincent Television broadcast system for selective transmission of viewer-chosen programs at viewer-requested times
JPH022441A (en) * 1988-06-14 1990-01-08 Nec Corp Communication control network calling system
JPH01318136A (en) * 1988-06-20 1989-12-22 Fuji Xerox Co Ltd Data processor
US4849811A (en) * 1988-07-06 1989-07-18 Ben Kleinerman Simultaneous audio and video transmission with restricted bandwidth
FR2634608B1 (en) 1988-07-22 1994-05-06 Etat Francais Cnet RECONFIGURABLE MULTI-POINT CABLE DOMOTIC NETWORK FOR SIMULTANEOUS AND / OR ALTERNATIVE DISTRIBUTION OF MULTIPLE SIGNAL TYPES, ESPECIALLY BASIC BAND IMAGE SIGNALS, AND METHOD FOR CONFIGURING SUCH A NETWORK
US5208745A (en) * 1988-07-25 1993-05-04 Electric Power Research Institute Multimedia interface and method for computer system
EP0354370A1 (en) 1988-07-28 1990-02-14 Siemens Aktiengesellschaft Video conference system
US4965819A (en) 1988-09-22 1990-10-23 Docu-Vision, Inc. Video conferencing system for courtroom and other applications
US5157491A (en) 1988-10-17 1992-10-20 Kassatly L Samuel A Method and apparatus for video broadcasting and teleconferencing
US4901346A (en) 1988-12-28 1990-02-13 American Telephone And Telegraph Co. Computer controlled speakerphone for adapting to a communication line
US4953159A (en) 1989-01-03 1990-08-28 American Telephone And Telegraph Company Audiographics conferencing arrangement
JP2823018B2 (en) 1989-01-20 1998-11-11 東京電力株式会社 Video transmission equipment
FR2642928B1 (en) 1989-02-03 1994-04-15 Guichard Jacques SOUND AND VISUAL COMMUNICATION TERMINAL WITH IMPROVED SETTINGS
DE69032361T2 (en) 1989-03-16 1998-10-29 Fujitsu Ltd VIDEO / AUDIO MULTIPLEX TRANSMISSION SYSTEM
US5109519A (en) 1989-03-28 1992-04-28 Wang Laboratories, Inc. Local computer participating in mail delivery system abstracts from directory of all eligible mail recipients only served by local computer
CA2015215C (en) 1989-06-16 1994-01-11 Bradley James Beitel Computer-based, audio/visual creation and presentation system and method
EP0410378A3 (en) 1989-07-26 1992-08-05 Hitachi, Ltd. Multimedia telemeeting terminal device, communication system and manipulation method thereof
US5206934A (en) * 1989-08-15 1993-04-27 Group Technologies, Inc. Method and apparatus for interactive computer conferencing
JP2771858B2 (en) 1989-08-23 1998-07-02 富士通株式会社 Multi-screen synthesis device
US5799320A (en) * 1989-08-23 1998-08-25 John R. Klug Remote multiple-user editing system and method
US5042062A (en) 1989-10-23 1991-08-20 At&T Bell Laboratories Method and apparatus for providing real-time switching of high bandwidth transmission channels
US5276860A (en) * 1989-12-19 1994-01-04 Epoch Systems, Inc. Digital data processor with improved backup storage
US5061046A (en) * 1989-12-19 1991-10-29 The University Of Rochester Gradient index liquid crystal devices and method of fabrication thereof
US5305097A (en) 1989-12-20 1994-04-19 Canon Kabushiki Kaisha Communicating apparatus having a calling party number display device
US5155761A (en) 1990-01-26 1992-10-13 Intervoice, Inc. Automatic call back system and method of operation
US5253362A (en) 1990-01-29 1993-10-12 Emtek Health Care Systems, Inc. Method for storing, retrieving, and indicating a plurality of annotations in a data cell
US5072442A (en) 1990-02-28 1991-12-10 Harris Corporation Multiple clock rate teleconferencing network
US5056136A (en) 1990-03-09 1991-10-08 The United States Of America As Represented By The United States Department Of Energy Secure video communications system
JPH0713146B2 (en) 1990-05-22 1995-02-15 信越化学工業株式会社 Method for producing organopolysiloxane
US5282270A (en) * 1990-06-06 1994-01-25 Apple Computer, Inc. Network device location using multicast
CA2022302C (en) * 1990-07-30 1995-02-28 Douglas J. Ballantyne Method and apparatus for distribution of movies
JP2865827B2 (en) 1990-08-13 1999-03-08 株式会社日立製作所 Data storage method in conference system
US5239466A (en) 1990-10-04 1993-08-24 Motorola, Inc. System for selectively routing and merging independent annotations to a document at remote locations
US5239492A (en) * 1990-10-11 1993-08-24 Spacelabs Medical, Inc. Automatic internal calibration circuit and method
US5181200A (en) * 1990-10-29 1993-01-19 International Business Machines Corporation Handoff method and apparatus for mobile wireless workstation
US5260941A (en) 1990-10-31 1993-11-09 Rose Communications, Inc. Digital radio telephone system
US5204947A (en) * 1990-10-31 1993-04-20 International Business Machines Corporation Application independent (open) hypermedia enablement services
US5177290A (en) * 1991-01-10 1993-01-05 Exxon Chemical Patents Inc. Isoprene process
EP0497022B1 (en) 1991-01-31 1999-04-07 Hewlett-Packard Company Conference system
US5351276A (en) 1991-02-11 1994-09-27 Simpact Associates, Inc. Digital/audio interactive communication network
DE4105554A1 (en) * 1991-02-22 1992-08-27 Bayer Ag METHOD FOR PRODUCING DIALKYL CARBONATES
US5341374A (en) * 1991-03-01 1994-08-23 Trilan Systems Corporation Communication network integrating voice data and video with distributed call processing
US5218327A (en) * 1991-03-27 1993-06-08 Hughes Aircraft Company Variable/switchable coupler
GB2255972A (en) * 1991-04-12 1992-11-25 Davy Res & Dev Ltd Production of diaryl carbonates.
US5317732A (en) * 1991-04-26 1994-05-31 Commodore Electronics Limited System for relocating a multimedia presentation on a different platform by extracting a resource map in order to remap and relocate resources
US5146325A (en) * 1991-04-29 1992-09-08 Rca Thomson Licensing Corporation Video signal decompression apparatus for independently compressed even and odd field data
JP3292248B2 (en) 1991-05-29 2002-06-17 富士通株式会社 Teleconferencing system
DE69222580T2 (en) * 1991-07-15 1998-04-16 Hitachi Ltd Image encoder decoder and teleconferencing terminal
JP3308562B2 (en) 1991-07-15 2002-07-29 株式会社日立製作所 Video conference terminal
EP0523617B1 (en) * 1991-07-15 1997-10-01 Hitachi, Ltd. Teleconference terminal equipment
DE59208887D1 (en) 1991-09-30 1997-10-16 Siemens Ag Method for establishing conference connections in a computer-controlled communication system
GB9127404D0 (en) * 1991-12-24 1992-02-19 Ncr Co Local area network system
CA2078714A1 (en) 1991-12-20 1993-06-21 Robert A. Pascoe Automated audio/visual presentation
US5333299A (en) 1991-12-31 1994-07-26 International Business Machines Corporation Synchronization techniques for multimedia data streams
JP2792311B2 (en) 1992-01-31 1998-09-03 日本電気株式会社 Method and apparatus for removing multi-channel echo
US5526024A (en) 1992-03-12 1996-06-11 At&T Corp. Apparatus for synchronization and display of plurality of digital video data streams
US5243596A (en) 1992-03-18 1993-09-07 Fischer & Porter Company Network architecture suitable for multicasting and resource locking
JP3200932B2 (en) 1992-03-24 2001-08-20 株式会社日立製作所 Electronic dialogue system
US5249218A (en) * 1992-04-06 1993-09-28 Spectrum Information Technologies, Inc. Programmable universal interface system
US5467288A (en) * 1992-04-10 1995-11-14 Avid Technology, Inc. Digital audio workstations providing digital storage and display of video information
JP2892901B2 (en) 1992-04-27 1999-05-17 インターナショナル・ビジネス・マシーンズ・コーポレイション Automation system and method for presentation acquisition, management and playback
US5333133A (en) 1992-04-28 1994-07-26 Teloquent Communications Corporation Call processing control system
US5262875A (en) * 1992-04-30 1993-11-16 Instant Video Technologies, Inc. Audio/video file server including decompression/playback means
JP2680777B2 (en) 1992-05-08 1997-11-19 三洋電機株式会社 Videophone equipment
US5335321A (en) 1992-06-19 1994-08-02 Intel Corporation Scalable multimedia platform architecture
US5440624A (en) * 1992-11-10 1995-08-08 Netmedia, Inc. Method and apparatus for providing adaptive administration and control of an electronic conference
US5444476A (en) 1992-12-11 1995-08-22 The Regents Of The University Of Michigan System and method for teleinteraction
US5367629A (en) 1992-12-18 1994-11-22 Sharevision Technology, Inc. Digital video compression system utilizing vector adaptive transform
US5373549A (en) 1992-12-23 1994-12-13 At&T Corp. Multi-level conference management and notification
CA2102857C (en) * 1992-12-31 1998-06-23 Alexander Feiner Technique for reducing echoes in conference communications
US5430872A (en) * 1993-03-12 1995-07-04 Asymetrix Corporation Verifying multimedia linking for a multimedia presentation
US5565910A (en) 1993-03-26 1996-10-15 Vionx, Inc. Data and television network for digital computer workstations
JPH06324298A (en) * 1993-03-31 1994-11-25 Citizen Watch Co Ltd Optical device
US5689553A (en) 1993-04-22 1997-11-18 At&T Corp. Multimedia telecommunications network and service
US5471318A (en) 1993-04-22 1995-11-28 At&T Corp. Multimedia communications network
US5796727A (en) * 1993-04-30 1998-08-18 International Business Machines Corporation Wide-area wireless lan access
DE69405388T2 (en) 1993-05-10 1998-03-19 Taligent Inc MULTIMEDIA SYNCHRONIZATION SYSTEM
WO1994030015A1 (en) 1993-06-03 1994-12-22 Target Technologies, Inc. Data and television network for digital computer workstations
US5561736A (en) 1993-06-04 1996-10-01 International Business Machines Corporation Three dimensional speech synthesis
US6594688B2 (en) 1993-10-01 2003-07-15 Collaboration Properties, Inc. Dedicated echo canceler for a workstation
US5524110A (en) 1993-11-24 1996-06-04 Intel Corporation Conferencing over multiple transports
US5473679A (en) 1993-12-09 1995-12-05 At&T Corp. Signaling system for broadband communications networks
US5581702A (en) 1993-12-20 1996-12-03 Intel Corporation Computer conferencing system for selectively linking and unlinking private page with public page by selectively activating linked mode and non-linked mode for each participant
JP3125552B2 (en) * 1993-12-28 2001-01-22 三菱電機株式会社 Video conference system
US5751338A (en) * 1994-12-30 1998-05-12 Visionary Corporate Technologies Methods and systems for multimedia communications via public telephone networks
TW339353B (en) * 1995-12-22 1998-09-01 Sumitomo Bakelite Co Epoxy resin composition
US5742769A (en) * 1996-05-06 1998-04-21 Banyan Systems, Inc. Directory with options for access to and display of email addresses
US5907604A (en) * 1997-03-25 1999-05-25 Sony Corporation Image icon associated with caller ID
US6219415B1 (en) * 1998-07-08 2001-04-17 Lucent Technologies Inc. Call hold state notification across a communications network
WO2001077824A1 (en) * 2000-04-07 2001-10-18 Agora Interactive, Inc. Method for managing the simultaneous utilization of diverse real-time collaborative software applications
US6392078B1 (en) * 2000-06-12 2002-05-21 Catalytic Distillation Technologies Process and catalyst for making dialkyl carbonates
US6768722B1 (en) * 2000-06-23 2004-07-27 At&T Corp. Systems and methods for managing multiple communications

Patent Citations (99)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US563507A (en) * 1896-07-07 nachmias
US3723653A (en) * 1967-10-24 1973-03-27 Matsushita Electric Ind Co Ltd Television telephone system
US4005265A (en) * 1970-11-13 1977-01-25 U.S. Philips Corporation Videophone system synchronizer
US3873771A (en) * 1972-04-11 1975-03-25 Telescan Communications System Simultaneous transmission of a video and an audio signal through an ordinary telephone transmission line
US4210927A (en) * 1977-05-11 1980-07-01 Hitachi, Ltd. Method for transmitting a color video signal on a narrow-band transmission line
US4441180A (en) * 1979-06-01 1984-04-03 Licentia Patent-Verwaltungs-Gmbh Service integrated communication transmission and interchange system
US4430526A (en) * 1982-01-25 1984-02-07 Bell Telephone Laboratories, Incorporated Interactive graphics transmission system employing an adaptive stylus for reduced bandwidth
US4516156A (en) * 1982-03-15 1985-05-07 Satellite Business Systems Teleconferencing method and system
US4645872A (en) * 1982-04-01 1987-02-24 John Hopkins University Videophone network system
US4451705A (en) * 1982-05-28 1984-05-29 Bell Telephone Laboratories, Incorporated Call completion circuit and method
US4500960A (en) * 1982-06-28 1985-02-19 At&T Bell Laboratories Geographically distributed multiprocessor time-shared communication processing system
US4529839A (en) * 1983-10-25 1985-07-16 At&T Bell Laboratories Multilocation video conference terminal including an arrangement to reduce disruption in video switching
US4574374A (en) * 1983-10-25 1986-03-04 At&T Bell Laboratories Multilocation video conference terminal including rapid video switching
US4577065A (en) * 1983-11-03 1986-03-18 At&T Bell Laboratories Meet-me conference arrangement
US4650929A (en) * 1984-02-29 1987-03-17 Heinrich-Hertz-Institut Fur Nachrichtentechnik Berlin Gmbh Communication system for videoconferencing
US4672442A (en) * 1984-07-17 1987-06-09 Kokusai Denshin Denwa Co., Ltd. Frame rate conversion system in television signal
US4922523A (en) * 1984-12-29 1990-05-01 Hashimoto Corporation Videotex with telephone call screening system
US4741025A (en) * 1985-03-01 1988-04-26 Nec Corporation Speech additive distribution equipment for conferencing system
US4800344A (en) * 1985-03-21 1989-01-24 And Yet, Inc. Balun
US4653090A (en) * 1985-12-16 1987-03-24 American Telephone & Telegraph (At&T) Graphics based call management
US4720850A (en) * 1986-03-14 1988-01-19 American Telephone And Telegraph Company At&T Bell Laboratories Communication system control arrangement
US4739509A (en) * 1986-03-21 1988-04-19 American Telephone And Telegraph Company Automated message chronicling system
US4931872A (en) * 1986-11-24 1990-06-05 Gte Laboratories Incorporated Methods of and apparatus for the generation of split-screen video displays
US4817018A (en) * 1987-01-29 1989-03-28 International Business Machines Corporation Electronic calendaring method which provides for automatic assignment of alternates in requested events
US4893326A (en) * 1987-05-04 1990-01-09 Video Telecom Corp. Video-telephone communications system
US4827798A (en) * 1987-06-04 1989-05-09 Anchor Tech., Inc. Apparatus and method for exerting a braking torque upon a vehicle
US4987492A (en) * 1987-09-28 1991-01-22 Stults Robert A User interface control for communication system
US5109515A (en) * 1987-09-28 1992-04-28 At&T Bell Laboratories User and application program transparent resource sharing multiple computer interface architecture with kernel process level transfer of user requested services
US4914586A (en) * 1987-11-06 1990-04-03 Xerox Corporation Garbage collector for hypermedia systems
US4796293A (en) * 1987-12-18 1989-01-03 Communications Network Enhancement Inc. Enhanced dedicated teleconferencing system
US4995078A (en) * 1988-06-09 1991-02-19 Monslow H Vincent Television broadcast system for selective transmission of viewer-chosen programs at viewer-requested times
US5200989A (en) * 1988-06-16 1993-04-06 Italtel Societa Italiana Wide band communication system transmitting video and audio signals among a plurality of users
US4995071A (en) * 1988-07-08 1991-02-19 Telenorma Telefonbau Und Normalzeit Gmbh Video conference installation
US5027400A (en) * 1988-08-19 1991-06-25 Hitachi Ltd. Multimedia bidirectional broadcast system
US5382972A (en) * 1988-09-22 1995-01-17 Kannes; Deno Video conferencing system for courtroom and other applications
US5012509A (en) * 1988-12-23 1991-04-30 Hitachi, Ltd. Communication conference apparatus and method for remote conferencing
US5386581A (en) * 1989-03-28 1995-01-31 Matsushita Electric Industrial Co., Ltd. Multimedia data editing apparatus including visual graphic display of time information
US5191611A (en) * 1989-04-03 1993-03-02 Lang Gerald S Method and apparatus for protecting material on storage media and for transferring material on storage media to various recipients
US5014267A (en) * 1989-04-06 1991-05-07 Datapoint Corporation Video conferencing network
US5003532A (en) * 1989-06-02 1991-03-26 Fujitsu Limited Multi-point conference system
US5010399A (en) * 1989-07-14 1991-04-23 Inline Connection Corporation Video transmission and control system utilizing internal telephone lines
US5303343A (en) * 1989-09-27 1994-04-12 Kabushiki Kaisha Toshiba Multi-medium store-and-forward exchange apparatus and method of controlling the apparatus
US4998243A (en) * 1989-10-10 1991-03-05 Racal Data Communications Inc. ISDN terminal adapter with teleconference provision
US5190586A (en) * 1989-11-28 1993-03-02 Nippon Paint Co., Ltd. Pigment composite particles prepared by suspension polymorization
US5121426A (en) * 1989-12-22 1992-06-09 At&T Bell Laboratories Loudspeaking telephone station including directional microphone
US5323472A (en) * 1990-03-27 1994-06-21 The Boeing Company Optical image analyzer using optical correlation and opto-electronic feedback
US5016976A (en) * 1990-03-30 1991-05-21 The United States Of America As Represented By The Secretary Of The Air Force Modified compact 2f optical correlator
US5195086A (en) * 1990-04-12 1993-03-16 At&T Bell Laboratories Multiple call control method in a multimedia conferencing system
US5224094A (en) * 1990-04-16 1993-06-29 Motorola, Inc. Communication system network that includes full duplex conference calling
US5517652A (en) * 1990-05-30 1996-05-14 Hitachi, Ltd. Multi-media server for treating multi-media information and communication system empolying the multi-media server
US5127041A (en) * 1990-06-01 1992-06-30 Spectrum Information Technologies, Inc. System and method for interfacing computers to diverse telephone networks
US5099510A (en) * 1990-06-11 1992-03-24 Communications Network Enhancement Inc. Teleconferencing with bridge partitioning and other features
US5202957A (en) * 1990-08-09 1993-04-13 Future Communications Full motion video telephone system
US5283637A (en) * 1990-08-20 1994-02-01 Christine Holland Trustee/Goolcharan Trust Telecommunication system for transmitting full motion video
US5390316A (en) * 1990-08-31 1995-02-14 International Business Machines Corporation Multicomputer complex having a distributed shared memory system for providing a single system view from multiple consoles
US5195087A (en) * 1990-08-31 1993-03-16 At&T Bell Laboratories Telephone system with monitor on hold feature
US5109517A (en) * 1990-10-09 1992-04-28 Ast Research, Inc. System for selectively controlling slots in an IBM-AT/NEC 9801 dual-compatible computer
US5379374A (en) * 1990-11-21 1995-01-03 Hitachi, Ltd. Collaborative information processing system and workstation
US5216633A (en) * 1990-11-30 1993-06-01 Samsung Electronics Co., Ltd. Nonvolatile semiconductor memory device including access code circuitry
US5218627A (en) * 1990-12-19 1993-06-08 U S West Advanced Technologies Decentralized video telecommunication system
US5319795A (en) * 1991-02-27 1994-06-07 Nec Corporation Cellular mobile communications system using frequencies commonly shared by neighboring cells for handoff operations
US5323445A (en) * 1991-03-07 1994-06-21 Mitsubishi Denki Kabushiki Kaisha Multi-location television conference system
US5202828A (en) * 1991-05-15 1993-04-13 Apple Computer, Inc. User interface system having programmable user interface elements
US5293619A (en) * 1991-05-30 1994-03-08 Sandia Corporation Method and apparatus for collaborative use of application program
US5491695A (en) * 1991-07-17 1996-02-13 Digital Access Corporation Means and method of dial up bridging of network for high bandwidth digital communication
US5404435A (en) * 1991-07-29 1995-04-04 International Business Machines Corporation Non-text object storage and retrieval
US5485504A (en) * 1991-08-07 1996-01-16 Alcatel N.V. Hand-held radiotelephone with video transmission and display
US5315633A (en) * 1991-12-20 1994-05-24 Unisys Corporation Digital video switch for video teleconferencing
US5392346A (en) * 1992-02-25 1995-02-21 At&T Corp. Mobile log-in capability featuring fixed physical (terminal-dependent) translations and portable logical (user-dependent) translations
US5610972A (en) * 1992-03-05 1997-03-11 Bell Atlantic Network Services, Inc. Personal communications service using wireline/wireless integration
US5881168A (en) * 1992-03-30 1999-03-09 Canon Kabushiki Kaisha Image processing apparatus method and system for converting character codes of character images extracted from an original image and positions of the images for transmission
US5594495A (en) * 1992-06-03 1997-01-14 Digital Equipment Corporation Video data scaling for video teleconferencing workstations communicating by digital data network
US5623690A (en) * 1992-06-03 1997-04-22 Digital Equipment Corporation Audio/video storage and retrieval for multimedia workstations by interleaving audio and video data in data file
US5608653A (en) * 1992-06-03 1997-03-04 Digital Equipment Corporation Video teleconferencing for networked workstations
US6577324B1 (en) * 1992-06-03 2003-06-10 Compaq Information Technologies Group, L.P. Video and audio multimedia pop-up documentation by performing selected functions on selected topics
US5392400A (en) * 1992-07-02 1995-02-21 International Business Machines Corporation Collaborative computing system using pseudo server process to allow input from different server processes individually and sequence number map for maintaining received data sequence
US5408662A (en) * 1992-07-17 1995-04-18 Fuji Xerox Co., Ltd. System for performing a cooperative operation on common data
US5392223A (en) * 1992-07-29 1995-02-21 International Business Machines Corp. Audio/video communications processor
US5404316A (en) * 1992-08-03 1995-04-04 Spectra Group Ltd., Inc. Desktop digital video processing system
US5422883A (en) * 1992-10-16 1995-06-06 International Business Machines Corporation Call setup and channel allocation for a multi-media network bus
US5384598A (en) * 1992-10-20 1995-01-24 International Business Machines Corporation System and method for frame differencing video compression and decompression with frame rate scalability
US5408526A (en) * 1992-10-29 1995-04-18 At&T Corp. Conference calling system
US5416618A (en) * 1992-11-10 1995-05-16 The United States Of America As Represented By The Administrator Of The National Aeronautics & Space Administration Full complex modulation using two one-parameter spatial light modulators
US5410595A (en) * 1992-11-12 1995-04-25 Motorola, Inc. Apparatus and method for noise reduction for a full-duplex speakerphone or the like
US5325423A (en) * 1992-11-13 1994-06-28 Multimedia Systems Corporation Interactive multimedia communication system
US5491797A (en) * 1992-11-30 1996-02-13 Qwest Communications Schedulable automatically configured video conferencing system
US5515491A (en) * 1992-12-31 1996-05-07 International Business Machines Corporation Method and system for managing communications within a collaborative data processing system
US5719786A (en) * 1993-02-03 1998-02-17 Novell, Inc. Digital media data stream network management system
US5864844A (en) * 1993-02-18 1999-01-26 Apple Computer, Inc. System and method for enhancing a user interface with a computer based training tool
US5872923A (en) * 1993-03-19 1999-02-16 Ncr Corporation Collaborative video conferencing system
US5392277A (en) * 1993-06-11 1995-02-21 At&T Corp. Routing to intelligence
US5384772A (en) * 1993-09-01 1995-01-24 Intel Corporation Method and apparatus for audio flow control during teleconferencing
US5602580A (en) * 1993-09-17 1997-02-11 Tseng; Ling-Yuan Video communication controller using FM sideband transmission
US6336143B1 (en) * 1993-09-27 2002-01-01 International Business Machines Corporation Method and apparatus for multimedia data interchange with pacing capability in a distributed data processing system
US6343314B1 (en) * 1993-10-01 2002-01-29 Collaboration Properties, Inc. Remote participant hold and disconnect during videoconferencing
US5529024A (en) * 1993-10-29 1996-06-25 Daimler-Benz A.G. Fuel injection system for an internal-combustion engine
US5506954A (en) * 1993-11-24 1996-04-09 Intel Corporation PC-based conferencing system
US6181867B1 (en) * 1995-06-07 2001-01-30 Intervu, Inc. Video storage and retrieval system
US5751336A (en) * 1995-10-12 1998-05-12 International Business Machines Corporation Permutation based pyramid block transmission scheme for broadcasting in video-on-demand storage systems

Cited By (29)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20090070420A1 (en) * 2006-05-01 2009-03-12 Schuyler Quackenbush System and method for processing data signals
US20080181218A1 (en) * 2007-01-31 2008-07-31 Gorzynski Mark E Coordinated media control system
US7911955B2 (en) 2007-01-31 2011-03-22 Hewlett-Packard Development Company, L.P. Coordinated media control system
US20080228876A1 (en) * 2007-03-13 2008-09-18 Byron Johnson System and method for online collaboration
US20080229405A1 (en) * 2007-03-16 2008-09-18 Yoshihiro Fujii Communication System, Communication System Management Apparatus, Terminal Connection Control Method, and Program
US8510799B2 (en) * 2007-03-16 2013-08-13 Sony Corporation Communication system, communication system management apparatus, terminal connection control method, and program
US8627211B2 (en) 2007-03-30 2014-01-07 Uranus International Limited Method, apparatus, system, medium, and signals for supporting pointer display in a multiple-party communication
US8060887B2 (en) 2007-03-30 2011-11-15 Uranus International Limited Method, apparatus, system, and medium for supporting multiple-party communications
US20080244461A1 (en) * 2007-03-30 2008-10-02 Alexander Kropivny Method, Apparatus, System, Medium, and Signals For Supporting Pointer Display In A Multiple-Party Communication
US20080244615A1 (en) * 2007-03-30 2008-10-02 Uranus International Limited Method, Apparatus, System, Medium, and Signals for Supporting a Multiple-Party Communication on a Plurality of Computer Servers
US7765261B2 (en) 2007-03-30 2010-07-27 Uranus International Limited Method, apparatus, system, medium and signals for supporting a multiple-party communication on a plurality of computer servers
US7765266B2 (en) 2007-03-30 2010-07-27 Uranus International Limited Method, apparatus, system, medium, and signals for publishing content created during a communication
US20080244702A1 (en) * 2007-03-30 2008-10-02 Uranus International Limited Method, Apparatus, System, Medium, and Signals for Intercepting a Multiple-Party Communication
US10963124B2 (en) 2007-03-30 2021-03-30 Alexander Kropivny Sharing content produced by a plurality of client computers in communication with a server
US10180765B2 (en) 2007-03-30 2019-01-15 Uranus International Limited Multi-party collaboration over a computer network
US9579572B2 (en) 2007-03-30 2017-02-28 Uranus International Limited Method, apparatus, and system for supporting multi-party collaboration between a plurality of client computers in communication with a server
US7950046B2 (en) 2007-03-30 2011-05-24 Uranus International Limited Method, apparatus, system, medium, and signals for intercepting a multiple-party communication
US20080244013A1 (en) * 2007-03-30 2008-10-02 Alexander Kropivny Method, Apparatus, System, Medium, and Signals for Publishing Content Created During a Communication
US8702505B2 (en) 2007-03-30 2014-04-22 Uranus International Limited Method, apparatus, system, medium, and signals for supporting game piece movement in a multiple-party communication
US20080243994A1 (en) * 2007-03-30 2008-10-02 Alexander Kropivny Method, Apparatus, System, and Medium for Supporting Multiple-Party Communications
US20080242422A1 (en) * 2007-03-30 2008-10-02 Uranus International Limited Method, Apparatus, System, Medium, and Signals for Supporting Game Piece Movement in a Multiple-Party Communication
CN102474548A (en) * 2009-10-20 2012-05-23 阿瓦雅公司 Display of persona information for peer-to-peer sessions
CN102474508A (en) * 2009-10-20 2012-05-23 阿瓦雅公司 Hierarchal Structuring Of Nodes In A Peer-To-Peer Network
US8706888B2 (en) * 2009-10-20 2014-04-22 Avaya, Inc. Hierarchal structuring of nodes in a peer-to-peer network
US20110093598A1 (en) * 2009-10-20 2011-04-21 Avaya Inc. Display of persona information for peer-to-peer sessions
US9614905B2 (en) 2009-10-20 2017-04-04 Avaya Inc. Determination of persona information availability and delivery on peer-to-peer networks
US20110090901A1 (en) * 2009-10-20 2011-04-21 Avaya Inc. Determination of persona information availability and delivery on peer-to-peer networks
US10681121B2 (en) 2009-10-20 2020-06-09 Avaya, Inc. Profile information based on participant identity for peer-to-peer sessions
US20110093599A1 (en) * 2009-10-20 2011-04-21 Avaya Inc. Hierarchal structuring of nodes in a peer-to-peer network

Also Published As

Publication number Publication date
US20070078933A1 (en) 2007-04-05
US7730132B2 (en) 2010-06-01
US20060041616A1 (en) 2006-02-23
US7437412B2 (en) 2008-10-14
US20070078931A1 (en) 2007-04-05
US7487210B2 (en) 2009-02-03
US20060041617A1 (en) 2006-02-23
US20070083593A1 (en) 2007-04-12
US7412482B2 (en) 2008-08-12
US20070078932A1 (en) 2007-04-05
US20070083595A1 (en) 2007-04-12
US7433921B2 (en) 2008-10-07
US6594688B2 (en) 2003-07-15
US20060059266A1 (en) 2006-03-16
US20040107253A1 (en) 2004-06-03
US7398296B2 (en) 2008-07-08
US20070088782A1 (en) 2007-04-19
US20040107254A1 (en) 2004-06-03
US7822813B2 (en) 2010-10-26
US6959322B2 (en) 2005-10-25
US20040103152A1 (en) 2004-05-27
US7054904B2 (en) 2006-05-30
US7206809B2 (en) 2007-04-17
US7831663B2 (en) 2010-11-09
US7421470B2 (en) 2008-09-02
US7152093B2 (en) 2006-12-19
US20060075121A1 (en) 2006-04-06
US20030225832A1 (en) 2003-12-04
US20070168426A1 (en) 2007-07-19
US7441001B2 (en) 2008-10-21
US20070088781A1 (en) 2007-04-19
US20070083596A1 (en) 2007-04-12
US6789105B2 (en) 2004-09-07
US7908320B2 (en) 2011-03-15
US7437411B2 (en) 2008-10-14
US20070078930A1 (en) 2007-04-05
US20030187940A1 (en) 2003-10-02
US20040107255A1 (en) 2004-06-03
US20020124051A1 (en) 2002-09-05
US20070083594A1 (en) 2007-04-12
US20030158901A1 (en) 2003-08-21
US20010044826A1 (en) 2001-11-22
US7444373B2 (en) 2008-10-28

Similar Documents

Publication Publication Date Title
US7908320B2 (en) Tracking user locations over multiple networks to enable real time communications
US6237025B1 (en) Multimedia collaboration system
US7185054B1 (en) Participant display and selection in video conference calls
US6898620B1 (en) Multiplexing video and control signals onto UTP

Legal Events

Date Code Title Description
AS Assignment

Owner name: AVISTAR COMMUNICATIONS CORPORATION, CALIFORNIA

Free format text: MERGER;ASSIGNOR:COLLABORATION PROPERTIES, INC.;REEL/FRAME:019910/0032

Effective date: 20071001

Owner name: AVISTAR COMMUNICATIONS CORPORATION,CALIFORNIA

Free format text: MERGER;ASSIGNOR:COLLABORATION PROPERTIES, INC.;REEL/FRAME:019910/0032

Effective date: 20071001

AS Assignment

Owner name: BALDWIN ENTERPRISES, INC., AS COLLATERAL AGENT, UT

Free format text: SECURITY AGREEMENT;ASSIGNOR:AVISTAR COMMUNICATIONS CORPORATION;REEL/FRAME:020325/0091

Effective date: 20080104

Owner name: BALDWIN ENTERPRISES, INC., AS COLLATERAL AGENT,UTA

Free format text: SECURITY AGREEMENT;ASSIGNOR:AVISTAR COMMUNICATIONS CORPORATION;REEL/FRAME:020325/0091

Effective date: 20080104

AS Assignment

Owner name: VICOR, INC., CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:LUDWIG, LESTER F.;LAUWERS, J. CHRIS;LANTZ, KEITH A.;AND OTHERS;REEL/FRAME:023574/0004;SIGNING DATES FROM 19930927 TO 19930929

Owner name: COLLABORATION PROPERTIES, INC., NEVADA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:VICOR, INC.;REEL/FRAME:023574/0345

Effective date: 20000626

Owner name: VICOR, INC.,CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:LUDWIG, LESTER F.;LAUWERS, J. CHRIS;LANTZ, KEITH A.;AND OTHERS;SIGNING DATES FROM 19930927 TO 19930929;REEL/FRAME:023574/0004

Owner name: COLLABORATION PROPERTIES, INC.,NEVADA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:VICOR, INC.;REEL/FRAME:023574/0345

Effective date: 20000626

Owner name: VICOR, INC., CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:LUDWIG, LESTER F.;LAUWERS, J. CHRIS;LANTZ, KEITH A.;AND OTHERS;SIGNING DATES FROM 19930927 TO 19930929;REEL/FRAME:023574/0004

AS Assignment

Owner name: AVISTAR COMMUNICATIONS CORPORATION, CALIFORNIA

Free format text: RELEASE OF SECURITY INTEREST IN INTELLECTUAL PROPERTY;ASSIGNOR:BALDWIN ENTERPRISES, INC., AS COLLATERAL AGENT;REEL/FRAME:023708/0861

Effective date: 20091229

Owner name: AVISTAR COMMUNICATIONS CORPORATION,CALIFORNIA

Free format text: RELEASE OF SECURITY INTEREST IN INTELLECTUAL PROPERTY;ASSIGNOR:BALDWIN ENTERPRISES, INC., AS COLLATERAL AGENT;REEL/FRAME:023708/0861

Effective date: 20091229

AS Assignment

Owner name: AVISTAR COMMUNICATIONS CORPORATION,CALIFORNIA

Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:BALDWIN ENTERPRISES, INC.;REEL/FRAME:023928/0118

Effective date: 20090115

Owner name: INTELLECTUAL VENTURES FUND 61 LLC,NEVADA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:AVISTAR COMMUNICATIONS CORPORATION;REEL/FRAME:023928/0222

Effective date: 20091217

Owner name: AVISTAR COMMUNICATIONS CORPORATION, CALIFORNIA

Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:BALDWIN ENTERPRISES, INC.;REEL/FRAME:023928/0118

Effective date: 20090115

Owner name: INTELLECTUAL VENTURES FUND 61 LLC, NEVADA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:AVISTAR COMMUNICATIONS CORPORATION;REEL/FRAME:023928/0222

Effective date: 20091217

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION

AS Assignment

Owner name: PRAGMATUS AV LLC, VIRGINIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:INTELLECTUAL VENTURES FUND 61 LLC;REEL/FRAME:025339/0981

Effective date: 20100616