WO2012093330A1 - A method and system of controlling a remote controlled device in a remote controlled surgical procedure - Google Patents

A method and system of controlling a remote controlled device in a remote controlled surgical procedure Download PDF

Info

Publication number
WO2012093330A1
WO2012093330A1 PCT/IB2012/000009 IB2012000009W WO2012093330A1 WO 2012093330 A1 WO2012093330 A1 WO 2012093330A1 IB 2012000009 W IB2012000009 W IB 2012000009W WO 2012093330 A1 WO2012093330 A1 WO 2012093330A1
Authority
WO
WIPO (PCT)
Prior art keywords
program
control
client
computer
camera
Prior art date
Application number
PCT/IB2012/000009
Other languages
French (fr)
Inventor
Monroe M. Thomas
Matthew J. STEPHURE
Kevin G. ROBINSON
Original Assignee
Calgary Scientific, Inc.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Calgary Scientific, Inc. filed Critical Calgary Scientific, Inc.
Priority to CA2823615A priority Critical patent/CA2823615C/en
Priority to EP12731899.6A priority patent/EP2661654A4/en
Publication of WO2012093330A1 publication Critical patent/WO2012093330A1/en

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/18Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
    • H04N7/181Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast for receiving images from a plurality of remote sources
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/617Upgrading or updating of programs or applications for camera control
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/66Remote control of cameras or camera parts, e.g. by remote control devices
    • H04N23/661Transmitting camera control signals through networks, e.g. control via the Internet
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/04Synchronising
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/46Multiprogramming arrangements
    • G06F9/52Program synchronisation; Mutual exclusion, e.g. by means of semaphores
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/62Control of parameters via user interfaces

Definitions

  • a surgeon carrying out a surgical procedure on a patient in an operating room may require assistance from an expert that is unable to attend the surgery in the operating room.
  • Cameras mounted in the operating room may capture the surgical procedure and transmit video data thereof to a computer for display to the expert.
  • State of the art surveillance systems such as, for example, PUREACTIV provide remote control of camera functions such as, for example, pan, tilt, and zoom, via a server computer connected thereto.
  • Authorized users are enabled to access the surveillance system using a client computer connected to the server computer via a computer network such as the Internet for controlling the camera functions, receiving captured video data as well as accessing previously captured video data stored in a database.
  • wireless handheld devices such as, for example, IPHONE and ANDROID devices has rapidly increased over the last couple of years to the extent that now nearly every professional owns at least one wireless handheld device.
  • wireless mobile technology enables use of small wireless handheld devices to access the internet and download video streams.
  • a method and system for providing remote control of a device over a computer network by a remotely connected client device is provided.
  • the method and system provide remote control of a camera from a wireless handheld device via a wireless computer network.
  • Figure 1 is a simplified block diagram of a system for remotely controlling a device via a computer network
  • Figures 2a to 2d are simplified block diagrams illustrating communication between a client computer and a server computer of the system shown in Figure 1;
  • Figures 3a to 3c are simplified flow diagrams of a method for remotely controlling a device via a computer network. DETAILED DESCRIPTION OF THE DISCLOSURE
  • a system 100 for remotely controlling a device via a computer network comprises a plurality of client computers 112A, 112B connected via a communication network 110 to a server computer 102, for example, a central server computer of a surveillance system.
  • client computers 112A, 112B are shown as mobile devices, (e.g., wireless handheld devices), the client computers 112A, 112B may be any type of computing device, such as laptop/notebook computers, tablet devices, desktop computers, etc.
  • the server computer 102 may perform one or more of many functions, such as functions associated with a video server, a gateway server, a recording service, etc.
  • Cameras 108A, 108B may be connected to the server computer 102 using, for example, a Local Area Network (LAN) or the Internet.
  • LAN Local Area Network
  • the cameras 108A, 108B are mounted in an operating room for capturing video data of a surgery performed therein.
  • the system 100 may be made up of any combination of one or more of each of the server, the client device and the cameras.
  • the cameras 108A, 108B may be digital cameras such as, for example, digital Internet Protocol (IP) cameras, connected to a processor 104 of the server computer 102.
  • IP digital Internet Protocol
  • analog cameras may be connected to the processor 104 of the server computer 102, for example, via a capture card for digitizing the analog video signal.
  • one or more cameras for example, a web cam, or a camera of another mobile device such as an I PHONE are connected to a server computer via the Internet.
  • the system 100 may provide for video distribution, recording, and camera control capabilities.
  • the cameras 108A, 108B may be statically defined at configuration time, such that their capabilities and control functionalities are known to the server computer 102.
  • the server computer 102 may be a video server or a gateway to access other video servers, or the server computer 102 may perform both functions.
  • the cameras 108A, 108B may be defined as a "video source.”
  • Video sources represent static sources of video data, whether through an Internet Protocol (IP) camera, or through an analog camera that is digitized via a capture card installed in, e.g., the server computer 102.
  • IP Internet Protocol
  • the server computer 102 may enumerate available video sources and request the most recent frame captured from a particular video source, at a variety of resolutions and/or compression levels.
  • the system 100 enables a camera control channel to be associated with a video source.
  • the camera control channel is typically an S232/422/485 serial data connection between, e.g., cameras 108A, 108B and the server computer 102. Multiple cameras may be addressed by the server computer 102 over the camera control channels.
  • an IP connection may be provided over a LAN to a networked camera.
  • an abstract camera control protocol may be defined, which may be implemented on top of native protocols used by the various different models of cameras.
  • the abstract camera protocol may be used by the system 100 to define functions, such as a linear zoom model that allows point and click positioning and drag box zooming to be implemented on a remote user interface.
  • Such absolute positioning commands mitigate lag effects due to latency from which relative positioning commands suffer.
  • an add-on device may be interposed between the cameras 108A, 108B and the server computer 102 to control the cameras 108A, 108B. The add-on device may convert commands to the native camera protocols to implement functions of the cameras 108A, 108B.
  • the camera control facilities of the system 100 may be exposed through a TCP based network port for use by multiple users, as discussed in further detail with regard to
  • the camera functions such as, for example, pan, tilt, and zoom are controlled by executing commands, on the processor 104, a camera control program that stored in memory
  • the camera control program further controls capture and provision of video data in a substantially real-time fashion.
  • remote control of the camera functions using, for example, client computers 112A,
  • 112B is enabled by executing a server remote access program on the processor 104 of the server computer 102 and a respective client remote access program executed on a processor
  • the server remote access program may be performed by executing executable commands stored in the memory 106 of the server computer 102, while the client remote access program is performed by executing executable commands stored in memory 120A, 120B of the client computer 112A, 112B.
  • An example of the server remote access program is PUREWEB, available from Calgary Scientific, Inc. of Calgary, Alberta.
  • the following capabilities may be provided by the system 100: enumeration of the abstract camera control capabilities associated with each video source; requesting permission to acquire control of a camera, as only one user may control a camera at a time; sending various camera control commands according to the abstract protocol if control has been acquired; and releasing control of a camera after a timeout period or if a user with higher priority requests control
  • the video source (e.g., cameras 108A, 108B) may be recorded by a recording service running on, e.g., the server computer 102 or another computer in communication with the server computer 102.
  • the recording service may specify a maximum time to store video or a minimum disk space to maintain as free space. If the video stream is being recorded, a frame captured at certain time in the past may be requested and delivered to the recording service for storage.
  • the recording service may delete video in time-based chunks from oldest to newest across all video sources, define a policy to mark portions of video to prevent deletion, and/or export segments of recorded video as an AVI (or other video format) file.
  • the recording service may store video either in its native form as received from the cameras 108A, 108B or the computer network 110, or may transcode the video to use, e.g., JPEG encoding.
  • the frame frequency may be decimated to reduce storage requirements.
  • the system 100 of Figure 1 may be used to provide a dynamic point-to-multipoint video distribution system.
  • the system 100 may provide security and audit capabilities for, e.g., medical or other applications requiring audit trails.
  • the system 100 may provide security and audit capabilities for, e.g., medical or other applications requiring audit trails.
  • 100 may provide dynamic publishing of video from a variety of sources to the server computer
  • the sources may include dynamic video sources, such as client computers 112A, 112B (e.g., mobile phones, laptops), remote IP cameras, and third-party video systems.
  • client computers 112A, 112B e.g., mobile phones, laptops
  • remote IP cameras e.g., Internet Protocol
  • third-party video systems e.g., third-party video systems.
  • Such dynamic video sources may not require a static configuration that is predefined at the server computer 102; rather the dynamic video sources may pass configuration information in a client difference program to the server computer 102, as described below.
  • the server computer 102 may route the received video to the recording service for recording or to the video server for future distribution.
  • the system 100 may provide a mechanism for client computers to subscribe to video from the server computer 102. Such video may be recorded video available through the recording service, or live video being provided by the video server.
  • the client computers 112A, 112B may communicate with the server computer 102, as an interface, which will delegate communications to and from the video server and/or recording service.
  • the processor 104 may execute an application program that is stored in memory 106 of the server computer 102.
  • the application program may be any program.
  • the server computer 102 may enable remote access to the application program, such that the client computers 112A, 112B may interact separately or collaboratively with the application program over the communication network 110.
  • the application program may not directly support video.
  • the sever computer 102 may inject video frames into the communications involving the application program and the client computers 112A, 112B.
  • the server computer 102 may provide a security and authentication service.
  • the sever computer 102 may include security policies and audit provisions such that in, e.g., medical applications, video can be safely collected with confidence that no one will be able to see or replay the video without consent.
  • a subject may specify a policy regarding how long the recorded video may be kept, whether copies of recordings may be made, and who can view recordings in the future, as well as have an audit trail maintained of any copies made or viewings of recordings for forensic security purposes.
  • communication between the client computer 112A, 112B and the server computer 102 is provided as communication between the server remote access program and the client remote access program via, for example, a wireless or wired computer network.
  • the server remote access program communicates with the camera control program while the client remote access program communicates with a user interaction program such as, for example, a web browser for: displaying data such as, for example, video data and camera control data; and for receiving user input data for controlling the camera or video replay functions from, for example, a touch screen of a handheld wireless device, from a mouse or keyboard, or by voice command.
  • the server remote access program and the client remote access program may be implemented using standard programming languages and communication is enabled using standard communication technologies such as, for example, Hyper Text Transfer Protocol (HTTP), Virtual Private Networks (VPN), and Secure Socket Layers (SSL) which are well known to those skilled in the art.
  • Provision of the server remote access program and the client remote access program enable implementation of aspects of the disclosure as a retrofit to existing technologies on the server side as well as on the client side.
  • the server remote access program may control storage and replay of the captured video data.
  • the server remote access program may control access for controlling the cameras 108A, 108B as well as access to the real-time video data and the stored video data.
  • Communication between the server remote access program and the client remote access program may be based on transmission of difference programs having encoded data indicative of a change of state of the camera 108A, 108B, as will be described herein below.
  • the difference programs comprise executable commands for execution by a processor.
  • the encoded data are, for example, indicative of: a path to the change in the state of the camera; a type of the change; and a value of the change.
  • the path to the change is, for example, one of the camera functions such as pan, tilt, and zoom.
  • the type of change is, for example, a "modification", with a new value is for one of the camera function.
  • An example of a difference program is:
  • the server remote access program may generate "view data" of the video data captured by the camera and transmit the same in a fashion according to hardware capabilities of the client computer 112A, 112B.
  • the "view data” may be determined in accordance with a processing capacity, memory size, type of graphical display, and type of user interface of the client computer 112A, 112B.
  • "view data” generated and transmitted for a laptop computer are different from “view data” generated and transmitted for a handheld device such as, for example, an I PHONE.
  • Generation of "view data” enables a substantial reduction in the amount of data transmitted (i.e., results in a savings in bandwidth utilization) for display on the small display of a handheld wireless device.
  • a further reduction of the amount of transmitted data is realized by transmitting only image data related to pixel values that have changed between successive image frames of the video data using difference programs as illustrated in Figure 2a.
  • the values of only a portion of the image pixels change between successive image frames, thus enabling a substantial reduction in transmitted data.
  • the client computers may cache image data received from the server computer 102.
  • the client when requesting a most recent frame captured from a particular video source, may utilize locally cached image data to reduce the amount of video date transmitted by the server computer 102.
  • Logic may be provided on the client computer 112A, 112B to determine what cached image data may be reused to provide a complete image to the user of the client computer.
  • the video data are directly transmitted as a video stream of successive image frames between the server remote access program and the client remote access program or the camera control program and the user interaction program.
  • Direct transmission of the video data as a video stream reduces the latency between image capture and display at the cost of increased bandwidth needed for the data transmission to the client computer 112A, 112B.
  • the client computer 112A, 112B may access a TCP-based network port to retrieve video.
  • the video may be simultaneously accessed by multiple client computers at the enumerated TCP-based network port.
  • live video may be distributed as individual JPEG frames with configurable resolution and compression quality.
  • the frames are not streamed, rather each frame is requested by the client computer 112A, 112B. While this may limit the maximum frame rate achievable, it offers a substantially reduced latency between capture at the image source and distribution with no buffering required.
  • the video data may be stored in a video database 103 connected to the server computer 102 enabling replay of the video data.
  • the camera control program provides the video data to the recording service or video database for storage therein, retrieves the same in dependence upon a request received from the server remote access program and provides the retrieved video data to the server remote access program for transmission to the client computer 112A, 112B.
  • the server remote access program may directly retrieve the video data from the video database.
  • the server remote access program may store the data in the video database 103 upon receipt from the camera control program.
  • the video server may include a buffer that plays back the video at its original rate where the client requests sample the buffer as the requests are received.
  • a method for remotely controlling a device via a computer network is shown.
  • a client difference program having encoded data indicative of a change of a state of the device last received from the server computer 102 is generated (10) and transmitted (12) to the server computer 102.
  • the data are indicative of a change of pan, tilt, or zoom of the camera 108A, 108B.
  • the client difference program is executed and an updated state of the device is determined (14).
  • the processor 104 then generates control data indicative of the updated state of the device and provides the control data to the device (16).
  • the updated state of the device is determined by the server remote access program while the control data are determined by the camera control program.
  • the device executes the control data - for example, zooms in towards an object - and sends data indicative of the state of the device - for example, the new focal length - to the processor 104 of the server computer 102.
  • the processor 104 of the server computer 102 Upon receipt (18), the processor 104 of the server computer 102 generates (20) a server difference program having encoded a difference between the stale of the device and a state of the device last transmitted to the client computer 112A, 1128 and transmits (22) the same to the client computer 112A, 112B.
  • the processor 118A, 118B of the client computer 1 12A, 1128 executes (24) the server difference program, updates the state of the device last received from the server computer accordingly and displays the same on display 114A, 114B in a human comprehensible fashion.
  • the video data are transmitted as a succession of server difference programs having encoded data indicative of a change in pixel values between consecutive image frames of the video data.
  • the data indicative of a change of zoom of the camera 108A, 108B are based on a linear zoom model allowing point and click positioning and drag box zooming using the graphical user interface 114A of the client computer 112A.
  • FIG. 3b a method for managing access for remotely controlling a device via a computer network according to a preferred embodiment of the disclosure is shown.
  • user A of client computer 112A is controlling the device and user B of client computer 1128 provides input data indicative of a request to control the device.
  • the processor 118B of the client computer 112B Upon receipt of the input data the processor 118B of the client computer 112B generates (30) data indicative of a request to control the device and transmits (32) the request to the server computer 102.
  • the processor 104 of the server computer determines (34) a priority of the client computer 112A and 112B.
  • data indicative of a priority of each of a plurality of users may be stored in a database in the form of a look-up table which is managed by an administrator of the remote control system.
  • the processor accesses the look-up table and retrieves the data indicative of the priority of the user.
  • the processor 104 of the server computer 102 continues (38A) the communication with the client computer 112A allowing user A to control the device.
  • the processor 104 of the server computer 102 further generates (39A) data indicative of denied access to control the device and transmits the same to the client computer 112B.
  • the processor 104 of the server computer 102 enables (38B) the communication with the client computer 112B allowing user B 10 control the device.
  • the processor 104 of the server computer 102 further generates (39B) data indicative of denied access to control the device and transmits the same to the client computer 112A.
  • Functions related to manage user access i.e., the steps 34 to 39, are may be by executing the server remote access program, thus limiting the camera control program to communicate the commands to the camera and provide the captured video data.
  • user A or user B is controlling the device, other are able to view the video capture or provided by device by either accessing the video stream or requesting frames of image data, as noted above.
  • user control may be restricted to specific devices of a plurality of devices connected to the server computer such as: user A is enabled to control camera 108A while user B is enabled to control camera B.
  • user A is enabled to control the devices while user B is only enabled to receive the video data, etc.
  • user A and user B may receive video data.
  • One of ordinary skill in the art would now recognize that other variations of access and presentation of video are possible where multiple devices are communicating with the server computer 102.
  • FIG. 3c a method for remotely controlling a device via a computer network is shown.
  • the process of controlling the device and transmission of the video data is the same as illustrated in Figure 3a, with same reference numerals referring to same steps.
  • the video data are stored in memory, for example, in memory 106 of the server computer 102 or in video database 103 connected to the server computer 102.
  • the processor 118A, 118B of the client computer 112A, 112B receives input data indicative of a request to provide the stored video data and generates data indicative thereof, for example, encoded in a client difference program and transmits (42) the same to the server computer 102.
  • the request is, for example, indicative of video data captured by which camera and a start time of the video data.
  • the user is provided with a menu for the video data selection displayed on the user interaction interface 114A,
  • the processor 104 of the server computer 102 retrieves the stored video data and generates (44) a server difference program having encoded a difference between the stored state of the device and the stored state of the device last transmitted to the client computer 112A, 112B, i.e. a difference between consecutive image frames of the video data, and transmits (46) the same to the client computer 112A, 112B.
  • the processor 118A, 118B of the client computer 112A, 112B executes (48) the server difference program, updates the state of the device last received from the server computer accordingly and displays the same on display 114A, 114B in a human comprehensible fashion.
  • the video data are transmitted as a succession of server difference programs having encoded data indicative of a change in pixel values between consecutive image frames of the video data.
  • the camera control program and the server remote access program are integrated in one program for execution on the processor 104 of the server computer 102.
  • the client remote access program and the user interaction program are integrated into one program for execution on the processor 118A, 118B of the client computer 112A, 112B.
  • each client computer may communicate a respective client difference program to the server computer 102 that represents changes made at the particular client computer.
  • the client difference programs may be executed at the server computer 102 for determining control data to update a state of the device, followed by generating a second server difference program reflecting the updated state of the device.
  • the plurality of client computers may, thus simultaneously and collaboratively control the device.
  • the plurality of client computers may be prioritized, as in Figure 3b.
  • the second server difference program may be communicated to each of the plurality of client device to update the displays thereof.
  • each connected client computer receives live or recorded video while control operations are being performed.
  • Implementation of the methods for remotely controlling a device via a computer network is not limited to the computer network architecture illustrated in Figure 1 and 2a to 2d.
  • the server remote access program may be executed on a processor of a first server computer while the camera control program may be executed on a processor of a second server computer connected thereto, for example, via a computer network such as the Internet.
  • the system 100 may be used to control multiple cameras in communication with multiple server computers.
  • each of the multiple server computers may be accessible at a different network address (e.g., a Uniform Resource Locator (URL)) to which the client access program man connect.
  • URL Uniform Resource Locator
  • the video database may be connected to the first server computer and accessed by the server remote control program, leaving the second server computer and the camera control program for controlling the cameras and providing the captured video data.
  • remote control may be effected by any type of client computer, i.e., mobile device, desktop computer, laptop/notebook computer, a tablet device, etc.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Software Systems (AREA)
  • Studio Devices (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
  • Closed-Circuit Television Systems (AREA)

Abstract

A method and a system of controlling a remote controlled device in a remote controlled surgical procedure using a computer network. A client computer generates a client difference program indicative of a change of a state of the device last received from a server computer and transmits the same to the server computer. Upon receipt, the server computer executes the client difference program and determines an updated state of the device, generates control data indicative of the updated state of the device, provides the control data to the device, and generates and transmits a server difference program having encoded a difference between the state of the device and a state of the device last transmitted to the client computer. The client computer executes the server difference program for updating the state of the device last received from the server computer and for displaying the same in a human comprehensible fashion.

Description

A Method and System of Controlling a Remote Controlled Device in a Remote Controlled Surgical Procedure.
CROSS-REFERENCE TO RELATED APPLICATION
[0001] This application claims priority to U.S. Patent Application No. 61/429,573, entitled, "Method and System for Providing Remote Control from a Mobile Device," filed January 4, 2011, the which is incorporated herein by reference in its entirety.
BACKGROUND OF THE DISCLOSURE
[0002] In numerous present day scenarios it may be desired to enable remote control of devices, such as a camera. For example, in the healthcare sector, a surgeon carrying out a surgical procedure on a patient in an operating room may require assistance from an expert that is unable to attend the surgery in the operating room. Cameras mounted in the operating room may capture the surgical procedure and transmit video data thereof to a computer for display to the expert.
[0003] State of the art surveillance systems such as, for example, PUREACTIV provide remote control of camera functions such as, for example, pan, tilt, and zoom, via a server computer connected thereto. Authorized users are enabled to access the surveillance system using a client computer connected to the server computer via a computer network such as the Internet for controlling the camera functions, receiving captured video data as well as accessing previously captured video data stored in a database.
[0004] Use of wireless handheld devices such as, for example, IPHONE and ANDROID devices has rapidly increased over the last couple of years to the extent that now nearly every professional owns at least one wireless handheld device. State of the art wireless mobile technology enables use of small wireless handheld devices to access the internet and download video streams.
SUMMARY OF THE DISCLOSURE
[0005] A method and system for providing remote control of a device over a computer network by a remotely connected client device is provided. According to some implementations, the method and system provide remote control of a camera from a wireless handheld device via a wireless computer network.
[0006] Other systems, methods, features and/or advantages will be or may become apparent to one with skill in the art upon examination of the following drawings and detailed description. It is intended that all such additional systems, methods, features and/or advantages be included within this description and be protected by the accompanying claims.
BRIEF DESCRIPTION OF THE DRAWINGS
[0007] The accompanying drawings are not intended to be drawn to scale. In the drawings, each identical or nearly identical component that is illustrated in various figures is represented by a like numeral. For purposes of clarity, not every component may be labeled in every drawing. In the drawings:
[0008] Figure 1 is a simplified block diagram of a system for remotely controlling a device via a computer network;
[0009] Figures 2a to 2d are simplified block diagrams illustrating communication between a client computer and a server computer of the system shown in Figure 1; and
[0010] Figures 3a to 3c are simplified flow diagrams of a method for remotely controlling a device via a computer network. DETAILED DESCRIPTION OF THE DISCLOSURE
[0011] Unless defined otherwise, all technical and scientific terms used herein have the same meaning as commonly understood by one of ordinary skill in the art to which the disclosure belongs. Any methods and materials similar or equivalent to those described herein can be used in the practice or testing of the present disclosure.
[0012] While implementations of the disclosure will be described for remotely controlling functions of a camera in a clinical environment only for the sake of simplicity, it will become evident to those skilled in the art that the embodiments of the disclosure are not limited thereto, but are applicable in numerous other fields such as, for example, controlling of cameras of surveillance systems and cameras of telerobotic equipment, of remotely operated vehicles as well as the operation of the vehicle. Furthermore, while implementations of the disclosure will be described for video data transmission, it will become evident that the embodiments of the disclosure are also applicable for transmission of other data such as, for example, audio data or data captured by various types of sensors such as temperature sensors.
[0013] Referring to Figure 1, a system 100 for remotely controlling a device via a computer network is shown. The system comprises a plurality of client computers 112A, 112B connected via a communication network 110 to a server computer 102, for example, a central server computer of a surveillance system. Although the client computers 112A, 112B are shown as mobile devices, (e.g., wireless handheld devices), the client computers 112A, 112B may be any type of computing device, such as laptop/notebook computers, tablet devices, desktop computers, etc. As will be described below, the server computer 102 may perform one or more of many functions, such as functions associated with a video server, a gateway server, a recording service, etc. Cameras 108A, 108B may be connected to the server computer 102 using, for example, a Local Area Network (LAN) or the Internet. For example, in a clinical environment the cameras 108A, 108B are mounted in an operating room for capturing video data of a surgery performed therein. The system 100 may be made up of any combination of one or more of each of the server, the client device and the cameras.
[0014] The cameras 108A, 108B may be digital cameras such as, for example, digital Internet Protocol (IP) cameras, connected to a processor 104 of the server computer 102. Alternatively, analog cameras may be connected to the processor 104 of the server computer 102, for example, via a capture card for digitizing the analog video signal. Further alternatively, one or more cameras, for example, a web cam, or a camera of another mobile device such as an I PHONE are connected to a server computer via the Internet.
[0015] In some implementations, the system 100 may provide for video distribution, recording, and camera control capabilities. For instance, the cameras 108A, 108B may be statically defined at configuration time, such that their capabilities and control functionalities are known to the server computer 102. In such a system 100, the server computer 102 may be a video server or a gateway to access other video servers, or the server computer 102 may perform both functions.
[0016] The cameras 108A, 108B may be defined as a "video source." Video sources represent static sources of video data, whether through an Internet Protocol (IP) camera, or through an analog camera that is digitized via a capture card installed in, e.g., the server computer 102. The server computer 102 may enumerate available video sources and request the most recent frame captured from a particular video source, at a variety of resolutions and/or compression levels. The system 100 enables a camera control channel to be associated with a video source. The camera control channel is typically an S232/422/485 serial data connection between, e.g., cameras 108A, 108B and the server computer 102. Multiple cameras may be addressed by the server computer 102 over the camera control channels. In some implementations, an IP connection may be provided over a LAN to a networked camera.
[0017] In some implementations, an abstract camera control protocol may be defined, which may be implemented on top of native protocols used by the various different models of cameras. As such, the abstract camera protocol may be used by the system 100 to define functions, such as a linear zoom model that allows point and click positioning and drag box zooming to be implemented on a remote user interface. Such absolute positioning commands mitigate lag effects due to latency from which relative positioning commands suffer. Optionally or alternatively, an add-on device may be interposed between the cameras 108A, 108B and the server computer 102 to control the cameras 108A, 108B. The add-on device may convert commands to the native camera protocols to implement functions of the cameras 108A, 108B.
[0018] The camera control facilities of the system 100 may be exposed through a TCP based network port for use by multiple users, as discussed in further detail with regard to
Figures 3a-3c. The camera functions such as, for example, pan, tilt, and zoom are controlled by executing commands, on the processor 104, a camera control program that stored in memory
106 of the server computer 102. The camera control program further controls capture and provision of video data in a substantially real-time fashion. According to aspects of the disclosure, remote control of the camera functions using, for example, client computers 112A,
112B is enabled by executing a server remote access program on the processor 104 of the server computer 102 and a respective client remote access program executed on a processor
118A, 118B of the client computer 112A, 112B. The server remote access program may be performed by executing executable commands stored in the memory 106 of the server computer 102, while the client remote access program is performed by executing executable commands stored in memory 120A, 120B of the client computer 112A, 112B. An example of the server remote access program is PUREWEB, available from Calgary Scientific, Inc. of Calgary, Alberta.
[0019] In addition to the above camera control functions, the following capabilities may be provided by the system 100: enumeration of the abstract camera control capabilities associated with each video source; requesting permission to acquire control of a camera, as only one user may control a camera at a time; sending various camera control commands according to the abstract protocol if control has been acquired; and releasing control of a camera after a timeout period or if a user with higher priority requests control
[0020] In some implementations, the video source (e.g., cameras 108A, 108B) may be recorded by a recording service running on, e.g., the server computer 102 or another computer in communication with the server computer 102. For example, the recording service may specify a maximum time to store video or a minimum disk space to maintain as free space. If the video stream is being recorded, a frame captured at certain time in the past may be requested and delivered to the recording service for storage. The recording service may delete video in time-based chunks from oldest to newest across all video sources, define a policy to mark portions of video to prevent deletion, and/or export segments of recorded video as an AVI (or other video format) file. The recording service may store video either in its native form as received from the cameras 108A, 108B or the computer network 110, or may transcode the video to use, e.g., JPEG encoding. The frame frequency may be decimated to reduce storage requirements.
[0021] In some implementations, the system 100 of Figure 1 may be used to provide a dynamic point-to-multipoint video distribution system. The system 100 may provide security and audit capabilities for, e.g., medical or other applications requiring audit trails. The system
100 may provide dynamic publishing of video from a variety of sources to the server computer
102. The sources may include dynamic video sources, such as client computers 112A, 112B (e.g., mobile phones, laptops), remote IP cameras, and third-party video systems. Such dynamic video sources may not require a static configuration that is predefined at the server computer 102; rather the dynamic video sources may pass configuration information in a client difference program to the server computer 102, as described below. The server computer 102 may route the received video to the recording service for recording or to the video server for future distribution.
[0022] The system 100 may provide a mechanism for client computers to subscribe to video from the server computer 102. Such video may be recorded video available through the recording service, or live video being provided by the video server. The client computers 112A, 112B may communicate with the server computer 102, as an interface, which will delegate communications to and from the video server and/or recording service.
[0023] Additionally or alternatively, the processor 104 may execute an application program that is stored in memory 106 of the server computer 102. For purposes of the present disclosure, the application program may be any program. The server computer 102 may enable remote access to the application program, such that the client computers 112A, 112B may interact separately or collaboratively with the application program over the communication network 110. The application program, however, may not directly support video. In accordance with some implementations, the sever computer 102 may inject video frames into the communications involving the application program and the client computers 112A, 112B.
[0024] In some implementations, the server computer 102 may provide a security and authentication service. The sever computer 102 may include security policies and audit provisions such that in, e.g., medical applications, video can be safely collected with confidence that no one will be able to see or replay the video without consent. Furthermore, a subject may specify a policy regarding how long the recorded video may be kept, whether copies of recordings may be made, and who can view recordings in the future, as well as have an audit trail maintained of any copies made or viewings of recordings for forensic security purposes.
[0025] With reference now to Figures 2a to 2d, communication between the client computer 112A, 112B and the server computer 102 is provided as communication between the server remote access program and the client remote access program via, for example, a wireless or wired computer network. The server remote access program communicates with the camera control program while the client remote access program communicates with a user interaction program such as, for example, a web browser for: displaying data such as, for example, video data and camera control data; and for receiving user input data for controlling the camera or video replay functions from, for example, a touch screen of a handheld wireless device, from a mouse or keyboard, or by voice command. The server remote access program and the client remote access program may be implemented using standard programming languages and communication is enabled using standard communication technologies such as, for example, Hyper Text Transfer Protocol (HTTP), Virtual Private Networks (VPN), and Secure Socket Layers (SSL) which are well known to those skilled in the art. Provision of the server remote access program and the client remote access program enable implementation of aspects of the disclosure as a retrofit to existing technologies on the server side as well as on the client side. Thus, the server remote access program may control storage and replay of the captured video data. Further, the server remote access program may control access for controlling the cameras 108A, 108B as well as access to the real-time video data and the stored video data.
[0026] Communication between the server remote access program and the client remote access program may be based on transmission of difference programs having encoded data indicative of a change of state of the camera 108A, 108B, as will be described herein below. The difference programs comprise executable commands for execution by a processor. The encoded data are, for example, indicative of: a path to the change in the state of the camera; a type of the change; and a value of the change. The path to the change is, for example, one of the camera functions such as pan, tilt, and zoom. The type of change is, for example, a "modification", with a new value is for one of the camera function. An example of a difference program is:
Path="Camera Function/Zoom"
Type="Modification"
Value="50"
[0027] The server remote access program may generate "view data" of the video data captured by the camera and transmit the same in a fashion according to hardware capabilities of the client computer 112A, 112B. For example, the "view data" may be determined in accordance with a processing capacity, memory size, type of graphical display, and type of user interface of the client computer 112A, 112B. A such, "view data" generated and transmitted for a laptop computer are different from "view data" generated and transmitted for a handheld device such as, for example, an I PHONE. Generation of "view data" enables a substantial reduction in the amount of data transmitted (i.e., results in a savings in bandwidth utilization) for display on the small display of a handheld wireless device. A further reduction of the amount of transmitted data is realized by transmitting only image data related to pixel values that have changed between successive image frames of the video data using difference programs as illustrated in Figure 2a. Typically, the values of only a portion of the image pixels change between successive image frames, thus enabling a substantial reduction in transmitted data. In addition, the client computers may cache image data received from the server computer 102. By providing a cache, the client, when requesting a most recent frame captured from a particular video source, may utilize locally cached image data to reduce the amount of video date transmitted by the server computer 102. Logic may be provided on the client computer 112A, 112B to determine what cached image data may be reused to provide a complete image to the user of the client computer.
[0028] Alternatively, the video data are directly transmitted as a video stream of successive image frames between the server remote access program and the client remote access program or the camera control program and the user interaction program. Direct transmission of the video data as a video stream reduces the latency between image capture and display at the cost of increased bandwidth needed for the data transmission to the client computer 112A, 112B. The client computer 112A, 112B may access a TCP-based network port to retrieve video. In some implementations, the video may be simultaneously accessed by multiple client computers at the enumerated TCP-based network port.
[0029] In some implementations, live video may be distributed as individual JPEG frames with configurable resolution and compression quality. Thus, the frames are not streamed, rather each frame is requested by the client computer 112A, 112B. While this may limit the maximum frame rate achievable, it offers a substantially reduced latency between capture at the image source and distribution with no buffering required.
[0030] Referring to Figure 2d, the video data may be stored in a video database 103 connected to the server computer 102 enabling replay of the video data. The camera control program provides the video data to the recording service or video database for storage therein, retrieves the same in dependence upon a request received from the server remote access program and provides the retrieved video data to the server remote access program for transmission to the client computer 112A, 112B. Alternatively, the server remote access program may directly retrieve the video data from the video database. Further alternatively, the server remote access program may store the data in the video database 103 upon receipt from the camera control program. In some implementations, the video server may include a buffer that plays back the video at its original rate where the client requests sample the buffer as the requests are received.
[0031] Referring to Figures 3a to 3c, a method for remotely controlling a device via a computer network is shown. Using a processor 118A, 118B of the client computer 112A, 112B, a client difference program having encoded data indicative of a change of a state of the device last received from the server computer 102 is generated (10) and transmitted (12) to the server computer 102. For example, the data are indicative of a change of pan, tilt, or zoom of the camera 108A, 108B. Using the processor 104 of the server computer 102, the client difference program is executed and an updated state of the device is determined (14). The processor 104 then generates control data indicative of the updated state of the device and provides the control data to the device (16). For example, the updated state of the device is determined by the server remote access program while the control data are determined by the camera control program. The device executes the control data - for example, zooms in towards an object - and sends data indicative of the state of the device - for example, the new focal length - to the processor 104 of the server computer 102. Upon receipt (18), the processor 104 of the server computer 102 generates (20) a server difference program having encoded a difference between the stale of the device and a state of the device last transmitted to the client computer 112A, 1128 and transmits (22) the same to the client computer 112A, 112B. Upon receipt, the processor 118A, 118B of the client computer 1 12A, 1128 executes (24) the server difference program, updates the state of the device last received from the server computer accordingly and displays the same on display 114A, 114B in a human comprehensible fashion.
[0032] Optionally, when transmitting video data to the client computer 112A, 112B generation and transmission of client difference programs are omitted, i.e. the video data are transmitted as a succession of server difference programs having encoded data indicative of a change in pixel values between consecutive image frames of the video data. For example, the data indicative of a change of zoom of the camera 108A, 108B are based on a linear zoom model allowing point and click positioning and drag box zooming using the graphical user interface 114A of the client computer 112A.
[0033] Referring to Figure 3b, a method for managing access for remotely controlling a device via a computer network according to a preferred embodiment of the disclosure is shown. For example, user A of client computer 112A is controlling the device and user B of client computer 1128 provides input data indicative of a request to control the device. Upon receipt of the input data the processor 118B of the client computer 112B generates (30) data indicative of a request to control the device and transmits (32) the request to the server computer 102. Upon receipt of the request the processor 104 of the server computer determines (34) a priority of the client computer 112A and 112B. For example, data indicative of a priority of each of a plurality of users may be stored in a database in the form of a look-up table which is managed by an administrator of the remote control system. At 34, the processor accesses the look-up table and retrieves the data indicative of the priority of the user.
[0034] If user A has priority (36) the processor 104 of the server computer 102 continues (38A) the communication with the client computer 112A allowing user A to control the device. The processor 104 of the server computer 102 further generates (39A) data indicative of denied access to control the device and transmits the same to the client computer 112B. If user B has priority (36) the processor 104 of the server computer 102 enables (38B) the communication with the client computer 112B allowing user B 10 control the device. The processor 104 of the server computer 102 further generates (39B) data indicative of denied access to control the device and transmits the same to the client computer 112A.
[0035] Functions related to manage user access, i.e., the steps 34 to 39, are may be by executing the server remote access program, thus limiting the camera control program to communicate the commands to the camera and provide the captured video data. In addition, while user A or user B is controlling the device, other are able to view the video capture or provided by device by either accessing the video stream or requesting frames of image data, as noted above.
[0036] Of course there are numerous variations of provision of user access possible. For example, user control may be restricted to specific devices of a plurality of devices connected to the server computer such as: user A is enabled to control camera 108A while user B is enabled to control camera B. In another example, user A is enabled to control the devices while user B is only enabled to receive the video data, etc. Yet further, user A and user B may receive video data. One of ordinary skill in the art would now recognize that other variations of access and presentation of video are possible where multiple devices are communicating with the server computer 102.
[0037] Referring to Figure 3c, a method for remotely controlling a device via a computer network is shown. The process of controlling the device and transmission of the video data is the same as illustrated in Figure 3a, with same reference numerals referring to same steps. At 19, the video data are stored in memory, for example, in memory 106 of the server computer 102 or in video database 103 connected to the server computer 102.
[0038] At 40, the processor 118A, 118B of the client computer 112A, 112B receives input data indicative of a request to provide the stored video data and generates data indicative thereof, for example, encoded in a client difference program and transmits (42) the same to the server computer 102. The request is, for example, indicative of video data captured by which camera and a start time of the video data. Optionally, the user is provided with a menu for the video data selection displayed on the user interaction interface 114A,
114B. Upon receipt, the processor 104 of the server computer 102 retrieves the stored video data and generates (44) a server difference program having encoded a difference between the stored state of the device and the stored state of the device last transmitted to the client computer 112A, 112B, i.e. a difference between consecutive image frames of the video data, and transmits (46) the same to the client computer 112A, 112B. Upon receipt, the processor 118A, 118B of the client computer 112A, 112B executes (48) the server difference program, updates the state of the device last received from the server computer accordingly and displays the same on display 114A, 114B in a human comprehensible fashion. Optionally, when transmitting video data to the client computer 112A, 112B generation and transmission of client difference programs are omitted, i.e. the video data are transmitted as a succession of server difference programs having encoded data indicative of a change in pixel values between consecutive image frames of the video data.
[0039] Alternatively, the camera control program and the server remote access program are integrated in one program for execution on the processor 104 of the server computer 102. Further alternatively, the client remote access program and the user interaction program are integrated into one program for execution on the processor 118A, 118B of the client computer 112A, 112B.
[0040] Further, as introduced above, the methods of Figures 3a-3c may enable remote control and/or video access to a plurality of client computers 112A, 112B, which are, for example, simultaneously, connected to the server computer 102 via the communication network 110. In such an implementation, each client computer may communicate a respective client difference program to the server computer 102 that represents changes made at the particular client computer. The client difference programs may be executed at the server computer 102 for determining control data to update a state of the device, followed by generating a second server difference program reflecting the updated state of the device. The plurality of client computers may, thus simultaneously and collaboratively control the device.
Alternatively, the plurality of client computers may be prioritized, as in Figure 3b. The second server difference program may be communicated to each of the plurality of client device to update the displays thereof. Thus, each connected client computer receives live or recorded video while control operations are being performed.
[0041] Implementation of the methods for remotely controlling a device via a computer network is not limited to the computer network architecture illustrated in Figure 1 and 2a to 2d. For example, the server remote access program may be executed on a processor of a first server computer while the camera control program may be executed on a processor of a second server computer connected thereto, for example, via a computer network such as the Internet. Further, the system 100 may be used to control multiple cameras in communication with multiple server computers. For example, each of the multiple server computers may be accessible at a different network address (e.g., a Uniform Resource Locator (URL)) to which the client access program man connect. Further, the video database may be connected to the first server computer and accessed by the server remote control program, leaving the second server computer and the camera control program for controlling the cameras and providing the captured video data. Still further, remote control may be effected by any type of client computer, i.e., mobile device, desktop computer, laptop/notebook computer, a tablet device, etc.
[0042] The present disclosure has been described herein with regard to several implementations. However, it will be obvious to persons skilled in the art that a number of variations and modifications can be made without departing from the scope of the disclosure as described herein.

Claims

WHAT IS CLAIMED IS:
1. A method of providing remote control of a device, comprising:
providing a remote connection to a server computer accessible by a client computer, the server computer being in communication with a control program that controls the device by communicating control data over a control communication link;
communicating, by the remote connection, a difference program containing data representing a change in a state of the device made by the client computer or the control program, the difference program capturing changes in the state of the device since the state of the device was last received by the server computing device or the client computing device; and
communicating, in accordance with information in the difference program, the control data to the device over the control communication link; and
synchronizing the state of the device using the difference program between the control program and the client computer.
2. The method of claim 1, wherein the device is a camera.
3. The method of claim 2, further comprising:
generating view data of image data captured by the camera; and
transmitting the view data to the client computer in accordance with hardware capabilities of the client computer.
4. The method of claim 3, wherein the hardware capabilities comprising one of processing capacity, memory size, type of graphical display, and type of user interface.
5. The method of any of claims 3-4, further comprising causing the client computer to display a predetermined portion of the state of the control program.
6. The method of any of claims 2-5, further comprising controlling functions of the camera using the control data.
7. The method of claim 6, where the functions of the camera comprise one of pan, tilt and zoom.
8. The method of any of claims 2-7, further comprising transmitting the image data as a video stream of successive image frames between a server remote access program executing on the server computer and a client remote access program executing on the client computer.
9. The method of any of claims 2-8, further comprising transmitting only image data related to pixel values that have been changed between successive image frames generated by the camera.
10. The method of any of claims 2-9, wherein plural client computers control functions of the camera, the method further comprising:
determining a prioritization of the plural client computers;
managing access to the camera in accordance with the prioritization; and
providing image data from the camera to each of the plural client computers during the control of the functions of the camera.
11. The method of any of claims 1-10, further comprising capturing changes in the state of the control program in the difference program since a last difference program was communicated by the client computer or the server computer.
12. The method of any of claims 1-11, further comprising: providing a server remote access program executing on a server computer, the server remote access program being in communication with the control program;
providing the remote connection to a client remote access program executing on a client computer;
communicating the difference program over the remote connection between the server remote access program and the client remote access program.
13. The method of any of claims 1-12, further comprising managing access of plural client computers that remotely control the device.
14. The method of claim 13, further comprising:
receiving a request from at least one of the plural client computers at the server computer; determining a priority of the plural client computers in accordance with information stored in a database; and
allowing a client computer having a predetermined priority to control the device.
15. The method of any of claims 13-14, further comprising providing a functionality of the device to the plural client computers while at least one of the plural client computers remotely controls the device.
16. A computer readable medium comprising computer executable instructions that when executed by a processor of a computing device perform a method of providing remote control of a device, comprising:
providing a remote connection to accessible by a client remote access program to enable remote access of the device using a control program; instantiating a server remote access program on a server computer that is in communication with the control program and the client remote access program;
communicating a difference program containing data representative of a state of the device over the remote connection;
communicating, in accordance with information in the difference program, control data over a control communication link to the device; and
synchronizing the state of the device using the difference program between the control program and a client computer.
17. The computer readable medium of claim 16, further comprising instructions for:
managing access of plural client computers that remotely control the device in accordance with a predetermined priority.
18. The computer readable medium of claim 17, further comprising instructions for providing a functionality of the device to the plural client computers while at least one of the plural client computers remotely controls the device.
19. The computer readable medium of claim 16, further comprising instructions for generating the control data from the data contained in the difference program representative of the state of the device.
20. A method of providing remote control of a camera, comprising:
providing a server remote access program on a server computer, the server remote access program being in communication with a camera control program that controls the camera via a control communication link; providing a remote connection to a client remote access program executing on a client computer, the remote connection enabling remote access to the control program;
communicating a difference program containing data representative of a state of the camera over the remote connection between the client computer and the server computer and between the server computer and the client computer;
communicating image data from the camera to the client computer;
communicating, in accordance with information in the difference program, control data to the camera to affect operation of the camera; and
synchronizing the state of the camera using the difference program between the control program and the client computer to enable concurrent control by the client computer and presentation of the image data on a display associated with the client computer.
21. The method of claim 20, further comprising managing access of plural client computers that remotely control the camera,
wherein a request from one of the plural client computers to control the camera is evaluated in accordance with prioritization information; and
wherein the image data is provided to each of the plural computers while the one of the plural computers controls the camera.
PCT/IB2012/000009 2011-01-04 2012-01-04 A method and system of controlling a remote controlled device in a remote controlled surgical procedure WO2012093330A1 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
CA2823615A CA2823615C (en) 2011-01-04 2012-01-04 A method and system of controlling a remote controlled device in a remote controlled surgical procedure
EP12731899.6A EP2661654A4 (en) 2011-01-04 2012-01-04 A method and system of controlling a remote controlled device in a remote controlled surgical procedure

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US201161429573P 2011-01-04 2011-01-04
US61/429,573 2011-01-04

Publications (1)

Publication Number Publication Date
WO2012093330A1 true WO2012093330A1 (en) 2012-07-12

Family

ID=46380437

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/IB2012/000009 WO2012093330A1 (en) 2011-01-04 2012-01-04 A method and system of controlling a remote controlled device in a remote controlled surgical procedure

Country Status (4)

Country Link
US (2) US9247120B2 (en)
EP (1) EP2661654A4 (en)
CA (1) CA2823615C (en)
WO (1) WO2012093330A1 (en)

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9294572B2 (en) 2011-11-11 2016-03-22 Calgary Scientific Inc. Session transfer and suspension in a remote access application framework
US9591100B2 (en) 2011-09-30 2017-03-07 Calgary Scientific Inc. Tiered framework for providing remote access to an application accessible at a uniform resource locator (URL)
US9648057B2 (en) 2011-11-23 2017-05-09 Calgary Scientific Inc. Methods and systems for collaborative remote application sharing and conferencing
US10334042B2 (en) 2008-11-26 2019-06-25 Calgary Scientific Inc. Method and system for providing remote access to a state of an application program
US10410306B1 (en) 2011-01-04 2019-09-10 Calgary Scientific Inc. Method and system for providing remote access to data for display on a mobile device
US10693940B2 (en) 2011-08-15 2020-06-23 Calgary Scientific Inc. Remote access to an application program
US11631496B2 (en) 2013-09-12 2023-04-18 Johnson & Johnson Surgical Vision, Inc. Computer-based operating room support system
US11715560B2 (en) * 2013-09-12 2023-08-01 Johnson & Johnson Surgical Vision, Inc. Computer-based operating room support system

Families Citing this family (32)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9631964B2 (en) 2011-03-11 2017-04-25 Intelligent Agricultural Solutions, Llc Acoustic material flow sensor
US10318138B2 (en) 2011-03-11 2019-06-11 Intelligent Agricultural Solutions Llc Harvesting machine capable of automatic adjustment
US9629308B2 (en) 2011-03-11 2017-04-25 Intelligent Agricultural Solutions, Llc Harvesting machine capable of automatic adjustment
US9474208B2 (en) 2011-11-15 2016-10-25 Appareo Systems, Llc System and method for determining material yield and/or loss from a harvesting machine using acoustic sensors
US9324197B2 (en) * 2011-03-11 2016-04-26 Intelligent Agricultural Soultions Method and system for managing the hand-off between control terminals
US9330062B2 (en) 2011-03-11 2016-05-03 Intelligent Agricultural Solutions, Llc Vehicle control and gateway module
US20160246296A1 (en) * 2011-03-11 2016-08-25 Intelligent Agricultural Solutions, Llc Gateway system and method
US10321624B2 (en) 2011-03-11 2019-06-18 Intelligent Agriculture Solutions LLC Air seeder manifold system
US20140118542A1 (en) * 2012-10-30 2014-05-01 Teleste Oyj Integration of Video Surveillance Systems
US9414004B2 (en) 2013-02-22 2016-08-09 The Directv Group, Inc. Method for combining voice signals to form a continuous conversation in performing a voice search
KR20150034061A (en) * 2013-09-25 2015-04-02 삼성전자주식회사 The method and apparatus for setting imaging environment by using signals received from a plurality of clients
EP2871871B1 (en) * 2013-11-12 2019-07-17 Telefonaktiebolaget LM Ericsson (publ) Remote socket connection
US9485420B2 (en) * 2014-02-03 2016-11-01 Point Grey Research Inc. Video imaging using plural virtual image capture devices
JP6305110B2 (en) * 2014-02-28 2018-04-04 キヤノン株式会社 Imaging apparatus and imaging system
US10412420B2 (en) * 2014-03-07 2019-09-10 Eagle Eye Networks, Inc. Content-driven surveillance image storage optimization apparatus and method of operation
US20160048366A1 (en) * 2014-08-14 2016-02-18 David John Cowin Video Streaming System and Method for a Wearable Display
US10085379B2 (en) 2014-09-12 2018-10-02 Appareo Systems, Llc Grain quality sensor
CN106470325A (en) * 2015-07-30 2017-03-01 中兴通讯股份有限公司 A kind of video front-end equipment and its control method
CN105187796B (en) * 2015-09-14 2019-06-21 北京奇虎科技有限公司 A kind of data transmission method, apparatus and system
KR102612815B1 (en) * 2016-10-24 2023-12-13 삼성전자주식회사 Display apparatus and control method thereof
CN107493460B (en) * 2017-09-21 2020-10-09 北京图森智途科技有限公司 Image acquisition method and system
CN110839139A (en) * 2018-08-17 2020-02-25 视联动力信息技术股份有限公司 Monitoring state sharing method and device
CN110661981B (en) * 2019-11-21 2021-10-15 深圳市圆周率软件科技有限责任公司 System for remotely controlling focusing system of camera module
CN113099104B (en) * 2020-01-09 2023-06-27 博泰车联网科技(上海)股份有限公司 Method for capturing an image, electronic device, computer storage medium and vehicle
US11368991B2 (en) 2020-06-16 2022-06-21 At&T Intellectual Property I, L.P. Facilitation of prioritization of accessibility of media
US11233979B2 (en) 2020-06-18 2022-01-25 At&T Intellectual Property I, L.P. Facilitation of collaborative monitoring of an event
US11037443B1 (en) 2020-06-26 2021-06-15 At&T Intellectual Property I, L.P. Facilitation of collaborative vehicle warnings
US11184517B1 (en) * 2020-06-26 2021-11-23 At&T Intellectual Property I, L.P. Facilitation of collaborative camera field of view mapping
US11411757B2 (en) 2020-06-26 2022-08-09 At&T Intellectual Property I, L.P. Facilitation of predictive assisted access to content
US11356349B2 (en) 2020-07-17 2022-06-07 At&T Intellectual Property I, L.P. Adaptive resource allocation to facilitate device mobility and management of uncertainty in communications
US11768082B2 (en) 2020-07-20 2023-09-26 At&T Intellectual Property I, L.P. Facilitation of predictive simulation of planned environment
CN114895613B (en) * 2022-07-12 2022-10-21 三一重型装备有限公司 Control method and device for mining equipment, storage medium and computer equipment

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5249121A (en) * 1989-10-27 1993-09-28 American Cyanamid Company Remote control console for surgical control system
EP1015068A1 (en) * 1996-12-12 2000-07-05 Intuitive Surgical, Inc. Multi-component telepresence system and method
US6602185B1 (en) * 1999-02-18 2003-08-05 Olympus Optical Co., Ltd. Remote surgery support system
US20050154288A1 (en) * 1996-06-24 2005-07-14 Computer Motion, Inc. Method and apparatus for accessing medical data over a network

Family Cites Families (85)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7079177B2 (en) * 1995-02-27 2006-07-18 Canon Kabushiki Kaisha Remote control system and access control method for information input apparatus with limitation by user for image access and camemremote control
GB2324175B (en) 1997-04-10 2002-07-31 Ibm Personal conferencing system
JP3667042B2 (en) * 1997-08-01 2005-07-06 キヤノン株式会社 Camera control system, control method thereof, camera client, and storage medium storing program for controlling camera client
US6342906B1 (en) 1999-02-02 2002-01-29 International Business Machines Corporation Annotation layer for synchronous collaboration
US7152092B2 (en) 1999-05-05 2006-12-19 Indeliq, Inc. Creating chat rooms with multiple roles for multiple participants
US7533146B1 (en) 1999-06-14 2009-05-12 Epiphany, Inc. Shared web browser apparatus and method for interactive communications
US6742015B1 (en) 1999-08-31 2004-05-25 Accenture Llp Base services patterns in a netcentric environment
US6698021B1 (en) * 1999-10-12 2004-02-24 Vigilos, Inc. System and method for remote control of surveillance devices
US7934251B2 (en) * 1999-12-02 2011-04-26 Western Digital Technologies, Inc. Managed peer-to-peer applications, systems and methods for distributed data access and storage
WO2001067245A2 (en) * 2000-03-08 2001-09-13 Marbles, Inc. Remote operation of real-time graphical applications
WO2001091482A1 (en) 2000-05-23 2001-11-29 Media Farm, Inc. Remote displays in mobile communication networks
US7299403B1 (en) 2000-10-11 2007-11-20 Cisco Technology, Inc. Methods and apparatus for obtaining a state of a browser
GB2368246B (en) 2000-10-17 2004-09-01 Hewlett Packard Co Overview subsystem for information page server
US20020051541A1 (en) 2000-10-30 2002-05-02 Glick Barry J. System and method for maintaining state between a client and server
JP3852742B2 (en) 2000-11-02 2006-12-06 インターナショナル・ビジネス・マシーンズ・コーポレーション Information processing system, terminal device, information processing method, and storage medium
US7133895B1 (en) 2001-02-20 2006-11-07 Siebel Systems, Inc. System and method of integrating collaboration systems with browser based application systems
US6981062B2 (en) 2001-04-20 2005-12-27 Sbc Technology Resources, Inc. World wide web content synchronization between wireless devices
GB0117077D0 (en) 2001-07-13 2001-09-05 Univ Strathclyde A generic object-based resource-sharing interface for distance co-working
US7191233B2 (en) 2001-09-17 2007-03-13 Telecommunication Systems, Inc. System for automated, mid-session, user-directed, device-to-device session transfer system
US7254634B1 (en) 2002-03-08 2007-08-07 Akamai Technologies, Inc. Managing web tier session state objects in a content delivery network (CDN)
US7346616B2 (en) * 2002-03-20 2008-03-18 Extended System, Inc. Synchronizing data shared between two devices independent of any other devices that may also share the data
US20030179230A1 (en) 2002-03-25 2003-09-25 Gerry Seidman Method and apparatus for providing remote peer-to-peer collaborative user interfaces
US7987491B2 (en) 2002-05-10 2011-07-26 Richard Reisman Method and apparatus for browsing using alternative linkbases
US7287054B2 (en) 2002-05-31 2007-10-23 Microsoft Corporation Systems and methods for shared browsing among a plurality of online co-users
US7356563B1 (en) 2002-06-06 2008-04-08 Microsoft Corporation Methods of annotating a collaborative application display
US7146616B2 (en) 2002-09-03 2006-12-05 Sap Aktiengesellschaft Content based messaging for e-learning including the extension of a remote procedure call
US7349831B2 (en) 2003-01-31 2008-03-25 Fmsm Engineers, Inc. River assessment, monitoring and design system
US7676506B2 (en) * 2003-06-20 2010-03-09 Innopath Software, Inc. Differential file compression of software image versions
US7363342B1 (en) 2003-07-08 2008-04-22 Microsoft Corporation Method and apparatus for providing web services in a collaborative computing system
US20070244930A1 (en) * 2003-07-18 2007-10-18 Bartlette Troy L System and method for utilizing profile information
US7401294B2 (en) 2003-11-24 2008-07-15 International Business Machines Corporation Method and system for collaborative web browsing
US7702730B2 (en) 2004-09-03 2010-04-20 Open Text Corporation Systems and methods for collaboration
JP4546202B2 (en) * 2004-09-28 2010-09-15 キヤノン株式会社 VIDEO RECEIVING DEVICE, ITS CONTROL METHOD, PROGRAM, AND STORAGE MEDIUM
US20060085245A1 (en) 2004-10-19 2006-04-20 Filenet Corporation Team collaboration system with business process management and records management
US7810089B2 (en) 2004-12-30 2010-10-05 Citrix Systems, Inc. Systems and methods for automatic installation and execution of a client-side acceleration program
JP2006222721A (en) * 2005-02-10 2006-08-24 Canon Inc Information processing apparatus, information processing method, and program thereof
US20110128378A1 (en) * 2005-03-16 2011-06-02 Reza Raji Modular Electronic Display Platform
US20070024706A1 (en) * 2005-08-01 2007-02-01 Brannon Robert H Jr Systems and methods for providing high-resolution regions-of-interest
US8280944B2 (en) 2005-10-20 2012-10-02 The Trustees Of Columbia University In The City Of New York Methods, media and systems for managing a distributed application running in a plurality of digital processing devices
US8024290B2 (en) * 2005-11-14 2011-09-20 Yahoo! Inc. Data synchronization and device handling
US7761591B2 (en) 2005-12-16 2010-07-20 Jean A. Graham Central work-product management system for coordinated collaboration with remote users
WO2007090235A1 (en) 2006-02-06 2007-08-16 Uiactive Ip Pty Ltd A system for conducting multi-media communication sessions
US8151323B2 (en) 2006-04-12 2012-04-03 Citrix Systems, Inc. Systems and methods for providing levels of access and action control via an SSL VPN appliance
US8073941B2 (en) * 2006-05-25 2011-12-06 AppToU Technologies Ltd. Method and system for providing remote access to applications
US20080052377A1 (en) 2006-07-11 2008-02-28 Robert Light Web-Based User-Dependent Customer Service Interaction with Co-Browsing
CA2657614C (en) 2006-07-19 2016-10-18 Mdatalink, Llc Method and system for remote review of clinical data
US7984115B2 (en) 2006-12-08 2011-07-19 Microsoft Corporation Extensible application platform
JP5439190B2 (en) 2007-01-16 2014-03-12 ギズモックス リミテッド Method and system for creating server-based web applications for IT
US20080320081A1 (en) 2007-06-19 2008-12-25 Microsoft Corporation Service componentization and composition architecture
US9239666B2 (en) 2007-09-12 2016-01-19 Citrix Systems, Inc. Methods and systems for maintaining desktop environments providing integrated access to remote and local resources
US8375126B2 (en) 2007-10-17 2013-02-12 Attachmate Corporation Methods, apparatus and techniques for suspending, resuming, and sharing sessions using object serialization
US8015496B1 (en) 2007-10-26 2011-09-06 Sesh, Inc. System and method for facilitating visual social communication through co-browsing
US8065166B2 (en) 2007-10-30 2011-11-22 Onemednet Corporation Methods, systems, and devices for managing medical images and records
US20090172100A1 (en) 2007-12-31 2009-07-02 International Business Machines Corporation Deriving and communicating attention spans in collaborative applications
US8099662B2 (en) 2008-01-17 2012-01-17 Seiko Epson Corporation Efficient image annotation display and transmission
US8909703B2 (en) 2008-01-24 2014-12-09 International Business Machines Corporation System and computer program product for facilitating a real-time virtual interaction
US8271887B2 (en) 2008-07-17 2012-09-18 The Boeing Company Systems and methods for whiteboard collaboration and annotation
US8239773B1 (en) * 2008-10-28 2012-08-07 United Services Automobile Association (Usaa) Systems and methods for co-browsing on a mobile device
BRPI0921978A2 (en) 2008-11-26 2019-01-15 Calgary Scient Inc method of providing remote access to application program state and storage media.
US20110119716A1 (en) * 2009-03-12 2011-05-19 Mist Technology Holdings, Inc. System and Method for Video Distribution Management with Mobile Services
US20100274858A1 (en) 2009-04-27 2010-10-28 Nokia Corporation Mid-service sharing
WO2011021909A2 (en) * 2009-08-21 2011-02-24 Samsung Electronics Co., Ltd. Method and apparatus for providing contents via network, method and apparatus for receiving contents via network, and method and apparatus for backing up data via network, backup data providing device, and backup system
US8675084B2 (en) * 2009-09-04 2014-03-18 Apple Inc. Systems and methods for remote camera control
US8290920B2 (en) * 2009-09-30 2012-10-16 Zynga Inc. System and method for remote updates
US8904343B2 (en) 2009-12-21 2014-12-02 Sap Se Adding services to application platform via extension
US8195146B2 (en) 2009-12-23 2012-06-05 Intel Corporation Remote management over a wireless wide-area network using short message service
US8531414B2 (en) 2010-02-03 2013-09-10 Bump Technologies, Inc. Bump suppression
US8577292B2 (en) 2010-02-03 2013-11-05 Google Inc. Bump validation
US9065532B2 (en) 2010-02-03 2015-06-23 Google Inc. Bump button
US8527591B2 (en) 2010-05-20 2013-09-03 Actual Works, Inc. Method and apparatus for the implementation of a real-time, sharable browsing experience on a guest device
US20120233555A1 (en) 2010-11-08 2012-09-13 Eyelead Sa Real-time multi-user collaborative editing in 3d authoring system
US20120210243A1 (en) 2011-02-11 2012-08-16 Gavin Andrew Ross Uhma Web co-navigation
US9367635B2 (en) 2011-02-12 2016-06-14 International Business Machines Corporation Contact center co-browsing for a mobile device
US8856259B2 (en) 2011-05-20 2014-10-07 International Business Machines Corporation Multi-domain co-browsing utilizing localized state management
US9323871B2 (en) 2011-06-27 2016-04-26 Trimble Navigation Limited Collaborative development of a model on a network
AU2012296247B2 (en) 2011-08-15 2017-06-22 Calgary Scientific Inc. Non-invasive remote access to an application program
US8935328B2 (en) 2011-09-15 2015-01-13 Ramakrishna J Tumuluri System and method for collaborative 3D visualization and real-time interaction on a computer network
JP6322140B2 (en) 2011-09-30 2018-05-09 カルガリー サイエンティフィック インコーポレイテッド Unconnected application extension including interactive digital surface layer for sharing and annotation of collaborative remote applications
CA2850425A1 (en) 2011-09-30 2013-04-04 Calgary Scientific Inc. Coupled application extensions for collaborative remote application sharing
US9294572B2 (en) 2011-11-11 2016-03-22 Calgary Scientific Inc. Session transfer and suspension in a remote access application framework
EP2780826B1 (en) 2011-11-15 2020-08-12 Trimble Inc. Browser-based collaborative development of a 3d model
WO2013076554A1 (en) 2011-11-23 2013-05-30 Calgary Scientific Inc. Methods ans systems for collaborative remote application sharing and conferencing
US9098592B2 (en) 2011-12-14 2015-08-04 International Business Machines Corporation Session completion through co-browsing
US8553965B2 (en) 2012-02-14 2013-10-08 TerraRecon, Inc. Cloud-based medical image processing system with anonymous data upload and download
WO2014033554A2 (en) 2012-08-15 2014-03-06 Calgary Scientific Inc. Methods and systems for collaborative browsing

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5249121A (en) * 1989-10-27 1993-09-28 American Cyanamid Company Remote control console for surgical control system
US20050154288A1 (en) * 1996-06-24 2005-07-14 Computer Motion, Inc. Method and apparatus for accessing medical data over a network
EP1015068A1 (en) * 1996-12-12 2000-07-05 Intuitive Surgical, Inc. Multi-component telepresence system and method
US6602185B1 (en) * 1999-02-18 2003-08-05 Olympus Optical Co., Ltd. Remote surgery support system

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
See also references of EP2661654A4 *

Cited By (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10334042B2 (en) 2008-11-26 2019-06-25 Calgary Scientific Inc. Method and system for providing remote access to a state of an application program
US10965745B2 (en) 2008-11-26 2021-03-30 Calgary Scientific Inc. Method and system for providing remote access to a state of an application program
US10410306B1 (en) 2011-01-04 2019-09-10 Calgary Scientific Inc. Method and system for providing remote access to data for display on a mobile device
US10693940B2 (en) 2011-08-15 2020-06-23 Calgary Scientific Inc. Remote access to an application program
US10284688B2 (en) 2011-09-30 2019-05-07 Calgary Scientific Inc. Tiered framework for proving remote access to an application accessible at a uniform resource locator (URL)
US9596320B2 (en) 2011-09-30 2017-03-14 Calgary Scientific Inc. Uncoupled application extensions including interactive digital surface layer for collaborative remote application sharing and annotating
US10904363B2 (en) 2011-09-30 2021-01-26 Calgary Scientific Inc. Tiered framework for proving remote access to an application accessible at a uniform resource locator (URL)
US9591100B2 (en) 2011-09-30 2017-03-07 Calgary Scientific Inc. Tiered framework for providing remote access to an application accessible at a uniform resource locator (URL)
US10083056B2 (en) 2011-11-11 2018-09-25 Calgary Scientific Inc. Session transfer and suspension in a remote access application framework
US9294572B2 (en) 2011-11-11 2016-03-22 Calgary Scientific Inc. Session transfer and suspension in a remote access application framework
US9648057B2 (en) 2011-11-23 2017-05-09 Calgary Scientific Inc. Methods and systems for collaborative remote application sharing and conferencing
US10454979B2 (en) 2011-11-23 2019-10-22 Calgary Scientific Inc. Methods and systems for collaborative remote application sharing and conferencing
US11631496B2 (en) 2013-09-12 2023-04-18 Johnson & Johnson Surgical Vision, Inc. Computer-based operating room support system
US11715560B2 (en) * 2013-09-12 2023-08-01 Johnson & Johnson Surgical Vision, Inc. Computer-based operating room support system

Also Published As

Publication number Publication date
CA2823615C (en) 2017-08-29
EP2661654A4 (en) 2014-07-09
US9247120B2 (en) 2016-01-26
US20160198081A1 (en) 2016-07-07
CA2823615A1 (en) 2012-07-12
US20120169874A1 (en) 2012-07-05
EP2661654A1 (en) 2013-11-13

Similar Documents

Publication Publication Date Title
CA2823615C (en) A method and system of controlling a remote controlled device in a remote controlled surgical procedure
ES2659346T3 (en) Video Intermediary Server for Video Surveillance
US20130219006A1 (en) Multiple media devices through a gateway server or services to access cloud computing service storage
ES2305980T3 (en) SYSTEMS AND PROCEDURE FOR THE NETWORK CONNECTION OF DIGITAL VIDEO RECORDING EQUIPMENT.
US20120084350A1 (en) Adaptive distributed medical image viewing and manipulating systems
US10645459B2 (en) Devices, systems, and methods for remote video retrieval
RU2518194C2 (en) Distributed node video surveillance system and control method thereof
US10162936B2 (en) Secure real-time healthcare information streaming
JP4774921B2 (en) File display method and system
US20130111353A1 (en) Medical coordination system
JP2013243544A (en) Imaging device, mask image setting method, and program
EP2452479A2 (en) Selectively distributing updates of changing images to client devices
US8504726B2 (en) User targeted medical imaging and information packaging, compression and distribution system
US10678569B2 (en) Enhanced virtual desktop infrastructures
US20240185698A1 (en) Devices, systems, and methods for video retrieval
JPWO2014112416A1 (en) Video supply device, video acquisition device, and program
WO2014085918A1 (en) System and method of viewing digital medical images
KR100792240B1 (en) Multi vision materialization method and system for the same
KR101116230B1 (en) Method for providing medical information to mobile terminals and computer readable recording media recording the same
KR20110137194A (en) System and method for providing medical information
US20090164562A1 (en) Contents recording and reproducing apparatus
JP2015222931A (en) Image processing device and external device
KR101492687B1 (en) Method for providing n-screen service using virtual aggregation gateway
JP2017212644A (en) Communication device, system, information processing method, and program
Inamdar et al. A Web Architecture for E-Health Applications Supporting the Efficient Multipath Transport of Medical Images

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 12731899

Country of ref document: EP

Kind code of ref document: A1

ENP Entry into the national phase

Ref document number: 2823615

Country of ref document: CA

WWE Wipo information: entry into national phase

Ref document number: 2012731899

Country of ref document: EP

NENP Non-entry into the national phase

Ref country code: DE