WO2005008431A2 - Software development kit for client server applications - Google Patents

Software development kit for client server applications Download PDF

Info

Publication number
WO2005008431A2
WO2005008431A2 PCT/US2004/022568 US2004022568W WO2005008431A2 WO 2005008431 A2 WO2005008431 A2 WO 2005008431A2 US 2004022568 W US2004022568 W US 2004022568W WO 2005008431 A2 WO2005008431 A2 WO 2005008431A2
Authority
WO
WIPO (PCT)
Prior art keywords
data
processes
computer systems
threads
object classes
Prior art date
Application number
PCT/US2004/022568
Other languages
French (fr)
Other versions
WO2005008431A3 (en
Inventor
Bradford C. Davis
Original Assignee
Computer Associates Think, Inc.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Computer Associates Think, Inc. filed Critical Computer Associates Think, Inc.
Priority to EP04778205A priority Critical patent/EP1652039A4/en
Publication of WO2005008431A2 publication Critical patent/WO2005008431A2/en
Publication of WO2005008431A3 publication Critical patent/WO2005008431A3/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/46Multiprogramming arrangements
    • G06F9/54Interprogram communication
    • G06F9/546Message passing systems or structures, e.g. queues
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F8/00Arrangements for software engineering
    • G06F8/20Software design
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/46Multiprogramming arrangements
    • G06F9/54Interprogram communication
    • G06F9/542Event management; Broadcasting; Multicasting; Notifications
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L49/00Packet switching elements
    • H04L49/90Buffering arrangements
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/01Protocols
    • H04L67/10Protocols in which an application is distributed across nodes in the network
    • H04L67/104Peer-to-peer [P2P] networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2209/00Indexing scheme relating to G06F9/00
    • G06F2209/54Indexing scheme relating to G06F9/54
    • G06F2209/548Queue

Definitions

  • the system in one aspect may include a plurality of public interfaces operable to move data among a plurality of computer systems.
  • At least some of the plurality of computer systems are in a heterogeneous networked environment and at least some other of the plurality of computer systems are in a homogeneous environment .
  • the plurality of public interfaces are operable to move data between threads, to move data between processes on the same computer, to move data between processes across a network between like operating systems, and/or to move data between processes across a network where the processes are running on different operating systems.
  • the system further may include object classes that operate in an event-driven manner for moving data between threads.
  • the object classes further may allow the processes to be aware of one another and allow transfer device between those processes.
  • the Object classes may be operable to reconnect after loss of connection, and to address sequential integrity of data after reconnection is established.
  • the object classes may further be operable to mutually exclude data being accessed by a processor.
  • the method in one aspect may include providing a plurality of public interfaces for moving data among a plurality of computer systems, at least some of the plurality of computer systems being in a heterogeneous networked environment and at least some other of the plurality of computer systems being in a homogeneous environment.
  • the method may also include enabling the plurality of public interfaces to move data between threads, to move data between processes on the same computer, to move data between processes across a network between like operating systems, and/or to move data between processes across a network where the processes are running on different operating systems.
  • the method may also include providing object classes that operate in an event-driven manner for moving data between threads.
  • Fig. 1 is a flow diagram illustrating a method of the present disclosure in one embodiment.
  • Fig. 2 is a block diagram illustrating components for moving arbitrary objects between threads in one embodiment of the present disclosure.
  • Fig. 3 is a block diagram illustrating components for moving arbitrary objects between processes on the same node in one embodiment of the present disclosure.
  • the method and system of the present disclosure in one embodiment provides software development kit ("sdk”) that enables development of efficient and scalable client server applications in heterogeneous networked environments as well as homogeneous environments and within local machines.
  • the sdk utilizes standard protocols of TCP/IP, Berkeley Sockets, POXIS Threads, and object oriented methodology to provide a set of public interfaces for moving data among computer systems, including moving data between threads, moving data between processes on the same node or computer, moving data between processes across a network between like operating systems, and moving data between processes across a network where the processes are running on different operating systems.
  • the sdk of the present disclosure comprises object classes that operate in an event driven manner for moving data between threads .
  • the sdk' s object classes allow processes to be aware of one another and allow transfer mechanism between those processes. Further, the processes through the sdk' s object classes are enabled to reconnect in case of loss of connection, and address sequential integrity of the data issues when the connection is reestablished.
  • Fig. 1 is a flow diagram illustrating a method of the present disclosure in one embodiment.
  • public interfaces are provided to allow moving data among a plurality of computer systems. The computer systems may be in heterogeneous as well as homogeneous environment .
  • object classes are provided for moving data between threads.
  • object classes are provided for moving data between processes on the same computer.
  • object classes are provided for moving data between processes across a network between like operating systems.
  • object classes are provided to move data between processes across a network where the processes are running on different operating systems.
  • the moving of data is made event-driven.
  • objects are moved between threads in an event driven manner. In the event driven manner, the arrival of the event may be realized in the destination thread at a point where the destination thread can wait while consuming the minimal CPU possible until the object is available. This is the notion of 'event driven' versus a polling algorithm where the destination thread continuously queries the source supplying the object for its readiness.
  • the system and method of the present application uses the application space on top of the known TCP/IP and socket layers.
  • the socket layer is shared between the application and the operating system. Above the socket layer is a space referred to herein as the application space. It is in this application space where the present application may run.
  • the present application in one embodiment employs a canonical client server model, which will be described below.
  • the well-known POSIX thread primitives may be used for UnixTM and LinuxTM based implementations.
  • the pthread primitives are encapsulated into a set of objects and their use has been made easier and robust.
  • the MS thread primitives are used and these are interfaced via the POSIX interface to localize the porting problem.
  • the system and method of the present application uses a number of other objects that make no reliance on POSIX, MicrosoftTM, BerkeleyTM or any other existing standards.
  • the components of the present application will now be described in more detail.
  • the components are packaged and referred to as "Libmsg.”
  • Libmsg For moving arbitrary object between threads, the following is provided.
  • Fig. 2 is a block diagram illustrating components for moving arbitrary objects between threads.
  • Mutexing means to protect an area of data that two or more processors may vie for simultaneously at runtime. While two or more processes may read the same data at the same time, they cannot be allowed to write to that address at the same time. If the application lets that happen the program may generally fault, that is, die, crash, cease to run.
  • the pthread library provides a number of primitives to accomplish mutexing, however, they are written in C and are not encapsulated. As such, the application programmer need to manage a number of variables to accomplish routine things that appear repeatedly in any application that deals with multi-threading. This makes writing threaded applications difficult and error prone.
  • Encapsulation is a fundamental notion in the Object Oriented ("00") paradigm and one of its goals is to create a simple public interface which exposes only what is needed by the user of the object.
  • the system and method of the present disclosure utilizes encapsulation.
  • the following class is provided for encapsulating the mutex functionality.
  • the msgMutex Class (Fig. 2 202) may include: Lock - attain a mutual exclusion Lock or get in line and wait if the lock is already held. TryLock - return rather than block if the lock is already held. Unlock - relinquish the lock.
  • ConditionedTimedWait wait this duration in milliseconds or until signaled.
  • ConditionedWait wait until signaled.
  • ClearPredicate clear the predicate condition set by the application that is a precondition to ConditionTimedWait or a ConditionedWait . Signal either wait method at which point they will unblock. Broadcast a signal to a number of *Wait methods simultaneously.
  • the system and method of the present application uses asynchronous decomposition, which means taking the problem from the outset and segregating into its natural asynchronous components or threads. If a given loop breaks a rule, the iteration is too complex and needs to be broken down further.
  • the Protectedlnt Class may include: IsSetMethod interrogates a mutexed integer for a set clear state. Set the integer. Clear the integer. Increment the integer. Decrement the integer. Test if the integer IsPositive and return 1 or 0 accordingly. Wait or block in the iteration until signaled. Signal the Wait method so that it will unblock. Return the Address of the integer. Evaluate and return the value of the integer.
  • the second Protected class deals with addresses that may need to be modified and interrogated across threads.
  • the ProtectedPointer Class (Fig. 2 206) may include: Set the address . Evaluate the address .
  • the msgQueue Class 208 operates on a contiguous chunk of memory that is available to one or more threads for reading and writing. Data is stored in a protection fashion and individual elements or blocks of data can be inserted or deleted from any thread that has access to the msgQueue. The contiguous block of data is resized automatically as needed, up to a value in megabytes (default 8MB) and this default can be reset at runtime. This class achieves the asynchronous decomposition. This class allows data to be moved in an event driven data manner between threads . Enqueue - put an obj ect at the end of the queue .
  • Dequeue - take an object off the head of the queue, block (wait) if empty.
  • Readqueue - peek the value at the head of the queue .
  • Rmelem - take the head element off the queue.
  • ReadBlock - peek a block from the head of the queue, block (wait) if empty.
  • RmBlock - take the bock from the head of the queue .
  • Unblock - unblock an empty queue .
  • Block - put this queue in a blocked state.
  • Entries - return the number of elements in the queue .
  • GetHeadPointer return the address of the head element . IsEmpty - return 1 empty, return 0 otherwise.
  • the msgQueue methods that introduce data into a thread are Dequeue, Readqueue, ReadBlock and GetHeadPointer.
  • data in the queue is not shared between the queue and the calling routine.
  • the queue makes a deep copy of that object from the address passed to it by the caller on insertion and likewise loads a copy of queue owned element into the address passed by the caller on retrieval.
  • these methods block when the queue is empty. Generally placed at the top of ' the iteration, they idle a particular iteration when no data is available.
  • the POSIX and MSTM primitives are used in the system and method of the present application such that the scheduler does not bring into context a thread blocked because no data is available .
  • asynchronous decomposition does not restrict the number of destinations that can be targeted from a single thread. Any number of threads or processes may be dispatched from any number of points in a thread. Queues are routinely arrayed and dispatching a set of targets is done through an indexing scheme, where the targets are related to the indices of the array.
  • a PersistentStore class 210 is provided that shares the same interface as the msgQueue and achieves the same functionality.
  • PersistentStore class 210 all elements are written and read from disk and are retrievable for subsequent runtime instances of the application. At the instantiation of this object, an identifier of path and filename are provided which locates the object, a flat binary file. This preserves data across runtime instances whereas data in the msgQueue is lost. To understand the tradeoff in real terms, consider the TCP/IP model that assumes data in transport as volatile, but acknowledges receipt of data back to the sender. In this model, only the endpoints need to employ the PersistentStore and intermediate processes involved in the transport of data would use the msgQueue.
  • the msgList class 212 is similar to the msgQueue 208 in that it is used to move data between threads. It differs from the msgQueue 208 in that the notion of head and tail do not restrict the access to the list.
  • msgList class 212 Add - put an object on the list if object is unique. Remove - object from list by matching the value of argument to key of element in list. RemoveEle based on the address. RegisterKey - register the location of this element in the object as a searchable key. FindByKey - Find an object on the list based on the value of predetermined key element . IsKeyRegistered - return one if true; zero otherwise . IsEmpty - return one if true; zero otherwise.
  • arbitrary objects may be moved between threads in an event driven fashion, in an application that has been decomposed into asynchronous components which then scales well, is intuitive and unencumbered by callbacks and signal handlers .
  • the canonical socket model using the AF_UNIX family of sockets is contemplated.
  • Fig. 3 is a block diagram illustrating components for moving arbitrary objects between processes on the same node.
  • the AF_UNIX family of sockets is used to move data between processes on the same machine, versus the AF_INET family that moves them across the network using TCP/IP, which is described below. This is very efficient means of data transfer between processes on the same node with all the robustness of Berkeley sockets.
  • the implementation at the socket level is done via shared memory but the application is shielded from those details .
  • the canonical socket model distinguishes a client and server.
  • servers listen for and accept clients on a known port made public to the clients and clients connect to the server. At the point a connection is established client and servers can talk to one another by means of send and recv. Send and recv are two examples of several calls, which can be used to achieve the sending and receiving of data.
  • the classes described above may be used to transfer data between processes .
  • the initial handshake of both the server and client are cleanly encapsulated and absolve the user of any knowledge of the socket model .
  • the application developer need only know if he is a sender or receiver of data. If a sender is needed, a Client object is used. If an application is to receive data, a Server object is used. If an application wants to send and receive data, then both a Client and a Server object are used.
  • These objects are detailed below with a number of intermediate objects that is developed to build the Client and Server. To instantiate the Client object, only the Server host and port need be known to the application. Similarly, the only information that the application supplies to the Server object is the port on which the Server is listening.
  • the system and method of the present disclosure provides guaranteed message delivery (GMD) .
  • GMD guaranteed message delivery
  • Data is sequenced and held until acknowledged and received in msgQueues . Since data is held in a msgQueue until acknowledged, data is not lost regardless of the state of the network or an outright loss due to catastrophic failure in the network infrastructure.
  • the system and method of the present disclosure provides automatic recovery. If the application containing the Server goes away, the Client will reconnect automatically when the Server object becomes available in a new process. Processes containing Client objects can reconnect as well with the destruction and re-instantiation of the Client object.
  • Client side reconnection is not restricted to object destruction and re-instantiation.
  • client and server are full peers. This means that automatic recovery happens from either side. If a Server is lost, the Clients quickly and automatically reconnect when the Server comes back. If a Client is lost, it can be brought back up and it will reconnect with the Server.
  • the system and method described provides fast transport . Data in msgQueues are evacuated in blocks up to 64K in size and uploaded on the receive side in similar blocks in a single function call. This takes advantage of the maximum 'TCP window' and reduces the overhead the layers under the application (socket and TCP/IP) need to move the application data.
  • the msgSocket Class 302 is private to the SDK and though integral to workings of the public classes of Libmsg, it is not part of the public interface. Users of the SDK do not need to understand it, nor do they need to be aware of its existence. Note the two constructors .
  • msgSocket - server constructor calls the following methods . CreateSocket - create and initialize the socket. Bind - encapsulates the Berkeley "bind" function. GetSockName - encapsulates the Berkeley "getsockname” function. Listen - encapsulates the Berkeley "listen” function.
  • msgSocket - client constructor calls the following methods . CreateSocket - create and initialize the socket. Connect - encapsulates the Berkeley "connect” function.
  • the remaining methods of the msgSocket are private methods, which are not called from either constructor.
  • IsSckValid - is the object in good condition. Accept - encapsulates the Berkeley "accept” function and is called only from the Server. Send - encapsulates the Berkeley "send” function called from both Client and Server. Receive - encapsulate the Berkeley "recv” function. CloseSocket - encapsulates the system "close” function.
  • the GetServerAddress Class 304 is private to the SDK and though integral to workings of the public classes of Libmsg, it is not part of the public interface. Users of the SDK do not need to understand it, nor do they need to be aware of its existence.
  • the GetServerAddress Class 304 accesses the Berkeley "gethostbyname, " and populates the "hostent" structure. It is pulled from the msgSocket because it need not be called at the same frequency of the msgSocket .
  • Gethostbyname is implemented as a thread, which is monitored with a msgMutex: : ConditionedTimedWait , since the Berkeley call is known to hang. IsValid - returns the condition of this object as 1/0.
  • GetHostByName - encapsulates, threads and times the potentially blocking Berkeley "gethostbyname.”
  • the Server and Client objects are two pubic classes used to move objects between processes.
  • the objects at this level do not care whether those objects are to move between two processes on the same machine, or between two processes on machines that are located on opposite sides of the planet .
  • the mechanics of these objects is the same, regardless.
  • the Server Class 306 owns several subordinate private objects.
  • the Server constructor creates a limited Client object solely for the purpose of shutting down.
  • the Berkeley "accept” is generally in a blocked state and a signal is not raised against it to unblock it in the event a shutdown is ordered. Therefore, to unblock the "accept” and allow the shutdown to proceed, a msgSocket :: Connect is called.
  • the Server starts the AcceptClientsThread that starts a RecvThread for every Client object that seeks to connect with the Server. RecvThreads handle the incoming data and come and go for a variety of reasons.
  • the AcceptClientsThread persists for the life of the object.
  • Recv - called from application with a pointer to an address at which data will be copied, blocks if no data is available.
  • Recv_NoBlock same as Recv but does not block in the event no data is available. Shutdown - shutdown the various threads which may exist and destroy this object.
  • RecvService - do the msgSocket : :Accept handle receive threads .
  • msgQueue - store data.
  • Peerld enable the Server to keep track of message sequences from incoming Clients.
  • msgThreadlndex enable the Server to keep track of simultaneous RecvThreads.
  • RecvThread - The work of the Server object is done in the RecvThread. there is a RecvThread for every incoming Client object that is sending data.
  • Clients that quit sending data are timed out in accordance with a variable set in the environment. Clients can also set an explicit "breakSig" that will terminate the RecvThread.
  • the RecvThread calls msgSocket ::Receive, tests for the sequential integrity of the data, sends the acknowledgement to the appropriate Client and puts the - data on the msgQueue .
  • the Peerld Class The Peerld 308 is private to the SDK and though integral to workings of the public classes of Libmsg, it is not part of the public interface. Users of the SDK do not need to understand it, nor do they need to be aware of its existence. This object exists solely to support the Server object. Clients are identified by information in the headers of their incoming messages and the Peerld 308 is accessed to track the sequential integrity of the message from that Client. The sequential integrity of messages is the central component to the guaranteed message delivery and will be discussed further in the Client. But sequential integrity is important to the Server as well . Clients hold messages in their respective msgQueues until they receive an acknowledgement.
  • the Server sends the acknowledgement on the condition that the message meets security protocol and the message is deemed in good condition. However, it is possible that the Client may have resent a message because it failed to receive an acknowledgement on a previously sent message. In other words, the Server received the message but the Client did not receive the acknowledgement . The Server is able to defend against this, so that it does not enqueue the same message twice.
  • the Peerld 308 owns a msgList and a msgMutex object.
  • the msgList stores an address and pid (process id) pair and the mutex is used internally to secure access to the object since more than a single RecvThread may generally contend for the Peerld.
  • the msgThreadlndex Class 310 is private to the SDK and though integral to workings of the public classes of Libmsg, it is not part of the public interface. Users of the SDK do not need to understand it, nor do they need to be aware of its existence.
  • This object manages the indexing the array of RecvThreads that are currently in existence within the Server. GetNextFreeIndex - get the next available (not in use) index counting from zero. GetThreadlndex - get the index which was assigned to this thread. SetThreadlndex - set this threads index to a value. GetLastlndex - get the last index in use.
  • the Client Class The Client 312 is primary counterpart to the Server object defined above. Note the distinction between public and private objects and methods. Only the public methods are visible to users of the SDK. The private components are listed to enable a broader understanding of the Client object. public methods: IsValid - test the condition of the object return 1/0. Send - application interface to send. SendWaitForAck - application interface to send and wait for acknowledgement . NinQ - return the number of elements in this msgQueue .
  • GetServerAddress get Server hostent struct information.
  • msgSocket do the necessary client side socket work.
  • the msgSocket is constructed and destructed in the Client object for every reconnection.
  • the Server terminates connections after a short duration of no incoming data.
  • Sends from the Client which follow a delay exceeding that timeout fail, despite the preexisting connection.
  • the cost of a failed send is sub- millisecond and a new connection can be re-established between nodes on opposite coasts in less than half a second (and within 20 milliseconds on nodes in the same segment) . Since data is safe in the msgQueue, this is a very efficient handshake mechanism that can be kept completely out of the application.
  • This mechanism provides automatic recovery in the Client.
  • msgQueue (2) one for outbound msgs and one for acks .
  • ProtectedPointer protected address access between threads .
  • Protectedlnt protected integer access between threads .
  • msgMutex (2) general protection for areas of contention.
  • RecycleThread recycle the ClientSendThread if it gets dropped.
  • ClientSendThread does the work of the of the Client object. Peek the head element on the msgQueue, send it, receive the ack, remove the element
  • the msgBase Class An application developer can define a message object in any way desired. In one embodiment, that object is derived from msgBase 314, which puts a 48-byte header at the front of the user's object. The Client and Server objects interrogate this header, but the application need not look at it. The application developer may need to be aware that their object does derive from msgBase 314 as well as the fact that their data starts + 48 bytes from the point of the object they have created from the msgBase class 314. MsgBase 314 also provide methods that serve as good examples in how to deal with the endian problems that integer transport invites . In one embodiment, none of the msgBase methods are public .
  • ToNetBaseAlignment - called from Client (send) side .
  • ToHostBaseAlignment - called from Server (recv) side .
  • SetElemsThisBlock number of elements sent in this packet, maximize to 64K.
  • the above-described class components allow implementing an event driven applications that communicate between processes on the same node with capability for GMD, automatic recovery, and fast data transport .
  • For moving data between processes on different nodes for example, connected by Ethernet, routers, TCP/IP, cable, fiber, etc., Berkeley sockets and TCP/IP layers may be utilized.
  • the system and method of the present disclosure provides a set of header files and static library or a shared object that compiles and links in particular environment.
  • the public SDK interface is identical across the Unix variants, Linux, the Microsoft platforms, VMS, as/400 and MVS . Porting issues are minimized in large part to the industry's use of the POSIX Draft 10 of pthreads and the universally accepted Berkeley sockets.
  • Libmsg described above, event driven applications that communicate between processes on different nodes running a full variety of operating systems from Solaris 2.5.1, OSF 4.
  • Od, AIX 4.3.2, HP 11, Linux 7.2, full range of Microsoft, VMS 6.2, as/400 may be implemented.
  • the system and method of the present disclosure may be implemented and run on a general-purpose computer.
  • the embodiments described above are illustrative examples and it should not be construed that the present invention is limited to these particular embodiments .
  • particular classes were provided as examples, it should be understood that the method and system disclosed in the present application may be implemented using different programming methodologies. Thus, various changes and modifications may be effected by one skilled in the art without departing from the spirit or scope of the invention as defined in the appended claims .

Landscapes

  • Engineering & Computer Science (AREA)
  • Software Systems (AREA)
  • Theoretical Computer Science (AREA)
  • General Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Signal Processing (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Multimedia (AREA)
  • Computer And Data Communications (AREA)
  • Synchronisation In Digital Transmission Systems (AREA)
  • Electric Clocks (AREA)
  • Data Exchanges In Wide-Area Networks (AREA)
  • Multi Processors (AREA)

Abstract

At (102) public interfaces are provided to allow moving data among a pulrality of computer systems. The computer systems may be in heterogeneous as well as homogeneous environment. At (104), object classes are provided for moving data between threads. At (106), object classes are provided for moving data between processes on the same computer. At (108), object classes are provided for moving data between processes across a network between like operating system. At (110), object classes are provided for moving data between processes across a network where the processes are running on different operating system. At (112), the moving of data is made event-driven.

Description

SOFTWARE DEVELOPMENT KIT FOR CLIENT SERVER APPLICATIONS
CROSS-REFERENCE TO RELATED APPLICATIONS This application claims the benefit of U.S. Provisional Patent Application No. 60/486,596 entitled SYSTEM AND METHOD FOR STANDARDIZING CLOCKS IN A HETEROGENEOUS NETWORKED ENVIRONMENT filed on July 11, 2003, the entire disclosure of which is incorporated herein by reference. TECHNICAL FIELD This application relates to software development kit, and particularly to software development kit that enables development of efficient and scalable client server applications in heterogeneous networked environment .
BACKGROUND Standard protocols such as TCP/IP, Berkely Sockets, POSIX Threads have been used to move data between computers. There is, however, a lack of a unified and scalable methodology that enables a developer or a user to easily and efficiently implement such functionalities in heterogenous networked environment as well as homogenous environment . SUMMARY System and method for development of client server applications for heterogeneous networked environments and homogeneous environments are provided. The system in one aspect may include a plurality of public interfaces operable to move data among a plurality of computer systems. At least some of the plurality of computer systems are in a heterogeneous networked environment and at least some other of the plurality of computer systems are in a homogeneous environment . The plurality of public interfaces are operable to move data between threads, to move data between processes on the same computer, to move data between processes across a network between like operating systems, and/or to move data between processes across a network where the processes are running on different operating systems. In one aspect, the system further may include object classes that operate in an event-driven manner for moving data between threads. The object classes further may allow the processes to be aware of one another and allow transfer device between those processes. The Object classes may be operable to reconnect after loss of connection, and to address sequential integrity of data after reconnection is established. The object classes may further be operable to mutually exclude data being accessed by a processor. , The method in one aspect may include providing a plurality of public interfaces for moving data among a plurality of computer systems, at least some of the plurality of computer systems being in a heterogeneous networked environment and at least some other of the plurality of computer systems being in a homogeneous environment. The method may also include enabling the plurality of public interfaces to move data between threads, to move data between processes on the same computer, to move data between processes across a network between like operating systems, and/or to move data between processes across a network where the processes are running on different operating systems. The method may also include providing object classes that operate in an event-driven manner for moving data between threads. Further features as well as the structure and operation of various embodiments are described in detail below with reference to the accompanying drawings . In the drawings, like reference numbers indicate identical or functionally similar elements.
BRIEF DESCRIPTION OF THE DRAWINGS Fig. 1 is a flow diagram illustrating a method of the present disclosure in one embodiment. Fig. 2 is a block diagram illustrating components for moving arbitrary objects between threads in one embodiment of the present disclosure. Fig. 3 is a block diagram illustrating components for moving arbitrary objects between processes on the same node in one embodiment of the present disclosure.
DETAILED DESCRIPTION The method and system of the present disclosure in one embodiment provides software development kit ("sdk") that enables development of efficient and scalable client server applications in heterogeneous networked environments as well as homogeneous environments and within local machines. In one aspect, the sdk utilizes standard protocols of TCP/IP, Berkeley Sockets, POXIS Threads, and object oriented methodology to provide a set of public interfaces for moving data among computer systems, including moving data between threads, moving data between processes on the same node or computer, moving data between processes across a network between like operating systems, and moving data between processes across a network where the processes are running on different operating systems. In one embodiment, the sdk of the present disclosure comprises object classes that operate in an event driven manner for moving data between threads . For moving objects between two processes running on the same node, the sdk' s object classes allow processes to be aware of one another and allow transfer mechanism between those processes. Further, the processes through the sdk' s object classes are enabled to reconnect in case of loss of connection, and address sequential integrity of the data issues when the connection is reestablished. Fig. 1 is a flow diagram illustrating a method of the present disclosure in one embodiment. At 102, public interfaces are provided to allow moving data among a plurality of computer systems. The computer systems may be in heterogeneous as well as homogeneous environment . At 104, object classes are provided for moving data between threads. At 106, object classes are provided for moving data between processes on the same computer. At 108, object classes are provided for moving data between processes across a network between like operating systems. At 110, object classes are provided to move data between processes across a network where the processes are running on different operating systems. At 112, the moving of data is made event-driven. In one embodiment, objects are moved between threads in an event driven manner. In the event driven manner, the arrival of the event may be realized in the destination thread at a point where the destination thread can wait while consuming the minimal CPU possible until the object is available. This is the notion of 'event driven' versus a polling algorithm where the destination thread continuously queries the source supplying the object for its readiness. In addition, data contention for the shared object is fully anticipated for threaded applications running in multi- cpu environment . In one embodiment, the system and method of the present application uses the application space on top of the known TCP/IP and socket layers. The TCP/IP seven layer protocol layer accordingly is preserved. The socket layer is shared between the application and the operating system. Above the socket layer is a space referred to herein as the application space. It is in this application space where the present application may run. The present application in one embodiment employs a canonical client server model, which will be described below. In one aspect, the well-known POSIX thread primitives may be used for Unix™ and Linux™ based implementations. In the system and method of the present application, the pthread primitives are encapsulated into a set of objects and their use has been made easier and robust. For the Microsoft™ platforms, the MS thread primitives are used and these are interfaced via the POSIX interface to localize the porting problem. In addition, the system and method of the present application uses a number of other objects that make no reliance on POSIX, Microsoft™, Berkeley™ or any other existing standards. The components of the present application will now be described in more detail. In one embodiment, the components are packaged and referred to as "Libmsg." For moving arbitrary object between threads, the following is provided. Fig. 2 is a block diagram illustrating components for moving arbitrary objects between threads. A functionality provided is mutexing (from mutex or mutual exclusion) . Mutexing means to protect an area of data that two or more processors may vie for simultaneously at runtime. While two or more processes may read the same data at the same time, they cannot be allowed to write to that address at the same time. If the application lets that happen the program may generally fault, that is, die, crash, cease to run. The pthread library provides a number of primitives to accomplish mutexing, however, they are written in C and are not encapsulated. As such, the application programmer need to manage a number of variables to accomplish routine things that appear repeatedly in any application that deals with multi-threading. This makes writing threaded applications difficult and error prone. However, these primitives lend themselves to an encapsulation that shields the application programmer from the cumbersome details required in the management of these variables. Encapsulation is a fundamental notion in the Object Oriented ("00") paradigm and one of its goals is to create a simple public interface which exposes only what is needed by the user of the object. The system and method of the present disclosure utilizes encapsulation. The following class is provided for encapsulating the mutex functionality. The msgMutex Class (Fig. 2 202) may include: Lock - attain a mutual exclusion Lock or get in line and wait if the lock is already held. TryLock - return rather than block if the lock is already held. Unlock - relinquish the lock. ConditionedTimedWait - wait this duration in milliseconds or until signaled. ConditionedWait - wait until signaled. ClearPredicate - clear the predicate condition set by the application that is a precondition to ConditionTimedWait or a ConditionedWait . Signal either wait method at which point they will unblock. Broadcast a signal to a number of *Wait methods simultaneously. For synchronization, the system and method of the present application uses asynchronous decomposition, which means taking the problem from the outset and segregating into its natural asynchronous components or threads. If a given loop breaks a rule, the iteration is too complex and needs to be broken down further. To break down a complex iteration means to find the largest contiguous subset of the iteration that satisfies the rules and put that in its own thread. Modern schedulers switch a thread in about 20 microseconds, a third of the time it takes to switch an application. That cost at runtime is insigni icant. Given the sufficient decomposition into asynchronous components, events are raised between threads, or interthread-signaling is performed. The following set of 'protected' classes whose characteristic includes that they safely operate on data between threads is provided. These are built on top of the msgMutex class and are defined and work as follows.
The Protectedlnt Class (Fig. 2 204) may include: IsSetMethod interrogates a mutexed integer for a set clear state. Set the integer. Clear the integer. Increment the integer. Decrement the integer. Test if the integer IsPositive and return 1 or 0 accordingly. Wait or block in the iteration until signaled. Signal the Wait method so that it will unblock. Return the Address of the integer. Evaluate and return the value of the integer.
The second Protected class deals with addresses that may need to be modified and interrogated across threads. The ProtectedPointer Class (Fig. 2 206) may include: Set the address . Evaluate the address .
With the classes provided above, the functionality of moving data between threads may be performed.
The msgQueue Class: The msgQueue class 208 operates on a contiguous chunk of memory that is available to one or more threads for reading and writing. Data is stored in a protection fashion and individual elements or blocks of data can be inserted or deleted from any thread that has access to the msgQueue. The contiguous block of data is resized automatically as needed, up to a value in megabytes (default 8MB) and this default can be reset at runtime. This class achieves the asynchronous decomposition. This class allows data to be moved in an event driven data manner between threads . Enqueue - put an obj ect at the end of the queue . Dequeue - take an object off the head of the queue, block (wait) if empty. Readqueue - peek the value at the head of the queue . Rmelem - take the head element off the queue. ReadBlock - peek a block from the head of the queue, block (wait) if empty. RmBlock - take the bock from the head of the queue . Unblock - unblock an empty queue . Block - put this queue in a blocked state. Entries - return the number of elements in the queue . GetHeadPointer - return the address of the head element . IsEmpty - return 1 empty, return 0 otherwise.
The msgQueue methods that introduce data into a thread are Dequeue, Readqueue, ReadBlock and GetHeadPointer. In one embodiment, data in the queue is not shared between the queue and the calling routine. The queue makes a deep copy of that object from the address passed to it by the caller on insertion and likewise loads a copy of queue owned element into the address passed by the caller on retrieval. In one embodiment, these methods block when the queue is empty. Generally placed at the top of ' the iteration, they idle a particular iteration when no data is available. In one aspect, the POSIX and MS™ primitives are used in the system and method of the present application such that the scheduler does not bring into context a thread blocked because no data is available . The arrival of data is an event that motivates the context switch of that thread by the scheduler and the processing of that data. In a well-constructed iteration, the complete iteration can generally be accomplished in the 20 milliseconds or so of time slice available to that context switch. As noted previously, asynchronous decomposition does not restrict the number of destinations that can be targeted from a single thread. Any number of threads or processes may be dispatched from any number of points in a thread. Queues are routinely arrayed and dispatching a set of targets is done through an indexing scheme, where the targets are related to the indices of the array. In one embodiment, a PersistentStore class 210 is provided that shares the same interface as the msgQueue and achieves the same functionality. In addition, in PersistentStore class 210, all elements are written and read from disk and are retrievable for subsequent runtime instances of the application. At the instantiation of this object, an identifier of path and filename are provided which locates the object, a flat binary file. This preserves data across runtime instances whereas data in the msgQueue is lost. To understand the tradeoff in real terms, consider the TCP/IP model that assumes data in transport as volatile, but acknowledges receipt of data back to the sender. In this model, only the endpoints need to employ the PersistentStore and intermediate processes involved in the transport of data would use the msgQueue. Since the network is inherently volatile, and a chain is only as strong as its weakest link, only the source and destination processes need employ the PersistentStore. The originating process uses the PersistentStore and holds on to its data until the acknowledgement from the final destination process is received by the originator. The receiver can defend against duplicate data sent in the case of a lost acknowledgement . Another class provided in the system and method of the present disclosure is the msgList class 212. The msgList 212 is similar to the msgQueue 208 in that it is used to move data between threads. It differs from the msgQueue 208 in that the notion of head and tail do not restrict the access to the list. Items are put on the list and chained in the order in which they are inserted but can be removed from any point in the list. In simple terms, a protected list class. The following methods are provided in the msgList class 212 in one embodiment: Add - put an object on the list if object is unique. Remove - object from list by matching the value of argument to key of element in list. RemoveEle based on the address. RegisterKey - register the location of this element in the object as a searchable key. FindByKey - Find an object on the list based on the value of predetermined key element . IsKeyRegistered - return one if true; zero otherwise . IsEmpty - return one if true; zero otherwise. Lock - preclude access to the list from any thread except that which holds this lock. Unlock - open access to the list to any thread. GetHeadPointer - return the head pointer of the list. With the above defined classes, arbitrary objects may be moved between threads in an event driven fashion, in an application that has been decomposed into asynchronous components which then scales well, is intuitive and unencumbered by callbacks and signal handlers . For moving arbitrary objects between processes on the same node, the canonical socket model using the AF_UNIX family of sockets is contemplated. Fig. 3 is a block diagram illustrating components for moving arbitrary objects between processes on the same node. The AF_UNIX family of sockets is used to move data between processes on the same machine, versus the AF_INET family that moves them across the network using TCP/IP, which is described below. This is very efficient means of data transfer between processes on the same node with all the robustness of Berkeley sockets. In one embodiment, the implementation at the socket level is done via shared memory but the application is shielded from those details . The canonical socket model distinguishes a client and server. In one aspect, servers listen for and accept clients on a known port made public to the clients and clients connect to the server. At the point a connection is established client and servers can talk to one another by means of send and recv. Send and recv are two examples of several calls, which can be used to achieve the sending and receiving of data. The classes described above may be used to transfer data between processes . In one embodiment, the initial handshake of both the server and client are cleanly encapsulated and absolve the user of any knowledge of the socket model . The application developer need only know if he is a sender or receiver of data. If a sender is needed, a Client object is used. If an application is to receive data, a Server object is used. If an application wants to send and receive data, then both a Client and a Server object are used. These objects are detailed below with a number of intermediate objects that is developed to build the Client and Server. To instantiate the Client object, only the Server host and port need be known to the application. Similarly, the only information that the application supplies to the Server object is the port on which the Server is listening. This is encapsulation of the canonical socket model from the standpoint of the application developer and greatly simplifies the encoding of a full-fledged socket application. In addition, the system and method of the present disclosure provides guaranteed message delivery (GMD) . Data is sequenced and held until acknowledged and received in msgQueues . Since data is held in a msgQueue until acknowledged, data is not lost regardless of the state of the network or an outright loss due to catastrophic failure in the network infrastructure. Further the system and method of the present disclosure provides automatic recovery. If the application containing the Server goes away, the Client will reconnect automatically when the Server object becomes available in a new process. Processes containing Client objects can reconnect as well with the destruction and re-instantiation of the Client object. However, Client side reconnection is not restricted to object destruction and re-instantiation. , . In one embodiment, client and server are full peers. This means that automatic recovery happens from either side. If a Server is lost, the Clients quickly and automatically reconnect when the Server comes back. If a Client is lost, it can be brought back up and it will reconnect with the Server. The system and method described provides fast transport . Data in msgQueues are evacuated in blocks up to 64K in size and uploaded on the receive side in similar blocks in a single function call. This takes advantage of the maximum 'TCP window' and reduces the overhead the layers under the application (socket and TCP/IP) need to move the application data. The msgSocket Class : In one embodiment, the msgSocket Class 302 is private to the SDK and though integral to workings of the public classes of Libmsg, it is not part of the public interface. Users of the SDK do not need to understand it, nor do they need to be aware of its existence. Note the two constructors . msgSocket - server constructor calls the following methods . CreateSocket - create and initialize the socket. Bind - encapsulates the Berkeley "bind" function. GetSockName - encapsulates the Berkeley "getsockname" function. Listen - encapsulates the Berkeley "listen" function. msgSocket - client constructor calls the following methods . CreateSocket - create and initialize the socket. Connect - encapsulates the Berkeley "connect" function.
The remaining methods of the msgSocket are private methods, which are not called from either constructor.
IsSckValid - is the object in good condition. Accept - encapsulates the Berkeley "accept" function and is called only from the Server. Send - encapsulates the Berkeley "send" function called from both Client and Server. Receive - encapsulate the Berkeley "recv" function. CloseSocket - encapsulates the system "close" function.
The GetServerAddress Class : The GetServerAddress Class 304 is private to the SDK and though integral to workings of the public classes of Libmsg, it is not part of the public interface. Users of the SDK do not need to understand it, nor do they need to be aware of its existence. The GetServerAddress Class 304 accesses the Berkeley "gethostbyname, " and populates the "hostent" structure. It is pulled from the msgSocket because it need not be called at the same frequency of the msgSocket . Gethostbyname is implemented as a thread, which is monitored with a msgMutex: : ConditionedTimedWait , since the Berkeley call is known to hang. IsValid - returns the condition of this object as 1/0. GetHostByName - encapsulates, threads and times the potentially blocking Berkeley "gethostbyname."
In one embodiment, the Server and Client objects are two pubic classes used to move objects between processes. The objects at this level do not care whether those objects are to move between two processes on the same machine, or between two processes on machines that are located on opposite sides of the planet . The mechanics of these objects is the same, regardless.
The Server Class : The Server Class 306 owns several subordinate private objects. The Server constructor creates a limited Client object solely for the purpose of shutting down. The Berkeley "accept" is generally in a blocked state and a signal is not raised against it to unblock it in the event a shutdown is ordered. Therefore, to unblock the "accept" and allow the shutdown to proceed, a msgSocket :: Connect is called. The Server starts the AcceptClientsThread that starts a RecvThread for every Client object that seeks to connect with the Server. RecvThreads handle the incoming data and come and go for a variety of reasons. A Client that has been quiet in excess of the timeout; Clients that send an explicit breakSig and Clients that send data whose header does not conform to security protocol are dropped and the RecvThread terminates . The AcceptClientsThread persists for the life of the object.
public : IsValid - return 1 if object is in good condition; 0 otherwise . Recv - called from application with a pointer to an address at which data will be copied, blocks if no data is available. Recv_NoBlock - same as Recv but does not block in the event no data is available. Shutdown - shutdown the various threads which may exist and destroy this object.
private : Client - a limited Client object existing only for the purposes of shutting down. RecvService - do the msgSocket : :Accept handle receive threads . msgQueue - store data. Peerld - enable the Server to keep track of message sequences from incoming Clients. msgThreadlndex - enable the Server to keep track of simultaneous RecvThreads. AcceptClientsThread - block on msgSocket ::Accept , return a socket when Client connects. RecvThread - The work of the Server object is done in the RecvThread. there is a RecvThread for every incoming Client object that is sending data. Clients that quit sending data are timed out in accordance with a variable set in the environment. Clients can also set an explicit "breakSig" that will terminate the RecvThread. The RecvThread calls msgSocket ::Receive, tests for the sequential integrity of the data, sends the acknowledgement to the appropriate Client and puts the - data on the msgQueue .
The Peerld Class: The Peerld 308 is private to the SDK and though integral to workings of the public classes of Libmsg, it is not part of the public interface. Users of the SDK do not need to understand it, nor do they need to be aware of its existence. This object exists solely to support the Server object. Clients are identified by information in the headers of their incoming messages and the Peerld 308 is accessed to track the sequential integrity of the message from that Client. The sequential integrity of messages is the central component to the guaranteed message delivery and will be discussed further in the Client. But sequential integrity is important to the Server as well . Clients hold messages in their respective msgQueues until they receive an acknowledgement. The Server sends the acknowledgement on the condition that the message meets security protocol and the message is deemed in good condition. However, it is possible that the Client may have resent a message because it failed to receive an acknowledgement on a previously sent message. In other words, the Server received the message but the Client did not receive the acknowledgement . The Server is able to defend against this, so that it does not enqueue the same message twice. The Peerld 308 owns a msgList and a msgMutex object. The msgList stores an address and pid (process id) pair and the mutex is used internally to secure access to the object since more than a single RecvThread may generally contend for the Peerld. Add - add this Client to the msgList of currently connected Clients. Remove - this Client from the msgList of currently connected Clients. GetPrevSeqno - get the previous sequence number for this Client. ReviseSeqno - revise the sequence number for this Client.
The msgThreadlndex Class: The msgThreadlndex Class 310 is private to the SDK and though integral to workings of the public classes of Libmsg, it is not part of the public interface. Users of the SDK do not need to understand it, nor do they need to be aware of its existence. This object manages the indexing the array of RecvThreads that are currently in existence within the Server. GetNextFreeIndex - get the next available (not in use) index counting from zero. GetThreadlndex - get the index which was assigned to this thread. SetThreadlndex - set this threads index to a value. GetLastlndex - get the last index in use.
The Client Class: The Client 312 is primary counterpart to the Server object defined above. Note the distinction between public and private objects and methods. Only the public methods are visible to users of the SDK. The private components are listed to enable a broader understanding of the Client object. public methods: IsValid - test the condition of the object return 1/0. Send - application interface to send. SendWaitForAck - application interface to send and wait for acknowledgement . NinQ - return the number of elements in this msgQueue .
private methods and objects: GetServerAddress - get Server hostent struct information. msgSocket - do the necessary client side socket work. The msgSocket is constructed and destructed in the Client object for every reconnection. The Server terminates connections after a short duration of no incoming data. Sends from the Client which follow a delay exceeding that timeout fail, despite the preexisting connection. The cost of a failed send is sub- millisecond and a new connection can be re-established between nodes on opposite coasts in less than half a second (and within 20 milliseconds on nodes in the same segment) . Since data is safe in the msgQueue, this is a very efficient handshake mechanism that can be kept completely out of the application. This mechanism provides automatic recovery in the Client. msgQueue (2) - one for outbound msgs and one for acks . ProtectedPointer - protected address access between threads . Protectedlnt - protected integer access between threads . msgMutex (2) - general protection for areas of contention. RecycleThread - recycle the ClientSendThread if it gets dropped. ClientSendThread - does the work of the of the Client object. Peek the head element on the msgQueue, send it, receive the ack, remove the element
The msgBase Class : An application developer can define a message object in any way desired. In one embodiment, that object is derived from msgBase 314, which puts a 48-byte header at the front of the user's object. The Client and Server objects interrogate this header, but the application need not look at it. The application developer may need to be aware that their object does derive from msgBase 314 as well as the fact that their data starts + 48 bytes from the point of the object they have created from the msgBase class 314. MsgBase 314 also provide methods that serve as good examples in how to deal with the endian problems that integer transport invites . In one embodiment, none of the msgBase methods are public . ToNetBaseAlignment - called from Client (send) side . ToHostBaseAlignment - called from Server (recv) side . SetElemsThisBlock - number of elements sent in this packet, maximize to 64K. The above-described class components allow implementing an event driven applications that communicate between processes on the same node with capability for GMD, automatic recovery, and fast data transport . For moving data between processes on different nodes, for example, connected by Ethernet, routers, TCP/IP, cable, fiber, etc., Berkeley sockets and TCP/IP layers may be utilized. When the Libmsg application sees that the host value at time of the Server instantiation is other than "localhost" it uses the AF_INET family sockets in the msgSocket in lieu of the AF_UNIX family sockets (both of which are Berkeley constructs) and populates a variant hostent structure . The same thing happens in the Client instantiation. Using the above-described components of Libmsg, event driven applications that communicate between processes on different nodes of the same operating system having capabilities for GMD, automatic recovery, and fast data transport, may be implemented. For moving arbitrary objects between processes across the network where the processes are running on different operating systems, the system and method of the present disclosure provides a set of header files and static library or a shared object that compiles and links in particular environment. The public SDK interface is identical across the Unix variants, Linux, the Microsoft platforms, VMS, as/400 and MVS . Porting issues are minimized in large part to the industry's use of the POSIX Draft 10 of pthreads and the universally accepted Berkeley sockets. Thus, using the components of Libmsg described above, event driven applications that communicate between processes on different nodes running a full variety of operating systems from Solaris 2.5.1, OSF 4. Od, AIX 4.3.2, HP 11, Linux 7.2, full range of Microsoft, VMS 6.2, as/400 may be implemented. The system and method of the present disclosure may be implemented and run on a general-purpose computer. The embodiments described above are illustrative examples and it should not be construed that the present invention is limited to these particular embodiments . Although particular classes were provided as examples, it should be understood that the method and system disclosed in the present application may be implemented using different programming methodologies. Thus, various changes and modifications may be effected by one skilled in the art without departing from the spirit or scope of the invention as defined in the appended claims .

Claims

I claim :
1. A system for enabling development of client server applications for heterogeneous networked environments and homogeneous environments, comprising: a plurality of public interfaces operable to move data among a plurality of computer systems, at least some of the plurality of computer systems being in a heterogeneous networked environment and at least some other of the plurality of computer systems being in a homogeneous environment .
2. The system of claim 1, wherein the plurality of public interfaces are operable to move data between threads .
3. The system of claim 1, wherein the plurality of public interfaces are operable to move data between processes on the same computer.
4. The system of claim 1, wherein the plurality of public interfaces are operable to move data between processes across a network between like operating systems .
5. The system of claim 1, wherein the plurality of public interfaces are operable to move data between processes across a network where the processes are running on different operating systems.
6. The system of claim 2, wherein the system further includes object classes that operate in an event- driven manner for moving data between threads.
7. The system of claim 3, wherein the system further includes object classes that allow the processes to be aware of one another and allow transfer device between those processes .
8. The system of claim 1, wherein the system further includes object classes operable to reconnect after loss of connection.
9. The system of claim 8, wherein the system further includes object classes operable to address sequential integrity of data after reconnection is established.
10. The system of claim 6, wherein the object classes are operable to mutually exclude data being accessed by a processor.
11. A method for enabling development of client server applications for heterogeneous networked environments and homogeneous environments, comprising: providing a plurality of public interfaces for moving data among a plurality of computer systems, at least some of the plurality of computer systems being in a heterogeneous networked environment and at least some other of the plurality of computer systems being in a homogeneous environment .
12. The method of claim 11, further including: enabling the plurality of public interfaces to move data between threads .
13. The method of claim 11, further including: enabling the plurality of public interfaces to move data between processes on the same computer.
14. The method of claim 11, further including: enabling the plurality of public interfaces to move data between processes across a network between like operating systems .
15. The method of claim 11, further including: enabling the plurality of public interfaces to move data between processes across a network where the processes are running on different operating systems.
16. The method of claim 12, further including: providing object classes that operate in an event- driven manner for moving data between threads .
17. The method of claim 13, further including: providing object classes that allow the processes to be aware of one another and allow transfer device between those processes.
18. The method of claim 11, further including: enabling reconnecting after loss of connection.
19. The method of claim 18, further including: addressing sequential integrity of data after reconnection is established.
20. A program storage device readable by machine, tangibly embodying a program of instructions executable by the machine to perform a method for enabling development of client server applications for heterogeneous networked environments and homogeneous environments, comprising: providing a plurality of public interfaces for moving data among a plurality of computer systems, at least some of the plurality of computer systems being in a heterogeneous networked environment and at least some other of the plurality of computer systems being in a homogeneous environment ; and providing object classes that operate in an event- driven manner for moving data between threads .
PCT/US2004/022568 2003-07-11 2004-07-12 Software development kit for client server applications WO2005008431A2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
EP04778205A EP1652039A4 (en) 2003-07-11 2004-07-12 Software development kit for client server applications

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US48659603P 2003-07-11 2003-07-11
US60/486,596 2003-07-11

Publications (2)

Publication Number Publication Date
WO2005008431A2 true WO2005008431A2 (en) 2005-01-27
WO2005008431A3 WO2005008431A3 (en) 2006-11-02

Family

ID=34079258

Family Applications (3)

Application Number Title Priority Date Filing Date
PCT/US2004/022549 WO2005008430A2 (en) 2003-07-11 2004-07-12 System and method for standardizing clocks in a heterogeneous networked environment
PCT/US2004/022548 WO2005008429A2 (en) 2003-07-11 2004-07-12 Direct point-to-point communications between applications using a single port
PCT/US2004/022568 WO2005008431A2 (en) 2003-07-11 2004-07-12 Software development kit for client server applications

Family Applications Before (2)

Application Number Title Priority Date Filing Date
PCT/US2004/022549 WO2005008430A2 (en) 2003-07-11 2004-07-12 System and method for standardizing clocks in a heterogeneous networked environment
PCT/US2004/022548 WO2005008429A2 (en) 2003-07-11 2004-07-12 Direct point-to-point communications between applications using a single port

Country Status (3)

Country Link
US (4) US20060218231A1 (en)
EP (3) EP1652039A4 (en)
WO (3) WO2005008430A2 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109582400A (en) * 2018-11-30 2019-04-05 北京小米移动软件有限公司 Routine call method and device

Families Citing this family (54)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8108559B2 (en) * 2003-07-11 2012-01-31 Computer Associates Think, Inc. Standardizing clocks in a networked computing environment
US8086752B2 (en) 2006-11-22 2011-12-27 Sonos, Inc. Systems and methods for synchronizing operations among a plurality of independently clocked digital data processing devices that independently source digital data
US8234395B2 (en) 2003-07-28 2012-07-31 Sonos, Inc. System and method for synchronizing operations among a plurality of independently clocked digital data processing devices
US11106424B2 (en) 2003-07-28 2021-08-31 Sonos, Inc. Synchronizing operations among a plurality of independently clocked digital data processing devices
US11294618B2 (en) 2003-07-28 2022-04-05 Sonos, Inc. Media player system
US11650784B2 (en) 2003-07-28 2023-05-16 Sonos, Inc. Adjusting volume levels
US8290603B1 (en) 2004-06-05 2012-10-16 Sonos, Inc. User interfaces for controlling and manipulating groupings in a multi-zone media system
US10613817B2 (en) 2003-07-28 2020-04-07 Sonos, Inc. Method and apparatus for displaying a list of tracks scheduled for playback by a synchrony group
US11106425B2 (en) 2003-07-28 2021-08-31 Sonos, Inc. Synchronizing operations among a plurality of independently clocked digital data processing devices
US9374607B2 (en) 2012-06-26 2016-06-21 Sonos, Inc. Media playback system with guest access
US9977561B2 (en) 2004-04-01 2018-05-22 Sonos, Inc. Systems, methods, apparatus, and articles of manufacture to provide guest access
US8868698B2 (en) 2004-06-05 2014-10-21 Sonos, Inc. Establishing a secure wireless network with minimum human intervention
US8326951B1 (en) 2004-06-05 2012-12-04 Sonos, Inc. Establishing a secure wireless network with minimum human intervention
US7818746B2 (en) * 2005-03-30 2010-10-19 Hewlett-Packard Development Company, L.P. System and method for benchmarking using a multi-threaded load generator
US7975271B2 (en) * 2005-03-30 2011-07-05 Hewlett-Packard Development Company, L.P. System and method for dynamically determining a portion of a resource for which a thread is to obtain a lock
US7797704B2 (en) * 2005-03-30 2010-09-14 Hewlett-Packard Development Company, L.P. System and method for performing work by one of plural threads using a lockable resource
US8541007B2 (en) 2005-03-31 2013-09-24 Glaxosmithkline Biologicals S.A. Vaccines against chlamydial infection
US9202509B2 (en) 2006-09-12 2015-12-01 Sonos, Inc. Controlling and grouping in a multi-zone media system
US8788080B1 (en) 2006-09-12 2014-07-22 Sonos, Inc. Multi-channel pairing in a media system
US8483853B1 (en) 2006-09-12 2013-07-09 Sonos, Inc. Controlling and manipulating groupings in a multi-zone media system
KR101039480B1 (en) * 2010-10-29 2011-06-08 한화에스앤씨주식회사 Application store system for applying application development interoperated with unified device and method for management application store
US11429343B2 (en) 2011-01-25 2022-08-30 Sonos, Inc. Stereo playback configuration and control
US11265652B2 (en) 2011-01-25 2022-03-01 Sonos, Inc. Playback device pairing
US9729115B2 (en) 2012-04-27 2017-08-08 Sonos, Inc. Intelligently increasing the sound level of player
US9008330B2 (en) 2012-09-28 2015-04-14 Sonos, Inc. Crossover frequency adjustments for audio speakers
US9510055B2 (en) 2013-01-23 2016-11-29 Sonos, Inc. System and method for a media experience social interface
US9307508B2 (en) 2013-04-29 2016-04-05 Google Technology Holdings LLC Systems and methods for syncronizing multiple electronic devices
US9288596B2 (en) 2013-09-30 2016-03-15 Sonos, Inc. Coordinator device for paired or consolidated players
US9720576B2 (en) 2013-09-30 2017-08-01 Sonos, Inc. Controlling and displaying zones in a multi-zone system
US9654545B2 (en) 2013-09-30 2017-05-16 Sonos, Inc. Group coordinator device selection
US20150095679A1 (en) 2013-09-30 2015-04-02 Sonos, Inc. Transitioning A Networked Playback Device Between Operating Modes
US9300647B2 (en) 2014-01-15 2016-03-29 Sonos, Inc. Software application and zones
US9313591B2 (en) 2014-01-27 2016-04-12 Sonos, Inc. Audio synchronization among playback devices using offset information
US20150220498A1 (en) 2014-02-05 2015-08-06 Sonos, Inc. Remote Creation of a Playback Queue for a Future Event
US9226073B2 (en) 2014-02-06 2015-12-29 Sonos, Inc. Audio output balancing during synchronized playback
US9226087B2 (en) 2014-02-06 2015-12-29 Sonos, Inc. Audio output balancing during synchronized playback
US9679054B2 (en) 2014-03-05 2017-06-13 Sonos, Inc. Webpage media playback
US10587693B2 (en) 2014-04-01 2020-03-10 Sonos, Inc. Mirrored queues
US20150324552A1 (en) 2014-05-12 2015-11-12 Sonos, Inc. Share Restriction for Media Items
US20150356084A1 (en) 2014-06-05 2015-12-10 Sonos, Inc. Social Queue
US9874997B2 (en) 2014-08-08 2018-01-23 Sonos, Inc. Social playback queues
US9690540B2 (en) 2014-09-24 2017-06-27 Sonos, Inc. Social media queue
EP3114625A1 (en) 2014-09-24 2017-01-11 Sonos, Inc. Social media connection recommendations based on playback information
US9723038B2 (en) 2014-09-24 2017-08-01 Sonos, Inc. Social media connection recommendations based on playback information
US10645130B2 (en) 2014-09-24 2020-05-05 Sonos, Inc. Playback updates
US9860286B2 (en) 2014-09-24 2018-01-02 Sonos, Inc. Associating a captured image with a media item
US9667679B2 (en) 2014-09-24 2017-05-30 Sonos, Inc. Indicating an association between a social-media account and a media playback system
US9959087B2 (en) 2014-09-24 2018-05-01 Sonos, Inc. Media item context from social media
US10230670B1 (en) * 2014-11-10 2019-03-12 Google Llc Watermark-based message queue
US10248376B2 (en) 2015-06-11 2019-04-02 Sonos, Inc. Multiple groupings in a playback system
US10303422B1 (en) 2016-01-05 2019-05-28 Sonos, Inc. Multiple-device setup
US9886234B2 (en) 2016-01-28 2018-02-06 Sonos, Inc. Systems and methods of distributing audio to one or more playback devices
US10712997B2 (en) 2016-10-17 2020-07-14 Sonos, Inc. Room association based on name
CN108234057A (en) * 2018-01-24 2018-06-29 郑州云海信息技术有限公司 Method for synchronizing time, device and the storage medium of server based on BMC

Family Cites Families (34)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5481721A (en) * 1991-07-17 1996-01-02 Next Computer, Inc. Method for providing automatic and dynamic translation of object oriented programming language-based message passing into operation system message passing using proxy objects
US5371852A (en) * 1992-10-14 1994-12-06 International Business Machines Corporation Method and apparatus for making a cluster of computers appear as a single host on a network
US5428645A (en) * 1992-11-03 1995-06-27 International Business Machines Corporation Anonymous time synchronization method
US5630066A (en) * 1994-12-20 1997-05-13 Sun Microsystems, Inc. System and method for locating object view and platform independent object
US6292820B1 (en) * 1996-07-29 2001-09-18 At& T Corp. Porting POSIX-conforming operating systems to Win32 API-conforming operating systems
US6175879B1 (en) * 1997-01-29 2001-01-16 Microsoft Corporation Method and system for migrating connections between receive-any and receive-direct threads
US6192514B1 (en) * 1997-02-19 2001-02-20 Unisys Corporation Multicomputer system
US20010054064A1 (en) * 1997-07-02 2001-12-20 Pallipuram V. Kannan Method system and computer program product for providing customer service over the world-wide web
US7237036B2 (en) * 1997-10-14 2007-06-26 Alacritech, Inc. Fast-path apparatus for receiving data corresponding a TCP connection
US6148402A (en) * 1998-04-01 2000-11-14 Hewlett-Packard Company Apparatus and method for remotely executing commands using distributed computing environment remote procedure calls
US6496871B1 (en) * 1998-06-30 2002-12-17 Nec Research Institute, Inc. Distributed agent software system and method having enhanced process mobility and communication in a computer network
US6324586B1 (en) * 1998-09-17 2001-11-27 Jennifer Wallace System for synchronizing multiple computers with a common timing reference
US6725278B1 (en) * 1998-09-17 2004-04-20 Apple Computer, Inc. Smart synchronization of computer system time clock based on network connection modes
US6311283B1 (en) * 1998-09-17 2001-10-30 Apple Computer, Inc. Need based synchronization of computer system time clock to reduce loading on network server
CN1166249C (en) * 1999-01-19 2004-09-08 西门子公司 Method for the time synchronisation of a computer network and computer network with time synchronisation
US7206805B1 (en) * 1999-09-09 2007-04-17 Oracle International Corporation Asynchronous transcription object management system
EP1113361A1 (en) * 2000-01-03 2001-07-04 Wimba.Com S.A. Process of communication between an applet and a local agent using a socket communication channel
AT5327U1 (en) * 2000-03-06 2002-05-27 Keroe Nikolaus Dipl Ing METHOD FOR SYNCHRONIZING COMPUTER WATCHES IN NETWORKS FOR INFORMATION TRANSFER, DEVICE FOR CARRYING OUT THIS METHOD AND DATA PACKAGE SUITABLE FOR SYNCHRONIZING COMPUTER WATCHES
US7051337B2 (en) * 2000-04-08 2006-05-23 Sun Microsystems, Inc. Method and apparatus for polling multiple sockets with a single thread and handling events received at the sockets with a pool of threads
US6665541B1 (en) * 2000-05-04 2003-12-16 Snaptrack, Incorporated Methods and apparatuses for using mobile GPS receivers to synchronize basestations in cellular networks
US6922685B2 (en) * 2000-05-22 2005-07-26 Mci, Inc. Method and system for managing partitioned data resources
GB0019341D0 (en) * 2000-08-08 2000-09-27 Easics Nv System-on-chip solutions
US6829769B2 (en) * 2000-10-04 2004-12-07 Microsoft Corporation High performance interprocess communication
US6718395B1 (en) * 2000-10-10 2004-04-06 Computer Access Technology Corporation Apparatus and method using an inquiry response for synchronizing to a communication network
US6957357B2 (en) * 2000-10-30 2005-10-18 International Business Machines Corporation Clock synchronization with removal of clock skews through network measurements in derivation of a convext hull
US7058955B2 (en) * 2000-12-06 2006-06-06 Microsoft Corporation Method and system for passing messages between threads
US7023816B2 (en) * 2000-12-13 2006-04-04 Safenet, Inc. Method and system for time synchronization
NO20006684D0 (en) * 2000-12-28 2000-12-28 Abb Research Ltd Time Synchronization
US7035246B2 (en) * 2001-03-13 2006-04-25 Pulse-Link, Inc. Maintaining a global time reference among a group of networked devices
US6915353B2 (en) * 2001-08-01 2005-07-05 Hewlett-Packard Development Company, L.P. Method and apparatus for avoiding unnecessary computer peripheral calibration activities
US7283568B2 (en) * 2001-09-11 2007-10-16 Netiq Corporation Methods, systems and computer program products for synchronizing clocks of nodes on a computer network
US20030084190A1 (en) * 2001-10-25 2003-05-01 Kimball Robert H. Apparatus and system for maintaining accurate time in a wireless environment
US7139346B2 (en) * 2002-08-09 2006-11-21 The Boeing Company Mobile network time distribution
US7372875B2 (en) * 2002-09-30 2008-05-13 Lucent Technologies Inc. Systems and methods for synchronization in asynchronous transport networks

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
See references of EP1652039A4 *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109582400A (en) * 2018-11-30 2019-04-05 北京小米移动软件有限公司 Routine call method and device
CN109582400B (en) * 2018-11-30 2023-02-21 北京小米移动软件有限公司 Program calling method and device

Also Published As

Publication number Publication date
WO2005008430A3 (en) 2007-11-15
WO2005008431A3 (en) 2006-11-02
WO2005008429A2 (en) 2005-01-27
EP1652059A2 (en) 2006-05-03
WO2005008429A3 (en) 2005-12-01
EP1652039A2 (en) 2006-05-03
EP1652039A4 (en) 2007-11-14
US20050188082A1 (en) 2005-08-25
US20060101472A1 (en) 2006-05-11
US20090157905A1 (en) 2009-06-18
WO2005008430A2 (en) 2005-01-27
EP1652038A4 (en) 2010-05-12
EP1652059A4 (en) 2007-12-19
US20060218231A1 (en) 2006-09-28
EP1652038A2 (en) 2006-05-03

Similar Documents

Publication Publication Date Title
US20060101472A1 (en) Software development kit for client server applications
Jha et al. Derecho: Fast state machine replication for cloud services
US7010586B1 (en) System and method for event subscriptions for CORBA gateway
US6839748B1 (en) Synchronous task scheduler for corba gateway
US8769127B2 (en) Cross-domain solution (CDS) collaborate-access-browse (CAB) and assured file transfer (AFT)
US6950935B1 (en) Pluggable authentication modules for telecommunications management network
US8804488B2 (en) Endpoint-based parallel data processing in a parallel active messaging interface of a parallel computer
Fiuczynski et al. An Extensible Protocol Architecture for Application-Specific Networking.
US20120185873A1 (en) Data Communications In A Parallel Active Messaging Interface Of A Parallel Computer
US20140229536A1 (en) Technique for enabling a plurality of software components to communicate in a software component matrix environment
JPH07160518A (en) Event architecture for system management of operating system
Allcock et al. The globus extensible input/output system (xio): A protocol independent io system for the grid
US7478403B1 (en) Secure access to managed network objects using a configurable platform-independent gateway providing individual object-level access control
Rashid An inter-process communication facility for UNIX
US7228346B1 (en) IDL event and request formatting for corba gateway
Bryce et al. Lana: An approach to programming autonomous systems
Adamczewski-Musch et al. Dataflow engine in DAQ backbone DABC
Hiesgen et al. A configurable transport layer for CAF
Franke et al. MPI on IBM SP1/SP2: Current status and future directions
Birman et al. ISIS documentation: release 1
Narasimhakumar Benchmarking and Optimizations of Data Shuffling on High-performance Networks
Vella CSP/occam on networks of workstations
Stanton A users guide to spread version 0.11
Lee Distributed Procedure Call
Tanenbaum et al. The Amoeba Microkennel

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A2

Designated state(s): AE AG AL AM AT AU AZ BA BB BG BR BW BY BZ CA CH CN CO CR CU CZ DE DK DM DZ EC EE EG ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX MZ NA NI NO NZ OM PG PH PL PT RO RU SC SD SE SG SK SL SY TJ TM TN TR TT TZ UA UG US UZ VC VN YU ZA ZM ZW

AL Designated countries for regional patents

Kind code of ref document: A2

Designated state(s): BW GH GM KE LS MW MZ NA SD SL SZ TZ UG ZM ZW AM AZ BY KG KZ MD RU TJ TM AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IT LU MC NL PL PT RO SE SI SK TR BF BJ CF CG CI CM GA GN GQ GW ML MR NE SN TD TG

121 Ep: the epo has been informed by wipo that ep was designated in this application
WWE Wipo information: entry into national phase

Ref document number: 2004778205

Country of ref document: EP

WWP Wipo information: published in national office

Ref document number: 2004778205

Country of ref document: EP