US10268516B2 - Merging connection pools to form a logical pool of connections during a preset period of time thereby more efficiently utilizing connections in connection pools - Google Patents
Merging connection pools to form a logical pool of connections during a preset period of time thereby more efficiently utilizing connections in connection pools Download PDFInfo
- Publication number
- US10268516B2 US10268516B2 US15/292,828 US201615292828A US10268516B2 US 10268516 B2 US10268516 B2 US 10268516B2 US 201615292828 A US201615292828 A US 201615292828A US 10268516 B2 US10268516 B2 US 10268516B2
- Authority
- US
- United States
- Prior art keywords
- connections
- cluster
- members
- resource
- connection
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active, expires
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F9/00—Arrangements for program control, e.g. control units
- G06F9/06—Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
- G06F9/46—Multiprogramming arrangements
- G06F9/50—Allocation of resources, e.g. of the central processing unit [CPU]
- G06F9/5005—Allocation of resources, e.g. of the central processing unit [CPU] to service a request
- G06F9/5011—Allocation of resources, e.g. of the central processing unit [CPU] to service a request the resources being hardware resources other than CPUs, Servers and Terminals
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F9/00—Arrangements for program control, e.g. control units
- G06F9/06—Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
- G06F9/44—Arrangements for executing specific programs
- G06F9/455—Emulation; Interpretation; Software simulation, e.g. virtualisation or emulation of application or operating system execution engines
- G06F9/45533—Hypervisors; Virtual machine monitors
- G06F9/45558—Hypervisor-specific management and integration aspects
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F9/00—Arrangements for program control, e.g. control units
- G06F9/06—Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
- G06F9/46—Multiprogramming arrangements
- G06F9/50—Allocation of resources, e.g. of the central processing unit [CPU]
- G06F9/5061—Partitioning or combining of resources
- G06F9/5077—Logical partitioning of resources; Management or configuration of virtualized resources
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L65/00—Network arrangements, protocols or services for supporting real-time applications in data packet communication
- H04L65/10—Architectures or entities
- H04L65/1063—Application servers providing network services
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L65/00—Network arrangements, protocols or services for supporting real-time applications in data packet communication
- H04L65/1066—Session management
- H04L65/1069—Session establishment or de-establishment
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F9/00—Arrangements for program control, e.g. control units
- G06F9/06—Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
- G06F9/44—Arrangements for executing specific programs
- G06F9/455—Emulation; Interpretation; Software simulation, e.g. virtualisation or emulation of application or operating system execution engines
- G06F9/45533—Hypervisors; Virtual machine monitors
- G06F9/45558—Hypervisor-specific management and integration aspects
- G06F2009/45595—Network integration; Enabling network access in virtual machine instances
Definitions
- the present invention relates generally to connection pools utilized in the Java EE Connector Architecture (JCA), and more particularly to merging connection pools to form a logical pool of connections during a preset period of time thereby more efficiently utilizing connections in connection pools.
- JCA Java EE Connector Architecture
- the Java 2 Platform, Enterprise Edition is a software standard for developing multitier enterprise applications.
- the J2EE architecture has a client tier, middle tier, and back-end tier.
- the client tier supports a variety of client types, and the middle tier supports client services and business logic through web and Enterprise Java Beans (EJB) containers in the middle tier.
- the back-end tier includes the Enterprise Information Systems (EIS) in the EIS tier and many standard Application Programming Interfaces (APIs) for accessing databases.
- EIS Enterprise Information Systems
- APIs Application Programming Interfaces
- An Enterprise Information System is generally any kind of computing system that offers high quality of service, deals with large volumes of data and capable of supporting some large organization (“an enterprise”).
- An enterprise can accordingly alter the objects and components resident on the several tiers.
- “Containers” are standardized runtime environments that provide services to components on the platform. All containers provide runtime support for responding to client requests and returning results to clients. The containers also provide APIs to support user session management.
- J2EE Connector Architecture JCA
- JCA J2EE Connector Architecture
- connection management contract which allows applications to connect to an EIS, and enables the application server to utilize pooling (pooling of resource connections);
- transaction management contract which allows an application server to manage and perform transactional access across one to many EIS resource managers;
- security contract which provides support for secure access to the EIS.
- connection management contract allows the application server to pool resource connections.
- an object commonly referred to as a connection factory, creates the resource connections to a resource (e.g., database) based on the properties configured on the connection factory instance.
- An EIS system may service different number of requests during different times of the day and therefore need a different number of resource connections to a database (resource) to store or retrieve data in connection with servicing these requests.
- a database e.g., a tax collection application, online banking application
- the tax collection application is busy at the end of the work day between 4:00 pm-6:00 pm and also at the end of the month (e.g., 29 th , 30 th and 31 st day) as tax submissions peak at this time.
- this application requires more than 100 connections to the database (resource) to be available for the requests to be serviced.
- the application needs an average of 30-40 connections to the database.
- the administrator configures the connection factory to create a maximum pool size of 100 connections. In case there are more than 100 concurrent users accessing the tax collection application, they would have to wait for free connections to become available.
- an online banking application uses a separate connection pool to access the same database (resource).
- the online banking application is busy from 9:00 am-4:00 pm on Monday through Friday, and especially busy during 9:00 am-1:00 pm (peak-time), where 250 connections to the database are required to service the requests.
- peak-time 9:00 am-1:00 pm
- the administrator configures the connection factory to create a maximum pool size of 200 connections for this application.
- the online banking application would not be able to service all the requests concurrently.
- the tax collection application only needs an average of 30-40 connections to the database. Hence, it has about 60-70 connections unused during that time since it is allocated 100 connections to the database to service requests. As a result, the resource connections are not being optimally utilized.
- the online banking application users will have to wait until the connections are released in order to be served despite the fact that tax collection application has about 60-70 connections unused during the time that the online banking application could use additional connections.
- connection pool for the applications While an administrator may attempt to manually reconfigure the size of the connection pool for the applications, such as reducing the size of the connection pool for the tax collection application and increasing the size of the connection pool for the online banking application, applications may unpredictably receive an increase in the number of requests to be serviced, including for the application whose pool of connections was reduced, and therefore, may no longer be able to service all the requests concurrently.
- a method for efficiently utilizing connections in connection pools comprises identifying a period of time a first application running on a first virtual machine needs a greater number of resource connections to a resource than allocated in its first pool of connections of a first fixed size.
- the method further comprises merging, by a processor, the first pool of connections with a second pool of connections of a second fixed size utilized by a second application of a second virtual machine to access the resource to form a logical pool of connections to be shared by the first and second applications of the first virtual machine and the second virtual machine, respectively, during the period of time, where the first and second pools of connections contain resource connections to the resource.
- a method for efficiently utilizing connections in connection pools comprises assigning a weight to each member of a cluster of members based on a size of a load serviced by each cluster member.
- the method further comprises computing a maximum connection pool size for each member of the cluster of members based on the weight assigned to each member of the cluster of members and a total number of connections allocated to a resource.
- the method comprises configuring, by a processor, a connection pool size for each member of the cluster of members to correspond with its corresponding computed maximum connection pool size.
- FIG. 1 illustrates a virtualization environment for a conventional server system in accordance with an embodiment of the present invention
- FIG. 2 is a flowchart of a method for merging pools of connections to form a logical pool of connections in accordance with an embodiment of the present invention
- FIG. 3 illustrates merging pools of connections utilized by the applications running on the virtual machines to form a logical pool of connections in accordance with an embodiment of the present invention
- FIG. 4 illustrates configuring connection pool sizes for members of a cluster in an application server cluster environment in accordance with an embodiment of the present invention
- FIG. 5 is a flowchart of a method for configuring connection pool sizes for members of a cluster in an application server cluster environment in accordance with an embodiment of the present invention.
- the present invention comprises a method, system and computer program product for efficiently utilizing connections in connection pools.
- a period of time an application running on a virtual machine needs a greater number of resource connections to an external resource (e.g., database) than allocated in its pool of connections is identified.
- the connection pool for this application as well as the connection pools for the other applications containing the resource connections to the same external resource are merged to form a logical pool of connections to be shared by those applications of the virtual machines during the identified period of time.
- an application needs more connections at its peak load than the number of connections allocated in its pool of connections, then those connections that are not being utilized in the connection pools by the other applications can be utilized by the application.
- connection pools utilized by the application servers to access the external resource may be reconfigured based on the weight assigned to each member (or application server) of the cluster which is based on the load size serviced by the member. In this manner, the resource connections in these pools of connections will be more efficiently utilized.
- FIG. 1 illustrates a virtualization environment for a conventional server system 100 in accordance with an embodiment of the present invention.
- Server 100 includes a virtual operating system 101 , which can be an IBMTM z/VM operating system.
- Operating system 101 executes on a real or physical computer 102 .
- Real computer 102 includes one or more processors 103 , a memory 104 (also referred to herein as the host physical memory), one or more disk drives 105 and the like.
- Real computer 102 further includes a processor controller 106 that provides communication between virtual operating system 101 and physical computer 102 .
- Other components of real computer 102 are not discussed herein for the sake of brevity.
- virtual operating system 101 includes user portions 107 A- 107 C (identified as “Guest 1 ,” “Guest 2 ,” and “Guest 3 ,” respectively in FIG. 1 ), referred to herein as “guests.”
- Guests 107 A- 107 C may collectively or individually be referred to as guests 107 or guest 107 , respectively.
- Each guest 107 is capable of functioning as a separate system. That is, each guest 107 A- 107 C can be independently reset, host a guest operating system 108 A- 108 C (identified as “Guest O/S 1 ,” “Guest O/S 2 ,” and “Guest O/S 3 ,” respectively in FIG. 1 ), respectively, and operate with different programs.
- Guest operating systems 108 A- 108 C may collectively or individually be referred to as guest operating systems 108 or guest operating system 108 , respectively.
- Each guest operating system 108 may host one or more virtual machine applications 109 A- 109 E (identified as “VM 1 ,” “VM 2 ,” “VM 3 ,” “VM 4 ,” and “VM 5 ,” respectively in FIG. 1 ), such as JavaTM virtual machines.
- virtual machine applications 109 A- 109 E such as “VM 1 ,” “VM 2 ,” “VM 3 ,” “VM 4 ,” and “VM 5 ,” respectively in FIG. 1
- guest operating system 108 A hosts virtual machine applications 109 A- 109 B.
- Guest operating system 108 B hosts virtual machine applications 109 C- 109 D
- guest operating system 108 C hosts virtual machine application 109 E.
- Virtual machine applications 109 A- 109 E may collectively or individually be referred to as virtual machine applications 109 or virtual machine application 109 , respectively.
- FIG. 1 is not to be limited in scope to a particular number of guests 107 executing a particular number of guest operating systems 108 hosting a particular number of virtual machine applications 109 .
- Virtual operating system 101 further includes a common base portion 110 referred to as a hypervisor or as a control program.
- Hypervisor or control program 110 may be implemented in microcode running on processor 103 or it may be implemented in software as part of virtual operating system 101 .
- Control program 110 is configured to manage and enable guests 107 to run on a single host.
- Control program 110 includes a virtual processor controller 111 A- 111 C associated with each guest 107 A- 107 C, respectively.
- Virtual processor controllers 111 A- 111 C may collectively or individually be referred to as virtual processor controllers 111 or virtual processor controller 111 , respectively.
- Virtual processor controller 111 provides communication between guest operating system 108 and control program 110 .
- Control program 110 further implements virtual processors 112 A- 112 C, one for each virtual processor controller 111 A- 111 C, respectively. That is, control program 110 implements virtual processors 112 A- 112 C, one for each guest 107 A- 107 C, respectively. Virtual processors 112 A- 112 C may collectively or individually be referred to as virtual processors 112 or virtual processor 112 , respectively.
- virtual operating system 101 and its components execute on physical or real computer 102 . These software components may be loaded into memory 104 for execution by processor 103 . Furthermore, an application for efficiently utilizing connections in connection pools, as discussed below in connection with FIGS. 2-5 , may be loaded into memory 104 for execution by processor 103 .
- Server 100 may include applications running on virtual machines 109 , including an application server that includes applications. Server 100 may provide interfaces to certain external resources, such as a database. In one embodiment, server 100 includes a J2EE compatible system. In operation of server 100 , connections may be required between the applications running on virtual machines 109 and the external resources. To produce such connections, an entity may make a request to generate an instance of a connection factory, the connection factory being a program object to generate a pool of connections between the application and the external resource. The principles of the present invention ensure that there are enough connections available to the external resource by the application during a particular time, such as during the peak-load time of the application, as discussed further below.
- the present invention may be a system, a method, and/or a computer program product.
- the computer program product may include a computer readable storage medium (or media) having computer readable program instructions thereon for causing a processor to carry out aspects of the present invention.
- the computer readable storage medium can be a tangible device that can retain and store instructions for use by an instruction execution device.
- the computer readable storage medium may be, for example, but is not limited to, an electronic storage device, a magnetic storage device, an optical storage device, an electromagnetic storage device, a semiconductor storage device, or any suitable combination of the foregoing.
- a non-exhaustive list of more specific examples of the computer readable storage medium includes the following: a portable computer diskette, a hard disk, a random access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or Flash memory), a static random access memory (SRAM), a portable compact disc read-only memory (CD-ROM), a digital versatile disk (DVD), a memory stick, a floppy disk, a mechanically encoded device such as punch-cards or raised structures in a groove having instructions recorded thereon, and any suitable combination of the foregoing.
- RAM random access memory
- ROM read-only memory
- EPROM or Flash memory erasable programmable read-only memory
- SRAM static random access memory
- CD-ROM compact disc read-only memory
- DVD digital versatile disk
- memory stick a floppy disk
- a mechanically encoded device such as punch-cards or raised structures in a groove having instructions recorded thereon
- a computer readable storage medium is not to be construed as being transitory signals per se, such as radio waves or other freely propagating electromagnetic waves, electromagnetic waves propagating through a waveguide or other transmission media (e.g., light pulses passing through a fiber-optic cable), or electrical signals transmitted through a wire.
- Computer readable program instructions described herein can be downloaded to respective computing/processing devices from a computer readable storage medium or to an external computer or external storage device via a network, for example, the Internet, a local area network, a wide area network and/or a wireless network.
- the network may comprise copper transmission cables, optical transmission fibers, wireless transmission, routers, firewalls, switches, gateway computers and/or edge servers.
- a network adapter card or network interface in each computing/processing device receives computer readable program instructions from the network and forwards the computer readable program instructions for storage in a computer readable storage medium within the respective computing/processing device.
- Computer readable program instructions for carrying out operations of the present invention may be assembler instructions, instruction-set-architecture (ISA) instructions, machine instructions, machine dependent instructions, microcode, firmware instructions, state-setting data, or either source code or object code written in any combination of one or more programming languages, including an object oriented programming language such as Smalltalk, C++ or the like, and conventional procedural programming languages, such as the “C” programming language or similar programming languages.
- the computer readable program instructions may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server.
- the remote computer may be connected to the user's computer through any type of network, including a local area network (LAN) or a wide area network (WAN), or the connection may be made to an external computer (for example, through the Internet using an Internet Service Provider).
- electronic circuitry including, for example, programmable logic circuitry, field-programmable gate arrays (FPGA), or programmable logic arrays (PLA) may execute the computer readable program instructions by utilizing state information of the computer readable program instructions to personalize the electronic circuitry, in order to perform aspects of the present invention.
- These computer readable program instructions may be provided to a processor of a general purpose computer, special purpose computer, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks.
- These computer readable program instructions may also be stored in a computer readable storage medium that can direct a computer, a programmable data processing apparatus, and/or other devices to function in a particular manner, such that the computer readable storage medium having instructions stored therein comprises an article of manufacture including instructions which implement aspects of the function/act specified in the flowchart and/or block diagram block or blocks.
- the computer readable program instructions may also be loaded onto a computer, other programmable data processing apparatus, or other device to cause a series of operational steps to be performed on the computer, other programmable apparatus or other device to produce a computer implemented process, such that the instructions which execute on the computer, other programmable apparatus, or other device implement the functions/acts specified in the flowchart and/or block diagram block or blocks.
- each block in the flowchart or block diagrams may represent a module, segment, or portion of instructions, which comprises one or more executable instructions for implementing the specified logical function(s).
- the functions noted in the block may occur out of the order noted in the figures.
- two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved.
- the administrator configures the connection factory to create a maximum connection pool size for each application.
- the resource connections are not being optimally utilized.
- connection pool While an administrator may attempt to manually reconfigure the size of the connection pool for the applications, applications may unpredictably receive an increase in the number of requests to be serviced, including for the application whose pool of connections was reduced, and therefore, may no longer be able to service all the requests concurrently. Hence, there is not currently a means for efficiently utilizing the resource connections in the connection pools.
- FIG. 2 is a flowchart of a method for merging pools of connections to form a logical pool of connections.
- FIG. 3 illustrates merging pools of connections utilized by the applications running on the virtual machines to form a logical pool of connections.
- FIG. 4 illustrates configuring connection pool sizes for members of a cluster in an application server cluster environment.
- FIG. 5 is a flowchart of a method for configuring connection pool sizes for members of a cluster in an application server cluster environment.
- FIG. 2 is a flowchart of a method 200 for merging pools of connections to form a logical pool of connections in accordance with an embodiment of the present invention.
- server 100 identifies a period of time an application running on a virtual machine 109 needs a greater number of resource connections to an external resource (e.g., database) than allocated in its pool of connections.
- an external resource e.g., database
- server 100 identifies a period of time an application running on a virtual machine 109 needs a greater number of resource connections to an external resource (e.g., database) than allocated in its pool of connections.
- an external resource e.g., database
- the maximum connection pool size for the online banking application is limited to 200 connections.
- the period of time may be based on the day, week, month, morning, evening and/or specific hours.
- server 100 identifies the connection factory object involved with the application (application discussed in step 201 ) which will be part of the resource reference declaration.
- the connection factory object creates the resource connections to the external resource, such as a database, based on the properties configured on the connection factory instance.
- Each application may be associated with a connection factory object, which is used to create a pool of connections to the external resource.
- the resource is referenced by a resource reference declaration which will include the connection factory that is used to create the pool of connections utilized by the application discussed in step 201 to connect to the external resource.
- step 203 server 100 clusters the identified connection factory object of step 202 with the other connection factory objects connecting to the same external resource to be accessed by the other virtual machines 109 .
- the connection pools created by these connection factory objects will be able to be merged as discussed below.
- server 100 merges the connection pools for the cluster connection factory objects for those connection pools containing the resource connections to the same external resource to be accessed by virtual machines 109 during the period of time identified in step 201 thereby forming a logical pool of connections to be shared by the applications of virtual machines 109 during the period of time as illustrated in FIG. 3 .
- FIG. 3 illustrates merging pools of connections utilized by the applications running on virtual machines 109 to form a logical pool of connections in accordance with an embodiment of the present invention.
- applications of virtual machines 109 such as application 301 A (identified as “Application 1 ” in FIG. 3 ) of virtual machine 109 A, application 301 B (identified as “Application 2 ” in FIG. 3 ) of virtual machine 109 B and application 301 C (identified as “Application 3 ” in FIG. 3 ) of virtual machine 109 C, are connected to an external resource 302 , such as a database, via connection pools 303 A- 303 C, respectively, established by connection factory objects.
- Applications 301 A- 301 C may collectively or individually be referred to as applications 301 or application 301 , respectively.
- Connection pools 303 A- 303 C may collectively or individually be referred to as connection pools 303 or connection pool 303 , respectively.
- FIG. 3 is not to be limited in scope to the depicted number of applications 301 , virtual machines 109 and connection pools 303 .
- the principles of the present invention may be used to merge any number of connection pools 303 into a logical pool of connections, as discussed below, to be shared by any number of applications 301 running on any number of virtual machines 109 .
- each of these connection pools 303 A- 303 C are merged to form a logical pool of connections 304 to be shared by applications 301 A- 301 C for a period of time determined in step 201 .
- an application 301 such as application 301 A
- those connections that are not being utilized in the other connection pool(s) 303 by the other applications 301 can be utilized by application 301 A.
- the resource connections are being more efficiently utilized in connection pools 303 .
- step 205 a determination is made by server 100 as to whether the period of time identified in step 201 elapsed. If the period of time identified in step 201 did not elapse, then server 100 continues to determine whether the period of time identified in step 201 elapsed in step 205 .
- server 100 returns the merged connection pools 303 to the connection pools 303 of fixed size. For example, referring to FIG. 3 , if connection pool 303 A had a fixed size of 200 resource connection, connection pool 303 B had a fixed size of 50 resource connections and connection pool 303 C had a fixed size of 50 resource connections, then after merging connection pools 303 A- 303 C, a logical pool of connections 304 of 300 resource connection to be shared among applications 301 would be formed for the period of time identified in step 201 . After the period of time elapses, then the merged connection pools 303 (i.e., the logical pool of connections 304 ) would be returned to connection pools 303 A- 303 C of fixed size of 200, 50 and 50 resource connections, respectively.
- the merged connection pools 303 i.e., the logical pool of connections 304
- server 100 may include applications running on virtual machines 109 , including an application server that includes applications.
- an application server may need a greater number of resource connections to the external resource than allocated in its pool of connections in a similar manner as discussed above.
- the connection pools utilized by the application servers to access the external resource may be reconfigured based on the weight assigned to each member (or application server) of the cluster which is based on the load size serviced by the member as discussed further below in connection with FIGS. 4-5 .
- FIG. 4 illustrates members (e.g., application servers) of a cluster 400 in an application server cluster environment whose connection pool sizes will be reconfigured in accordance with an embodiment of the present invention.
- application servers 401 A- 401 C (identified as “application server 1 ,” “application server 2 ,” and “application server 3 ,” respectively in FIG. 4 ) running on virtual machines 109 A- 109 C, respectively, in cluster 400 are connected to an external resource 302 , such as a database, via connection pools 402 A- 402 C, respectively, established by connection factory objects.
- Application servers 401 A- 401 C may collectively or individually be referred to as application servers 401 or application server 401 , respectively.
- Connection pools 402 A- 402 C may collectively or individually be referred to as connection pools 402 or connection pool 402 , respectively.
- each virtual machine 109 (e.g., virtual machines 109 A- 109 C) is located on a node of cluster 400 .
- a software component referred to as a “load balancer” 403 , which may be loaded into memory 104 for execution by processor 103 , is configured to reconfigure the number of resource connections in connections pools 402 based on the weight assigned to each member (or application server) of cluster 400 as discussed further below in connection with FIG. 5 .
- FIG. 4 is not to be limited in scope to the depicted number of application servers 401 , virtual machines 109 and connection pools 402 .
- FIG. 5 is a flowchart of a method 500 for configuring connection pool sizes for members of a cluster in an application server cluster environment in accordance with an embodiment of the present invention.
- load balancer 403 assigns a weight to a cluster member based on the size of the load serviced by the cluster member (e.g., application server 401 ). For example, load balancer 403 may assign a weight of 20 to application server 401 A, assign a weight of 10 to application server 401 B and assign a weight of 5 to application server 401 C based on the size of the load serviced by these application servers 401 . For instance, application server 401 B may be assigned a weight of twice the weight of application server 401 C since the size of its load is twice the size of the load serviced by application server 401 C. Similarly, application server 401 A may be assigned a weight of twice the weight of application server 401 B since the size of its load is twice the size of the load serviced by application server 401 B.
- load balancer 403 computes the maximum connection pool size (i.e., the number of connections available in connection pool 402 ) for each cluster member (e.g., application server 401 ) based on the weights assigned to the cluster members and the total number of connections allocated to external resource 302 .
- the maximum connection pool size for application server 401 A (i.e., the maximum number of resource connections in connection pool 402 A) corresponds to 171 ((20/35)*300) connections using EQ 1 as illustrated in FIG. 4 .
- the maximum connection pool size for application server 401 B (i.e., the maximum number of resource connections in connection pool 402 B) corresponds to 86 ((10/35)*300) connections using EQ 1 as illustrated in FIG. 4 .
- the maximum connection pool size for application server 401 C (i.e., the maximum number of resource connections in connection pool 402 C) corresponds to 43 ((5/35)*300) connections using EQ 1 as illustrated in FIG. 4 .
- load balancer 403 configures the connection pool size (i.e., the number of connections available in connection pool 402 ) for each cluster member (e.g., application server 401 ) to correspond with the maximum connection pool size computed in step 502 .
- Method 500 may be reiterated in response to any changes in the weight assigned to an application server 401 which occurs based on changes in the size of the load handled by application server 401 .
- load balancer 403 reconfiguring the sizes of the pools of connections 402 for application servers 401 in cluster 400 based on the size of the loads serviced by application servers 401 , the resource connections in these pools of connections 402 will be more efficiently utilized.
Landscapes
- Engineering & Computer Science (AREA)
- Software Systems (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- Computer Networks & Wireless Communication (AREA)
- Signal Processing (AREA)
- Multimedia (AREA)
- Business, Economics & Management (AREA)
- General Business, Economics & Management (AREA)
- Stored Programmes (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
Abstract
A method, system and computer program product for efficiently utilizing connections in connection pools. A period of time an application running on a virtual machine needs a greater number of connections to an external resource than allocated in its pool of connections is identified. The connection pool for this application as well as the connection pools for the other applications containing connections to the same external resource are merged to form a logical pool of connections to be shared by those applications during the identified period of time. Alternatively, in an application server cluster environment, the connection pools utilized by the application servers to access the external resource may be reconfigured based on the weight assigned to each member (or application server) of the cluster which is based on the member's load size. In these manners, the resource connections in these pools of connections will be more efficiently utilized.
Description
The present invention relates generally to connection pools utilized in the Java EE Connector Architecture (JCA), and more particularly to merging connection pools to form a logical pool of connections during a preset period of time thereby more efficiently utilizing connections in connection pools.
The Java 2 Platform, Enterprise Edition (J2EE) is a software standard for developing multitier enterprise applications. The J2EE architecture has a client tier, middle tier, and back-end tier. The client tier supports a variety of client types, and the middle tier supports client services and business logic through web and Enterprise Java Beans (EJB) containers in the middle tier. The back-end tier includes the Enterprise Information Systems (EIS) in the EIS tier and many standard Application Programming Interfaces (APIs) for accessing databases. An Enterprise Information System is generally any kind of computing system that offers high quality of service, deals with large volumes of data and capable of supporting some large organization (“an enterprise”). One of skill in the art can accordingly alter the objects and components resident on the several tiers. “Containers” are standardized runtime environments that provide services to components on the platform. All containers provide runtime support for responding to client requests and returning results to clients. The containers also provide APIs to support user session management.
Resident within the J2EE architecture is a “resource adapter” that plays a central role in the integration and connectivity between an EIS and an application server (software framework dedicated to the efficient execution of procedures for supporting the construction of applications) and serves as the point of contact between application components, application servers and enterprise information systems. A resource adapter and other components, must communicate with one another based on a standard referred to as the J2EE Connector Architecture (JCA). To enable seamless integration with an application server, the resource adapter abides by system-level “contracts” defined by the JCA. These contracts exist between the application server and the EIS, and are implemented through the resource adapter. The contracts specify how a system external to the J2EE platform integrates with it by supporting basic functions handled by the J2EE container. There are three major contracts: the “connection management contract” which allows applications to connect to an EIS, and enables the application server to utilize pooling (pooling of resource connections); the “transaction management contract” which allows an application server to manage and perform transactional access across one to many EIS resource managers; and the “security contract” which provides support for secure access to the EIS.
As discussed above, the connection management contract allows the application server to pool resource connections. In connection with pooling resource connections, an object, commonly referred to as a connection factory, creates the resource connections to a resource (e.g., database) based on the properties configured on the connection factory instance.
Applications running by an EIS system (e.g., tax collection application, online banking application) may service different number of requests during different times of the day and therefore need a different number of resource connections to a database (resource) to store or retrieve data in connection with servicing these requests. For example, suppose a tax collection application is used to collect tax information from individuals and corporations on behalf of the government. The tax collection application is busy at the end of the work day between 4:00 pm-6:00 pm and also at the end of the month (e.g., 29th, 30th and 31st day) as tax submissions peak at this time. At peak time, this application requires more than 100 connections to the database (resource) to be available for the requests to be serviced. Normally, during non-peak times, the application needs an average of 30-40 connections to the database. Hence, in order to ensure that there are enough connections to the database (resource) during any particular time, the administrator configures the connection factory to create a maximum pool size of 100 connections. In case there are more than 100 concurrent users accessing the tax collection application, they would have to wait for free connections to become available.
Furthermore, suppose a second application, an online banking application, uses a separate connection pool to access the same database (resource). The online banking application is busy from 9:00 am-4:00 pm on Monday through Friday, and especially busy during 9:00 am-1:00 pm (peak-time), where 250 connections to the database are required to service the requests. However, if the database can only handle 300 connections concurrently, then the administrator configures the connection factory to create a maximum pool size of 200 connections for this application. Hence, during the peak-time, the online banking application would not be able to service all the requests concurrently.
However, during 9:00 am-4:00 pm on Monday through Friday (except the last few days of the month), the tax collection application only needs an average of 30-40 connections to the database. Hence, it has about 60-70 connections unused during that time since it is allocated 100 connections to the database to service requests. As a result, the resource connections are not being optimally utilized. The online banking application users will have to wait until the connections are released in order to be served despite the fact that tax collection application has about 60-70 connections unused during the time that the online banking application could use additional connections.
While an administrator may attempt to manually reconfigure the size of the connection pool for the applications, such as reducing the size of the connection pool for the tax collection application and increasing the size of the connection pool for the online banking application, applications may unpredictably receive an increase in the number of requests to be serviced, including for the application whose pool of connections was reduced, and therefore, may no longer be able to service all the requests concurrently.
Hence, there is not currently a means for efficiently utilizing the resource connections in the connection pools.
In one embodiment of the present invention, a method for efficiently utilizing connections in connection pools comprises identifying a period of time a first application running on a first virtual machine needs a greater number of resource connections to a resource than allocated in its first pool of connections of a first fixed size. The method further comprises merging, by a processor, the first pool of connections with a second pool of connections of a second fixed size utilized by a second application of a second virtual machine to access the resource to form a logical pool of connections to be shared by the first and second applications of the first virtual machine and the second virtual machine, respectively, during the period of time, where the first and second pools of connections contain resource connections to the resource.
Other forms of the embodiment of the method described above are in a system and in a computer program product.
In another embodiment of the present invention, a method for efficiently utilizing connections in connection pools comprises assigning a weight to each member of a cluster of members based on a size of a load serviced by each cluster member. The method further comprises computing a maximum connection pool size for each member of the cluster of members based on the weight assigned to each member of the cluster of members and a total number of connections allocated to a resource. Furthermore, the method comprises configuring, by a processor, a connection pool size for each member of the cluster of members to correspond with its corresponding computed maximum connection pool size.
Other forms of the embodiment of the method described above are in a system and in a computer program product.
The foregoing has outlined rather generally the features and technical advantages of one or more embodiments of the present invention in order that the detailed description of the present invention that follows may be better understood. Additional features and advantages of the present invention will be described hereinafter which may form the subject of the claims of the present invention.
A better understanding of the present invention can be obtained when the following detailed description is considered in conjunction with the following drawings, in which:
The present invention comprises a method, system and computer program product for efficiently utilizing connections in connection pools. In one embodiment of the present invention, a period of time an application running on a virtual machine needs a greater number of resource connections to an external resource (e.g., database) than allocated in its pool of connections is identified. The connection pool for this application as well as the connection pools for the other applications containing the resource connections to the same external resource are merged to form a logical pool of connections to be shared by those applications of the virtual machines during the identified period of time. In this manner, when an application needs more connections at its peak load than the number of connections allocated in its pool of connections, then those connections that are not being utilized in the connection pools by the other applications can be utilized by the application. As a result, the resource connections are being more efficiently utilized in the connection pools. Alternatively, in an application server cluster environment, the connection pools utilized by the application servers to access the external resource may be reconfigured based on the weight assigned to each member (or application server) of the cluster which is based on the load size serviced by the member. In this manner, the resource connections in these pools of connections will be more efficiently utilized.
In the following description, numerous specific details are set forth to provide a thorough understanding of the present invention. However, it will be apparent to those skilled in the art that the present invention may be practiced without such specific details. In other instances, well-known circuits have been shown in block diagram form in order not to obscure the present invention in unnecessary detail. For the most part, details considering timing considerations and the like have been omitted inasmuch as such details are not necessary to obtain a complete understanding of the present invention and are within the skills of persons of ordinary skill in the relevant art.
Referring now to the Figures in detail, FIG. 1 illustrates a virtualization environment for a conventional server system 100 in accordance with an embodiment of the present invention. Server 100 includes a virtual operating system 101, which can be an IBM™ z/VM operating system. Operating system 101 executes on a real or physical computer 102. Real computer 102 includes one or more processors 103, a memory 104 (also referred to herein as the host physical memory), one or more disk drives 105 and the like. Real computer 102 further includes a processor controller 106 that provides communication between virtual operating system 101 and physical computer 102. Other components of real computer 102 are not discussed herein for the sake of brevity.
Referring again to FIG. 1 , virtual operating system 101 includes user portions 107A-107C (identified as “Guest 1,” “Guest 2,” and “Guest 3,” respectively in FIG. 1 ), referred to herein as “guests.” Guests 107A-107C may collectively or individually be referred to as guests 107 or guest 107, respectively. Each guest 107 is capable of functioning as a separate system. That is, each guest 107A-107C can be independently reset, host a guest operating system 108A-108C (identified as “Guest O/S 1,” “Guest O/S 2,” and “Guest O/S 3,” respectively in FIG. 1 ), respectively, and operate with different programs. An operating system or application program running in guest 107 appears to have access to a full and complete system, but in reality, only a portion of it is available. Guest operating systems 108A-108C may collectively or individually be referred to as guest operating systems 108 or guest operating system 108, respectively.
Each guest operating system 108 may host one or more virtual machine applications 109A-109E (identified as “VM 1,” “VM2,” “VM3,” “VM4,” and “VM5,” respectively in FIG. 1 ), such as Java™ virtual machines. For example, guest operating system 108A hosts virtual machine applications 109A-109B. Guest operating system 108B hosts virtual machine applications 109C-109D and guest operating system 108C hosts virtual machine application 109E. Virtual machine applications 109A-109E may collectively or individually be referred to as virtual machine applications 109 or virtual machine application 109, respectively.
As discussed above, virtual operating system 101 and its components execute on physical or real computer 102. These software components may be loaded into memory 104 for execution by processor 103. Furthermore, an application for efficiently utilizing connections in connection pools, as discussed below in connection with FIGS. 2-5 , may be loaded into memory 104 for execution by processor 103.
The present invention may be a system, a method, and/or a computer program product. The computer program product may include a computer readable storage medium (or media) having computer readable program instructions thereon for causing a processor to carry out aspects of the present invention.
The computer readable storage medium can be a tangible device that can retain and store instructions for use by an instruction execution device. The computer readable storage medium may be, for example, but is not limited to, an electronic storage device, a magnetic storage device, an optical storage device, an electromagnetic storage device, a semiconductor storage device, or any suitable combination of the foregoing. A non-exhaustive list of more specific examples of the computer readable storage medium includes the following: a portable computer diskette, a hard disk, a random access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or Flash memory), a static random access memory (SRAM), a portable compact disc read-only memory (CD-ROM), a digital versatile disk (DVD), a memory stick, a floppy disk, a mechanically encoded device such as punch-cards or raised structures in a groove having instructions recorded thereon, and any suitable combination of the foregoing. A computer readable storage medium, as used herein, is not to be construed as being transitory signals per se, such as radio waves or other freely propagating electromagnetic waves, electromagnetic waves propagating through a waveguide or other transmission media (e.g., light pulses passing through a fiber-optic cable), or electrical signals transmitted through a wire.
Computer readable program instructions described herein can be downloaded to respective computing/processing devices from a computer readable storage medium or to an external computer or external storage device via a network, for example, the Internet, a local area network, a wide area network and/or a wireless network. The network may comprise copper transmission cables, optical transmission fibers, wireless transmission, routers, firewalls, switches, gateway computers and/or edge servers. A network adapter card or network interface in each computing/processing device receives computer readable program instructions from the network and forwards the computer readable program instructions for storage in a computer readable storage medium within the respective computing/processing device.
Computer readable program instructions for carrying out operations of the present invention may be assembler instructions, instruction-set-architecture (ISA) instructions, machine instructions, machine dependent instructions, microcode, firmware instructions, state-setting data, or either source code or object code written in any combination of one or more programming languages, including an object oriented programming language such as Smalltalk, C++ or the like, and conventional procedural programming languages, such as the “C” programming language or similar programming languages. The computer readable program instructions may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server. In the latter scenario, the remote computer may be connected to the user's computer through any type of network, including a local area network (LAN) or a wide area network (WAN), or the connection may be made to an external computer (for example, through the Internet using an Internet Service Provider). In some embodiments, electronic circuitry including, for example, programmable logic circuitry, field-programmable gate arrays (FPGA), or programmable logic arrays (PLA) may execute the computer readable program instructions by utilizing state information of the computer readable program instructions to personalize the electronic circuitry, in order to perform aspects of the present invention.
Aspects of the present invention are described herein with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems), and computer program products according to embodiments of the invention. It will be understood that each block of the flowchart illustrations and/or block diagrams, and combinations of blocks in the flowchart illustrations and/or block diagrams, can be implemented by computer readable program instructions.
These computer readable program instructions may be provided to a processor of a general purpose computer, special purpose computer, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks. These computer readable program instructions may also be stored in a computer readable storage medium that can direct a computer, a programmable data processing apparatus, and/or other devices to function in a particular manner, such that the computer readable storage medium having instructions stored therein comprises an article of manufacture including instructions which implement aspects of the function/act specified in the flowchart and/or block diagram block or blocks.
The computer readable program instructions may also be loaded onto a computer, other programmable data processing apparatus, or other device to cause a series of operational steps to be performed on the computer, other programmable apparatus or other device to produce a computer implemented process, such that the instructions which execute on the computer, other programmable apparatus, or other device implement the functions/acts specified in the flowchart and/or block diagram block or blocks.
The flowchart and block diagrams in the Figures illustrate the architecture, functionality, and operation of possible implementations of systems, methods, and computer program products according to various embodiments of the present invention. In this regard, each block in the flowchart or block diagrams may represent a module, segment, or portion of instructions, which comprises one or more executable instructions for implementing the specified logical function(s). In some alternative implementations, the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved. It will also be noted that each block of the block diagrams and/or flowchart illustration, and combinations of blocks in the block diagrams and/or flowchart illustration, can be implemented by special purpose hardware-based systems that perform the specified functions or acts or carry out combinations of special purpose hardware and computer instructions.
As stated in the Background section, in order to ensure that there are enough connections to the database (external resource) by the application during any particular time, including the peak-load time of the application, the administrator configures the connection factory to create a maximum connection pool size for each application. However, in certain situations, there may be connections in the connection pool to the resource unused by the application during its non-peak times; whereas, other applications may have an insufficient number of connections to the resource during its peak-times, which may correspond to the non-peak times of the application with excess resource connections. As a result, the resource connections are not being optimally utilized. While an administrator may attempt to manually reconfigure the size of the connection pool for the applications, applications may unpredictably receive an increase in the number of requests to be serviced, including for the application whose pool of connections was reduced, and therefore, may no longer be able to service all the requests concurrently. Hence, there is not currently a means for efficiently utilizing the resource connections in the connection pools.
The principles of the present invention provide a means for more efficiently utilizing the resource connections in the connection pools by merging the pools of connections utilized by the applications running on virtual machines 109 (FIG. 1 ) containing connections to the same resource to form a logical pool of connections to be shared by the applications during a period of time, which may be based on business requirements as discussed below in association with FIGS. 2-5 . FIG. 2 is a flowchart of a method for merging pools of connections to form a logical pool of connections. FIG. 3 illustrates merging pools of connections utilized by the applications running on the virtual machines to form a logical pool of connections. FIG. 4 illustrates configuring connection pool sizes for members of a cluster in an application server cluster environment. FIG. 5 is a flowchart of a method for configuring connection pool sizes for members of a cluster in an application server cluster environment.
As stated above, FIG. 2 is a flowchart of a method 200 for merging pools of connections to form a logical pool of connections in accordance with an embodiment of the present invention.
Referring to FIG. 2 , in conjunction with FIG. 1 , in step 201, server 100 identifies a period of time an application running on a virtual machine 109 needs a greater number of resource connections to an external resource (e.g., database) than allocated in its pool of connections. For example, suppose that an online banking application running on virtual machine 109 needs 250 connections to the external resource (e.g., database) from 9:00 am-1:00 pm (peak-time) on Monday-Friday. However, the maximum connection pool size for the online banking application is limited to 200 connections. Hence, based on the business requirements for this application, there is a need for a greater number of resource connections to the external resource during that period of time. The period of time, as used herein, may be based on the day, week, month, morning, evening and/or specific hours.
In step 202, server 100 identifies the connection factory object involved with the application (application discussed in step 201) which will be part of the resource reference declaration. As discussed above, the connection factory object creates the resource connections to the external resource, such as a database, based on the properties configured on the connection factory instance. Each application may be associated with a connection factory object, which is used to create a pool of connections to the external resource. In one embodiment, the resource is referenced by a resource reference declaration which will include the connection factory that is used to create the pool of connections utilized by the application discussed in step 201 to connect to the external resource.
In step 203, server 100 clusters the identified connection factory object of step 202 with the other connection factory objects connecting to the same external resource to be accessed by the other virtual machines 109. By clustering the connection factory objects connecting to the same external resource, the connection pools created by these connection factory objects will be able to be merged as discussed below.
In step 204, server 100 merges the connection pools for the cluster connection factory objects for those connection pools containing the resource connections to the same external resource to be accessed by virtual machines 109 during the period of time identified in step 201 thereby forming a logical pool of connections to be shared by the applications of virtual machines 109 during the period of time as illustrated in FIG. 3 .
Referring to FIG. 3 , in conjunction with FIGS. 1-2 , applications of virtual machines 109, such as application 301A (identified as “Application 1” in FIG. 3 ) of virtual machine 109A, application 301B (identified as “Application 2” in FIG. 3 ) of virtual machine 109B and application 301C (identified as “Application 3” in FIG. 3 ) of virtual machine 109C, are connected to an external resource 302, such as a database, via connection pools 303A-303C, respectively, established by connection factory objects. Applications 301A-301C may collectively or individually be referred to as applications 301 or application 301, respectively. Connection pools 303A-303C may collectively or individually be referred to as connection pools 303 or connection pool 303, respectively. FIG. 3 is not to be limited in scope to the depicted number of applications 301, virtual machines 109 and connection pools 303. The principles of the present invention may be used to merge any number of connection pools 303 into a logical pool of connections, as discussed below, to be shared by any number of applications 301 running on any number of virtual machines 109.
During a period of time, such as when application 301A needs a greater number of resource connections to external resource 302 than allocated in its pool of connections, each of these connection pools 303A-303C are merged to form a logical pool of connections 304 to be shared by applications 301A-301C for a period of time determined in step 201. In this manner, when an application 301, such as application 301A, needs more connections at its peak load than the number of connections allocated in its pool of connections 303, then those connections that are not being utilized in the other connection pool(s) 303 by the other applications 301, such as application 301B, can be utilized by application 301A. As a result, the resource connections are being more efficiently utilized in connection pools 303. After the period of time (identified in step 201) elapses, those connections that were utilized by application 301 that came from a different connection pool 303 than its own connection pool 303 would be returned to its original connection pool 303.
Returning to FIG. 2 , in conjunction with FIGS. 1 and 3 , in step 205, a determination is made by server 100 as to whether the period of time identified in step 201 elapsed. If the period of time identified in step 201 did not elapse, then server 100 continues to determine whether the period of time identified in step 201 elapsed in step 205.
If, however, the period of time identified in step 201 elapsed, then, in step 206, server 100 returns the merged connection pools 303 to the connection pools 303 of fixed size. For example, referring to FIG. 3 , if connection pool 303A had a fixed size of 200 resource connection, connection pool 303 B had a fixed size of 50 resource connections and connection pool 303C had a fixed size of 50 resource connections, then after merging connection pools 303A-303C, a logical pool of connections 304 of 300 resource connection to be shared among applications 301 would be formed for the period of time identified in step 201. After the period of time elapses, then the merged connection pools 303 (i.e., the logical pool of connections 304) would be returned to connection pools 303A-303C of fixed size of 200, 50 and 50 resource connections, respectively.
As discussed above, server 100 may include applications running on virtual machines 109, including an application server that includes applications. In an application server cluster environment, where each application server runs on a virtual machine 109 that corresponds to a node of the cluster, an application server may need a greater number of resource connections to the external resource than allocated in its pool of connections in a similar manner as discussed above. To address such a need, the connection pools utilized by the application servers to access the external resource may be reconfigured based on the weight assigned to each member (or application server) of the cluster which is based on the load size serviced by the member as discussed further below in connection with FIGS. 4-5 .
Referring to FIG. 4 , application servers 401A-401C (identified as “application server 1,” “application server 2,” and “application server 3,” respectively in FIG. 4 ) running on virtual machines 109A-109C, respectively, in cluster 400 are connected to an external resource 302, such as a database, via connection pools 402A-402C, respectively, established by connection factory objects. Application servers 401A-401C may collectively or individually be referred to as application servers 401 or application server 401, respectively. Connection pools 402A-402C may collectively or individually be referred to as connection pools 402 or connection pool 402, respectively. In one embodiment, each virtual machine 109 (e.g., virtual machines 109A-109C) is located on a node of cluster 400.
As further illustrated in FIG. 4 , a software component, referred to as a “load balancer” 403, which may be loaded into memory 104 for execution by processor 103, is configured to reconfigure the number of resource connections in connections pools 402 based on the weight assigned to each member (or application server) of cluster 400 as discussed further below in connection with FIG. 5 . FIG. 4 is not to be limited in scope to the depicted number of application servers 401, virtual machines 109 and connection pools 402.
Referring to FIG. 5 , in conjunction with FIGS. 1 and 4 , in step 501, load balancer 403 assigns a weight to a cluster member based on the size of the load serviced by the cluster member (e.g., application server 401). For example, load balancer 403 may assign a weight of 20 to application server 401A, assign a weight of 10 to application server 401B and assign a weight of 5 to application server 401C based on the size of the load serviced by these application servers 401. For instance, application server 401B may be assigned a weight of twice the weight of application server 401C since the size of its load is twice the size of the load serviced by application server 401C. Similarly, application server 401A may be assigned a weight of twice the weight of application server 401B since the size of its load is twice the size of the load serviced by application server 401B.
In step 402, load balancer 403 computes the maximum connection pool size (i.e., the number of connections available in connection pool 402) for each cluster member (e.g., application server 401) based on the weights assigned to the cluster members and the total number of connections allocated to external resource 302. In one embodiment, the maximum connection pool size for each cluster member equals:
MaxConnectionPoolSizecm=(Wcm/(TWcms))*MaxTotalSize (EQ 1)
where the parameter “MaxConectionPoolSizecm” corresponds to the maximum connection pool size for a cluster member (e.g.,application server 401A), the parameter “Wcm” corresponds to the weight assigned to the cluster member (e.g., weight of 20 assigned to application server 401A), the parameter “TWcms” corresponds to the total weight assigned to all the cluster members (e.g., weights of 20 plus 10 plus 5) and the parameter of MaxTotalSize corresponds to the total number of resource connections (e.g., 300 connections) allocated to this external resource (e.g., database 302).
MaxConnectionPoolSizecm=(Wcm/(TWcms))*MaxTotalSize (EQ 1)
where the parameter “MaxConectionPoolSizecm” corresponds to the maximum connection pool size for a cluster member (e.g.,
For example, suppose that application server 401A is assigned a weight of 20, application server 401B is assigned a weight of 10 and application server 401C is assigned a weight of 5. Suppose further that the total number of resource connections allocated to database 302 is 300. Hence, the maximum connection pool size for application server 401A (i.e., the maximum number of resource connections in connection pool 402A) corresponds to 171 ((20/35)*300) connections using EQ 1 as illustrated in FIG. 4 . The maximum connection pool size for application server 401B (i.e., the maximum number of resource connections in connection pool 402B) corresponds to 86 ((10/35)*300) connections using EQ 1 as illustrated in FIG. 4 . Furthermore, the maximum connection pool size for application server 401C (i.e., the maximum number of resource connections in connection pool 402C) corresponds to 43 ((5/35)*300) connections using EQ 1 as illustrated in FIG. 4 .
In step 503, load balancer 403 configures the connection pool size (i.e., the number of connections available in connection pool 402) for each cluster member (e.g., application server 401) to correspond with the maximum connection pool size computed in step 502.
By load balancer 403 reconfiguring the sizes of the pools of connections 402 for application servers 401 in cluster 400 based on the size of the loads serviced by application servers 401, the resource connections in these pools of connections 402 will be more efficiently utilized.
The descriptions of the various embodiments of the present invention have been presented for purposes of illustration, but are not intended to be exhaustive or limited to the embodiments disclosed. Many modifications and variations will be apparent to those of ordinary skill in the art without departing from the scope and spirit of the described embodiments. The terminology used herein was chosen to best explain the principles of the embodiments, the practical application or technical improvement over technologies found in the marketplace, or to enable others of ordinary skill in the art to understand the embodiments disclosed herein.
Claims (3)
1. A method for efficiently utilizing connections in connection pools, the method comprising:
assigning a weight to each member of a cluster of members based on a size of a load serviced by each cluster member, wherein each member of said cluster of members corresponds to an application server running on a virtual machine, wherein said virtual machine is located on a node of said cluster of members;
computing a maximum connection pool size for each member of said cluster of members based on said weight assigned to each member of said cluster of members and a total number of connections allocated to a resource; and
configuring, by a processor, a connection pool size for each member of said cluster of members to correspond with its corresponding computed maximum connection pool size.
2. A computer program product for efficiently utilizing connections in connection pools, the computer program product comprising a computer readable storage medium having program code embodied therewith, the program code comprising the programming instructions for:
assigning a weight to each member of a cluster of members based on a size of a load serviced by each cluster member, wherein each member of said cluster of members corresponds to an application server running on a virtual machine, wherein said virtual machine is located on a node of said cluster of members;
computing a maximum connection pool size for each member of said cluster of members based on said weight assigned to each member of said cluster of members and a total number of connections allocated to a resource; and
configuring a connection pool size for each member of said cluster of members to correspond with its corresponding computed maximum connection pool size.
3. A system, comprising:
a memory unit for storing a computer program for efficiently utilizing connections in connection pools; and
a processor coupled to the memory unit, wherein the processor is configured to execute the program instructions of the computer program comprising:
assigning a weight to each member of a cluster of members based on a size of a load serviced by each cluster member, wherein each member of said cluster of members corresponds to an application server running on a virtual machine, wherein said virtual machine is located on a node of said cluster of members;
computing a maximum connection pool size for each member of said cluster of members based on said weight assigned to each member of said cluster of members and a total number of connections allocated to a resource; and
configuring a connection pool size for each member of said cluster of members to correspond with its corresponding computed maximum connection pool size.
Priority Applications (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US15/292,828 US10268516B2 (en) | 2014-09-30 | 2016-10-13 | Merging connection pools to form a logical pool of connections during a preset period of time thereby more efficiently utilizing connections in connection pools |
US16/260,565 US10740147B2 (en) | 2014-09-30 | 2019-01-29 | Merging connection pools to form a logical pool of connections during a preset period of time thereby more efficiently utilizing connections in connection pools |
US16/868,895 US11429443B2 (en) | 2014-09-30 | 2020-05-07 | Merging connection pools to form a logical pool of connections during a preset period of time thereby more efficiently utilizing connections in connection pools |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/502,226 US9563475B2 (en) | 2014-09-30 | 2014-09-30 | Merging connection pools to form a logical pool of connections during a preset period of time thereby more efficiently utilizing connections in connection pools |
US15/292,828 US10268516B2 (en) | 2014-09-30 | 2016-10-13 | Merging connection pools to form a logical pool of connections during a preset period of time thereby more efficiently utilizing connections in connection pools |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/502,226 Division US9563475B2 (en) | 2014-09-30 | 2014-09-30 | Merging connection pools to form a logical pool of connections during a preset period of time thereby more efficiently utilizing connections in connection pools |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US16/260,565 Continuation US10740147B2 (en) | 2014-09-30 | 2019-01-29 | Merging connection pools to form a logical pool of connections during a preset period of time thereby more efficiently utilizing connections in connection pools |
Publications (2)
Publication Number | Publication Date |
---|---|
US20170031716A1 US20170031716A1 (en) | 2017-02-02 |
US10268516B2 true US10268516B2 (en) | 2019-04-23 |
Family
ID=55584514
Family Applications (4)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/502,226 Expired - Fee Related US9563475B2 (en) | 2014-09-30 | 2014-09-30 | Merging connection pools to form a logical pool of connections during a preset period of time thereby more efficiently utilizing connections in connection pools |
US15/292,828 Active 2035-02-12 US10268516B2 (en) | 2014-09-30 | 2016-10-13 | Merging connection pools to form a logical pool of connections during a preset period of time thereby more efficiently utilizing connections in connection pools |
US16/260,565 Expired - Fee Related US10740147B2 (en) | 2014-09-30 | 2019-01-29 | Merging connection pools to form a logical pool of connections during a preset period of time thereby more efficiently utilizing connections in connection pools |
US16/868,895 Active US11429443B2 (en) | 2014-09-30 | 2020-05-07 | Merging connection pools to form a logical pool of connections during a preset period of time thereby more efficiently utilizing connections in connection pools |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/502,226 Expired - Fee Related US9563475B2 (en) | 2014-09-30 | 2014-09-30 | Merging connection pools to form a logical pool of connections during a preset period of time thereby more efficiently utilizing connections in connection pools |
Family Applications After (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US16/260,565 Expired - Fee Related US10740147B2 (en) | 2014-09-30 | 2019-01-29 | Merging connection pools to form a logical pool of connections during a preset period of time thereby more efficiently utilizing connections in connection pools |
US16/868,895 Active US11429443B2 (en) | 2014-09-30 | 2020-05-07 | Merging connection pools to form a logical pool of connections during a preset period of time thereby more efficiently utilizing connections in connection pools |
Country Status (1)
Country | Link |
---|---|
US (4) | US9563475B2 (en) |
Families Citing this family (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106021544B (en) * | 2016-05-27 | 2020-02-07 | 北京京东尚科信息技术有限公司 | Database distributed connection pool management method and system |
US10146563B2 (en) * | 2016-08-03 | 2018-12-04 | International Business Machines Corporation | Predictive layer pre-provisioning in container-based virtualization |
US11243970B2 (en) * | 2017-06-30 | 2022-02-08 | Paypal, Inc. | Intelligent database connection management |
US11204792B2 (en) * | 2018-04-04 | 2021-12-21 | International Business Machines Corporation | Attaching storage resources to virtual machine instances |
CN108845860B (en) * | 2018-04-26 | 2022-06-03 | 北京奇艺世纪科技有限公司 | Method and device for managing quota and electronic equipment |
US11176155B2 (en) * | 2019-12-13 | 2021-11-16 | Paypal, Inc | Self-adjusting connection pool in a distributed system |
US11799941B2 (en) | 2021-09-07 | 2023-10-24 | Red Hat, Inc. | Handling connection pool sizing with heterogeneous concurrency |
CN114996342B (en) * | 2022-06-27 | 2024-09-03 | 平安银行股份有限公司 | Pool dividing method and device for connecting pools |
Citations (28)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20050188088A1 (en) | 2004-01-13 | 2005-08-25 | International Business Machines Corporation | Managing escalating resource needs within a grid environment |
US20050268276A1 (en) | 2004-06-01 | 2005-12-01 | Tankov Nikolai D | Program object to support connection generation |
US20050278439A1 (en) * | 2004-06-14 | 2005-12-15 | Ludmila Cherkasova | System and method for evaluating capacity of a heterogeneous media server configuration for supporting an expected workload |
US20060031388A1 (en) | 2004-06-03 | 2006-02-09 | Tankov Nikolai D | Connection resource system |
US20060209718A1 (en) | 2005-03-17 | 2006-09-21 | Microsoft Corporation | Enhanced network system through the combination of network objects |
US20100333111A1 (en) | 2009-06-29 | 2010-12-30 | Software Ag | Systems and/or methods for policy-based JMS broker clustering |
US20110149737A1 (en) | 2009-12-23 | 2011-06-23 | Manikam Muthiah | Systems and methods for managing spillover limits in a multi-core system |
US20110178790A1 (en) | 2010-01-20 | 2011-07-21 | Xyratex Technology Limited | Electronic data store |
US20110202925A1 (en) | 2010-02-18 | 2011-08-18 | International Business Machines Corporation | Optimized capacity planning |
US20120226788A1 (en) | 2004-03-13 | 2012-09-06 | Cluster Resources, Inc. | System and method for providing multi-resource management support in a compute environment |
US8286174B1 (en) | 2006-04-17 | 2012-10-09 | Vmware, Inc. | Executing a multicomponent software application on a virtualized computer platform |
US20120284712A1 (en) | 2011-05-04 | 2012-11-08 | Chitti Nimmagadda | Systems and methods for sr-iov pass-thru via an intermediary device |
US8423998B2 (en) | 2010-06-04 | 2013-04-16 | International Business Machines Corporation | System and method for virtual machine multiplexing for resource provisioning in compute clouds |
US8429652B2 (en) | 2009-06-22 | 2013-04-23 | Citrix Systems, Inc. | Systems and methods for spillover in a multi-core system |
US20130290957A1 (en) * | 2012-04-26 | 2013-10-31 | International Business Machines Corporation | Efficient execution of jobs in a shared pool of resources |
US20140173113A1 (en) | 2012-12-19 | 2014-06-19 | Symantec Corporation | Providing Optimized Quality of Service to Prioritized Virtual Machines and Applications Based on Quality of Shared Resources |
US20140173595A1 (en) | 2012-12-17 | 2014-06-19 | International Business Machines Corporation | Hybrid virtual machine configuration management |
US20140222889A1 (en) | 2013-02-05 | 2014-08-07 | International Business Machines Corporation | Collaborative negotiation of system resources among virtual servers running in a network computing environment |
US20150040128A1 (en) | 2013-08-05 | 2015-02-05 | International Business Machines Corporation | Utilizing Multiple Memory Pools During Mobility Operations |
US20150244643A1 (en) | 2014-02-21 | 2015-08-27 | Hitachi, Ltd. | Data center resource allocation system and data center resource allocation method |
US20150263982A1 (en) | 2014-03-12 | 2015-09-17 | International Business Machines Corporation | Software defined infrastructures that encapsulate physical server resources into logical resource pools |
US20150331707A1 (en) | 2012-12-21 | 2015-11-19 | Telefonaktiebolaget Lm Ericsson (Publ) | Method and cloud management node for enabling a virtual machine |
US20160004551A1 (en) | 2013-10-04 | 2016-01-07 | Hitachi, Ltd. | Resource management system and resource management method |
US20160092277A1 (en) | 2014-09-25 | 2016-03-31 | International Business Machines Corporation | Overcommitting virtual machine hosts |
US9407692B2 (en) * | 2013-11-27 | 2016-08-02 | Avi Networks | Method and system for distributed load balancing |
US9558039B2 (en) * | 2015-03-13 | 2017-01-31 | International Business Machines Corporation | Managing resources of a shared pool of configurable computing resources |
US9692820B2 (en) * | 2013-04-06 | 2017-06-27 | Citrix Systems, Inc. | Systems and methods for cluster parameter limit |
US9870244B2 (en) * | 2014-12-29 | 2018-01-16 | International Business Machines Corporation | Optimized migration of virtual objects across environments in a cloud computing environment |
Family Cites Families (14)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6745221B1 (en) * | 1999-03-01 | 2004-06-01 | Mitel, Inc. | Dynamic resource reallocation |
US7266823B2 (en) * | 2002-02-21 | 2007-09-04 | International Business Machines Corporation | Apparatus and method of dynamically repartitioning a computer system in response to partition workloads |
US9798560B1 (en) * | 2008-09-23 | 2017-10-24 | Gogrid, LLC | Automated system and method for extracting and adapting system configurations |
US8239526B2 (en) * | 2008-11-14 | 2012-08-07 | Oracle International Corporation | System and method for performance data collection in a virtual environment |
US8572677B2 (en) * | 2010-07-14 | 2013-10-29 | William G. Bartholomay | Devices, systems, and methods for enabling reconfiguration of services supported by a network of devices |
US8548790B2 (en) * | 2011-01-07 | 2013-10-01 | International Business Machines Corporation | Rapidly determining fragmentation in computing environments |
WO2012147116A1 (en) * | 2011-04-25 | 2012-11-01 | Hitachi, Ltd. | Computer system and virtual machine control method |
US9471373B2 (en) * | 2011-09-24 | 2016-10-18 | Elwha Llc | Entitlement vector for library usage in managing resource allocation and scheduling based on usage and priority |
US10136165B2 (en) * | 2011-09-14 | 2018-11-20 | Mobitv, Inc. | Distributed scalable encoder resources for live streams |
US9244742B2 (en) * | 2012-05-31 | 2016-01-26 | Vmware, Inc. | Distributed demand-based storage quality of service management using resource pooling |
US9507612B1 (en) * | 2012-08-31 | 2016-11-29 | United Services Automobile Association (Usaa) | Managing dedicated and floating pool of virtual machines based on demand |
US9841991B2 (en) * | 2014-05-12 | 2017-12-12 | Netapp, Inc. | Techniques for virtual machine migration |
US9619268B2 (en) * | 2014-08-23 | 2017-04-11 | Vmware, Inc. | Rapid suspend/resume for virtual machines via resource sharing |
US9672074B2 (en) * | 2015-10-19 | 2017-06-06 | Vmware, Inc. | Methods and systems to determine and improve cost efficiency of virtual machines |
-
2014
- 2014-09-30 US US14/502,226 patent/US9563475B2/en not_active Expired - Fee Related
-
2016
- 2016-10-13 US US15/292,828 patent/US10268516B2/en active Active
-
2019
- 2019-01-29 US US16/260,565 patent/US10740147B2/en not_active Expired - Fee Related
-
2020
- 2020-05-07 US US16/868,895 patent/US11429443B2/en active Active
Patent Citations (28)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20050188088A1 (en) | 2004-01-13 | 2005-08-25 | International Business Machines Corporation | Managing escalating resource needs within a grid environment |
US20120226788A1 (en) | 2004-03-13 | 2012-09-06 | Cluster Resources, Inc. | System and method for providing multi-resource management support in a compute environment |
US20050268276A1 (en) | 2004-06-01 | 2005-12-01 | Tankov Nikolai D | Program object to support connection generation |
US20060031388A1 (en) | 2004-06-03 | 2006-02-09 | Tankov Nikolai D | Connection resource system |
US20050278439A1 (en) * | 2004-06-14 | 2005-12-15 | Ludmila Cherkasova | System and method for evaluating capacity of a heterogeneous media server configuration for supporting an expected workload |
US20060209718A1 (en) | 2005-03-17 | 2006-09-21 | Microsoft Corporation | Enhanced network system through the combination of network objects |
US8286174B1 (en) | 2006-04-17 | 2012-10-09 | Vmware, Inc. | Executing a multicomponent software application on a virtualized computer platform |
US8429652B2 (en) | 2009-06-22 | 2013-04-23 | Citrix Systems, Inc. | Systems and methods for spillover in a multi-core system |
US20100333111A1 (en) | 2009-06-29 | 2010-12-30 | Software Ag | Systems and/or methods for policy-based JMS broker clustering |
US20110149737A1 (en) | 2009-12-23 | 2011-06-23 | Manikam Muthiah | Systems and methods for managing spillover limits in a multi-core system |
US20110178790A1 (en) | 2010-01-20 | 2011-07-21 | Xyratex Technology Limited | Electronic data store |
US20110202925A1 (en) | 2010-02-18 | 2011-08-18 | International Business Machines Corporation | Optimized capacity planning |
US8423998B2 (en) | 2010-06-04 | 2013-04-16 | International Business Machines Corporation | System and method for virtual machine multiplexing for resource provisioning in compute clouds |
US20120284712A1 (en) | 2011-05-04 | 2012-11-08 | Chitti Nimmagadda | Systems and methods for sr-iov pass-thru via an intermediary device |
US20130290957A1 (en) * | 2012-04-26 | 2013-10-31 | International Business Machines Corporation | Efficient execution of jobs in a shared pool of resources |
US20140173595A1 (en) | 2012-12-17 | 2014-06-19 | International Business Machines Corporation | Hybrid virtual machine configuration management |
US20140173113A1 (en) | 2012-12-19 | 2014-06-19 | Symantec Corporation | Providing Optimized Quality of Service to Prioritized Virtual Machines and Applications Based on Quality of Shared Resources |
US20150331707A1 (en) | 2012-12-21 | 2015-11-19 | Telefonaktiebolaget Lm Ericsson (Publ) | Method and cloud management node for enabling a virtual machine |
US20140222889A1 (en) | 2013-02-05 | 2014-08-07 | International Business Machines Corporation | Collaborative negotiation of system resources among virtual servers running in a network computing environment |
US9692820B2 (en) * | 2013-04-06 | 2017-06-27 | Citrix Systems, Inc. | Systems and methods for cluster parameter limit |
US20150040128A1 (en) | 2013-08-05 | 2015-02-05 | International Business Machines Corporation | Utilizing Multiple Memory Pools During Mobility Operations |
US20160004551A1 (en) | 2013-10-04 | 2016-01-07 | Hitachi, Ltd. | Resource management system and resource management method |
US9407692B2 (en) * | 2013-11-27 | 2016-08-02 | Avi Networks | Method and system for distributed load balancing |
US20150244643A1 (en) | 2014-02-21 | 2015-08-27 | Hitachi, Ltd. | Data center resource allocation system and data center resource allocation method |
US20150263982A1 (en) | 2014-03-12 | 2015-09-17 | International Business Machines Corporation | Software defined infrastructures that encapsulate physical server resources into logical resource pools |
US20160092277A1 (en) | 2014-09-25 | 2016-03-31 | International Business Machines Corporation | Overcommitting virtual machine hosts |
US9870244B2 (en) * | 2014-12-29 | 2018-01-16 | International Business Machines Corporation | Optimized migration of virtual objects across environments in a cloud computing environment |
US9558039B2 (en) * | 2015-03-13 | 2017-01-31 | International Business Machines Corporation | Managing resources of a shared pool of configurable computing resources |
Non-Patent Citations (8)
Title |
---|
A. Makela, Concept for providing guaranteed service level over an array of unguaranteed commodity connections, 6 pages (Year: 2010). * |
Clayman et al., INOX: a managed service platform for inter-connected smart objects, 8 paged (Year: 2011). * |
List of IBM Patents or Patent Applications Treated as Related, dated Oct. 13, 2016, pp. 1-2. |
Meng et al., "Efficient Resource Provisioning in Compute Clouds via VM Multiplexing," Proceeding ICAC'10 Proceedings of the 7th lntemational Conference on Autonomic Computing, Jun. 7-11, 2010, pp. 11-20, ISBN: 978-1-4503-00742. |
Mick Jordan, "Policy-Based Management of a JDBC Connection Pool," Feb. 2006, pp. 1-21. |
Paturde et al., "Integration of Heterogeneous Databases and Creating Materialized View," International Journal of Advanced Electronics & Communication Systems, vol. 3, Issue 3, 2014, pp. 1-4. |
Soror et al., "Automatic Virtual Machine Configuration for Database Workloads," ACM Transactions on Database Systems, vol. 35, No. 1, Article 7, Feb. 2010, pp. 1-47. |
Xiaoli et al., "XML-Based Heterogeneous Database Integration System Design and Implementation," 3rd IEEE International Conference on Computer Science and Information Technology, Chengdu, 2010, pp. 547-550. |
Also Published As
Publication number | Publication date |
---|---|
US11429443B2 (en) | 2022-08-30 |
US20160092271A1 (en) | 2016-03-31 |
US9563475B2 (en) | 2017-02-07 |
US20190155654A1 (en) | 2019-05-23 |
US20170031716A1 (en) | 2017-02-02 |
US20200264929A1 (en) | 2020-08-20 |
US10740147B2 (en) | 2020-08-11 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11429443B2 (en) | Merging connection pools to form a logical pool of connections during a preset period of time thereby more efficiently utilizing connections in connection pools | |
US11593149B2 (en) | Unified resource management for containers and virtual machines | |
US10514960B2 (en) | Iterative rebalancing of virtual resources among VMs to allocate a second resource capacity by migrating to servers based on resource allocations and priorities of VMs | |
US9760395B2 (en) | Monitoring hypervisor and provisioned instances of hosted virtual machines using monitoring templates | |
US10169028B2 (en) | Systems and methods for on demand applications and workflow management in distributed network functions virtualization | |
US11068361B2 (en) | Cluster file system support for extended network service addresses | |
US10394475B2 (en) | Method and system for memory allocation in a disaggregated memory architecture | |
US9588797B2 (en) | Returning terminated virtual machines to a pool of available virtual machines to be reused thereby optimizing cloud resource usage and workload deployment time | |
US9413819B1 (en) | Operating system interface implementation using network-accessible services | |
US9571374B2 (en) | Dynamically allocating compute nodes among cloud groups based on priority and policies | |
US11461149B1 (en) | Capacity management in provider networks using dynamic host device instance model reconfigurations | |
US10594781B2 (en) | Dynamic healthchecking load balancing gateway | |
US10171445B2 (en) | Secure virtualized servers | |
US20170134339A1 (en) | Management of clustered and replicated systems in dynamic computing environments | |
KR20190117801A (en) | Virtualizing graphics processing in a provider network | |
US9736231B1 (en) | Representational state transfer request routing | |
US20170285964A1 (en) | Tuning memory across database clusters for distributed query stability | |
US11307889B2 (en) | Schedule virtual machines | |
US20240012692A1 (en) | Dynamic light-weighted multi-tenancy |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: INTERNATIONAL BUSINESS MACHINES CORPORATION, NEW Y Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:JAIN, RISPNA;RAMACHANDRA, ANOOP GOWDA MALALY;SIGNING DATES FROM 20140911 TO 20140912;REEL/FRAME:040008/0310 |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT VERIFIED |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 4 |