WO2008155124A2 - Load balancing - Google Patents

Load balancing Download PDF

Info

Publication number
WO2008155124A2
WO2008155124A2 PCT/EP2008/004963 EP2008004963W WO2008155124A2 WO 2008155124 A2 WO2008155124 A2 WO 2008155124A2 EP 2008004963 W EP2008004963 W EP 2008004963W WO 2008155124 A2 WO2008155124 A2 WO 2008155124A2
Authority
WO
WIPO (PCT)
Prior art keywords
cpus
load
physical
virtual
cpu
Prior art date
Application number
PCT/EP2008/004963
Other languages
French (fr)
Other versions
WO2008155124A3 (en
Inventor
Vladimir Grouzdev
Original Assignee
Virtuallogix Sa
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Virtuallogix Sa filed Critical Virtuallogix Sa
Publication of WO2008155124A2 publication Critical patent/WO2008155124A2/en
Publication of WO2008155124A3 publication Critical patent/WO2008155124A3/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/46Multiprogramming arrangements
    • G06F9/50Allocation of resources, e.g. of the central processing unit [CPU]
    • G06F9/5083Techniques for rebalancing the load in a distributed system
    • G06F9/5088Techniques for rebalancing the load in a distributed system involving task migration

Definitions

  • a virtual machine is a self-contained execution environment that behaves as if it is a separate computer and which can run its own operating system.
  • Virtual machines provide Virtual CPUs (VCPUs) to clients or "guests", and each VCPU runs on a dedicated physical CPU.
  • VPU Virtual CPUs
  • a VCPU is a representation of a physical processor within a Virtual Machine. In conventional systems, the mapping between virtual and physical CPUs is static.
  • the guest tasks are spread between the CPUs. This is preferably done such that the available resources are used in the most efficient way and computing time is decreased. This process is generally referred to as "load balancing”.
  • Fig. 1 illustrates schematically the architecture of a system to which the present invention can be applied.
  • Figs. 2 to 5 illustrate screen shots showing CPU usage statistics with and without load balancing in accordance with an embodiment of the invention.
  • Fig. 1 illustrates schematically the architecture of a system comprising physical
  • CPUs virtual CPUs
  • client (guest) applications to which the present invention can be applied.
  • the present invention is based on the realisation that the overall performance of a system as shown in Fig. 1 can be improved by automatically balancing the load on physical CPUs attributed to the same SMP guest.
  • the idea is to balance the physical CPUs load by migrating a virtual CPU (VCPU) from one physical CPU to another.
  • VCPU virtual CPU
  • the VCPUs to CPUs mapping within a fixed set of CPUs is transparent for SMP guests such that a scheduler which implements the present invention can decide which VCPU is to run on which physical CPU within the fixed CPU set.
  • the scheduler creates two VCPUs (x & y) and maps them to the physical CPUs.
  • two equivalent mappings are possible:
  • the scheduler is able to dynamically choose one of these mappings depending on CPU loads.
  • the mapping switch results in a "swapping" of VCPUs, i.e. in two VCPUs migrating from one physical CPU to another. Such an operation is fully transparent for the guest and does not change a fixed physical set of CPUs assigned to the guest.
  • the scheduler calculates the load of each physical
  • the positive load for a given VCPU n m is equal to the actual VCPU load:
  • VPLOAD n)m VLOAD n , m
  • the negative load for a given VCPU n m is equal to a sum of the loads of all other VCPUs running on the same physical CPU:
  • a physical CPU is considered overloaded if its load is above a predetermined threshold:
  • Load balancing is only applied to a pair of CPUs in which one CPU is overloaded and other CPU is underloaded:
  • the load balancing comprises finding two unbalanced VCPUs of the same SMP guest running on CPUj and CPU j such that:
  • the migration criteria is adjusted by introducing a positive migration threshold: VPLOAD i;k - VPLO ADy > MIGR _P OS JVTMARK
  • the migration criteria takes into account the negative load of the overloaded emigrant:
  • the negative load water mark avoids unnecessary migrations when the CPU overloading is not caused by a simultaneous activity of multiple guests, but rather by a single guest monopolizing the physical CPU.
  • a mymips program has been used to demonstrate skewing in the SMP load balancing of a guest operating system.
  • the mymips program permanently calculates the program execution speed (MIPS) and prints out the calculation results on console.
  • MIPS program execution speed
  • mymips provides the following result when running on a SMP Linux guest with a dedicated single physical CPU:
  • a basic configuration which can be used to implement the load balancing mechanism in accordance with an embodiment of the invention comprises two SMP Linux guests sharing two physical CPUs, hi order to obtain an unbalanced load on such a configuration, guests have been running on a conventional system without load balancing mechanism. Two mymips programs running simultaneously on each guest provide the following results:
  • the performance hit is due to sporadic mymips migrations from one CPU to another. Such a migration randomly runs mymips on the same processor.
  • FIGs. 2 illustrates a screen shot showing CPU usage statistics without load balancing in accordance with an embodiment of the invention. The screen shot represents the scenario described above.
  • the load balancing compensates sporadic migrations o ⁇ mymips (from one VCPU to another) caused by the Linux SMP scheduler. In other words, the scheduler tries to execute heavy loaded VCPUs on different physical CPUs.
  • Figs. 3 illustrates a screen shot showing CPU usage statistics with load balancing in accordance with an embodiment of the invention.
  • the screen shot represents the scenario described above.
  • the performance hit on the system without load balancing (2) is about 40%, while the performance hit on the system with load balancing (3) is about 20%. Accordingly, the load balancing improves a performance degradation caused by a transparent CPU sharing among multiple SMP guests.
  • Figs. 4 and 5 illustrate screen shots showing CPU usage statistics with and without load balancing in accordance with an embodiment of the invention.
  • the screen shots represent the scenarios described above.
  • the above results were obtained using the following load balancing parameters:

Landscapes

  • Engineering & Computer Science (AREA)
  • Software Systems (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Multi Processors (AREA)
  • Stored Programmes (AREA)

Abstract

In a preferred embodiment, the present invention provides a method of load balancing in a data processing system comprising a plurality of physical CPUs and a plurality of virtual CPUs, the method comprising: mapping one or more virtual CPUs to each of said physical CPUs; and dynamically adapting the mapping depending on the load of said physical CPUs.

Description

LOAD BALANCING
Background
A virtual machine is a self-contained execution environment that behaves as if it is a separate computer and which can run its own operating system. Virtual machines provide Virtual CPUs (VCPUs) to clients or "guests", and each VCPU runs on a dedicated physical CPU. A VCPU is a representation of a physical processor within a Virtual Machine. In conventional systems, the mapping between virtual and physical CPUs is static.
If multiple CPUs are available to a client or "guest", the guest tasks are spread between the CPUs. This is preferably done such that the available resources are used in the most efficient way and computing time is decreased. This process is generally referred to as "load balancing".
Conventional load balancing algorithms may be insufficient. Let us consider, for example, the sharing of a plurality of physical CPUs between dedicated real-time software and generic server software. Let us assume an UP (uniprocessor) execution environment (e.g. LINUX) running the real-time software on CPU 0, and an SMP (symmetric multiprocessing) execution environment (LINUX) running a server software on CPUs 0-3. In this example, CPU 0 is shared between the the real-time software and the server software. The dedicated real-time has a higher scheduling priority. In this example, the SMP load balancer does not take into account the realtime activity on CPU 0. This may skew the SMP load balancing. The present invention aims to address this and other problems of conventional load balancing. In particular, but not exclusively, the present invention is concerned with better balancing the load of physical CPUs in a computer system comprising physical and virtual CPUs.
Summary of the Invention
The invention is recited by the independent claims. Preferred features are recited by the dependent claims.
Brief Description of the Drawings
Fig. 1 illustrates schematically the architecture of a system to which the present invention can be applied; and
Figs. 2 to 5 illustrate screen shots showing CPU usage statistics with and without load balancing in accordance with an embodiment of the invention.
Description of Exemplary Embodiments of the Invention
Introduction
Fig. 1 illustrates schematically the architecture of a system comprising physical
CPUs, virtual CPUs, and client (guest) applications, to which the present invention can be applied.
The present invention is based on the realisation that the overall performance of a system as shown in Fig. 1 can be improved by automatically balancing the load on physical CPUs attributed to the same SMP guest. In particular, the idea is to balance the physical CPUs load by migrating a virtual CPU (VCPU) from one physical CPU to another. Although this affects a statical VCPU to CPU assignment for a UP guest, the VCPUs to CPUs mapping within a fixed set of CPUs is transparent for SMP guests such that a scheduler which implements the present invention can decide which VCPU is to run on which physical CPU within the fixed CPU set. In other words, if an SMP guest is running on two physical CPUs (a & b), the scheduler creates two VCPUs (x & y) and maps them to the physical CPUs. In this example, two equivalent mappings are possible:
(1) VCPUx → CPU3 and VCPUy → CPUb
(2) VCPUx → CPUb and VCPUy → CPUa
The scheduler is able to dynamically choose one of these mappings depending on CPU loads. The mapping switch results in a "swapping" of VCPUs, i.e. in two VCPUs migrating from one physical CPU to another. Such an operation is fully transparent for the guest and does not change a fixed physical set of CPUs assigned to the guest.
By implementing such a load balancing mechanism, it is possible to at least partially resolve the above described SMP load balancing skewing problem by migrating a server VCPU running on CPUo, f°r example, to another CPU when the real-time activity is high and this VCPU is loaded. In order to improve the overall performance by such a migration, an underloaded CPUn (n > 0, in this example)
must be found and VCPU running on CPUn must migrate to CPUQ. This solution is partial only in that it does not work when the system is heavily loaded, i.e. when all physical CPUs are fully loaded. However, as such a situation is rare in practice, it is acceptable.
Migration criteria
Regularly, at a given time period, the scheduler calculates the load of each physical
CPU (LOADn, n=0..N) and all VCPUs running on each physical CPU (VLOAD^m, m=0..M). More particularly, two loads are calculated for each VCPU: a "positive" load and a "negative" load.
The positive load for a given VCPUn m is equal to the actual VCPU load:
VPLOADn)m = VLOADn,m
The negative load for a given VCPUn m is equal to a sum of the loads of all other VCPUs running on the same physical CPU:
VNLOADn>m = ∑VLOADn)i i=0..M, i ≠ m
A physical CPU is considered overloaded if its load is above a predetermined threshold:
LOADn > LOADover A physical CPU is considered underloaded if its load is below a predetermined threshold:
LOADn < LOADunder
Load balancing is only applied to a pair of CPUs in which one CPU is overloaded and other CPU is underloaded:
CPUi ^ CPUj where CPUi ≥ LOADover and CPUj < LOADunder
The load balancing comprises finding two unbalanced VCPUs of the same SMP guest running on CPUj and CPUj such that:
VPLOADi)k > VPLOADμ
and swapping these VCPUs across physical CPUs.
Because a VCPU migration, in terms of processing power, is a quite expensive operation, the migration criteria is adjusted by introducing a positive migration threshold: VPLOADi;k - VPLO ADy > MIGR _P OS JVTMARK
In addition, the migration criteria takes into account the negative load of the overloaded emigrant:
VNL0ADi>k > MlGRJNEG JVTMARK
The negative load water mark avoids unnecessary migrations when the CPU overloading is not caused by a simultaneous activity of multiple guests, but rather by a single guest monopolizing the physical CPU.
Specific Implementation
A mymips program has been used to demonstrate skewing in the SMP load balancing of a guest operating system. The mymips program permanently calculates the program execution speed (MIPS) and prints out the calculation results on console.
mymips provides the following result when running on a SMP Linux guest with a dedicated single physical CPU:
min/max/ave: 258/258/258
The results above and below were obtained on a DELL D820 Dual Core 1.8 MHz Laptop. Two mymips programs provide the following result when running on an SMP Linux with two dedicated CPUs:
min/max/ave: 257/259/258 min/max/ave: 258/259/258
A basic configuration which can be used to implement the load balancing mechanism in accordance with an embodiment of the invention comprises two SMP Linux guests sharing two physical CPUs, hi order to obtain an unbalanced load on such a configuration, guests have been running on a conventional system without load balancing mechanism. Two mymips programs running simultaneously on each guest provide the following results:
min/max/ave: 101/258/190 min/max/ave: 92/257/190
This shows about 25% of performance hit comparing to a single SMP Linux which performs a load balancing (at OS level) across multiple CPUs. The performance hit is due to sporadic mymips migrations from one CPU to another. Such a migration randomly runs mymips on the same processor.
This practical result is fully in line with a theoretical determination. Because of a random nature of migrations, the probability of running both mymips on the same CPU is 0.5. Thus, an expected performance hit is 0.25 because when running two programs on the same CPU, only a half of the CPU power is available. Figs. 2 illustrates a screen shot showing CPU usage statistics without load balancing in accordance with an embodiment of the invention. The screen shot represents the scenario described above.
When running the same load with load balancing enabled, the performance is close to a single SMP Linux.
min/max/ave: 254/257/256 min/max/ave: 250/256/255
The load balancing compensates sporadic migrations oϊmymips (from one VCPU to another) caused by the Linux SMP scheduler. In other words, the scheduler tries to execute heavy loaded VCPUs on different physical CPUs.
Figs. 3 illustrates a screen shot showing CPU usage statistics with load balancing in accordance with an embodiment of the invention. The screen shot represents the scenario described above.
In order to confirm the above theoretical conclusion, a Linux kernel compilation was used as a variable load. Two compilations were running in parallel on two Linux guests in the following three configurations:
(1) Two Linux guests each running on a dedicated CPU (2) Two Linux guests sharing two CPUs without load balancing (3) Two Linux guests sharing two CPUs with load balancing
Each time, the duration of compilation was measured. Results corresponding to different Linux kernel compilations are shown below.
(1.1) l lm21.046s
(1.2) 8m34.204s
(2.1) 16m4.272s
(2.2) 12m20.575s
(3.1) 13m51.974s
(3.2) 10m32.467s
The performance hit on the system without load balancing (2) is about 40%, while the performance hit on the system with load balancing (3) is about 20%. Accordingly, the load balancing improves a performance degradation caused by a transparent CPU sharing among multiple SMP guests.
Figs. 4 and 5 illustrate screen shots showing CPU usage statistics with and without load balancing in accordance with an embodiment of the invention. The screen shots represent the scenarios described above. The above results were obtained using the following load balancing parameters:
PERIOD = 10 milliseconds LOADunder = 80%
LOADOver = 100%
MIGR_POS_WTMARK= 5% MIGR_NEG JVTMARK = 5%
It may be possible to achieve even better result on a variable load by modifying these parameters.
Other aspects and embodiments
It will be clear from the forgoing that the above-described embodiments are only examples, and that other embodiments are possible and included within the scope of the invention as determined from the claims.

Claims

Claims
1. A method of load balancing in a data processing system comprising a plurality of physical CPUs and a plurality of virtual CPUs, the method comprising: mapping one or more virtual CPUs to each of said physical CPUs; and dynamically adapting the mapping depending on the load of said physical CPUs.
2. The method of claim 1 , comprising: running a multiprocessor operation involving at least two physical CPUs; and assigning at least two virtual CPUs to said multiprocessor operation, for executing said multiprocessor operation; and mapping the at least two virtual CPUs to said at least two physical CPUs, respectively, wherein the mapping depends on the load of said at least two physical CPUs.
3. The method of claim 2, comprising: swapping the mapping of the two virtual CPUs to the at least two physical CPUs, respectively, in response to a change of load of at least one of the two physical CPUs.
4. The method of claim 2 or 3, comprising: determining the load of said two virtual CPUs; determining whether the load of said two physical CPUs is above a first threshold or below a second threshold; and swapping the mapping of the two virtual CPUs to the two physical CPUs, respectively, if the following conditions are met:
(i) the load of one of the two physical CPUs is above said first threshold, (ii) the load of the other one of the two physical CPUs is below said second threshold, and
(iii) the load of a virtual CPU mapped to the physical CPU whose load is above said first threshold is higher than the load of a virtual CPU mapped to the physical CPU whose load is below said second threshold.
5. The method of claim 4, comprising: swapping the mapping only if the difference of loads of said two virtual CPUs is above a predetermined threshold.
6. The method of claim 4 or 5, comprising: determining, for a given virtual CPU, a negative load indicative of the load of all other virtual CPUs that are allocated to the same physical CPU as the given virtual CPU; and swapping the mapping only if the negative load is above a predetermined threshold.
7. The method of claim 2, comprising: running an operation of a high scheduling priority on one of said at least two physical CPUs, wherein said multiprocessor operation has a relatively lower scheduling priority.
8. The method of claim 7, wherein the operation of a higher scheduling priority is executed by a dedicated real-time software.
9. The method of claim 7 or 8, wherein said multiprocessor operation is executed by a generic server software.
10. Computer system, arranged to perform the method of any preceding claim.
11. Computer program product, comprising machine-readable code which, when executed by a data processing device, executes the method of any of claims 1 to 9.
PCT/EP2008/004963 2007-06-19 2008-06-19 Load balancing WO2008155124A2 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
EP07290793.4 2007-06-19
EP07290793.4A EP2006770B1 (en) 2007-06-19 2007-06-19 Load balancing

Publications (2)

Publication Number Publication Date
WO2008155124A2 true WO2008155124A2 (en) 2008-12-24
WO2008155124A3 WO2008155124A3 (en) 2009-03-05

Family

ID=38657663

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/EP2008/004963 WO2008155124A2 (en) 2007-06-19 2008-06-19 Load balancing

Country Status (3)

Country Link
US (1) US8341630B2 (en)
EP (2) EP2006770B1 (en)
WO (1) WO2008155124A2 (en)

Families Citing this family (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8970603B2 (en) 2010-09-30 2015-03-03 Microsoft Technology Licensing, Llc Dynamic virtual device failure recovery
US9069622B2 (en) 2010-09-30 2015-06-30 Microsoft Technology Licensing, Llc Techniques for load balancing GPU enabled virtual machines
US20120137295A1 (en) * 2010-11-30 2012-05-31 Alcatel-Lucent Canada Inc. Method for displaying cpu utilization in a multi-processing system
US8830245B2 (en) * 2010-12-14 2014-09-09 Amazon Technologies, Inc. Load balancing between general purpose processors and graphics processors
CN102323897B (en) * 2011-05-12 2013-06-12 中山爱科数字科技股份有限公司 Method for internal business processing and load balancing for platform
US9104508B2 (en) 2012-01-18 2015-08-11 International Business Machines Corporation Providing by one program to another program access to a warning track facility
US9110878B2 (en) 2012-01-18 2015-08-18 International Business Machines Corporation Use of a warning track interruption facility by a program
US8850450B2 (en) 2012-01-18 2014-09-30 International Business Machines Corporation Warning track interruption facility
US20150081400A1 (en) * 2013-09-19 2015-03-19 Infosys Limited Watching ARM
US10715587B2 (en) * 2014-04-11 2020-07-14 Maxeler Technologies Ltd. System and method for load balancing computer resources
US11996988B1 (en) 2023-04-27 2024-05-28 Dell Products Lp Reinforced computer learning system and method for minimizing power consumed by underutilized data center hardware components

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6370560B1 (en) * 1996-09-16 2002-04-09 Research Foundation Of State Of New York Load sharing controller for optimizing resource utilization cost

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6298370B1 (en) * 1997-04-04 2001-10-02 Texas Instruments Incorporated Computer operating process allocating tasks between first and second processors at run time based upon current processor load
US7028302B2 (en) * 2002-04-24 2006-04-11 Hewlett-Packard Development Company, L.P. System and method for automatically tuning a multiprocessor computer system
US20050055694A1 (en) * 2003-09-04 2005-03-10 Hewlett-Packard Development Company, Lp Dynamic load balancing resource allocation

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6370560B1 (en) * 1996-09-16 2002-04-09 Research Foundation Of State Of New York Load sharing controller for optimizing resource utilization cost

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
E. PIEL, P. MARQUET, J. SOULA, AND J.-L. DEKEYSER: "LOAD-BALANCING FOR A REAL-TIME SYSTEM BASED ON ASYMMETRIC MULTI-PROCESSING" 16TH EUROMICRO CONFERENCE ON REAL-TIME SYSTEMS, WIP SESSION, June 2004 (2004-06), XP002458252 Catania, Italy *
JACOB J C ET AL: "A spanning tree based recursive refinement algorithm for fast task mapping" PROCEEDINGS OF THE FOURTH IEEE INTERNATIONAL SYMPOSIUM ON HIGH PERFORMANCE DISTRIBUTED COMPUTING (CAT. NO.95TB8075) IEEE COMPUT. SOC. PRESS LOS ALAMITOS, CA, USA, 1995, pages 58-65, XP002458251 ISBN: 0-8186-7088-6 *

Also Published As

Publication number Publication date
EP2527980A2 (en) 2012-11-28
EP2006770A1 (en) 2008-12-24
EP2527980A3 (en) 2013-01-09
US20080320489A1 (en) 2008-12-25
US8341630B2 (en) 2012-12-25
EP2006770B1 (en) 2014-01-01
WO2008155124A3 (en) 2009-03-05

Similar Documents

Publication Publication Date Title
US8341630B2 (en) Load balancing in a data processing system having physical and virtual CPUs
JP5473530B2 (en) Multi-core processor for configuring core functions based on executing instructions and method of use
US11681544B2 (en) Interference-aware scheduling service for virtual GPU enabled systems
US20100077185A1 (en) Managing thread affinity on multi-core processors
US11876731B2 (en) System and methods for sharing memory subsystem resources among datacenter applications
US9354934B2 (en) Partitioned shared processor interrupt-intensive task segregator
Guo et al. Flexslot: Moving hadoop into the cloud with flexible slot management
JP6778130B2 (en) Virtual computer system and its resource allocation method
EP2090983A1 (en) Determining an architecture for executing code in a multi architecture environment
US20170286168A1 (en) Balancing thread groups
Chanchio et al. Time-bound, thread-based live migration of virtual machines
US9753770B2 (en) Register-type-aware scheduling of virtual central processing units
Wu et al. Dynamic CPU allocation for Docker containerized mixed-criticality real-time systems
Pan et al. Hypervisor support for efficient memory de-duplication
Garg et al. Empirical analysis of hardware-assisted gpu virtualization
KR101330609B1 (en) Method For Scheduling of Mobile Multi-Core Virtualization System To Guarantee Real Time Process
Kim et al. Platform and co-runner affinities for many-task applications in distributed computing platforms
Ermakov et al. Testing docker performance for HPC applications
Cheng et al. A user-level numa-aware scheduler for optimizing virtual machine performance
Sun et al. A virtual machine based task scheduling approach to improving data locality for virtualized Hadoop
Ghit et al. Tyrex: Size-based resource allocation in mapreduce frameworks
Blagodurov et al. Towards the contention aware scheduling in hpc cluster environment
US11256540B2 (en) Server-to-container migration
US10628223B2 (en) Optimized allocation of tasks in heterogeneous computing systems
Lee et al. Minimizing scheduling delay for multimedia in xen hypervisor

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 08784523

Country of ref document: EP

Kind code of ref document: A2

NENP Non-entry into the national phase

Ref country code: DE

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 16/04/2010)

122 Ep: pct application non-entry in european phase

Ref document number: 08784523

Country of ref document: EP

Kind code of ref document: A2