TW201102805A - Hypervisor-based facility for communicating between a hardware management console and a logical partition - Google Patents

Hypervisor-based facility for communicating between a hardware management console and a logical partition Download PDF

Info

Publication number
TW201102805A
TW201102805A TW99106799A TW99106799A TW201102805A TW 201102805 A TW201102805 A TW 201102805A TW 99106799 A TW99106799 A TW 99106799A TW 99106799 A TW99106799 A TW 99106799A TW 201102805 A TW201102805 A TW 201102805A
Authority
TW
Taiwan
Prior art keywords
hypervisor
target
endpoint
request
primitive
Prior art date
Application number
TW99106799A
Other languages
Chinese (zh)
Other versions
TWI463304B (en
Inventor
Gary D Anderson
Curtis S Eide
Jonathan L Kaus
Steven E Royer
Original Assignee
Ibm
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from US12/403,402 external-priority patent/US8230077B2/en
Application filed by Ibm filed Critical Ibm
Publication of TW201102805A publication Critical patent/TW201102805A/en
Application granted granted Critical
Publication of TWI463304B publication Critical patent/TWI463304B/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F11/00Error detection; Error correction; Monitoring
    • G06F11/07Responding to the occurrence of a fault, e.g. fault tolerance
    • G06F11/0703Error or fault processing not based on redundancy, i.e. by taking additional measures to deal with the error or fault not making use of redundancy in operation, in hardware, or in data representation
    • G06F11/0706Error or fault processing not based on redundancy, i.e. by taking additional measures to deal with the error or fault not making use of redundancy in operation, in hardware, or in data representation the processing taking place on a specific hardware platform or in a specific software environment
    • G06F11/0712Error or fault processing not based on redundancy, i.e. by taking additional measures to deal with the error or fault not making use of redundancy in operation, in hardware, or in data representation the processing taking place on a specific hardware platform or in a specific software environment in a virtual computing platform, e.g. logically partitioned systems

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Mathematical Physics (AREA)
  • Quality & Reliability (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Computer And Data Communications (AREA)

Abstract

A hypervisor-based facility is provided for communicating between a hardware management console (HMC) and a logical partition of a data processing system. The facility includes: packaging a request or response of a source endpoint as cargo in a generic transport primitive, the source endpoint being either an HMC or a logical partition of the data processing system; and forwarding the generic transport primitive from the source endpoint to a target endpoint via the hypervisor. The forwarding includes receiving the transport primitive at the hypervisor and forwarding the cargo of the transport primitive to the target endpoint. The cargo includes the request or response from the source endpoint, and the hypervisor forwards the cargo absent inspection or parsing of that cargo. The target endpoint is the other one of the logical partition or the hardware management console of the data processing system.

Description

201102805 六、發明說明: 【發明所屬之技術領域】 本發明大體而言係關於資料處理系统,且更特定言之, 係關於-種在-經邏肖分割㈣料處理系統之硬體管理控 制台與邏輯分割區之間的以超管理器為基礎之通用通信設 施0 * 此申請案主張2008年6月6曰申請之題為^ Memory」之美國臨時申請案第61/〇59,492號之權利,該案 之全部内容以引用的方式併入本文中。 【先前技術】 在複雜電腦系統資源之㈣方㈣—新近發展為對系統 資源之邏輯分割。在概念上,邏輯分割意謂建立多個離散 分割區’且將特定類型之系統資源指派至各別分割區。舉 例而言’可藉由將不同處理器指派至不同分割區、藉由在 -些分割區(而非其他)之間共用處理器、藉由指定對於正 :用:組處理器之每一分割區而言可用的處理資源度量的 1等等來分割多處理器系統之處理器資源。在一邏輯分割 區内執行之任務僅可使用經指派至彼分割區之資源而非 經指派至另一分割區之資源。 一般地’藉由一體現為低層級經編石馬可執行指令及資料 之分割管理器來強制執行邏輯分割,雖然可能存在一定量 的對邏輯分割之硬體支援(諸如,保存狀態資訊之專用: 體暫存器)。低層級程式碼函式及/或硬體阻止對分配至不 同分割區之資源的存取。一般地,邏輯分割管理器之某一 146527.doc 201102805 部分包括一用於管理強制執行邏輯分割之低層級程式碼函 式的使用者介面。此邏輯分割管理器介面意欲供單一或一 小群授權使用者(亦即,系統管理員)使用。在本文中使用 時,此低層級邏輯分割程式碼稱為超管理器,且分割管理 器介面稱為硬體管理控制台(HMC)。在資料處理系統之 HMC與邏輯分割區之間的通信對於(例 >)同時硬體維護、 動態邏輯分割、清查收集(invent〇ry c〇Uecti〇n)、虛擬輸入/ 輸出(I/O)器件映射等可為理想的。 在資料處理系統之HMC與邏輯分割區之間的一通信方法 利用用於HMC與高強運算架構平台要求(pApR)(亦即, AIX®分割區及LINUX分割區)之間之通信的以資源監視及 控制(RMC)為基礎之設施。(ΑΙχ®為位於美國紐約阿蒙克 (Annonk,New York,U.s.A·)之國際商業機器公司之註冊商 標。本文中使用之其他名稱可為國際商業機器公司或其他 公司之註冊商標、商標或產品名稱遺憾的是,rmc解 決方案要求HMC與分割區之間的真實LAN連接。與真實 LAN連接相關聯的是額外硬體要求(LAN配接器及電纜 線)、額外組態任務(網路管理)及額外的潛在故障點(lan 連接)。 【發明内容】 在一經邏輯分割的資料 邏輯分割區之間通信的 源端點將該源端點之一 在本文中’在一態樣中提供一種在一 泛用傳送基元中,該源端 處理系統之一硬體管理控制台與一邏 電腦實施方法。該方法包括:由一源 請求或一回應作為貨物封裝於一泛用 146527.doc 201102805 點為該資料處理系統之一硬體管理控制台或一邏輯分割區 中之一者,其中該硬體管理控制台為用於分割管理之一使 用者;I面,及經由έ玄資料處理糸統之一超管理器將該泛用 傳送基元自該源端點轉遞至一目標端點,其中該超管理器 接收在該源端點處經封裝之該泛用傳送基元且將該泛用傳 送基元之該貨物轉遞至該目標端點,該貨物包含該請求或 該回應,且其中在由該超管理器進行之該接收及該轉遞中 未由該超官理器對該貨物進行檢驗或剖析,且該目標端點 為該資料處理系統之該邏輯分割區或該硬體管理控制台中 之另一者。 在另一態樣中,提供一種經邏輯分割的資料處理系統。 該經邏輯分割的資料處理系統包括:至少一處理器,其包 含至少一邏輯分割區;至少一外部硬體管理控制台;及一 超管理器,其使該至少一硬體管理控制台與該至少一邏輯 分割區介面連接。每一硬體管理控制台為用於分割管理之 一使用者介面。該超管理器包括用於經由該超管理器在該 至少一硬體管理控制台與該至少一邏輯分割區之間通信的 一通信設施。該通信包括:由一源端點將一請求或一回應 作為貨物封裝於一泛用傳送基元中,該源端點為該至少一 硬體管理控制台中之一硬體管理控制台或該至少一邏輯分 割區中之一邏輯分割區;及經由該超管理器將該泛用傳送 基元自該源端點轉遞至一目標端點,其中該超管理器接收 在該源端點處經封裝之該泛用傳送基元且將該泛用傳送基 元之該貨物轉遞至該目標端點,該貨物包括該源端點之該 146527.doc 201102805 = ’且其中在由該超管理器進行之該接收及轉 遞中未由該超管理器對該貨物進行檢驗或剖析,且該目桿 端點為該至少—邏輯分割區中之該邏輯分割區或該至少一 硬體管理控制台中之該硬體管理控制台中的另一者。 在另一態樣中,提供一猶句虹 ’、 至〉、一電腦可讀媒體之製 品’該至少-電腦可讀媒體具有用以促進_經邏輯分割的 資料處理糸統之-硬體管理控制台與—邏輯分割區之間之 通信的電腦可讀程式碼邏輯。當在一處理器上執行時,該 電腦可讀程式碼邏輯執行以下叙彳七.丄 钒仃下動作·由一源端點將該源端 點之-請求或1應作為貨物封裝於—泛用傳送基元中, 該源端點為該資料處理系統之一硬體管理控制台或一邏輯 分割區中之-者’纟中該硬體管理控制台為用於分割管理 之一使用者介面;及經由該資料處理系統之—超管理器將 該泛用傳送基元自該源端點轉遞至—目標端點其中該超 管理器接收在該源端點處經封裝之該泛用傳送基元且將古玄 泛用傳送基7G之該貨物轉遞至該目標端點,該貨物包含該 源端點之該請求或該回應,且其中在由該超管理器進行之 該接收及該轉遞中未由該超管理器對該貨物進行檢驗或剖201102805 VI. Description of the Invention: [Technical Field of the Invention] The present invention relates generally to a data processing system, and more particularly to a hardware management console for a system-based processing system. A hypervisor-based universal communication facility between the logical partitions and the logical partitions. * This application claims the benefit of U.S. Provisional Application No. 61/59,492, filed on Jun. 6, 2008. The entire contents of this application are incorporated herein by reference. [Prior Art] In the fourth (four) side of the complex computer system resources (four) - newly developed as a logical division of system resources. Conceptually, logical segmentation means establishing multiple discrete partitions' and assigning certain types of system resources to individual partitions. For example, 'by assigning different processors to different partitions, by sharing processors between some partitions (but not others), by specifying for each: using: group processor each partition The processor resources of the multiprocessor system are partitioned by the processing resource metrics 1 and so on available for the zone. Tasks performed within a logical partition may only use resources assigned to one partition rather than resources assigned to another partition. Generally, 'logical segmentation is enforced by a segmentation manager that is embodied as a low-level warp-horse executable instruction and data, although there may be a certain amount of hardware support for logical segmentation (such as the preservation of state information). : Body register). Low level code functions and/or hardware block access to resources allocated to different partitions. In general, one of the logical partition managers 146527.doc 201102805 includes a user interface for managing low-level code functions that enforce logical partitioning. This logical split manager interface is intended for use by a single or a small group of authorized users (i.e., system administrators). When used in this article, this low-level logical split code is called a hypervisor, and the split manager interface is called the Hardware Management Console (HMC). Communication between the HMC and the logical partition of the data processing system for (example >) simultaneous hardware maintenance, dynamic logic partitioning, inventory collection (invent〇ry c〇Uecti〇n), virtual input/output (I/O) ) Device mapping and the like may be desirable. A communication method between the HMC and the logical partition of the data processing system utilizes resource monitoring for communication between the HMC and high-intensity computing architecture platform requirements (pApR) (ie, AIX® partition and LINUX partition) And control (RMC) based facilities. (ΑΙχ® is a registered trademark of International Business Machines Corporation, Annonk, New York, Usa.) Other names used herein may be registered trademarks, trademarks or products of International Business Machines Corporation or other companies. The name is unfortunate that the rmc solution requires a true LAN connection between the HMC and the partition. Associated with the real LAN connection is additional hardware requirements (LAN adapters and cables), additional configuration tasks (network management) And additional potential points of failure (lan connection). SUMMARY OF THE INVENTION A source endpoint that communicates between logically partitioned data logical partitions provides one of the source endpoints in this context. In a general-purpose transport primitive, the source processing system is a hardware management console and a logic computer implementation method. The method includes: by a source request or a response as a package of goods in a general use 146527.doc 201102805 points One of the hardware management consoles or one of the logical partitions of the data processing system, wherein the hardware management console is used for one of the split management The I-side, and the hyper-manager of the data processing system, transfers the general-purpose transport primitive from the source endpoint to a target endpoint, wherein the hyper-manipulator receives the source endpoint Encapsulating the generic transport primitive and forwarding the shipment of the generic transport primitive to the target endpoint, the shipment containing the request or the response, and wherein the receiving is performed by the hypervisor The goods are not inspected or parsed by the super-president in the transfer, and the target endpoint is the logical partition of the data processing system or the other of the hardware management consoles. Provided is a logically segmented data processing system. The logically divided data processing system includes: at least one processor including at least one logical partition; at least one external hardware management console; and a hypervisor, The at least one hardware management console is connected to the at least one logical partition interface. Each hardware management console is a user interface for split management. The hypervisor includes a In the a communication facility that communicates between the hardware management console and the at least one logical partition. The communication includes: packaging, by a source endpoint, a request or a response as a commodity in a general-purpose transport primitive, the source The endpoint is a hardware management console in the at least one hardware management console or one of the at least one logical partition; and the generic transport primitive is from the source endpoint via the hypervisor Forwarding to a target endpoint, wherein the hypervisor receives the generalized transport primitive encapsulated at the source endpoint and forwards the shipment of the generic transport primitive to the target endpoint, the cargo 146527.doc 201102805 = ' including the source endpoint and wherein the hypervisor does not verify or parse the shipment in the receiving and forwarding by the hypervisor, and the target endpoint is The at least one of the logical partitions in the logical partition or the other of the hardware management consoles in the at least one hardware management console. In another aspect, an article of a computer readable medium is provided, the at least one computer readable medium having a data processing system for facilitating logical segmentation - hardware management Computer readable code logic for communication between the console and the logical partition. When executed on a processor, the computer readable code logic performs the following operations: 丄 丄 仃 · · · · · · · · · 由 由 由 由 由 由 由 由 由 由 由 由 由 由 由 由 由 由 由 由 由 由In the transport primitive, the source endpoint is one of the data processing system hardware management console or a logical partition, wherein the hardware management console is a user interface for split management; And via the data processing system - the hypervisor forwards the generic transport primitive from the source endpoint to the target endpoint, wherein the hypervisor receives the generic transport base encapsulated at the source endpoint Transmitting the goods of the ancient metaphysical transport base 7G to the target endpoint, the cargo containing the request or the response of the source endpoint, and wherein the receiving and the forwarding are performed by the hypervisor The goods are not inspected or cut by the hypervisor in the delivery.

析,且該目標端點為該資料處理系統之該邏輯分割區戈今 硬體管理控制台中之另一者。 X 另外,經由本發明之技術來實現額外特徵及優點。本發 明之其他實施例及態樣在本文中詳細描述,且將其視為所 主張之本發明的一部分。 【實施方式】 146527.doc 201102805 特定地指出被視為本發明之標的物,且在說明書之結尾 處的申請專利範圍t清楚地^以主張。自以下結合隨附圖 式進行之詳細描述將顯而易見本發明之前述及其他目標、 特徵及優點。 ' 邏輯分割為-種用於將單—大型電腦系統劃分為多個分 割區之技術’該等分割區中之每一者在一些方面像獨立電 腦祕-樣運作。可以各種方式中之任—種來分配電腦系 統資源以供該等分龍使用。—給定資源可經分配以供單 -分割區專用,或可基於時間交錯或其他方式在所有分割 區(或分割區之某-子群)之間共用。可將一些資源分配給 各別特定分㈣,而共用其他資源。可分割之資源的實例 為中央處理n、主記憶體、1/0處理器及配接器,及τ㈣ 件。將在經邏輯分割的電腦系統中執行之 指派至該等邏輯分割區中之—者(「在該分二: 灯」),此意謂著其僅可使用經指派至彼分割區之系統資 源或資源部份,而非經指派至其他分割區之資源。 邏輯分割區實際上為邏輯的而非實體的。通用電腦通常 具有實體資料連接,諸如在不同硬體組件之間延伸之匯流 排,從而允許該等不同硬體組件彼此通信。此等硬體資源 可由不同分割區共用及/或經分配至不同分割區。自實體 組態之觀點而言,it常不關於邏輯分割區作出區別。一般 地,藉由一體現為低層級經編碼可執行指令及資料之分割 管理器來強制執行邏輯分割區,雖然可存在_定量的對邏 輯分割區之硬體支援(諸如,保存狀態資訊之專用硬體暫 146527.doc 201102805 存益)系統之實體器件及其子組件通常實體地連接以允 午通l而不考慮邏輯分割區,且自此硬體之觀點而言,無 在刀割區A中執行之任務寫入至經分配至分割區 B的°己隐體或1/0器件。低層級程式碼函式及/或硬體阻止 對經分配至其他分割區之資源的存取。 ▲邏輯分割區約束之程式碼強制執行大體上意謂有可能改 變經邏輯分割的電腦系統之邏輯組態,亦即,&變邏輯分 彳品之數目或重新指派資源至不同分割區而不重新組態硬 體。-般地’邏輯分割管理器之某一部分包含一用於管理 強制執仃邏輯分割區之低層級程式碼函式的使用者介面。 此邏輯分割管理器介面意欲供單一或一小群授權使用者 (在本文中指明為系統管理員)使用。在本文中使用時,此 低層級邏輯分割程式碼稱為「超管理器」,且分割管理器 介面稱為「硬體管理控制台」。 ° 對大型電腦系統之邏輯分割區具有若干潛在優點。如上 文所提,彡靈活性在於易於實現對資源之重新組態及重新 分配而不改變硬體。其將任務或任務群組隔離,從而有助 於防止任-任務或任務群㈣斷系統資源q促進對提供 至特定使用者之資源的調節;此在電腦系統由服務提供者 所有之情況下為重要的’該服務提供者以按每次使用資源 時計費的方式將電腦服務提供至不同使用者。其可使得單 -電腦系統能夠同時支援多個作業系統及/或環境,因為 每-邏輯分割區可執行不同之作業系統或環境。最後,任 務及資源之隔離使得在一分割區中執行之處理程序更加難 146527.doc 201102805 以存取另一分割區中之資源,由 田此提供車父大之安全性及資 料完整性。 參考諸圖式’其中相同數字首 j数子貫穿若干視圖表示相同部 分,圖1為具有多個實體碌辦知 貝菔硬肢組件之可經邏輯分割的電腦 系統100之特定硬體組件的高層 J问層級表不。在功能層級,系 統10 0之主要組件在圖1中經展千盔南成μ 士 展不為虛線輪廓;此等组件包 括:-或多個中央處理單元(CPU)101、主記憶體102、服 務處理器1〇3、終端機介面106、儲存器介面1〇7、立他1/〇 器件介面1〇8及通信/網路介面⑽,所有該等組件經由一 或多個匯流排105耦接以用於進行組件間通信。 CPU 101為一或多個遇闲 通用可程式化處理器,其執行儲存 於記憶體1G2中之指令;系統刚可含有單—CPU或多個 CPU(其中任-情況共同地由圖i中之特徵cpu 表示), 且可包括-或多個層級之機載快取記憶體(未圖示)。通 常,-經邏輯分割的系統將含有多個CPU。記憶體102為 -用於儲存㈣及程式之隨機存取半"記憶體。記憶體 102在概念上為單—單塊實體,應理解,記憶體常常配置 於快取記憶體及其他記憶體器件之階層架構中。另外,可 將記憶體102劃分為與特定CPU或CPU組及特定匯流排相關 騷之邛分,如在各種所謂的非一致記憶體存取電 月白糸統架構之任·一者中時。 服務處理器103為一用於初始化系統、維護及其他低層 級功能之專用功能單元。A體而·^,其不執行使用者應用 程式,而CPU 101執行使用者應用程式。在一實施例中, 146527.doc 201102805 服務處理器1 03及附接之硬體管理控制台(HMC) 114尤其為 系統官理員或類似人員提供一介面,從而允許人管理對系 統10 0之邏輯分割區。 終端機介面106提供一用於附接一或多個使用者終端機 1A至i2ic(統稱為丨21)之連接,且可以各種方式來實施 該終端機介面1〇6。許多大型伺服器電腦系統(大型主機)經 由終端機介面I/O處理器(通常在一或多個電子電路卡上)而 支援夕個終端機之直接附接。或者,介面1〇6可提供一至 區域網路之連接,終端機121附接至該區域網路。各種其 他替代方案係可能的。資料儲存器介面107提供至一或多 個資料儲存器件122八至1220:(統稱為122)之介面,該一或 多個資料儲存器料為旋轉磁性硬碟機f元,_然可使用 其他類型之資料儲存器件。1/〇及其他器件介面1〇8提供至 各種其他輸入/輸出器件或其他類型之器件中之任一者之 介面。在圖1之例示性實施例中展示兩個此等器件:印表 機123及傳真機,應 ^ ^可存在目午多其他此等器件 \可為不同類型通信介面1〇9提供自系統⑽至其他數 位盗件及電腦系統之—或多個通信路徑^等路徑可包括 (例如)—或多個網路126(諸如,網際網路、區域網路或其 他網路)’或可包括遠端器件通信線路、無線連接等等。 S 1 ^ ^ ^ ^ ^ ^ Μ # ^ ^ ^ 圆1肀表不% —概念性匯流 ..^ 腦系姑可目士 桥貫體105 ’但應瞭解,典型電 腩系..先了具有多個匯流 ^ pkb « , s ,、逋常以一複雜拓撲配置,諸 P白層式星形或網狀組態中的 町.·占對點鏈路、多個階層式匯 146527.doc -12- 201102805 流排、平行及冗餘路徑等,且可存在用於傳達特定資訊 (諸如’位址或狀態資訊)的單獨匯流排。在_實施例中, 除用於作為正常資料處理操作之部分之資料通信的各種言 速貧料匯流排以外’使用I2C協定之特殊服務匯流排連接 各種硬體單元’從而允許服務處理器或其他低層級處理程 序獨立於高速資料隨排而執行各種功能,諸如開機及關 機、讀取識別硬體單元之資料等等。 通常自-或多個現場可替換單域構主要實體單元。通 常’此現場可替換單元(FRU)為一電子電路卡套組。然 而,實體单兀無需為電子電路卡套組。其可替代地為諸如 磁碟機儲存器件122、、終端機121、電源供應器等等之組 件另外單貫體單元可在其中具有一或多個。對 於較大系統’單-主要功能組件(諸如,cpu⑻或記憶體 1〇2)將通常包含呈電子電路卡套組之形式的多個實體單 ^,雖㈣代性地,-個以上主要功能組件有可能駐留於 早一實體單元中》在圖1中,CPU 1〇1被表示為含有四個電 路卡111A至1UD,每—電路卡可含有—或多個處理器;記 憶體102被表示為含有六個卡U2A至n2F ;服務處理器⑻ 被表不為含有單一卡113 ;匯流排1〇5被表示為含有三個卡 115A至11 5C ,終端機介面106被表示為含有三個卡u 6A至 U6C ;儲存器介面1〇7被表示為含有兩個卡117八至117B ; I/O及其他介面108被表示為含有兩個卡118A至;且通 k介面109被表示為含有兩個卡H9A至ii9B。 應瞭解’圖1思欲在南層級描繪例示性資料處理系統】〇 〇 146527.doc -13- 201102805 之代表性組件’個別組件可具#比圖丨所表示之組件更大 的複雜度此等功能單元及實體單元之數目、類型及植 態可顯著不同。進-步應瞭解,並非圖1中所展示之所有 組件可存在於一特定電腦系統中’且除彼等所展示組件以 外亦可存在其他組件。儘管系、統_經描料-具有多個 終端機之多使用者系統,但系統⑽可替代性地為單—使 用者系統,其通常僅含有單一使用者顯示器及鍵盤輸入。 圖2為展示電腦系,统!咐之在不同硬體及軟體抽象層級 之邏輯分割區的存在的概念說明。圖2表示一具有可用於 使用者應用程式之四個邏輯分割區2〇4至2〇7(指明為「分 割區1」、「分割區2」等)的系統,應瞭解,分割區之數2 可變化°如所熟知的’電腦系統為執行處理程序之順序狀 態機。可在不同抽象層級處表示此等處理程序。在高抽象 層級’使用者指定-處卿序及輸人,且接彳卜輸出。當 進行至較低層級時,可發現此等處理程序為以某一程式設 計語言編寫之指令序列,在繼續向下時該等指令序列被轉 譯成較低層級指令序列,且通過經授權之内碼,且最終成 為資料位元,該等資料位元被輸入機器暫存器中以促使執 仃特定動作。在一極低層級處,改變的電位引起各種電晶 體接通及關斷。在圖2中’「較高」抽象層級被表示為朝向 圖式之頂部’而較低層級被表示為朝向底部。 如圖2中所展示及早先所解釋,邏輯分割為一程式碼強 制執行之概念。在硬體層級201處,不存在邏輯分割區。. 在本文中使用時,硬體層級2〇1表示圖i中所展示之實體器 146527.doc • 14- 201102805 件(而非儲存於器件中之資料 排、I/O器件等)的隼人 _处理器、記憶體、匯流 他硬™二處: 令。在-實施例中,每—處理二,其:執行機器層級指 的。雖然程式碼可指引特定分 換 上勃并Υθ . . J £中之任務在特定處理器 上執仃,但在處理器自身中蛊 …法彳日疋此指派,且事實上可 藉由程式碼來改變該指派。 _ . „ — 因此’在圖2中硬體層級被表 不為早一貫體2〇1,其自 和 身在邏輯分割區之間並無區別。 错由一分割管理器(稱為「超管理器」,其由-不可重定 位不可WR部分202(亦稱為「不可分派超管理器」或「分 割授權内碼」或「pLIr 、 C」)及一可重定位可分派部分2〇3組 成)來強制執行分割。超管理器為超級權限可執行程式 二,、此夠存取指派至任何分割區之資源(諸如,處理器 資原及。己隐體)。超官理器維護在各種專用硬體暫存器中 及在通用5己憶體令的表或其它結構中之狀態資料,該等資 料控管邏輯分割區之邊界及行為。此狀態資料尤其定義了 對邏輯分割區之資源分配’且藉由改變該狀態資料而非對 硬體之實體重組態來改變該分配。 在—實施例中,不可分派超管理器2〇2包含由CPU 1〇1執 仃之不可重定位指令,正如用於在分割區中執行之任務的 扣令。該程式碼為不可重定位的,其意謂構成不可分派超 s理器之程式碼位於記憶體中之固定真實位址處。不可分 派超官理器202可存取系統100之整個真實記憶體範圍且 可知縱真實記憶體位址。可分派超管理器程式碼2〇3(以及 146527.doc •15- 201102805 所有分割區)被包含在與一邏輯分割區指派有關之位址 處,且因此此程式碼為可重定位的。可分派超管理器以與 使用者分割區大致相同之方式運作(且因&,有時被指明 為「分割區〇」),但其對使用者而言^隱藏的且不可用於 執打使用者應用程式。一般地,不可分派超管理器處 置對實體處理器之任務指派、記憶體映射及分割區強制執 行’以及在經分割區系統中執行應用程式碼所需的類似必 要分割區任務,而可分派超管理器2〇3處置以維護為導向 之任務’諸如建立及改變分割區定義。 如圖2中所表示,較高層級(在不可分派超管理器之 上的層級)與硬體層級201之間不存在直接路徑。雖然在較 高層級處執行之任務之機器指令可直接在處理器上執行, 但對硬體資源的存取由不可分派超管理器控制。不可分派 超管理器202強制執行對處理器資源之邏輯分割區。亦 即’較高層級處之任務分派器(各別作冑系統)將任務分派 至由邏輯分參數定義之虛擬處理器,騎管理器又將 虛擬處理器分派至硬體層級201處之實體處理器以用於執 行基礎任務。超管理器亦強制執行對其他資源之分割區 (諸如,將記憶體分配至分割區),及將1/0投送至與恰當分 割區相關聯之I/O器件。 可分派超管理器203執行並非任何分割區之管轄範圍 (prcm叫的許多輔助系統f理功能。可分派超管理器通 常管理較高層級之分割管理操作,諸如建立及刪除分割 區、同時硬體維護、將處理器、記憶體及其他硬體資源分 146527.doc 201102805 配至各種分割區等。可分派超管理器尤其可處置對實體指 示燈(indicator light)之存取。可分派超管理器203可包括視 覺指示器22 1之狀態資料結構、至分割區之可分割實體分 配222的狀態資料結構,及可分割實體位置223之狀態資料 結構’其結合可分派超管理器程式碼而用以調節對實體指 示燈之存取以及啟動及撤銷實體指示燈。 將特殊使用者互動介面提供至可分派超管理器203中, 以供系統管理員、服務人員或類似有權限之使用者使用。 在一貫施例中,亦即’在系統1 〇〇含有一服務處理器1 〇3及 附接之硬體管理控制台114之情況下,HMC 114提供一至 可分派超管理器的介面以用於服務及分割管理,且在本文 中之描述中將如此假定。 在不可分派超管理器202之上為複數個邏輯分割區204至 207。每一邏輯分割區(自在其内執行之處理程序的觀點而 言)如同一獨立電腦系統(其具有其自身之記憶體空間及其 他資源)而運作。每一邏輯分割區因此含有一各別作業系 統核心,其在本文中識別為r 〇s核心」211至214。在〇s 核心層級及以_L ’每—分割區以不同方式運作,且因此圖 2將0S核心表示為對應於四個不同分割區之四個不同實體 211至214。一般地,每一 〇8核心21丨至214執行大致等效之 功能。然而,不—定所有〇S核心211至214為彼此之相同複 本,且其可為架構上等效之作業系統的不同版本,或甚至 可為架構上不同之作業系統模組。QS核心211至214執行各 種^矛"理功成’諸如任務分派、傳呼、強制多個任務間 146527.doc 201102805 之資料完整性及安全性等等。 在每一各別分割區中之os核心之上,可存在一組高層 級作業系統函式,及使用者應用裎式碼及資料(未圖示)。 使用者可在OS核心、層級之上建立程式碼,該程式碼調用高 層級作業系統函式以存取0S核心,或該程式碼可直接存取 os核心。在聰iTM作業系統中,一使用者可存取的在架 構上固定之「機器介面」形成os核心(os核心稱為 SLIC」)之上邊界,但應理解,不同之作業系統架構可 以不同方式定義此介面,且將有可能使用邏輯分割區來在 共同硬體平台上操作不同之作業系統。 視情況將邏輯分割區中之一者指明為服務分割區。在圖 2中,將分割區205指明為服務分割區,其為經指派以供系 統維護人員使用以執行各種管理及維護功能之分割區。該 服務分割區可專門用於管理及維護功能,或其亦可用於使 用者應用程式。然而,在系統丨〇〇含有硬體管理控制台 114(如圓1之所說明實施例中展示)之情況下,自硬體管理 控制台而非服務分割區來執行大部分服務及維護功能。 如最初所提及,由本發明處理之基本問題為需要在經邏 輯为割的資料處理系統之硬體管理控制台(HMC)與邏輯分 割區之間的通用通信設施’該經邏輯分割的資料處理系統 諸如為由美國紐約阿蒙克(Armonk,New York,U.S.A.)之國 際商業機器公司提供的Power®計算系統。作為特定市售實 例’實施以超管理器為基礎之通信設施(諸如下文中所描 述)的資料處理系統可基於見於IBM之p/i Series產品系列的 146527.doc 201102805 韌體及系統體(systemware)中的技術(如在Power.org (http://www.power.org/members/developers/specs/PAPR_Version_ 2.7_090ct07.pdf)處之「高強運算架構平台參考」(PAPR) 材料中所描述,該材料特此以引用的方式併入本文中)來 建置。(IBM®、pSeries®、iSeries®及 Power® 為國際商業機 器公司(Armonk,New York, U.S.A.)之註冊商標。)本文中 所呈現之通信設施可用於系統管理任務,諸如同時硬體維 護、動態邏輯分割、清查收集(inventory collection)、虛擬 I/O器件映射等。此以超管理器為基礎之通信設施為在 HMC與邏輯分割區之間的通用、低等待時間、完全非同步 通信方法。此與現有通信設施形成對比,現有通信設施為 非通用的,抑或要求在HMC與邏輯分割區之間的額外實體 連接(諸如,真實LAN連接及額外組態任務)。 大體而言,本文中呈現在HMC與邏輯分割區之間的通 用、零維護、低等待時間、完全非同步之穩健通信設施, 其在超管理器中實施。此超管理器實施之通信設施在本文 中稱為超管理器管道。超管理器管道之通用性在於其並非 特定針對HMC與超管理器之間(或超管理器與邏輯分割區 之間)的任何特定類型或類別之命令或流(諸如動態邏輯分 割區對虛擬I/O '非同步對同步等)。 在管道自身未意識到正於HMC與邏輯分割區之間流動之 特定命令(亦即,請求及回應)的意義上,超管理器管道為 零維護的。將該等特定命令看作駐留於流經超管理器管道 之基本或泛用傳送基元中之貨物。超管理器辨識泛用傳送 146527.doc •19· 201102805 基7L ’但不檢驗或剖析包含於泛用傳送基元内之貨物。因 此’當在HMC與邏輯分割區之間引入新命令流時,超管理 器不受影響。 *超s理器官道經设計以對通過管道之通信流造成最小的 等待時間。由超管理器造成之等待時間主要為將該等通信 流直接記憶體存取(DMA)至靈活服務處理器(FSp)(其為 HMC與超管理器藉以進行通信之構件)或自靈活服務處理 益(FSP)直接§己憶體存取(DMa)該等通信流及將該等通信 流DMAS邏輯分割區或自邏輯分割區該等通信流所 需的時間。由超管理器管道進行之處理主要為在一或多個 HMC與一或多個邏輯分割區之間的投送、調整進度 (pacing)及統計資料收集。 為了在最小化超管理器管道之系統資源要求(亦即,記 憶體、緩衝器、任務)之同時最大化超管理器管道之輸送 量,且防止一不良運作(亦即,未及時地回應)之邏輯分割 區或HMC負面地影響其他分割區及HMC之輸送量,超管 理益官道本質上為完全非同步的且針對每一邏輯分割區及 每-HMC使用單獨之緩衝器集區(ρ〇〇ι)。換言之,若脱〔 H1起始一至分割區?1之通信流,但ρι極為忙碌或無反應 (hung)且不回應,則超管理器在卩丨緩衝器集區未耗盡之情 況下將通信流一直投送至P1,且接著對m應答(之後若?1 確實不回應,則H1可因等待來自!>丨之回應流而逾時),抑 或在超官理器處之P1緩衝器集區耗盡的情況下,超管理器 用忙碌狀態應答HMC H1。管道任務及超管理器不封鎖 146527.doc •20· 201102805 且經由 之責任 dk)對來自HMC或分割區之回應或應答的等待, 忙碌狀態或通信流自身之逾時而將處置無回應目標 傳遞回至通信流之來源。 本文中呈現之超管理器管道的穩健性在於不存在輸入/ 輸出(I/O)配接器及通信電纜’從而減少發生組態問題 體故障之機會。 μ因此,以超管理器為基礎之通信設施(亦即,超管理器 =道)介接於經邏輯分割的資料處理系統2HMc(外部平^ 管理器件)與邏輯分割區之間。超管理器管道為一通用: L «又施,使彳于其可處置只^1(:與邏輯分割區之間的許多類型 或類別之流/命令。超管理器管道僅意識到基本傳送基 疋,而非在HMC與邏輯分割區之間流動之特定命令(請求 或回應)(其被看作簡單貨物或額外資料),且因此每當將在 HMC與邏輯分割區之間之新命令、冑求或回應流引入至資 料處理系統時,超管理器管道不受影響。超管理器管道能 夠在或多個HMC與一或多個邏輯分割區之間投送流,且 不允許無回應之分割區或HMC阻止流之傳送或負面地影 響彼等來往於其他分割區或HMC之流的效能。 本文中呈現之通信設施優於上文所描述之習知RMC資源 管理機制,因為其不要求每一 HMC與邏輯分割區之間的單 獨貫體LAN連接,藉此降低硬體要求(無lan配接器)、 組態努力(無網路管理、佈線)及故障率(無配接器或電纜故 障)。然而,在每一 HMC與由彼HMC管理之系統之間仍存 在一網路連接。對於具有多個邏輯分割區之所管理系統, I46527.doc 201102805 本文中呈現之通信設施與RMC資源管理機制相比表現出在 網路額外負荷方面之潛在顯著降低。 圖3至圖5描繪根據本發明之態樣的超管理器管道的三個 不同視圖。 在圖3中,呈現實體框架視圖,其中經邏輯分割的資料 處理系統300包括多個邏輯分割區31〇(標記為ρι、?2及 P3)、包含本文中所描述之超管理器管道(或通信設施)325 的超管理器320、靈活服務處理器33〇,及多個硬體管理控 制台(HMC)340,該等HMC 34〇經由靈活服務處理器 (FSP)330而網路連接至超管理器32〇。Fsp 33〇與超管理器 320之間的通信係經由直接記憶體存取(DMA),且超管理 器與邏輯分割區亦經由一形式之分割區間DMA而通信。超 g理器中之超B理器管道組件進行流投送、調整進度及統 計資料收集。 圖4描繪資料處理系統3〇〇之較低層級邏輯視圖,其中 HMC 34G與超官理器32G(或更特定地,超管理器管道325) 經由FSP 330而建立邏輯HMC命令會期4〇〇。如所展示,在 每一聽與超管理器之間存在單-通信會期400。與每一 HMC對超管理器會期_相關聯的是超管理器管道内之 §fl息緩衝器集區4 〇 5。如下文谁一 + ^ r又進步解釋’此專訊息緩衝 器用於實施HMC與邏輯分割區之間的以超管理器為基礎之 通信設施。類似地,邏輯分割區與超管理器建立邏輯分割 事件會期❻,在該等邏輯分割事件會期指上經由泛用傳 送基元而交換通信流。針對每—邏輯分割對超管理器會期 146527.doc •22· 201102805 410再次定義各別訊息缓衝器集區415。在操作中,超管理 器剖析該等通用基元傳送命令以判定目標且在各別HMC與 邏輯分割區之間移動貨物。 在一實施中’除泛用傳送基元以外’使用三種類型之會 期管理請求,亦即,可在HMC與超管理器管道之間(以及 超管理器管道中之每一邏輯分割區之間)使用開啟通信會 期請求、關閉通信會期請求及交換能力請求。因此,在一 實施例中,來自HMC或邏輯分割區之請求可包含開啟請 求、交換能力請求、關閉請求或泛用傳送基元(在下文描 述之邏輯流程中稱為HypePipe請求)。 圖5描繪資料處理系統3〇〇之較高層級之邏輯表示,其中 本文中呈現之邏輯管道可被認為是經由資料 管理器320而建立每一聽340與每—分割區= 對點、邏輯HMC對分割區通信會期5〇〇。此圖為該通信框 架在其最高層級處之概念視圖。HMC中之端點將通信流發 送至邏輯分割區中之端點,且反之亦然。 圖6至圓10描繪用於實施一以超管理器為基礎之通俨嗖 施(諸如本文中所呈現)之超管㊣器(或超管理器管道)邏輯 的一貫施例,而圖U至圖13描綠根據本發明之態樣的可在 HMC或邏輯分割區内實施的特定端點邏輯流程。 首先參看圖6,超管理器普& 务丄 g里裔S道專待來自源端點之請求 600。在本文中使用時, 原螭點為HMC或邏輯分割區, =求可包含會期管理請求(諸如,開啟會期請求、交 換此切求或_會期請峨超道(或办咐㈣ 146527.doc -23· 201102805 請求。超管理器管道請求為本文中所描述之泛用或基本傳 送基元。端點正是使用此基元來經由超管理器管道進行通 信。在一實施例中,泛用傳送基元具有以下資料結構: 目標端點ID|貨物長度丨貨物(經囊封之請求或回應通信)丨 超管理器初始判定接收到之源端點請求605是一開啟會 期管理請求或是一交換能力會期管理請求6丨〇 ^若為任_ 者’則將源端點與超管理器之間的會期狀態設定為開啟 615 ’且在超管理器内設定由源端點要求之能力62〇。將應 答狀態(ACK狀態)設定為成功625,且將該應答(包括其狀 態及經調整之能力)傳回至源端點630。若源端點請求並非 開啟會期管理或交換能力會期管理請求,則超管理器判定 該s青求是否為關閉會期管理請求6 3 5。若是,則將源端點 與超管理器之間的會期狀態設定為關閉64〇,且取消針對 彼會期之任何排入符列或待決請求645。超管理器將應又 狀態設定為成功650,且將具有隨附狀態之應答發送回至 源端點630,之後返回以等待下一源端點請求6〇〇。若源端 點請求並非會期管理請求,則超管理器判定其是否為基本 傳送基元(亦即,超管理器管道(或Hypepipe)請求)。若 「是」,則自查詢655起超管理器執行圖7之處理66〇(描述 於下文)’之後返回以等待下一源端點請求。若源端點請 求並非會期管理請求或超管理器管道請求,則該請求為一 未辨識之請求,且因此將應答狀態設定為錯誤665,且傳 回具有一錯誤狀態之應答630。 假定超管理器管道接收到來自源端點之超管理器管道請 146527.doc •24· 201102805 求(亦即,泛用傳送基兀),則執行圖7之邏輯流程。超管理 器管道判定該泛用傳送基元是否具有可接受之貨物大小 (亦即,具有小於或等於最大傳送單元(MTU)大小的大 小)700。若「否」,則將應答狀態設定為]^[11;違規7〇5,且 超管理器返回770至圖6之邏輯。否則,超管理器判定目標 端點識別是否為資料處理系統内之有效端點識別71〇。若 「否」,則將應答狀態設定為一無效參數7丨5,且超管理器 返回770至圖6之邏輯。假定貨物大小及目標端點識別為可 接受的,則超管理器判定目標訊息緩衝器集區是否為空 720。若「是」,則將應答狀態設定為忙碌725,且超管理 器返回770。否則’自目標訊息緩衝器集區取出一目標訊 息緩衝器730 ’且在該目標訊息緩衝器中建置一新的超管 理器管道請求(亦即,目標傳送基元),其包括將貨物自源 超管理器管道請求(亦即,源傳送基元)複製至目標超管理 器管道凊求735。一旦建置了此目標傳送基元,超管理器 就判定目標端點會期是否處於開啟狀態74〇。若「否」,則 將應答狀態設定為會期關閉745,且超管理器返回770至圖 6之邏輯。否則’超管理器判定其是否處於與目標端點交 ' 換能力之過程中(亦即,該會期狀態是需要交換能力或正 • 等候能力應答)750。若「是」’則將應答狀態設定為忙碌 755’且超管理器返回770至圖6之邏輯。否則,非同步地 將目標傳送基元發送至目標端點760,且將應答狀態設定 為成功765,之後處理返回770。 圖8描繪用於處理來自端點之應答之超管理器邏輯的一 146527.doc 25· 201102805 實施例。超管理器等待來自端點之應答800。一旦接收到 應J 805,超管理器就判定該應答是否為交換能力應答 81〇。若「是」,則超管理器判定該端點是否履行(h〇n〇r)超 笞理器能力睛求8 1 5。若「是」,則將端點對超管理器會期 狀態設定為開啟820,且將在交換能力請求中使用之訊息 緩衝器傳回至各別端點訊息緩衝器集區825。若端點並未 履行超管理器能力,則僅將訊息緩衝器傳回至各別端點訊 息緩衝器集區’而不將端點對超管理器會期狀態設定為開 啟。在將訊息緩衝器傳回至超管理器管道中之對應端點^ 息緩衝器集區之後,超管理器判定該會期狀態是否為需要 交換能力會期狀態830 ^若「否」,則處理返回以等待來自 端點之下一應答800。否則,執行圖9之邏輯835(描述於 文)。 右接收到之應答並非交換能力應答,則超管理器判定气 應答是否為超管理器管道應答840。若「是」,則檢查該^ 答之隨附傳回狀態以查看其是否為—需要交換能:狀^ 討5。若「是」,則將會期狀態設定為需要發送能力 之後將端點訊息緩衝器傳回至對應端點訊息緩衝器集區 若接收到之應答並非交換能力或超管理器管道應答^ °°。 管理器未能辨識該應答855,且返回以等待下—應笈則与 假定接收到之應答之會期狀態為需要交換能力,則自 詢830起,執行圖9之邏輯835。此邏輯執行超管理器管查 與對應端點(與之建立了通信會期)之間的能力交換 γ遵 於經設定為需要交換能力之一通信會期 回應 U狀態,超管理器自 146527.doc • 26· 201102805 對應訊息緩衝器集區獲棋 设件—端點訊息緩衝器9〇〇,且在彼 訊息緩衝器中建置—交換 、月b力清求905。接著將超管理器 之通信會期狀態設定為等候能力應答91〇,且超管理写非 同步地將交換能力請求發送至通信會期之各別端點915, 之後返回至該邏輯被呼叫時所在之邏輯流程920。 。圖10描纟會用於處理資料處理系統内之能力改變事件之邏 輯的實知例。在發生了一能力改變事件時i咖,考慮每 HMC會期1〇〇5。舉例而言’能力改變事件可為同時勒體 更新超e理益判定是否仍存在待考慮之hmc會期⑻〇, 且若「是」,貝1j獲得下—HMC會㈣15。將此HMC會期之 會期狀態設定為需要交換能力咖,且超管理器判定目標 HMC訊息緩衝器集區是否為空助。若「是」,則考慮下 一 HMC會期1〇〇5。否則,執行圖9之邏輯ι〇3〇,之後超管 理器返回以評估下-HMC會期…旦已處理所有職會 期’則自查詢1G10起’超管理器評估每—邏輯分割(Lp)會 期1035。超管理器判定是否存在另一開啟的^通信會期 1040’且若「否」,則能力交換邏輯已完成1()45。若存在 另一待考慮之LP會期,則獲得下一 ^通信會期1〇5〇,且 將其會期狀態設定為需要交換能力1〇55。超管理器判定目 標邏輯分割區之訊息緩衝器集區是否為空1〇6〇,且若 「是」,則處理下一 LP會期。若訊息緩衝器集區不為空, 則針對所獲得之LP會期而執行圖9之邏輯1〇65。 圖11描繪用於在邏輯分割區或HMC内處理超管理器管道 請求(泛用傳送基元)之端點邏輯的一實施例。該端點等待 146527.doc 27· 201102805 一超管理器管道請求蘭,且在接收到時11〇5將貨物複製 至一本端緩衝器中1110。該端點接著判定貨物是否含有一 有效端點請求1115。若「是」,則該端點用成功狀態應答 該超管理器管道請求i i 2 G。該端點接著判定是否需要斑請 求源交換能力1125。若「是」,則將一以回應作為貨物之 新超管理器管道請求發送至源m中該回應含有一指 示需要交換能力之狀態1130。若不需要與源端點交換能 力’則目標端點處理該請求1135且判定是否需要對該請求 之回應1140。若「否」’則該端點返回以等待下一超管理 器管道請求。否貝,!,產生一回應且將其封裝為-自目標端 點發送回至源端點之新超管理器管道請求之貨物,該回應 具有一基於在目標端點處處理第一超管理器管道請求之結 果而設定的狀態1145 » 右貨物不含有一有效請求,則該端點判定該貨物是否含 有有效回應11 5 0。若「是」,則發送對超管理器管道請 求之應答,其中狀態經設定為成功丨丨55,且該端點判定該 回應是否針對交換能力請求116〇β若「否」,則處理該回 應1165,且若「是」,則目標端點記錄已與源端點交換能 力之事實1170。若貨物不含有有效請求或有效回應,則發 送一具有錯誤狀態之對超管理器管道請求之應答1175。 圖12及圖13描繪用於回應於在一端點處之能力改變事件 而交換能力的兩種方法。圖12為主動方法,而圖13為被動 方法。 首先參看圖12,端點邏輯判定一可能已改變端點中之能 146527.doc -28- 201102805 力的事件已發生12Q()。舉例而言,端點通電或同時勃體更 新已發生。端點邏輯設定能力交換狀態,以指示尚未對於 正,追蹤能力交換狀態之所有端點交換能力12〇5。接著將 一交換能力請求發送至超管理器121G。在—實施例中,對 ;:點通電’可㉟由開啟會期請求而非交換能力請求來交 換能力。考慮了所有可能目標端點1215後,主端點 ㈣卟ct endpoint)判定是否存在另一待考慮之端點122〇。 右:是」,則獲得下一端點1225’且將一以交換能力請求 為貨物之超官理器管道請求發送至彼端點1咖。—旦已考 慮所有端點,處理完成1235。 在替代方法t,可使用圖13之邏輯來回應於一端點中 之能力改變事件而交換能力。在此方法中,發生—可能已 改交主知點中之能力的事件,諸如同時勒體更新n〇〇。改 變主端點處之能力交換記錄以指示尚未對於正被追蹤能力 交換狀態之所有端點交換能力測。端點接著將一交換能 力請求發駐超管理n131G。再找,對㈣點通電,可 經由開啟會期請求而非交換能力請求來與超管理器交換能 力。一旦已將交換能力請求轉遞至超f理器,處理完成 1315。在此方法中,端點並非主動地將交換能力請求推至 其他端點,而是將等待彼等其他端點起始一通信流,且將 用一需要交換能力狀態來回應彼通信流。 」為進一步解釋,下文中描述用於在(例如)IBM P0赠⑧ 计算系統中實施上文所述之以超管理器為基礎之通信設施 的各種命令結構。僅作為實例提供以下論述,且應理2, 146527.doc -29- 201102805 與此說明書一起呈現之申請專利範圍不限於下文所呈現之 特定實施例。 術語及概念 在本文中使用時’使用以下術語: •上游-在此上下文中,基於具有位於韌體之上的分割區 的經邏輯分割系統(包括HMC)的共同觀點,上游係指一 自HMC至分割區之流。 •下游-在此上下文中,基於具有位於韌體之上的分割區 的經邏輯分割系統(包括HMC)的共同觀點,下游係指一 自分割區至HMC之流。 •入埠-在進出分割區及HMC之流的上下文中,入埠係指 一進入分割區或HMC之流,應自HMC或分割區而非超 管理器(HYP)之觀點來解擇該術語。 •出埠-在進出分割區及HMC之流的上下文中,入埠係指 一離開分割區或HMC之流《應自HMC或分割區而非ΗΥΡ 之觀點來解譯該術語。 高層級流 此部分論述在超管理器管道流中所涉及之子系統之間及 子系統中發生的主要互動及事件。 用於封包化與自HMC至分割區之上游請求相關聯之回應 資料的兩種方法係可能的。第一種方法為「拉」方法,藉 此請求之源端點在取得對初始請求之回應之後將一或多個 GetRemainingResp〇nseData(取得剩餘回應資料)流發送至 請求之目標,直至已擷取所有相關聯之回應資料為止。第 146527.doc -30- 201102805 二種方法為「推」方法,藉此請求之目標重複地發送針對 該請求之回應流,直至與該請求相關聯之所有回應資料被 傳遞回至請求之源。 拉方法之益處在於其允許初始請求之源按其自身之速率 擷取資料。 推方法之益處在於其涉及總體上較少的通過超管理器管 道之流,且其允許目標將所有回應資料立即發送回至源而 不必應對源直接放棄剩餘資料之情形。一潛在缺點在於若 目標緊接地發送回若干回應流,則其可自超管理器得到 「忙碌」狀態。若有其他分割區正同時發送流至同一 HMC,則此情況可加重。另外,在理論上封包有可能丟 失,其意謂HMC應監視此情形且在發生丟失時應重試初始 請求。藉由拉方法,HMC將僅重試失敗之GetRemainingResponseData (取得剩餘回應資料)流。 開啟會期 此為一 HMC訊息介面,HMC經由該HMC訊息介面而開 啟一與超管理器之會期且與超管理器交換能力。 可針對每一附接之HMC開啟一個會期。對於單一 HMC,不能開啟多個會期。 在下文之參數中,入埠及出埠係自HMC之觀點而言。 開啟會期請求參數: 146527.doc •31 - 201102805 名稱 内容描述 最大入埠訊息 未處理之入埠訊息的最大數目 最大出埠訊息 未處理之出埠事件的最大數目 能力位元組之數目 以下能力位元組之數目 能力位元組 每一位元表示一能力。HMC啟動表示其支援之能力的位 元,超管理器關閉表示其不支援之能力的彼等位元。 開啟會期回應參數: 除了「最大入埠訊息」及「最大出埠訊息」將含有超管 理器之反建議(counter proposal)(其將始終小於或等於HMC 所指定之數目)之外,回應參數與請求參數相同,且除了 表示不被超管理器支援之能力的位元被關閉之外,能力位 元遮罩將與在請求中所傳遞之遮罩相同。 超管理器將對「最大入埠事件」及「最大出埠事件」反 建議之最大值分別為8及1。若HMC指定較高值,則超管理 器將反建議此等值。若HMC出於某種原因指定較低值,則 超管理器將履行該等較低值,但管道輸送量可受到負面影 響。初始值零將導致GenericError(—般性錯誤)。 開啟會期狀態碼: 在共同標頭中傳遞回。無命令特定之傳回碼。可能值 • Good(良好) • InvalidOpcode(無效作業碼) • GenericError(—般性錯誤) • InvalidState(無效狀態)-在一會期開啟的同時,發布一開 啟會期。 146527.doc -32- 201102805 關閉會期 此為一介面,HMC經由該介面而關閉與超管理器之命令 會期。不存在請求或回應參數,且不存在會期特定的傳回 碼。 交換能力: 此為一 HMC訊息介面,HMC與超管理器經由該HMC訊 息介面而交換能力。在會期之持續時間/壽命内,可多次 發送此命令。該命令可緊接於開啟會期之後,及/或可稍 後在實際上將使用會期時發送該命令。HMC及HYP必須作 為源與目標皆支援該命令。 若HMC或超管理器之能力由於同時韌體更新而改變,則 可使用此命令來交換能力(假定非同時韌體更新將引起會 期關閉且在更新之後使一開啟會期發生)。兩種實體每當 接收到一以其為目標的並非交換能力之流時必須驗證其當 前能力已與源交換。若尚未交換,則必須傳回ExchangeCapabilities (交換能力)傳回碼。若一改變或可能已改變目標之能力的 事件發生,則目標必須再次與所有源交換能力。 請求參數‘· 名稱 内容描述 能力位元組之數目 以下能力位元組之數目 能力位元組 每一位元表示一能力。源啟動表示其支援之能力的位元, 目標關閉表示其不支援之能力的彼等位元。 回應參數: 除了表示不被目標支援之能力的位元為關閉之外,回應 參數與請求參數相同。 146527.doc •33- 201102805 傳回碼: 在共同標頭中傳遞回。無命令特定的傳回碼。可能值 為. • Good(良好) • InvalidOpcode(無效作業碼) • GenericError(—般性錯誤) 超管理器管道請求: 此為一 HMC訊息介面,HMC經由該HMC訊息介面而與 分割區直接通信。超管理器充當一簡單管道。其不檢驗經 由該管道傳送(pipe)至分割區或自分割區傳送之資料。其 精確地複製資料區(貨物)而不作修改。 此為一完全非同步命令。當HMC將該命令發送至超管理 器時,超管理器將此命令内之貨物投送至指定分割區。當 一分割區將一以HMC為目標之流發送至超管理器時,超管 理器將該命令連同來自分割區之貨物一起發送至指定 HMC。 對此等命令之所有計時由HMC及分割區負責。 請求參數: 名稱 内容描述 源/目標ID 對於出埠(自HMC至HYP)流,HMC設定為目標ID。對於入 埠(自HYP至HMC)流,HYP設定為源ID。 貨物大小 以下貨物之大小(以位元組為單位)。對於以分割區為目標 之出埠流,此值不可超過160位元組,且對於以HMC為目 標之入埠流,此值不可超過1488位元組。 貨物 貨物 146527.doc -34- 201102805And the target endpoint is the other one of the logical partitions of the data processing system. X Additionally, additional features and advantages are realized through the techniques of the present invention. Other embodiments and aspects of the invention are described in detail herein and are considered as part of the claimed invention. [Embodiment] 146527. Doc 201102805 specifically states that it is regarded as the subject matter of the present invention, and the scope of patent application t at the end of the specification is clearly claimed. The foregoing and other objects, features and advantages of the present invention will become apparent from 'Logical segmentation is a technique for dividing a single-large computer system into multiple partitions'. Each of these partitions operates in some ways like an independent computer. Computer system resources can be allocated in any of a variety of ways for use by such dragons. - A given resource may be allocated for single-partition-specific use, or may be shared among all partitions (or some-sub-group of partitions) based on time interleaving or other means. Some resources can be allocated to specific points (4), while other resources are shared. Examples of severable resources are central processing n, main memory, 1/0 processor and adapter, and τ(4). Assignment to a logically partitioned computer system to those logical partitions ("in the sub-two: lights"), which means that only system resources assigned to the partitions can be used Or a resource component, rather than a resource assigned to another partition. Logical partitions are actually logical rather than physical. A general purpose computer typically has a physical data connection, such as a busbar that extends between different hardware components, allowing the different hardware components to communicate with one another. These hardware resources may be shared by different partitions and/or assigned to different partitions. From the point of view of physical configuration, it often does not make a distinction about logical partitions. In general, the logical partition is enforced by a partition manager embodied as a low-level encoded executable instruction and data, although there may be a _quantitative hardware support for the logical partition (such as the preservation of state information) Hardware temporarily 146527. Doc 201102805 The physical device of the system and its sub-components are usually physically connected to allow the pass-through without regard to the logical partition, and from the point of view of this hardware, there is no task written in the knife-cut area A Enter into a hidden or 1/0 device that is assigned to partition B. Low level code functions and/or hardware block access to resources allocated to other partitions. ▲ Logic partitioning of the code partitioning of the code generally means that it is possible to change the logical configuration of the logically segmented computer system, that is, to change the number of logical products or reassign resources to different partitions without Reconfigure the hardware. Typically, a portion of the logical partition manager contains a user interface for managing the low level code functions of the mandatory logical partition. This logical partition manager interface is intended for use by a single or a small group of authorized users (indicated as system administrators in this document). When used in this article, this low-level logical split code is called the HyperManager, and the split manager interface is called the Hardware Management Console. ° There are several potential advantages to the logical partition of a large computer system. As mentioned above, the flexibility is that it is easy to reconfigure and redistribute resources without changing the hardware. It isolates tasks or task groups, thereby helping to prevent any task or task group (4) breaking system resources q to facilitate the adjustment of resources provided to specific users; this is where the computer system is owned by the service provider Important 'The service provider provides computer services to different users in a way that charges each time a resource is used. It enables a single-computer system to simultaneously support multiple operating systems and/or environments, as each logical partition can execute a different operating system or environment. Finally, the separation of tasks and resources makes it more difficult to execute procedures in a partition. 146527. Doc 201102805 To access resources in another partition, Tian provides the safety and integrity of the car. Referring to the drawings, wherein the same number of first digits represents the same parts throughout the several views, FIG. 1 is a high level J of a particular hardware component of a logically separable computer system 100 having a plurality of physical components. Ask the level table not. At the functional level, the main components of the system 10 are not shown as dashed outlines in Figure 1; these components include: - or a plurality of central processing units (CPU) 101, main memory 102, The service processor 101, the terminal interface 106, the memory interface 1〇7, the vertical 1/〇 device interface 1〇8, and the communication/network interface (10), all of which are coupled via one or more bus bars 105 Used for communication between components. The CPU 101 is one or more idle general-purpose programmable processors that execute instructions stored in the memory 1G2; the system may just contain a single-CPU or multiple CPUs (any of which - the situation is collectively The feature cpu represents) and may include - or multiple levels of onboard cache memory (not shown). Typically, a logically partitioned system will contain multiple CPUs. The memory 102 is - for storing (four) and program random access half " memory. Memory 102 is conceptually a single-monolithic entity, it being understood that memory is often configured in a hierarchical architecture of cache memory and other memory devices. In addition, the memory 102 can be divided into a specific CPU or CPU group and a specific bus bar, as in any of the so-called non-coherent memory access systems. Service processor 103 is a specialized functional unit for initializing system, maintenance, and other low-level functions. The body A does not execute the user application, and the CPU 101 executes the user application. In an embodiment, 146527. Doc 201102805 Service Processor 1 03 and the attached Hardware Management Console (HMC) 114 provide an interface for system administrators or similar personnel to allow a person to manage the logical partitions of the system 10 . The terminal interface 106 provides a connection for attaching one or more user terminals 1A to i2ic (collectively referred to as 丨 21), and the terminal interface 1 〇 6 can be implemented in various ways. Many large server computer systems (mainframes) support direct attachment of evening terminals via a terminal interface I/O processor (usually on one or more electronic circuit cards). Alternatively, interface 1 〇 6 provides a connection to the local area network to which terminal 112 is attached. Various other alternatives are possible. The data storage interface 107 is provided to one or more of the data storage devices 122 to 1220: (collectively referred to as 122), the one or more data storage materials are rotating magnetic hard disk drives, and other Type of data storage device. 1/〇 and other device interfaces 1〇8 are provided to any of a variety of other input/output devices or other types of devices. Two such devices are shown in the exemplary embodiment of FIG. 1 : printer 123 and fax machine, there may be other such devices in the future, and can provide self-system for different types of communication interfaces 1 〇 9 (10) Paths to other digital pirates and computer systems - or multiple communication paths ^ may include, for example, or multiple networks 126 (such as the Internet, a regional network, or other network)' or may include far End device communication lines, wireless connections, and the like. S 1 ^ ^ ^ ^ ^ ^ Μ # ^ ^ ^ Round 1 肀 Table No % - Conceptual Confluence . . ^ Brain system Gu Keshi Bridge body 105 ' But should know, typical electric system. . First, there are multiple confluences ^ pkb « , s , , 逋 often in a complex topology configuration, P white layered star or mesh configuration in the town. · Occupy point links, multiple hierarchical sinks 146527. Doc -12- 201102805 Streaming, parallel and redundant paths, etc., and there may be separate busses for communicating specific information such as 'address or status information. In the embodiment, in addition to various speedy busbars for data communication as part of normal data processing operations, 'I2C protocol special service busbars are used to connect various hardware units' to allow service processors or other Low-level handlers perform various functions independently of high-speed data, such as powering on and off, reading data identifying hardware units, and so on. Usually the primary entity unit is replaced by one or more fields. Typically, this Field Replaceable Unit (FRU) is an electronic circuit card set. However, the physical unit does not need to be an electronic circuit card set. It may alternatively be a component such as a disk drive storage device 122, a terminal machine 121, a power supply, etc., and another single body unit may have one or more therein. For larger systems, the 'single-primary functional component (such as cpu(8) or memory1〇2) will typically contain multiple entities in the form of an electronic circuit card set, although (four) generationally, more than one main function It is possible for a component to reside in a previous physical unit. In Figure 1, CPU 1〇1 is represented as containing four circuit cards 111A through 1UD, each of which may contain - or multiple processors; memory 102 is represented To contain six cards U2A to n2F; the service processor (8) is represented as containing a single card 113; the busbar 1〇5 is represented as containing three cards 115A to 11 5C, and the terminal interface 106 is represented as containing three cards u 6A to U6C; the memory interface 1〇7 is shown to contain two cards 117 8 to 117B; the I/O and other interfaces 108 are shown as containing two cards 118A to; and the k interface 109 is represented as containing two Cards H9A to ii9B. It should be understood that 'Figure 1 is intended to depict an exemplary data processing system at the southern level 〇 146 146527. Doc -13- 201102805 Representative components 'Individual components can have a greater complexity than the components represented by the diagrams. The number, type and organization of these functional units and physical units can vary significantly. Further, it should be understood that not all of the components shown in Figure 1 may be present in a particular computer system' and other components may be present in addition to those shown. Although the system is multi-user system with multiple terminals, the system (10) may alternatively be a single-user system, which typically only contains a single user display and keyboard input. Figure 2 shows the computer system, system! A conceptual description of the existence of logical partitions at different hardware and software abstraction levels. Figure 2 shows a system with four logical partitions 2〇4 to 2〇7 (designated as "segment 1", "segment 2", etc.) that can be used in the user application. It should be understood that the number of partitions 2 Can be changed ° As is well known, the computer system is a sequential state machine that executes the processing program. These handlers can be represented at different levels of abstraction. At the high abstraction level, the user specifies - the order and the input, and the output is output. When proceeding to a lower level, it can be found that the handlers are sequences of instructions written in a programming language that are translated into lower level instruction sequences as they continue downward, and are authorized The code, and eventually becomes a data bit, which is entered into the machine register to cause a particular action to be performed. At a very low level, the changing potential causes various dielectrics to turn "on" and "off". In Fig. 2, the 'higher' abstraction level is represented as the top of the pattern' and the lower level is shown as facing the bottom. As shown in Figure 2 and explained earlier, the logical division is a concept of code execution. At the hardware level 201, there is no logical partition. .  As used herein, the hardware level 2 〇 1 represents the entity 146527 shown in Figure i. Doc • 14- 201102805 (not the data row, I/O device, etc. stored in the device) _ processor, memory, sinking his hard TM two: order. In the embodiment, each is processed two, which: the execution of the machine level refers to. Although the code can guide the specific switching on the Υ Υ θ.  .  The task in J £ is executed on a specific processor, but this assignment is made in the processor itself, and in fact the assignment can be changed by the code. _ .  „ — Therefore, the hardware level in Figure 2 is not the same as the early consistency 2〇1, and there is no difference between the self-contained and the logical partition. The fault is divided by a split manager (called “super-manager”). , consisting of - non-relocatable non-WR part 202 (also known as "non-dispatchable hypervisor" or "split authorization inner code" or "pLIr, C") and a relocatable dispatchable part 2〇3) Force splitting. The hypervisor is a super privileged executable program 2, which is sufficient to access resources assigned to any partition (such as processor resources and implicit entities). The super-status maintains state data in various dedicated hardware registers and in tables or other structures of the general-purpose memory, which control the boundaries and behavior of the logical partition. This status profile specifically defines the resource allocation to the logical partition' and changes the allocation by changing the status data rather than reconfiguring the entity of the hardware. In an embodiment, the non-dispatchable hypervisor 2〇2 contains non-relocatable instructions executed by the CPU 1〇1, just as a deduction for tasks performed in the partition. The code is non-relocatable, which means that the code that constitutes the undispatchable super processor is located at a fixed real address in the memory. The non-dispatching super-orientation device 202 can access the entire real memory range of the system 100 and can be aware of the vertical real memory address. The hypervisor code can be assigned 2〇3 (and 146527. Doc •15- 201102805 All partitions are included in the address associated with a logical partition assignment, and therefore this code is relocatable. The hypervisor can be dispatched to operate in much the same way as the user partition (and sometimes referred to as "partition" because of &, but it is hidden from the user and is not available for execution) User application. In general, the non-dispatchable hypervisor handles task assignments to the physical processor, memory mapping and partition enforcement 'and the necessary necessary partitioning tasks required to execute the application code in the partitioned system, and can assign super Manager 2〇3 handles maintenance-oriented tasks such as establishing and changing partition definitions. As shown in Figure 2, there is no direct path between the higher level (the level above the non-dispatchable hypervisor) and the hardware level 201. While machine instructions for tasks performed at higher levels can be executed directly on the processor, access to hardware resources is controlled by an undispatchable hypervisor. The non-distributable hypervisor 202 enforces a logical partition of processor resources. That is, the task dispatcher at the higher level (the individual task system) dispatches the task to the virtual processor defined by the logical sub-parameter, and the ride manager dispatches the virtual processor to the entity processing at the hardware level 201. Used to perform basic tasks. The hypervisor also enforces partitioning of other resources (such as allocating memory to partitions) and delivering 1/0 to the I/O devices associated with the appropriate partition. The dispatchable hypervisor 203 performs the jurisdiction of not any partition (the many auxiliary system functions called by prcm. The dispatchable hypervisor usually manages higher level partition management operations, such as creating and deleting partitions, while hardware Maintenance, processor, memory and other hardware resources are divided into 146527. Doc 201102805 is equipped with various partitions and so on. The dispatchable hypervisor can specifically handle access to the indicator light. The assignable hypervisor 203 can include a status data structure of the visual indicator 22 1 , a status data structure to the partitionable entity allocation 222 of the partition, and a status data structure of the separable physical location 223 'which can be associated with the hypervisor The code is used to adjust access to the physical indicator and to activate and deactivate the physical indicator. A special user interaction interface is provided to the dispatchable hypervisor 203 for use by system administrators, service personnel, or similar authorized users. In a consistent example, where the system 1 includes a service processor 1 〇 3 and an attached hardware management console 114, the HMC 114 provides an interface to the dispatchable hypervisor for use in Service and partition management, and will be assumed in this description. Above the non-dispatchable hypervisor 202 is a plurality of logical partitions 204 through 207. Each logical partition (from the point of view of the processing program executing within it) operates as an independent computer system with its own memory space and other resources. Each logical partition thus contains a separate operating system core, which is identified herein as r 〇s cores 211 through 214. The 层s core level and _L 'per-partitions operate differently, and thus Figure 2 represents the OS core as four different entities 211 to 214 corresponding to four different partitions. In general, each 〇8 core 21丨 to 214 performs a substantially equivalent function. However, it is not intended that all of the 〇S cores 211 to 214 are identical copies of each other, and that they may be different versions of an architecturally equivalent operating system, or even architecturally different operating system modules. The QS cores 211 to 214 perform various kinds of tactics such as task assignment, paging, and enforcement of multiple tasks. Doc 201102805 data integrity and security, and so on. Above each of the os cores in each of the individual partitions, there may be a set of high level operating system functions, and the user application code and data (not shown). The user can create code on the OS core and level. The code calls the high-level operating system function to access the OS core, or the code can directly access the os core. In the CongyiTM operating system, a user-accessible architecturally fixed "machine interface" forms the upper boundary of the os core (os core called SLIC), but it should be understood that different operating system architectures can be differently This interface is defined and it will be possible to use logical partitions to operate different operating systems on a common hardware platform. One of the logical partitions is designated as a service partition as appropriate. In Figure 2, partition 205 is designated as a service partition, which is a partition that is assigned for use by system maintenance personnel to perform various management and maintenance functions. This service partition can be used exclusively for management and maintenance functions, or it can be used for user applications. However, where the system contains a hardware management console 114 (as shown in the illustrated embodiment of Circle 1), most of the service and maintenance functions are performed from the hardware management console rather than the service partition. As mentioned initially, the basic problem addressed by the present invention is the need for a generalized communication facility between the hardware management console (HMC) and the logical partition of the logically cut data processing system. The system is for example by Armonk, New York, U. S. A. ) The Power® computing system offered by International Business Machines Corporation. A data processing system that implements a hypervisor-based communication facility (such as described below) as a specific commercially available example can be based on 146527 found in IBM's p/i Series product line. Doc 201102805 Firmware and technology in systemware (as in Power.) Org (http://www. Power. Org/members/developers/specs/PAPR_Version_ 2. 7_090ct07. This material is hereby incorporated by reference in its reference to the "High Strength Computing Architecture Platform Reference" (PAPR) material, which is hereby incorporated by reference. (IBM®, pSeries®, iSeries®, and Power® are international commercial machine companies (Armonk, New York, U. S. A. ) registered trademark. The communication facilities presented in this paper can be used for system management tasks such as simultaneous hardware maintenance, dynamic logic partitioning, inventory collection, virtual I/O device mapping, and more. This hypervisor-based communication facility is a general-purpose, low-latency, fully asynchronous communication method between the HMC and the logical partition. This is in contrast to existing communication facilities that are non-universal or require additional physical connections between the HMC and the logical partition (such as real LAN connections and additional configuration tasks). In general, a robust, zero-maintenance, low-latency, fully asynchronous, robust communication facility between the HMC and the logical partition is presented herein, which is implemented in the hypervisor. The communication facility implemented by this hypervisor is referred to herein as a hypervisor pipeline. The versatility of the hypervisor pipeline is that it is not specific to any particular type or class of commands or flows between the HMC and the hypervisor (or between the hypervisor and the logical partition) (such as dynamic logical partition versus virtual I) /O 'non-synchronous pair synchronization, etc.). The hypervisor pipeline is zero-maintained in the sense that the pipeline itself is unaware of the specific commands (i.e., requests and responses) that are flowing between the HMC and the logical partition. These particular commands are treated as goods that reside in a basic or general transport primitive that flows through the hypervisor pipeline. Hypervisor identification of general use transmission 146527. Doc •19· 201102805 Base 7L 'but does not examine or dissect the goods contained in the generic transport primitive. Therefore, when a new command stream is introduced between the HMC and the logical partition, the hypervisor is not affected. * Super-organic channels are designed to minimize waiting time for communication through the pipeline. The waiting time caused by the hypervisor is mainly the direct memory access (DMA) of these communication streams to the flexible service processor (FSp), which is the component by which the HMC communicates with the hypervisor, or the flexible service processing. Benefit (FSP) Direct § Memory Access (DMa) The time required for the communication streams and the communication streams to be logically partitioned from the DAS or from the logical partition. The processing by the hypervisor pipeline is mainly the delivery, pacing and statistics collection between one or more HMCs and one or more logical partitions. In order to minimize the system resource requirements (ie, memory, buffers, tasks) of the hypervisor pipeline while maximizing the throughput of the hypervisor pipeline and preventing a bad operation (ie, not responding in time) The logical partition or HMC negatively affects the throughput of other partitions and HMCs. The super-management is essentially non-synchronous and uses a separate buffer pool for each logical partition and per-HMC (ρ 〇〇ι). In other words, if [H1 starts from one to the partition? 1 communication flow, but ρι is extremely busy or hung and does not respond, then the hypervisor will always deliver the communication flow to P1 without buffering the buffer buffer, and then respond to m (After the ?1 does not respond, then H1 can wait for the response flow from the time!>), or if the P1 buffer pool at the super-experimental device is exhausted, the hypervisor is busy. The status responds to HMC H1. Pipeline tasks and hypervisors are not blocked 146527. Doc •20· 201102805 and via dk) The wait for a response or response from the HMC or partition, the busy state or the timeout of the communication flow itself, passes the disposition-free target back to the source of the communication stream. The robustness of the hypervisor pipeline presented in this article is that there are no input/output (I/O) adapters and communication cables' to reduce the chance of a configuration problem. μ Therefore, a hypervisor-based communication facility (ie, hypervisor = track) is interfaced between the logically partitioned data processing system 2HMc (external management device) and the logical partition. The hypervisor pipeline is a generic: L «reconciliation, so that it can handle only ^1 (: many types or categories of streams/commands between logical partitions. Hypervisor pipelines only recognize the basic transport base疋, rather than a specific command (request or response) that flows between the HMC and the logical partition (which is considered a simple cargo or additional material), and therefore whenever a new command between the HMC and the logical partition, The hypervisor pipeline is unaffected when a request or response flow is introduced to the data processing system. The hypervisor pipeline can deliver traffic between one or more HMCs and one or more logical partitions, and does not allow no response. The partition or HMC blocks the transmission of traffic or negatively affects the performance of their flow to and from other partitions or HMCs. The communication facilities presented herein are superior to the conventional RMC resource management mechanisms described above because they are not required Separate cross-LAN connection between each HMC and the logical partition, thereby reducing hardware requirements (no lan adapter), configuration effort (no network management, cabling), and failure rate (no adapter or cable) Failure). , Still connected to the management system having a plurality of logical partitions, I46527 a network between each Held HMC HMC and management of the system. Doc 201102805 The communication facilities presented in this paper show a potentially significant reduction in network extra load compared to the RMC resource management mechanism. Figures 3 through 5 depict three different views of a hypervisor tube in accordance with aspects of the present invention. In FIG. 3, a solid frame view is presented, wherein the logically partitioned data processing system 300 includes a plurality of logical partitions 31 (labeled ρι, ?2, and P3), including the hypervisor pipeline described herein (or a hypervisor 320 of the communication facility 325, a flexible service processor 33A, and a plurality of hardware management consoles (HMCs) 340 that are networked to the super via a flexible service processor (FSP) 330 Manager 32〇. The communication between the Fsp 33〇 and the hypervisor 320 is via direct memory access (DMA), and the hypervisor and logical partitions also communicate via a form of partitioned interval DMA. The super-B processor pipeline component in the super-g processor performs flow delivery, adjustment progress, and statistical data collection. 4 depicts a lower level logical view of the data processing system 3, where the HMC 34G and the hypervisor 32G (or more specifically, the hypervisor pipeline 325) establish a logical HMC command session via the FSP 330. . As shown, there is a single-communication session 400 between each listener and hypervisor. Associated with each HMC to the hypervisor session is the §fl buffer pool 4 〇 5 in the hypervisor pipeline. As follows, whoever + ^ r progresses and explains this message buffer is used to implement a hypervisor-based communication facility between the HMC and the logical partition. Similarly, the logical partitioning and the hypervisor establish a logical segmentation event, and the communication flow is exchanged via the general-purpose transport primitives on the logical segmentation event. For each-logical partition to the hypervisor session 146527. Doc • 22· 201102805 410 again defines a separate message buffer pool 415. In operation, the hypervisor parses the generic primitive transfer commands to determine the target and move the goods between the respective HMCs and the logical partitions. In one implementation, 'in addition to the general purpose transport primitives', three types of session management requests are used, that is, between the HMC and the hypervisor pipeline (and each logical partition in the hypervisor pipeline). Between the use of open communication session request, close communication session request and exchange capability request. Thus, in one embodiment, a request from an HMC or a logical partition may include an open request, a switch capability request, a close request, or a general transfer primitive (referred to as a HypePipe request in the logic flow described below). 5 depicts a logical representation of a higher level of data processing system 3, wherein the logical pipeline presented herein can be considered to establish each listen 340 and each partition-to-point, logical HMC via data manager 320. The communication period for the partition is 5〇〇. This figure is a conceptual view of the communication frame at its highest level. The endpoint in the HMC sends the communication stream to the endpoint in the logical partition and vice versa. 6 through 10 depict a consistent embodiment of implementing a super-manager-based device (such as the hypermanipulator pipe) logic of a hypervisor-based device, such as Figure U to Figure 13 depicts a particular endpoint logic flow that may be implemented within an HMC or logical partition in accordance with aspects of the present invention. Referring first to Figure 6, the hypervisor & When used in this article, the original point is HMC or logical partition, = can include a session management request (such as opening a session request, exchanging this request or _ session please 峨 super-channel (or office (four) 146527 . Doc -23· 201102805 Request. The hypervisor pipe request is a generic or basic transport primitive as described in this article. The endpoint uses this primitive to communicate via the hypervisor pipeline. In one embodiment, the general purpose transport primitive has the following data structure: target endpoint ID | cargo length 丨 cargo (encapsulated request or response communication) 丨 super manager initial decision to receive source endpoint request 605 is An open session management request or an exchange capability session management request 6 丨〇 ^ if 任 者 ', then set the session state between the source endpoint and the hypervisor to enable 615 'and in the hypervisor The capability set by the source endpoint is set to 62〇. The answer status (ACK status) is set to success 625 and the response (including its status and adjusted capabilities) is passed back to source endpoint 630. If the source endpoint request does not open the session management or exchange capability session management request, the hypervisor determines whether the s request is a closed session management request 6 3 5 . If so, the expiration state between the source endpoint and the hypervisor is set to off 64〇, and any queued or pending requests 645 for the duration of the session are canceled. The hypervisor sets the state to success 650 and sends a reply with the attached state back to the source endpoint 630, then returns to wait for the next source endpoint request. If the source endpoint request is not a session management request, the hypervisor determines if it is a basic transport primitive (i.e., a hypervisor pipe (or Hypepipe) request). If "yes", the hypervisor returns from the query 655 to perform the process 66 of Figure 7 (described below) to wait for the next source endpoint request. If the source endpoint request is not a session management request or a hypervisor pipe request, then the request is an unrecognized request, and thus the response status is set to error 665 and a response 630 with an error status is returned. Assume that the hypervisor pipeline receives the hypervisor pipe from the source endpoint 146527. Doc •24· 201102805 Seeking (ie, general-purpose transport basis), the logic flow of Figure 7 is performed. The hypervisor pipeline determines if the generic transport primitive has an acceptable cargo size (i.e., has a size less than or equal to the maximum transfer unit (MTU) size) 700. If "No", the response status is set to ]^[11; violation 7〇5, and the hypervisor returns 770 to the logic of Figure 6. Otherwise, the hypervisor determines if the target endpoint identification is valid endpoint identification within the data processing system. If "No", the response status is set to an invalid parameter 7丨5, and the hypervisor returns 770 to the logic of Figure 6. Assuming the cargo size and target endpoint are recognized as acceptable, the hypervisor determines if the target message buffer pool is empty 720. If "yes", the answer status is set to busy 725 and the hypervisor returns 770. Otherwise 'take out a target message buffer 730' from the target message buffer pool and create a new hypervisor pipe request (ie, target transport primitive) in the target message buffer, which includes the goods from The source hypervisor pipe request (ie, the source transfer primitive) is copied to the target hypervisor pipe solicitation 735. Once the target transport primitive is built, the hypervisor determines if the target endpoint's duration is on. If "No", the response status is set to the expiration of 745, and the hypervisor returns 770 to the logic of Figure 6. Otherwise, the hypervisor determines whether it is in the process of interacting with the target endpoint (i.e., the session state is an exchange capability or a wait capability response) 750. If "yes" then the response status is set to busy 755' and the hypervisor returns 770 to the logic of Figure 6. Otherwise, the target transport primitive is sent asynchronously to the target endpoint 760, and the reply state is set to success 765, after which processing returns to 770. Figure 8 depicts a 146527 for processing hypervisor logic from a response from an endpoint. Doc 25· 201102805 Example. The hypervisor waits for a reply 800 from the endpoint. Upon receiving the J 805, the hypervisor determines if the response is an exchange capability response 81〇. If "yes", the hypervisor determines whether the endpoint fulfills (h〇n〇r) the processor's ability to look at 8 1 5 . If YES, the endpoint-to-hypervisor session status is set to 820, and the message buffer used in the exchange capability request is passed back to the respective endpoint message buffer pool 825. If the endpoint does not fulfill the hypervisor capability, then only the message buffer is passed back to the respective endpoint message buffer pool' without setting the endpoint to hypervisor session state to on. After the message buffer is returned to the corresponding endpoint buffer pool in the hypervisor pipeline, the hypervisor determines whether the session status is required to exchange the duration status 830. If "No", then the processing is processed. Return to wait for a response 800 from the endpoint. Otherwise, logic 835 of Figure 9 (described in the text) is executed. The response received by the right is not an exchange capability response, and the hypervisor determines if the air response is a hypervisor pipe response 840. If "Yes", check the status of the attached return to see if it is - you need to exchange: #^. If yes, the status is set to the required transmit capability and the endpoint message buffer is sent back to the corresponding endpoint message buffer pool. If the response received is not the exchange capability or the hypervisor pipe response ^ ° ° . The manager fails to recognize the response 855 and returns to wait for the next-and-successive state to assume that the status of the response to the received response is the need to exchange, then from 830, the logic 835 of Figure 9 is executed. This logic performs a hypervisor exchange with the corresponding endpoint (with which communication session is established). The capability exchange γ is compliant with one of the required exchange capabilities. The communication session responds to the U state, the hypervisor from 146527. Doc • 26· 201102805 The corresponding message buffer pool gets the game component—the endpoint message buffer 9〇〇, and it is built in the message buffer—exchange, month b force request 905. Then, the communication manager's communication session status is set to the waiting capability response 91〇, and the hyper management writes asynchronously sends the exchange capability request to the respective endpoint 915 of the communication session, and then returns to the time when the logic is called. Logic flow 920. . Figure 10 depicts an example of a logic that can be used to process a capability change event within a data processing system. In the event of a capacity change event, consider the time of each HMC 1〇〇5. For example, the capability change event may be a simultaneous update of the super-e-bene determination to determine whether there is still a hmc session (8) to be considered, and if "yes", Bj1 gets the next-HMC (4)15. The status of the session of the HMC session is set to require the exchange capability, and the hypervisor determines whether the target HMC message buffer pool is empty. If yes, consider the next HMC session period of 1〇〇5. Otherwise, perform the logic ι〇3〇 of Figure 9, and then the hypervisor returns to evaluate the next-HMC session...has processed all the term sessions' then since the query 1G10's hypervisor evaluation per-logical segmentation (Lp) The meeting will be 1035. The hypervisor determines if there is another open communication session 1040' and if "no", the capability exchange logic has completed 1 () 45. If there is another LP session to be considered, the next communication session is obtained 1〇5〇, and its session status is set to require the exchange capability 1〇55. The hypervisor determines whether the message buffer pool of the target logical partition is empty 1〇6〇, and if YES, processes the next LP session. If the message buffer pool is not empty, then the logic 1 〇 65 of Figure 9 is performed for the obtained LP session. Figure 11 depicts an embodiment of endpoint logic for processing hypervisor pipe requests (generic transfer primitives) within a logical partition or HMC. The endpoint waits for 146527. Doc 27· 201102805 A hypervisor pipe requests blue, and upon receipt, 11货物5 copies the goods to a local buffer 1110. The endpoint then determines if the shipment contains a valid endpoint request 1115. If YES, the endpoint responds to the hypervisor pipe request i i 2 G with a success status. The endpoint then determines if the spot request source switching capability 1125 is required. If "yes", a response is sent to the source m in response to a new hypervisor pipe request as a shipment. The response contains a status 1130 indicating the need for switching capability. If there is no need to exchange capabilities with the source endpoint, then the target endpoint processes the request 1135 and determines if a response to the request is required 1140. If "no" then the endpoint returns to wait for the next hypervisor pipe request. No shell,! Generating a response and encapsulating it as a new hypervisor pipe request sent back from the target endpoint to the source endpoint, the response having a result based on processing the first hypervisor pipe request at the target endpoint While the set state 1145 » The right cargo does not contain a valid request, the endpoint determines if the shipment contains a valid response 1950. If YES, a response to the hypervisor pipe request is sent, wherein the status is set to success 丨丨 55, and the endpoint determines whether the response is for the exchange capability request 116 〇 β if “No”, then the response is processed. 1165, and if "yes", the target endpoint records the fact 1170 that the capability has been exchanged with the source endpoint. If the shipment does not contain a valid request or a valid response, then a response 1175 to the hypervisor pipe request with an error status is sent. Figures 12 and 13 depict two methods for exchanging capabilities in response to a capability change event at an endpoint. Figure 12 is the active method and Figure 13 is the passive method. Referring first to Figure 12, the endpoint logic determines that one of the endpoints may have changed 146527. Doc -28- 201102805 The event of force has occurred 12Q(). For example, an endpoint power up or a simultaneous boost has occurred. The endpoint logic sets the capability exchange state to indicate that all endpoint switching capabilities 12〇5 have not been aligned, tracking capability exchange status. A switch capability request is then sent to the hypervisor 121G. In an embodiment, the pair: power on can be 35 to exchange capabilities by opening a session request instead of an exchange capability request. After considering all possible target endpoints 1215, the primary endpoint (4) 卟 ct endpoint determines if there is another endpoint 122 to be considered. Right: Yes, the next endpoint 1225' is obtained and a request for the transactional pipeline request for the goods is sent to the endpoint. Once all endpoints have been considered, processing is completed 1235. In alternative method t, the logic of Figure 13 can be used to exchange capabilities in response to a capability change event in an endpoint. In this method, an event occurs that may have been handed over to the ability in the master knowing point, such as a simultaneous update of n〇〇. The capability exchange record at the primary endpoint is changed to indicate that all endpoint exchange capabilities have not been exchanged for the capability being tracked. The endpoint then sends an exchange capability request to the hypervisor n131G. Looking again, powering on (4) points can exchange capabilities with the hypervisor by opening a session request instead of an exchange capability request. Once the exchange capability request has been forwarded to the super processor, processing is completed 1315. In this approach, the endpoint does not actively push the switching capability request to other endpoints, but will wait for other endpoints to initiate a traffic flow and will respond to the traffic flow with a required exchange capability state. For further explanation, various command structures for implementing the hypervisor-based communication facility described above in, for example, the IBM P0 Gift 8 computing system are described below. The following discussion is provided as an example only, and should be considered 2, 146527. Doc -29- 201102805 The scope of the patent application presented with this specification is not limited to the specific embodiments presented below. Terms and concepts are used herein to use the following terms: • Upstream - In this context, based on a common view of a logically partitioned system (including HMC) with partitions located above the firmware, the upstream refers to a self-HMC The flow to the partition. • Downstream - In this context, based on the common view of a logically partitioned system (including HMC) with partitions located above the firmware, downstream refers to a flow from the partition to the HMC. • Inbound - In the context of the ingress and egress and HMC flows, the entry refers to the flow into the partition or HMC, which should be resolved from the point of view of the HMC or partition rather than the hypervisor (HYP). . • Out – In the context of entering and leaving the partition and the flow of the HMC, the entry refers to the flow leaving the partition or HMC. The term should be interpreted from the perspective of HMC or partition rather than ΗΥΡ. High-level flow This section discusses the main interactions and events that occur between subsystems involved in the hypervisor pipeline flow and in subsystems. Two methods for packetizing the response data associated with the upstream request from the HMC to the partition are possible. The first method is a "pull" method whereby the source endpoint of the request sends one or more GetRemainingResp〇nseData streams to the target of the request after obtaining a response to the initial request until it has been retrieved. All relevant response information. No. 146527. Doc -30- 201102805 The two methods are the “push” method, whereby the target of the request repeatedly sends a response flow for the request until all response data associated with the request is passed back to the source of the request. The benefit of the pull method is that it allows the source of the initial request to retrieve data at its own rate. The benefit of the push method is that it involves a generally small flow through the hypervisor pipe, and it allows the target to immediately send all response data back to the source without having to deal with situations where the source directly discards the remaining data. A potential disadvantage is that if the target sends back a number of response streams in close proximity, it can get a "busy" status from the hypervisor. This can be aggravated if other partitions are sending streams to the same HMC at the same time. In addition, in theory, the packet may be lost, which means that the HMC should monitor this situation and should retry the initial request in the event of a loss. By pulling the method, the HMC will only retry the failed GetRemainingResponseData stream. Open Session This is an HMC message interface. The HMC uses the HMC message interface to open a session with the hypervisor and exchange with the hypervisor. A session can be opened for each attached HMC. For a single HMC, multiple sessions cannot be opened. In the parameters below, the entry and exit are from the perspective of HMC. Open the session request parameter: 146527. Doc •31 - 201102805 Name Description Maximum number of incoming messages The maximum number of incoming messages that are unprocessed Maximum number of outstanding outgoing messages The maximum number of outgoing events The number of capability bytes The number of capability bytes is the number of capability bits Each bit of the group represents a capability. The HMC initiates a bit indicating its ability to support, and the hypervisor closes its bits indicating its unsupported capabilities. Open session response parameters: In addition to the "Maximum incoming message" and "Maximum outgoing message" will contain the hyper-reporter counter proposal (which will always be less than or equal to the number specified by the HMC), the response parameters The capability bit mask will be the same as the mask passed in the request, except for the request parameters, and except that the bits representing the capabilities not supported by the hypervisor are turned off. The hyper-manager will have a maximum of 8 and 1 for the "maximum entry event" and "maximum exit event". If the HMC specifies a higher value, the hypervisor will deny this value. If the HMC specifies a lower value for some reason, the hypervisor will perform the lower value, but the pipeline throughput can be negatively affected. An initial value of zero will result in a GenericError. Open the session status code: Pass back in the common header. No command specific return code. Possible values • Good • InvalidOpcode • GenericError • InvalidState - Issues a start-up period at the same time as the session is started. 146527. Doc -32- 201102805 Close Session This is an interface through which the HMC closes the command period with the Hypervisor. There are no request or response parameters, and there are no session-specific return codes. Switching Capability: This is an HMC message interface. The HMC and HyperManager exchange capabilities through the HMC messaging interface. This command can be sent multiple times during the duration/life of the session. The command can be sent immediately after the opening session, and/or later when the session is actually used. HMC and HYP must support this command as both source and destination. If the capabilities of the HMC or Hypervisor change due to simultaneous firmware updates, this command can be used to exchange capabilities (assuming non-simultaneous firmware updates will cause the session to close and an open session to occur after the update). Each of the two entities must verify that their current capabilities have been exchanged with the source whenever they receive a flow that is not an exchange capability. If not already exchanged, the ExchangeCapabilities return code must be returned. If an event that changes or may have changed the ability of the target occurs, the target must exchange capabilities with all sources again. Request Parameter ‘· Name Content Description Number of Capacity Bytes Number of Capacity Bytes Below Capacity Bytes Each bit represents a capability. The source initiates a bit indicating its ability to support, and the target closes its bits indicating its unsupported capabilities. Response parameters: The response parameters are the same as the request parameters except that the bit indicating that the capability is not supported by the target is off. 146527. Doc •33- 201102805 Return Code: Pass back in the common header. No command specific return code. Possible values are .  • Good • InvalidOpcode • GenericError Hyper-Manager Pipeline Request: This is an HMC message interface that the HMC communicates directly with the partition via the HMC message interface. The hypervisor acts as a simple pipe. It does not examine the data that is piped through the pipe to the partition or from the partition. It accurately copies the data area (goods) without modification. This is a completely asynchronous command. When the HMC sends the command to the hypervisor, the hypervisor routes the goods within the command to the specified partition. When a partition sends a stream destined for the HMC to the hypervisor, the hypervisor sends the command along with the goods from the partition to the designated HMC. All timings for these orders are the responsibility of the HMC and the division. Request parameters: Name Description of content Source/target ID For outgoing (from HMC to HYP) streams, the HMC is set to the target ID. For incoming (from HYP to HMC) streams, HYP is set to the source ID. Cargo Size The size of the following goods (in bytes). For turbulences that target partitions, this value cannot exceed 160 bytes, and for inbound turbulence with HMC as the target, this value cannot exceed 1488 bytes. Goods goods 146,527. Doc -34- 201102805

Ack參數: 名稱 内容描述 傳回碼 參見下文之「傳回碼」 額外傳回碼資料 之長度 額外傳回碼特定的資料(若存在)之長度(以位元組為單位) 額外傳回碼資料 額外傳回碼特定的資料(若存在) 傳回碼: • Success(成功) • Failure(失敗) • Busy(忙碌) • InvalidParms(無效參數) • PipeClosed(管道關閉) • MtuViolation(:Mtu違規) • ExchangeCapabilities(交換能力) 傳回碼:Ack parameters: Name Description Return code See “Return Code” below. Length of extra return code data. Extra return code. Length of specific data (if any) (in bytes). Additional return code data. Additional return code specific data (if any) Return code: • Success • Failure • Busy • InvalidParms • PipeClosed • MtuViolation (: Mtu violation) • ExchangeCapabilities (return capability) return code:

Success(>§fe' ·* 成功地完成請求。 恢復:N/A Failure(失敗): 請求失敗,且無其他資訊。 恢復:重試該請求。若問題繼續存在,則開始標準分 析。Success(>§fe' ·* successfully completed the request. Recovery: N/A Failure: The request failed with no other information. Recovery: Retry the request. If the problem persists, start the standard analysis.

Busy(忙碌): 未執行請求,因為請求目標(或可能為超管理器)忙碌。 預期此狀況為暫時的。 146527.doc -35- 201102805 恢復:在短暫延遲之後,重試程序。若該狀況繼續存在 一延長時間段(亦即,若干分鐘),則目標或超管理器中有 可能存在錯誤。 invalidParms(無效參數): 未執行請求,因為命令含有無效或未辨識之參數。無效 目標ID為將導致此傳回碼之狀況的實例。 恢復:無Busy: The request was not executed because the request target (or possibly the hypervisor) is busy. This situation is expected to be temporary. 146527.doc -35- 201102805 Recovery: After a short delay, retry the program. If the condition persists for an extended period of time (i.e., several minutes), there may be an error in the target or hypervisor. invalidParms (invalid parameter): The request was not executed because the command contains invalid or unrecognized parameters. Invalid The target ID is an instance that will cause the status of this return code. Recovery: none

PipeClosed(管道關閉): 未執行請求,因為到達指定目標之管道關閉或不在作用 中。此可指示目標斷電或處於故障狀態。 恢復:通電或對目標進行IPL。PipeClosed: The request was not executed because the pipe that arrived at the specified target was closed or not active. This can indicate that the target is powered down or in a fault condition. Recovery: Power up or IPL the target.

MtuViolaiion (Mtuxi Μ): 指定貨物之大小超過最大傳送單元(MTU)之大小。 額外資料: 位元組 描述 0-3 所支援之MTU大小 恢復:使用一所支援之貨物大小重試。 ExchangeCapabilities(交換能力): 未執行請求,因為目標之能力自上次交換能力之後可能 已被改變。 恢復:發布ExchangeCapabi丨ities請求,且接著重試失 敗之請求。 邏輯分割(LP)事件: 開啟會期: 146527.doc -36- 201102805 此為一 LP事件介面,分割區經由該LP事件介面而在會 期開啟協定期間與超管理器協商能力及最大入埠/出埠 值。分割區始終為該事件之源/起始方,而超管理器始終 為目標。 在會期開啟之後,不可再次發送此事件。會期管理程式 碼監視該事件ID,且若其在會期開啟之後出現,則會期管 理程式碼將拒絕該發送。 在下文之參數中,入埠及出埠係自分割區之觀點而言。 請求參數: 名稱 内容描述 事件ID 事件碼。 最大入埠事件 未處理之入埠事件的最大數目。 最大出埠事件 未處理之出埠事件的最大數目。 能力位元組之數目 以下能力位元組之數目 能力位元組 每一位元表示一能力。分割區啟動表示其支援之能力的位 元,超管理器關閉表示其不支援之能力的彼等位元。 回應參數: 除了「最大入埠事件」及「最大出埠事件」將含有超管 理器之反建議(其將始終小於或等於分割區所指定的數目). 之外,回應參數與請求參數相同,且除了表示不被超管理 器支援之能力的位元為關閉之外,能力位元遮罩將與在請 求中傳遞之遮罩相同。 超管理器將對「最大入埠事件」及「最大出埠事件」反 建議之最大值分別為8及1。若分割區指定較高值,則超管 理器將反建議此等值。若分割區出於某種原因指定較低 值,分割區將履行該等較低值,但管道輸送量可受到負面 146527.doc -37- 201102805 影響。初始值零將導致GenericError(—般性錯誤)。 傳回碼: 在共同標頭中傳遞回。無特定事件傳回碼。可能值為: • Good(良好) • GenericError(—般性錯誤) 交換能力: 此為一 LP事件介面,分割區與超管理器在會期開啟之後 經由該LP事件介面而交換能力。不同於開啟會期,在會期 之持續時間/壽命内,可多次發送此命令。分割區及HYP必 須作為源與目標皆支援該命令。 若分割區或超管理器之能力由於同時韌體/程式碼更新 而改變,則可使用此命令來交換能力(假定非同時韌體/程 式碼更新將引起會期關閉且在更新之後使一開啟會期發 生)。兩種實體每當接收到一以其為目標的並非交換能力 之流時必須驗證其當前能力已與源交換。若尚未交換,則 必須傳回ExchangeCapabilities(交換能力)傳回碼。若一改 變或可能已改變目標之能力的事件發生,則目標必須再次 與所有源交換能力。 請求參數· 名稱 内容描述 事件ID 事件碼。 能力位元組之數目 以下能力位元組之數目 能力位元組 每一位元表示一能力。源啟動表示其支援之能力的位 元,目標關閉表示其不支援之能力的彼等位元。 回應參數: 146527.doc -38- 201102805 除了表示不被目標支援之能力的位元為關閉之外,回應 參數與請求參數相同。 傳回碼: 在共同標頭中傳遞回。無事件特定的傳回碼。可能值為: • Good(良好) • GenericError(〆般性錯誤)MtuViolaiion (Mtuxi Μ): The size of the specified item exceeds the maximum transfer unit (MTU). Additional Information: Bytes Description 0-3 Supported MTU Size Recovery: Retry with a supported cargo size. ExchangeCapabilities: The request was not executed because the capabilities of the target may have changed since the last exchange capability. Recovery: Issue an Exchange Capabi丨ities request and then retry the failed request. Logical Segmentation (LP) Event: Open Session: 146527.doc -36- 201102805 This is an LP event interface. The partition uses the LP event interface to negotiate with the hypervisor during the session open protocol and the maximum entry/ Depreciation. The partition is always the source/starter of the event, and the hypervisor is always the target. This event cannot be sent again after the session is turned on. The session management code monitors the event ID and if it occurs after the session is turned on, the term management code will reject the transmission. In the parameters below, the entry and exit are from the point of view of the partition. Request parameters: Name Description Event ID Event code. Maximum number of events The maximum number of unprocessed events. Maximum outbound event The maximum number of unprocessed outbound events. Number of capability bytes The number of capability bytes below The capability byte Each bit represents a capability. The partition initiates a bit indicating its ability to support, and the hypervisor closes its bits indicating its unsupported capabilities. Response parameters: In addition to the "maximum entry event" and "maximum exit event" will contain the hyper-reporter's counter-proposal (which will always be less than or equal to the number specified by the partition). The response parameters are the same as the request parameters. And in addition to the bit indicating that the capability not supported by the hypervisor is off, the capability bit mask will be the same as the mask passed in the request. The hyper-manager will have a maximum of 8 and 1 for the "maximum entry event" and "maximum exit event". If the partition specifies a higher value, the HyperManager will deny this value. If the partition specifies a lower value for some reason, the partition will perform the lower value, but the pipeline volume can be affected by the negative 146527.doc -37- 201102805. An initial value of zero will result in a GenericError. Return code: Pass back in the common header. There is no specific event to return the code. Possible values are: • Good • GenericError Switching capability: This is an LP event interface. The partition and hypervisor exchange capabilities via the LP event interface after the session is started. Unlike the opening period, this command can be sent multiple times during the duration/life of the session. The partition and HYP must support this command as both source and destination. If the ability of the partition or hypervisor changes due to simultaneous firmware/code update, you can use this command to exchange capabilities (assuming non-simultaneous firmware/code updates will cause the session to close and make an open after the update) The meeting will take place). Each of the two entities must verify that their current capabilities have been exchanged with the source whenever they receive a stream that is not a switching capability. If not already exchanged, the ExchangeCapabilities return code must be returned. If an event that changes or may have changed the ability of the target occurs, the target must exchange capabilities with all sources again. Request Parameters · Name Description Event ID Event Code. Number of capability bytes The number of capability bytes below The capability byte Each bit represents a capability. The source initiates a bit indicating its ability to support, and the target closes its bits indicating its unsupported capabilities. Response parameters: 146527.doc -38- 201102805 The response parameters are the same as the request parameters except that the bit indicating that the capability is not supported by the target is off. Return code: Pass back in the common header. No event-specific return code. Possible values are: • Good • GenericError

HypervisorPipeRequestlnbound(超管理器管道請求入埠> : 此為一 LP事件介面’超管理器經由該LP事件介面而將 一超管理器管道流傳送至一經指明為該流之目標的分割 區。自分割區之觀點而言’該流為入埠的。 此為一非同步事件。ACK僅應答分割區接收到該事件且 意欲處理該事件。在ACK中,應僅報告阻止分割區最終處 理請求之錯誤。換言之,若分割區由於某種如無效命令參 數之原因而將永不處理請求,則分割區應以一不良傳回碼 (諸如,InvalidParms(無效參數))ACK 該 HypervisorPipeRequestlnbound (超管理器管道請求入埠)。在HypervisorPipeRequestlnbound (超管理器管道請求入埠)經ACK之後,應在超管理器管道 請求出埠事件中經由貨物而將在處理該請求時發生之錯誤 報告回給源實體。 請求參數: 名稱 内容描述 事件ID Ϋ件碼。 源ID 源識別 貨物大小 以下貨物之大小(以位元組為單位)。 貨物 貨物 146527.doc •39· 201102805 ack參數: 名稱 内容描述 傳回碼 參見下文之「傳回碼」。 額外傳回碼資料之長度 額外傳回碼特定的資料(若存在)之長度 額外傳回碼資料 額外傳回碼特定的資料(若存在) 傳回碼: • Success(成功) • Failure(失敗) • InvalidParms(無效參數) • ExchangeCapabilities(交換能力) 超管理器管道請求出埠 此為一 LP事件介面,分割區經由該LP事件介面而作為 流之來源起始一超管理器管道流。自分割區之觀點而言, 該流為出埠的。超管理器將貨物輸送至目標實體。 此為一非同步事件。ACK僅應答超管理器接收到該事件 且意欲將其轉遞至目標。在ACK中,將僅報告阻止超管理 器將請求轉遞至目標之錯誤。 分割區可指定是否DMA該貨物或其是否包括於此事件 中。分割區可決定僅在貨物對於該事件而言過大之情況下 進行DMA或一直進行DMA,雖然當貨物小到足以裝入該 事件時,DMA與將貨物包括於事件自身中相比較為低效。 若指定DMA(「DMA required(需要DMA)」=1),則分割區 必須指定描述貨物資料緩衝器的一系列邏輯真實位址/長 度對。每一輸入項描述單一相連真實位址範圍。此範圍不 可跨頁。舉例而言,若分割區自系統堆積分配長度為4000 146527.doc -40· 201102805 位元組且跨越一頁邊界之貨物緩衝器,則分割區必須固定 (pin)儲存器且在緩衝器清單中建立兩個邏輯真實位址/長 度輸入項。此外,分割區在超管理器ACK該事件之前不能 釋放含有待DMA之資料的缓衝器。 請求參數: 名稱 内容描述 事件ID 事件識別 —— 目標ID 目標識別 需要DMA 〇-此事件中含有貨物,1=必須自分割區DMA貨物 ‘物大小 貨物之大小(以位兀組為單位)。若|需要DMAj欄位為〇,則 應反映貨物攔位中之貨物的大小,或若「需要DMA」欄位^ 1 ’則應反映待DMA之貨物的大小》 若「需要DMA」=0,則最大貨物大小為176位元組減去請求參 數之大小(176-16=160)。若「需要DMA」=1,則最大貨物大小 為1472位元組。 若「需要DMA , =1 ’則剩餘請求參數如下所展示: _ 緩衝器清單大小 緩衝器,月單之大小(以位元組為單位)。緩衝器清單之格式見下 文0 僅在「需要DMA」經設定時有效。 緩衝器清單位址 _·衝=清單在分割區位址空間中之真實位址(邏輯真實)。緩衝 器清單必須為固定的且駐留於相連真實記憶體中。其不可跨 頁。緩衝器清單之格式見下文。 、 僅在「需要DMA」經設定時有效》 右离要UMA」ϋ,則剩餘請求參數如下所展示: 貨物 丨貨物 -~~ 緩衝器清單: 緩衝器清單為一系列位址/長度對,其中每一對描述一 塊(chunk)相連的真實緩衝器空間。若緩衝器駐留於不相連 之真實記憶體中,則需要多個輸入項。 名稱 内容描述 ----- 緩衝器清單大小 緩衝器β單(包括此爛位)之大小(以位元組為單仇)。 對於每一位址/長; 变對,以下~~~---- 位址 大小 Ϊ、疋真實記憶體在分割區位址空間中之真實位址 Λ) ’自該位址複製一塊貨物》緩衝器不可终百。 |田位址」^定址之緩衝器「塊」的大小(以位元組為簞位 146527.doc -41 - 201102805 ack參數: 名稱 内容描述 傳回碼 參見下文之「傳回碼」。 額外傳回碼資料之大 小 額外傳回碼特定的資料(若存在)之大小(以位元組為單 位)。 額外傳回碼資料 額外傳回碼特定的資料(若存在) 傳回碼: • Success(成功) • Failure(失敗) • Busy(忙綠) • InvalidParms(無效參數) • PipeClosed(管道關閉) • MtuViolation(Mtu違規) • ExchangeCapabilities(交換能力) • BufferNotPinned(緩衝器不固定) •記憶體 Sue(MemorySue) 傳回碼 LP事件傳回碼與用於HMC命令之彼等傳回碼相同,附 加以下LP事件所獨有者。HypervisorPipeRequestlnbound: This is a LP event interface. The hypervisor streams a hypervisor pipeline to the partition specified as the target of the stream via the LP event interface. From the point of view of the zone, the stream is inbound. This is an asynchronous event. The ACK only acknowledges that the partition received the event and intends to handle the event. In the ACK, only the error that prevents the partition from processing the request should be reported. In other words, if the partition will never process the request due to some invalid command parameters, the partition should ACK the HypervisorPipeRequestlnbound with a bad return code (such as InvalidParms). After the ACK of HypervisorPipeRequestlnbound, the error that occurred when processing the request should be reported back to the source entity via the goods in the hypervisor pipe request event. Request parameters: Name Description Event ID ID Code Source ID Source identifies the size of the goods below the size of the goods (in bits) Group of goods.) Cargo goods 146527.doc •39· 201102805 ack parameters: Name Description Return code See “Return Code” below. Extra return code data length Additional code specific data (if any) The length of the extra return code data is additionally returned to the code specific data (if any). Return code: • Success • Failure • InvalidParms • ExchangeCapabilities • Super Manager Pipe Request This is an LP event interface, and the partition starts a hypervisor pipeline flow as a source of the stream via the LP event interface. From the point of view of the partition area, the stream is out of the way. The hypervisor transports the goods. To the target entity. This is an asynchronous event. The ACK only acknowledges that the hypervisor received the event and intends to forward it to the destination. In the ACK, only errors that prevent the hypervisor from forwarding the request to the destination will be reported. The partition may specify whether to DMA the goods or whether they are included in this event. The partition may decide to only proceed if the goods are too large for the event. DMA or DMA all the time, although the DMA is less efficient than including the goods in the event itself when the goods are small enough to load the event. If DMA is specified ("DMA required" = 1), then The partition must specify a series of logical real address/length pairs that describe the cargo data buffer. Each entry describes a single connected real address range. This range cannot be spread across pages. For example, if the partition is allocated from the system to a cargo buffer with a length of 4000 146527.doc -40·201102805 bytes and spans a page boundary, the partition must be pinned to the storage and in the buffer list. Create two logical real address/length entries. In addition, the partition cannot release the buffer containing the data to be DMA until the hypervisor ACKs the event. Request parameters: Name Description Event ID Event identification - Target ID Target identification Requires DMA 〇 - This event contains goods, 1 = DMA goods must be self-divided ‘ Object size The size of the goods (in units of 兀 group). If the DMAj field is required to be 〇, it should reflect the size of the goods in the cargo block, or if the "Require DMA" field ^ 1 ' should reflect the size of the goods to be DMA" If "Requires DMA" = 0, The maximum cargo size is 176 bytes minus the size of the request parameter (176-16=160). If "Requires DMA" = 1, the maximum cargo size is 1472 bytes. If "DMA is required, =1 ' then the remaining request parameters are as follows: _ buffer list size buffer, size of the monthly order (in bytes). The format of the buffer list is shown below 0 only in "requires DMA" It is valid when set. Buffer list address _·rush = the real address (logically true) of the list in the partition address space. The buffer list must be fixed and reside in the connected real memory. It cannot be spread across pages. The format of the buffer list is shown below. Only if "Requires DMA" is set to be valid, right away from UMA", the remaining request parameters are as follows: Goods/Goods-~~ Buffer List: The buffer list is a series of address/length pairs, where Each pair describes the real buffer space connected by a chunk. If the buffer resides in unconnected real memory, multiple entries are required. Name Description ----- Buffer list size The size of the buffer β (including this rogue) (in the case of a single bite). For each address/length; variable pair, the following ~~~---- address size Ϊ, 疋 real memory in the partition address space real address Λ) 'copy a piece of goods from the address' buffer The device can't be finished. |Field address"^ The size of the buffer "block" of the address (in bytes) 146527.doc -41 - 201102805 ack parameters: Name Description Return code See "Return Code" below. The size of the return data additionally returns the size of the specific data (if any) of the code (in bytes). The extra return code data is additionally returned to the code specific data (if any). Return code: • Success( Success) • Failure • Busy • InvalidParms • PipeClosed • MtuViolation • ExchangeCapabilities • BufferNotPinned • Memory Sue ( MemorySue) The return code LP event return code is the same as the return code used for the HMC command, and is unique to the following LP events.

BufferNotPinned(緩衝器不固定) 未執行請求,因為由分割區指定的自其DMA資料或 DMA資料至其之緩衝器不固定。 恢復:改變分割程式碼以固定緩衝器。此最有可能為分 割區中之程式碼錯誤。BufferNotPinned (buffer not fixed) The request was not executed because the buffer specified by the partition from its DMA data or DMA data is not fixed. Recovery: Change the split code to fix the buffer. This is most likely a code error in the partition.

MemorySue(記憶體 Sue) 146527.doc -42- 201102805 未執行^求,因為在存取分割區記憶體時發生記憶體 SUE。 恢復:無。 HMC-i5/OS 命令: 此部分列出作為超管理器管道流中之貨物而在HMc與分 割區之間流動的命令。應注意,此部分中之材料為架構流 經超管理H管道之貨物的彼等者提供實例及指南。 在以下部分中,請求參數為出現於起始該請求之超管理 窃官道流(源傳送基元)之貨物欄位中的參數,且回應參數 為出現於表不對原始請求之回應之超管理器管道流(目標 傳送基元)之貨物攔位中的參數。 命令類別,命令碼: 在下文之命令定義中,「命令類別」值表示一特定類別 之命令,諸如與一特定功能(亦即,會期管理)相關之所有 彼等命令。「命令碼」絲示在彼類別内之—特定命令。 類別值必須在所有類別中為唯—的。碼值僅必須在彼類別 内為唯一的。 針對與-特定請求相關聯之回應的命令碼為針對其上具 有高階位元之請求的命令碼。 交換能力 此為-交換,兩個端點(源及目標)經由該交換而交換定 義與該兩端點間之超管理器管道會期有關之期望及行為的 能力。 可根據兩個端點所需而協商最大入埠/出埠。應自請求 146527.doc -43- 201102805 、’、的觀點來解澤人蟑/出崞。喂CapabiHties(交換 能力)請求之源應將最大人料定為其同時可處理之請求 」、且將最大出埠設定為0。目標應將其出埠請求之 調1進度⑦定為小於或等於由源指定之最大人埠值的一 最大出埠值设定為其同時可處理之請求的數目。 =^著將其出崞請求之調整進度設定為小於或等於由目 —最大出蟀值的一值。作為一實例,假定最大入璋 及=阜皆經協商為值4,此意謂HMC可預期分割區同時支 '達個未處理之上游請求,且分割區可預期服C同時 援夕達4個未處理之下游請求。若存在一尚未被請求流 之源接收到之對應回應流,則認為該請求流為未處理的。 最大入崞/出崞之觀點而言,在通用超管理器管道上不 存:::回應流之請求流不會被認為是未處理的。 而點不具有對其可處置之未處理入埠之數目的可識 U 士可此為s亥端點動態地分配内部訊息及控制區塊 而非預分配該等内部訊息及控制區塊之狀況),則該端點 可將在交換能力中的對應於其可支援之最大未處理入蜂的 值叹疋為一指不無限制的特殊值(諸如,全部OxF)。另— 端:έ可接著將其所要數目之超管理器管道流發送至目標而 不周1進度’但應準備處置因目標可能暫時無法獲取必要 >源(諸如,戒自、4^* 心控制區塊空間等)之狀況而來自目標之 偶然的BUSy(忙碌)傳回碼。 請求參數: 146527.doc 201102805 名稱 —-_ 内容描述 命令類別 命令類-- — 命令碼 命令碼之指示^ —〜〜 需要回應 一指不此流之源是否預期一來自目標之回應流的位元。 請求ID $ 1睛求」超管理器管道流繫结至對應「回應」超香^· 官道流之符記。請求之目標簡單地在回應流中將該值傳遞 丐。請求之源將來自回應流之值與對應請求相關聯。 咅大入崞事件 禾渑理之入埠事件的最大數目。 最大出埠事件 未處理之出埠事件的最大數目。 能力位元組數目 以下此力之位元組數目 能力 ^位7L表不一能力。源啟動表示其支援之能力的_位元,目 標關閉表示其不支援之能力的彼等位元。 回應參數: 名稱 内容描述 - 命令類別 命令類別之指不。 命令碼 命令碼之指示。 傳回碼 參見下文之「傳回碼」 請求ID 來自相關聯之諳求的請求ID值。 最大入埠事件 目標對源之建議的反建議 最大出埠事件 目標對源之建議的反建議 能力位元組數目 以下能力之位元組數目 能力 除了表示不被目標支援之能力的位疋為關閉之外,與在請求 參數中相同。 傳回碼: • Success(成功) • Failure(失敗) • InvalidParms(無效參數) • PresentStateProhibits(目前狀態禁止) • Busy(忙碌) 取消請求 此為一交換,仍未處理之先前請求流之源/起始方可經 由該交換而取消彼請求。其意圖為在由源/起始方維持之 請求計時器過期或源僅僅出於某種原因不願意看到回應的 146527.doc • 45· 201102805 狀況下,使用此流。來自此請求之目標的肯定回應(成功 傳回碼)流意謂目標將不發送針對指定請求ID之回應流’ 且該指定請求ID可由請求之起始方再使用。來自目標之否 定回應(不成功傳回碼)意謂起始方仍可取得一針對指定凊 求ID之回應流,且在接收到回應流之前無法再使用該請求 ID。 請求參數· 名稱 ---内容描述 _ 命令類別 ---------—---- 命令類別之指示。_________ 命令碼 命今媽之指示0 ________r:---r- 需要回應 請求ID 知―n繫結至賴喊」超官理ts ί道求之9標簡單地在回應流中將值傳遞回。 關:________ 待取消之請求ID 與正嘖求對其之取消的先前請求相關聯的睛求比1。 回應參數: 名稱 内容描述 命令類別 命令類別之指示。 命令碼 命令碼之指示。 傳回碼 參見下文之「傳回碼」 請求ID 來自相關聯之請求的請求Π)值。 待取消之請求ID 與正請求對其之取消的先前請求相關聯的請求ID。 傳回碼: 146527.doc -46 - 1MemorySue (memory Sue) 146527.doc -42- 201102805 The execution is not performed because the memory SUE occurs when accessing the partition memory. Recovery: None. HMC-i5/OS command: This section lists the commands that flow between the HMC and the split zone as goods in the hypervisor pipe stream. It should be noted that the materials in this section provide examples and guidance for those who construct goods that flow through the Super Managed H pipeline. In the following section, the request parameter is a parameter that appears in the cargo field of the super-management sneak flow (source transport primitive) that initiated the request, and the response parameter is a super-management that occurs in the response of the table to the original request. The parameter in the cargo block of the pipe flow (target transfer primitive). Command Category, Command Code: In the command definitions below, the "Command Category" value indicates a specific category of commands, such as all of the commands associated with a particular function (i.e., session management). The "command code" is shown in the category - specific commands. Category values must be unique in all categories. The code value must only be unique within the category. The command code for the response associated with the -specific request is the command code for the request with the higher order bits on it. Switching Capability This is the exchange, through which the two endpoints (source and destination) exchange the ability to define expectations and behaviors related to the duration of the hypervisor pipeline between the two endpoints. The maximum entry/exit can be negotiated based on the needs of the two endpoints. It should be self-requested 146527.doc -43- 201102805, ', the point of view to explain the people / out. The source of the CapabiHties request should be set to the maximum number of requests that can be processed at the same time, and the maximum exit is set to zero. The goal should set its outbound request's 1st schedule 7 to be less than or equal to the maximum outbound value specified by the source. The maximum outbound value is set to the number of simultaneous requests that can be processed. =^ Set the progress of the outbound request to a value less than or equal to the target-maximum exit value. As an example, assume that the maximum entry and = 阜 are negotiated to a value of 4, which means that the HMC can expect the partition to simultaneously 'up to an unprocessed upstream request, and the partition can expect to serve C at the same time. Unprocessed downstream request. If there is a corresponding response stream that has not been received by the source of the requested stream, the request stream is considered unprocessed. In terms of maximum entry/exit, there is no deposit on the generic hypervisor pipeline::: The request stream for the response flow is not considered unprocessed. However, the point does not have the number of unprocessed entries that can be disposed of, and can dynamically allocate internal messages and control blocks to the endpoints instead of pre-allocating the internal messages and control blocks. ), the endpoint may sigh the value of the exchange capacity corresponding to the maximum unprocessed bee that it can support as a non-limiting special value (such as all OxF). Another - end: You can then send the required number of hypervisor pipeline flows to the target without the schedule 1 'but should be prepared for the target may be temporarily unavailable to obtain the necessary > source (such as, from, 4 ^ * heart The condition of the control block space, etc.) comes from the occasional BUSY (passive) return code of the target. Request parameters: 146527.doc 201102805 Name--_ Description of content Command class Command class --- Command code Command code indication ^ —~~ Need to respond to a source that does not expect the source of this stream to be expected from a target response stream . The request ID $1 eye-seeking" super-manager pipeline flow is tied to the corresponding "response" super-fragrance ^· official road flow. The target of the request simply passes the value in the response stream. The source of the request associates the value from the response stream with the corresponding request. The maximum number of incidents in the event. Maximum outbound event The maximum number of unprocessed outbound events. Number of capability bytes The number of bytes below this force Ability ^ 7L is not a capability. The source initiates a _bit indicating its ability to support, and the target closes its bits indicating its unsupported capabilities. Response parameters: Name Description of content - Command category Indicates the command category. Command code Command code indication. Return Code See "Return Code" below. Request ID The request ID value from the associated request. The maximum number of anti-recommendations for the source of the event's target is the maximum number of anti-suggestions of the target's proposed anti-recommendation. The number of the number of bytes below the ability to indicate the ability to be not supported by the target is off. Other than the same in the request parameters. Return code: • Success • Failure • InvalidParms • PresentStateProhibits • Busy cancels the request for this exchange, the source of the previous request stream that has not yet been processed The originator can cancel the request via the exchange. The intent is to use this stream if the request timer maintained by the source/initiator expires or the source is unwilling to see the response for some reason only 146527.doc • 45· 201102805. A positive response (successful return code) from the target of this request means that the target will not send a response stream for the specified request ID' and the specified request ID can be reused by the originator of the request. A negative response from the target (unsuccessful return of the code) means that the originator can still obtain a response stream for the specified request ID and can no longer use the request ID until the response stream is received. Request Parameters·Name ---Content Description _ Command Category -------------- Instructions for the command category. _________ Command code Commander's instructions for today's mother 0 ________r:---r- Requires response Request ID Known-n is tied to Lai." Super-rule ts _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ Off: ________ The request ID to be cancelled is associated with the previous request for the previous request for which cancellation is pending. Response parameters: Name Description of content Command category Indicates the command category. Command code Command code indication. Return Code See below for "Return Code" Request ID The request 来自) value from the associated request. The request ID to be canceled is the request ID associated with the previous request for which cancellation is being requested. Return code: 146527.doc -46 - 1

Success(成功) • Failure(失敗) • InvalidParms(無效參數) • PresentStateProhibits(目前狀態禁止) • RequestCannotBeCancelled(無法取消請求) • Busy(忙碌) 201102805 功能x請求 此部分說明對一虛構「功能X」請求之命令定義,該虛 構「功能X」請求為起源於HMC處之超管理器管道流之典 型請求且要求使用「拉」方法對回應資料進行封包化°該 請求為〜上游流。該回應為一下游流。 晴求參: 命令類別 命令碼 需要回應 請求 ID' 額外資 額外資#Γ _内容描述__ 命令類別之指示。_____ 命令碼之指示。___________ 一指示此流之源是否預期一來自目標之回應流的位元。_ 將一「請求j超管理器管道流繫結至對應「回應」超管理器管 道流之符記。請求之目標簡單地在回應流中將該值傳遞回。請 求之源將來自回應流之值與對應請求相關聯。__ 與此請求相關聯之任何額外資料的大小。 ~ 與此請求相關聯之額外資料。 "~~ 回應參敫: 名稱 命令類別 夺令碼 傳回碼 内容描述 命令類別之指示 命令碼之指示Success • Failure • InvalidParms • PresentStateProhibits • RequestCannotBeCancelled • Busy 201102805 Feature x Request this section to request a fictitious Feature X request The command defines that the fictitious "function X" request is a typical request originating from the hypervisor pipeline flow at the HMC and requires a "pull" method to encapsulate the response data. The request is an upstream stream. The response is a downstream stream. Clearly asked: Command Category Command Code Requires Response Request ID' Additional Resources Additional Resources #Γ _Content Description__ Instructions for the command category. _____ Command code indication. ___________ A bit indicating whether the source of this stream expects a response stream from the target. _ A "request j super manager pipeline flow" to the corresponding "response" hypervisor pipe flow. The target of the request simply passes the value back in the response stream. The source of the request associates the value from the response stream with the corresponding request. __ The size of any additional material associated with this request. ~ Additional information associated with this request. "~~ Response parameter: Name Command category Command code Return code Content description Command category indication Command code indication

請求ID 回應資料索引鍵 回應資料序號 總回應資料大小 剩餘回應資料大小 此命令中之回應資 料的大小 回應資料 傳回碼: 參見下文之「傳回碼 來自相關聯之請求的請求ID值 一符記,此回應之目標在對剩餘回應資料之請求中將該 遞回,且此回應之源接著使用該符記來識別正請求哪 料及在總回應資料中之何處繼續傳回資料。該值對於回夕^ 標而言巧簡單符記,但其可將一位元組位移編碼於用於回 源的一緩衝器中或將一索引編碼於用於回鹿之通沾一 “ ^:之 组 il«k T57 腌咨 Sa B a rir OJ& 二: --------- 葬之回應資料的總 g)此回應相藝之仍待傳送之^^料的大小 此回應流中之回應資料的大小 ^含於此回>^中之實際回應資i 為單 146527.doc •47· 201102805 • Success(成功) • Failure(失敗) • InvalidParms(無效參數) • PresentStateProhibits(目前狀態禁止) • ExchangeCapabilities(交換能力) • Busy(忙碌) 取得剩餘回應資料請求 此為一請求,其經由超管理器管道自—端點流至另一端 點以起始與早先請求相關聯之額外回應資料之傳送。要求 使用拉方法對回應資料進行封包化之每—類別之命令應定 義此命令之命令碼,以促進將該請求投送至目標内之恰當 組件。 亦可藉由經由超管理器管道發送多個回應流(推方法)來 達成對回應資料之封包化。亦即,請求之目標發送多個回 應直至不再存在回應資料為止。請求之源(回應之目標)知 曉何時接收到最後一個回應,因為「剩餘回應資料大小」 欄位為0。 請求參數· 名稱 内容描述 - 命令類別 命令類別之指示。 ------ 命令碼 命令碼之指示。 ---- 需要回應 指不此流之源是否預期一來自目標之回應流的彳立开。 請求ID 將 「請求」超管理器管道流繫結至對應「回應]管 道流之符記。請求之目標簡單地在回應流中將值傳遞回。請求 之源將來自回應流之值與對應請求相關聯。 ’/ 回應資料索引鍵 來自與正擷取之資料有關之最新近回應流幸引崎。 回應參數: 146527.doc -48- 201102805 名稱 内容描述 ---—-- 命令類別 命令類別之指示。 --—- 命令碼 命令碼之指示。 — 傳回碼 ¥見下文之 1回碼」 ------ 清求ID 來自相關聯之請求的請求ID值。 - 回應資料索引鍵 符^己’此回應之目標在對剩餘回應資料之請 ^回’且此回應之源接著使用該符記來識別正請 料及在總回應資料中之何處繼續傳回資料。該 3 ㊁ 標而言為簡單符記,但其可將_位元組位目 源的-緩衝器中或將一索引編祕丐d·於用f「回f之 回應資料序號 不按序進入之封包。此值基於1(自此口應之陶貞測 ^回^^黃料大小 ^棟取之資料相關聯之回應 剩餘回應貢料大小 此命令中之回應資 料的大小 此回應流中之回應資料的大---_______ ^------—— 貫際•田痛开中1 --- ------ ^ 151 Zff · 傳回碼: • Success(成功) • Failure(失敗)Request ID Response Data Index Key Response Data Serial Number Total Response Data Size Responsive Data Size The size of the response data in this command Response Data Return Code: See below "Returning Code from Request ID Value of Associated Request" The target of this response will be returned in the request for the remaining response data, and the source of this response will then use the token to identify what is being requested and where to continue to return the data in the total response data. In the case of the eve, it is a simple token, but it can encode a tuple displacement in a buffer for returning to the source or encode an index into the group for the return of the deer. Il«k T57 Picking Sa B a rir OJ& 2: --------- The total g of the response data of the funeral) This response is the size of the material that is still to be transmitted. The size of the response data^ is included in this back>^ The actual response is 146527.doc •47· 201102805 • Success • Failure • InvalidParms • PresentStateProhibits (current state is prohibited) • ExchangeCapabiliti Es (exchange capability) • Busy (Busy) Retrieve Responsive Data Request This is a request that flows from the endpoint to the other endpoint via the hypervisor pipeline to initiate the transfer of additional response data associated with the earlier request. A per-category command that requires the use of a pull method to encapsulate response data shall define the command code for this command to facilitate the delivery of the request to the appropriate component within the target. Encapsulation of response data can also be achieved by sending multiple response flows (push methods) via the hypervisor pipeline. That is, the target of the request sends multiple responses until there is no more response data. The source of the request (the target of the response) knows when the last response was received because the "Remaining Response Data Size" field is 0. Request Parameters · Name Description - Command Category Indicates the command category. ------ Command code Command code indication. ---- Need to respond Refers to whether the source of this flow is expected to open a response stream from the target. The request ID binds the "request" hypervisor pipeline to the corresponding "response" pipeline flow. The target of the request simply passes the value back in the response stream. The source of the request will come from the value of the response stream and the corresponding request. Relevant. '/ The response data index key comes from the latest response flow related to the data being retrieved. Response parameters: 146527.doc -48- 201102805 Name Description------ Command Category Command Category Indication. ---- Command code command code indication. — Return code ¥ see 1 code below. ------ Clear ID The request ID value from the associated request. - Respond to the data index key ^self' The goal of this response is to return the remaining response data and the source of this response is then used to identify the request and where to continue the data in the total response data. . The 3 2 standard is a simple token, but it can be used in the buffer of the _ byte group source or the index of the index can be used to enter the response data sequence number of f back to f. This value is based on 1 (from the mouth of the pottery test ^ back ^ ^ yellow material size ^ the data taken in response to the remaining response tribute size the size of the response data in this order in this response stream Responding to the big data ---- _______ ^------ - Continuation • Tian pain in the middle 1 --- ------ ^ 151 Zff · Return code: • Success (Success) • Failure ( failure)

InvalidParms(無效參數) PresentStateProhibits(目前狀態禁止)InvalidParms (invalid parameter) PresentStateProhibits (current state is prohibited)

ExchangeCapabilities(交換能力)ExchangeCapabilities

IrwalidResponseDataKey(無效回應資料索引鍵) Busy(忙綠) 傳回碼IrwalidResponseDataKey (invalid response data index key) Busy (busy green) return code

Success(成功)Success

成功地完成請求。 恢復:N/A -49- 146527.doc 201102805The request was successfully completed. Recovery: N/A -49- 146527.doc 201102805

Failure(失敗) 請求失敗’且無其他資訊。 恢復:重試該請求。^問題繼續存在,則開始標準問題 分析程序。Failure request failed 'and no other information. Recovery: Retry the request. ^ The problem continues to exist and the standard problem analysis process begins.

InvalidParms(無效參數) 目標偵測到命令中之無效資料/參數(諸如,未辨識之命 令類別或命令碼)。 恢復:無。可能為一錯誤。InvalidParms The target detects invalid data/parameters in the command (such as unrecognized command categories or command codes). Recovery: None. May be an error.

PresentStateProhibits(目前狀態禁止) 目標之目前狀態禁止處理此命令。 恢復:使目標進入一使其可處理該命令之狀態中。PresentStateProhibits (current state is disabled) The current state of the target is prohibited from processing this command. Recovery: Puts the target into a state in which it can process the command.

ExchangeCapabilities(^t ^ ^ Λ ) 未執行請求,因為目標之能力自上次交換能力之後可能 已被改變。 恢復:源應發布ExchangeCapabilities(交換能力)命令且 接著重試請求。ExchangeCapabilities(^t ^ ^ Λ ) The request was not executed because the ability of the target may have changed since the last exchange capability. Recovery: The source should issue the ExchangeCapabilities command and then retry the request.

InvalidResponseDataKey(無效回應資料索引鍵) 擷取額外回應資料之請求未由目標處理,因為命令中所 提供之回應資料索引鍵值並非有效的。 恢復:無。InvalidResponseDataKey The request to retrieve additional response data is not processed by the target because the response data index key value provided in the command is not valid. Recovery: None.

RequestCannotBeCancelled(無法取消請求) 無法取消指定請求。該指定請求可能處於傳送過程中 flight),或其可能已進行至一不再可能加以取消之境地。 回應流將來臨。 146527.doc -50· 201102805 恢復:無。RequestCannotBeCancelled The request could not be canceled. The specified request may be in the process of flight, or it may have proceeded to a situation where it is no longer possible to cancel. The response will come. 146527.doc -50· 201102805 Recovery: None.

Busy(忙綠) 未執行該請求,因為請求目標忙綠。預期此狀況為暫時 的。 恢復:在短暫延遲之後,重試程序。若該狀況繼續存在 一延長時間段(亦即,若干分鐘),則目標中有可能存在錯 誤,且應聯繫IBM技術維護部門。 設計細節· 命令流控制及調整進度 分割區-HYP LP事件會期 HYP必須為發送至分割區之每一 HypervisorPipeRequestlnbound (超管理器管道請求入埠)提供一 ACK訊息。HYP將為實際 存在之每一分割區(而非最大架構分割區)維護單獨之預分 配訊息集區。當建立或删除分割區時,將亦建立及刪除其 訊息集區及訊息。每一分割區具有一單獨集區連同立即以 忙碌狀態對至具有空訊息集區之目標的上游流進行ACK防 止了一不及時ACK事件之分割區影響至其他分割區之輸送 量。以忙碌狀態對自HMC至ACK訊息集區為空之分割區的 上游請求進行ACK會將處置目標分割區中之暫時忙碌時間 段及永久無反應的負擔置於HMC上。HMC應將一 Busy(忙 碌)傳回碼視作暫時狀況,且週期性地重試上游請求,直 至其起作用抑或一合理時間量已過去而仍然忙碌(在此狀 況下,HMC可假定目標無反應)。 對來自分割區之ACK的接收將不由HYP計時。並未正 146527.doc -51 - 201102805 ACK請求之分割區有可能亦非正在接收新請求,因此藉由 逾時而釋放ACK訊息以使得可發送額外請求有可能是無用 的0 HYP將用信號將超管理器管道請求事件非同步地發送至 分割區。此意謂HYP任務在接收到ACK之前未被封鎖,且 對於分割區而言,多個事件可同時為未處理的。由於可為 未處理之事件的數目受集區中之ACK訊息的數目限制,因 此HYP將與由該分割區建議之等於或小於每一分割區之預 分配訊息數目的任何最大入埠值一致。若分割區建議更大 值,則HYP將把該數目減小為開啟會期回應中的所組態數 目。HYP將對至分割區的上游請求發送進行調整進度,以 使得永不超過所協商之最大入埠值。HYP將始終在開啟會 期回應中將最大出埠值設定為1,因為在處理來自分割區 之下游超管理器管道請求出埠事件流之過程中,HYP將為 單線緒的,由此使得大於1之值無意義。 分割區應儘快ACK入埠流,以使得ACK訊息將被快速傳 回至HYP中之適當集區。此將最大化管道輸送量。分割區 ACK入埠流所耗費時間愈長,ACK訊息集區將耗盡的機會 愈大,在耗盡時,HYP將開始用忙碌狀態向源ACK以昏睡 (lethargic)分割區為目標之請求。 重新啟動一無反應且不ACK上游請求之分割區將迫使所 有未處理之ACK訊息回到HYP。 HMC-HYP命令會期 對HMC與HYP之間的會期之HYP處置類似於如先前部分 146527.doc -52- 201102805 中所描述之針對分割區與HYP之間之會期的處置。 HYP必須為其發送至HMC之每一超管理器管道請求命令 提供一 ACK訊息。HYP將為所連接之每一 HMC維護一訊息 集區。每一 HMC具有一單獨集區連同立即用忙碌狀態對至 具有空訊息集區之目標HMC的下游流進行ACK防止了一未 及時ACK事件之HMC影響至其他HMC之輸送量。用忙碌狀 態ACK自分割區至ACK訊息集區為空之HMC的下游請求會 將處置目標HMC中之暫時忙碌時間段及永久無反應的負擔 置於分割區上。分割區應將一 Busy(忙碌)傳回碼視為暫時 狀況,且週期性地重試下游請求,直至其起作用抑或一合 理時間量已過去而仍然忙碌(在此狀況下,分割區可假定 目標無反應)。 HYP將對至HMC的下游請求發送進行調整進度以使得永 不超過給定HMC之所協商之最大入埠值,該最大入琿值將 經協商為最大值(例如)8(若HMC初始建議一小於8之值,則 HYP將履行一較低值)。HYP將始終在開啟會期回應中將最 大出埠值設定為1,因為在處理來自HMC之上游超管理器 管道請求命令流之過程中,HYP將為單線緒的,由此使得 大於1之值無意義。 對來自HMC之ACK的接收將不由HYP計時。並未正ACK 請求之HMC有可能亦非正在接收新請求,因此藉由逾時而 釋放ACK訊息以使得可發送額外請求有可能是無用的。Busy (busy green) The request was not executed because the request target is busy green. This situation is expected to be temporary. Recovery: After a short delay, retry the program. If the condition persists for an extended period of time (i.e., several minutes), there may be an error in the target and contact IBM Technical Maintenance. Design Details · Command Flow Control and Adjustment Progress Partition - HYP LP Event Session HYP must provide an ACK message for each HypervisorPipeRequestlnbound sent to the partition. HYP maintains a separate pre-allocated message pool for each partition that is actually present (rather than the largest architectural partition). When a partition is created or deleted, its message set and message will also be created and deleted. Each partition has a separate pool along with an immediate upstream ACK to the upstream stream with a target of the empty message set to prevent the partition of the un-timed ACK event from affecting the throughput of the other partitions. An acknowledgment of an upstream request for a partition from which the HMC to ACK message pool is empty in a busy state places a temporary busy period in the target partition and a permanent unresponsive burden on the HMC. The HMC shall treat a Busy return code as a temporary condition and periodically retry the upstream request until it is active or a reasonable amount of time has passed and is still busy (in this case, the HMC can assume that the target is not reaction). Reception of ACKs from partitions will not be clocked by HYP. The 146527.doc -51 - 201102805 ACK request partition may or may not be receiving a new request, so the ACK message is released by timeout so that the extra request can be sent. 0 HYP will signal The hypervisor pipe request event is sent asynchronously to the partition. This means that the HYP task is not blocked before receiving the ACK, and for partitions, multiple events can be unprocessed at the same time. Since the number of unprocessable events is limited by the number of ACK messages in the pool, the HYP will be consistent with any maximum entry value suggested by the partition equal to or less than the number of pre-allocated messages for each partition. If the partition suggests a larger value, HYP will reduce the number to the configured number in the open session response. The HYP will adjust the progress of the upstream request transmission to the partition so that the negotiated maximum entry value is never exceeded. HYP will always set the maximum exit value to 1 in the open session response, because HYP will be single-threaded during processing of the upstream hypervisor pipe request flow from the partition, thereby making it larger than The value of 1 is meaningless. The partition should be ACKed as soon as possible so that the ACK message will be quickly passed back to the appropriate pool in the HYP. This will maximize the amount of pipe delivery. The longer it takes for the ACK to enter the turbulence, the greater the chance that the ACK message set will be exhausted. When exhausted, HYP will begin to request the source ACK with a lethargic partition in a busy state. Restarting a partition that is unresponsive and does not ACK upstream requests will force all unprocessed ACK messages back to HYP. HMC-HYP Command Session The HYP treatment for the session between HMC and HYP is similar to the treatment for the session between the partition and HYP as described in the previous section 146527.doc -52- 201102805. HYP must provide an ACK message for each Hypervisor Pipeline Request command sent to the HMC. HYP will maintain a message pool for each HMC connected. Each HMC has a separate pool and ACKs to the downstream stream of the target HMC with the empty message set immediately with the busy state to prevent the HMC of an unacknowledged ACK event from affecting the throughput of the other HMCs. The downstream request of the HMC with the busy status ACK from the partition to the ACK message set area will place the temporary busy time period and the permanent non-responsive burden in the disposal target HMC on the partition. The partition should treat a Busy return code as a temporary condition and periodically retry the downstream request until it is active or a reasonable amount of time has passed and is still busy (in this case, the partition can be assumed The target is not responding). The HYP will adjust the progress of the downstream request transmission to the HMC so as to never exceed the negotiated maximum entry value of the given HMC, which will be negotiated as the maximum value (for example) 8 (if the HMC initial recommendation one) A value less than 8, then HYP will perform a lower value). HYP will always set the maximum outbound value to 1 in the open session response, because HYP will be single-threaded during processing of the upstream hypervisor pipe request command stream from the HMC, thereby making the value greater than 1. Meaningless. Reception of ACKs from the HMC will not be timed by HYP. An HMC that is not a positive ACK request may or may not be receiving a new request, so releasing the ACK message by timeout may make it possible to send an additional request.

HMC應儘快ACK入埠流,以使得ACK訊息將被快速傳回 至HYP中之適當集區。此將最大化管道輸送量。HMC 146527.doc 53· 201102805 ACK入蟑流所耗費時間愈長,ack訊息集區將耗盡的機會 愈大在耗盡時,HYP將開始用忙碌狀態向源ACK以昏睡 HMC為目標之請求。 重新啟動或斷開一無反應且不ACK下游請求之HMC將迫 使所有未處理之ACK訊息回到HYP。 HMC-分割區會期 上文所描述之會期包含端點(HMC與分割區)之間的較高 層級會期,且以與基礎會期大致相同之方式來管理及控制 此等會期。能力交換允許兩個端點協商最大入埠/出埠 值,且判定彼此對在能力位元組中定義之特定能力的支 援。 為了防止超官理器管道會期之兩個端點在能力及/或最 大入埠/出埠方面不同步,並非ExchangeCapabilities(交換 能力)命令之任何超管理器管道流之目的地必須檢查自目 的地經歷一可能已改變其能力之事件(諸如,IPL,在其期 間應用程式碼更新,或同時韌體更新)以來源是否已與目 的地交換能力。若否,則目的地必須在回應流中傳回 ExchangeCapabilities(交換能力)傳回碼。 當请求之源接收到該ExchangeCapabilities(交換能力)傳 回碼時,其必須將一 ExchangeCapabUities(交換能力)命令 發至目的地’且接著可重新發送失敗之原始請求。 可根據兩個端點所需而協商最大入埠/出埠。應根據請 求之源的觀點解譯入埠/出埠。ExchangeCapabilities(交換 能力)s青求之源應將最大入槔設定為其同時可處理之請求 146527.doc •54- 201102805 的數目,且將最大出淳 調整進度…, 目標應將其出崞請求之 门金退度叹疋為小於或等於 值,且將最大出蟑值,定疋之最大入槔值的-诉摩接以… 疋為其同時可處理之請求的數目。 標指定之最大出璋值的:定為小於或等於由目 及出槔皆經協商為值4,此卜^實例’假定最大入璋 ,,+ 此意明HMC可預期分割區同時支 杈多達4個未處理之上 J卞夂 4. ^ ^ ^ °月求,且/刀割區可預期HMC同時 支援夕達4個未處理之 A , „ 荇印求。未處理意謂請求流之源 而未接收到來自源之對應回應流。 若-端點不具有對其可處置之未處理人埠之數目的可識 限制(如可此為该端點動態地分配内部訊息及控制區塊 t :刀配》亥等内部訊息及控制區塊之狀況),則該端點 Z父換月b力中的對應於其可支援之最大未處理人缚的值 定為私示無限制的特殊值(諸如,全部〇xF) ^另一端 點可接者將其所要數目之超#理器f道流發送至彼目標而 不調整進度’但應準備處置因目標可能暫時無法獲取必要 育源(諸如,訊息、控制區塊空間等)之狀況而來自目標之 偶然的Busy(忙碌)傳回碼。 應注意’遵守HMC與分割區之間所協商之最大入埠/出 埠值不會保證在此會期上的請求流之源可避免Busy(忙碌) 傳回碼。若目標之ACK訊息集區為空,則Busy(忙碌)傳回 碼可由HYP在HypervisorPipeReqUes《超管理器管道請求) 請求流ACK中產生。 檢測 146527.doc •55· 201102805 Η Υ P應提供一定量之檢測來追蹤諸如所產生之忙碌A c κ 及造成其之目標端點的數目、發送失敗之次數、流處於管 道(針對上游及下游兩者)中之平均時間量、上游流及下游 流之總數目等的事項。The HMC shall ACK into the turbulence as soon as possible so that the ACK message will be quickly transmitted back to the appropriate pool in the HYP. This will maximize the amount of pipe delivery. HMC 146527.doc 53· 201102805 The longer it takes for the ACK to enter the turbulence, the greater the chance that the ack message pool will be exhausted. When the acknowledgment is exhausted, HYP will start to use the busy state to request the source ACK to slumber the HMC. Restarting or disconnecting an HMC that is unresponsive and does not ACK downstream requests will force all unprocessed ACK messages back to HYP. HMC-Segmentation Period The duration described above includes a higher level period between endpoints (HMCs and partitions) and manages and controls these periods in much the same way as the base period. Capability exchange allows two endpoints to negotiate maximum entry/exit values and determine support for each other's specific capabilities defined in the capability byte. In order to prevent the two endpoints of the hypervisor pipeline from being out of sync in terms of capabilities and/or maximum entry/exit, the destination of any hypervisor pipeline flow that is not an ExchangeCapabilities command must be checked from the destination. The event experiences an event that may have changed its capabilities (such as IPL, during which application code updates, or firmware updates) to see if the source has exchanged capabilities with the destination. If not, the destination must return an ExchangeCapabilities return code in the response stream. When the source of the request receives the ExchangeCapabilities return code, it must send an Exchange CapabUities command to the destination and then resend the original request that failed. The maximum entry/exit can be negotiated based on the needs of the two endpoints. Interpretation/exit should be based on the source of the request. The source of ExchangeCapabilities should be set to the maximum number of requests that can be processed at the same time 146527.doc •54- 201102805, and the maximum will be adjusted..., the target should be sent out of the request The threshold of the gold retreat is less than or equal to the value, and the maximum value of the depreciation is determined. The maximum value of the depreciation is determined by the number of requests that can be processed at the same time. The maximum specified exit value of the target is set to be less than or equal to the value of 4 for both the target and the exit. This example is assumed to be the maximum entry, and + this means that the HMC can expect the partition to be supported at the same time. Up to 4 unprocessed above J卞夂4. ^ ^ ^ °, and / knife cutting zone can expect HMC to support 4 unprocessed A at the same time, „ 荇 求. Unprocessed means request flow The source does not receive the corresponding response stream from the source. If the endpoint does not have an identifiable limit on the number of unhandled persons it can handle (such as dynamically assigning internal messages and control blocks to the endpoint) t: the condition of the internal message and the control block of the knife, and the value of the maximum unprocessed person in the force of the end of the Z-month change is specified as a private unrestricted special Value (such as all 〇xF) ^Another endpoint can send its required number of super-processed f-channel streams to his target without adjusting the progress' but should be prepared for disposal because the target may not be able to obtain the necessary source of education temporarily ( The status of the message, control block space, etc., comes from the occasional Busy (passive) return code of the target. It should be noted that 'obeying the maximum entry/exit value negotiated between the HMC and the partition does not guarantee that the source of the request stream at this session avoids the Busy (busy) return code. If the target ACK message area If it is empty, the Busy (busy) return code can be generated by HYP in the HypervisorPipeReqUes request message stream ACK. Detection 146527.doc •55· 201102805 Η Υ P should provide a certain amount of detection to track such as generated The busy A c κ and the number of target endpoints, the number of failed transmissions, the average amount of time in the pipeline (for both upstream and downstream), the total number of upstream and downstream flows, and so on.

Busy傳回碼處置 在端點中對來自HYP之忙碌傳回碼的建議處置為在一短 暫延遲之後重試出埠流,且重複直至一合理時間量過去或 出琿流最後成功。合理量可基於正發送之命令或請求而不 同0 交換能力處置 對來自HYP或其他端點之ExchangeCapabiUties(交換能 力)傳回碼的建議處置為發布ExchangeCapabUities(交換能 力)。月求且接著重试被拒絕之請求。或者,端點可選擇發 布ExchangeCapabilities(交換能力)請求且中止而非重試被 拒絕之請求。使用者接著將必須手動地重試失敗之請求。 在難以自動重試失敗之請求的情況下,此替代方案可具有 吸引力。若實施此替代方案,則應知道儘管能力改變很少 有,但分割ϋ斷電/通電可相對頻繁地發生且將引起分割 區傳回EXChangeCapabilities(交換能力)傳回碼(直至再次與 請求之源/起始方交換能力為止)。㈣,可採取措施來降 低ExchangeCapabilities(交換能力)傳回碼發生之可能性。 端點中之一實施(選項D為每當一可能已改變能力之事 件(諸如’同時程式碼更新(或移除)或端點通電)發生時, 將ExchangeCapabUities(交換能力)請求發布至所有作用中 146527.doc •56· 201102805 端點。藉由此實施,端點快取所有其他可能端點之 且藉由在可能改變其他端點之能力之某事發生b牛地 通知該等其他端點來保持所快取之能力為當前的。此二法 不會完全防止ExChangeCapabilities(交換能力)傳㈣因 為在交換能力_,流可能處於傳送過程中,但其確實將發 生ExchangeCapabiHties(交換能力)傳回碼的機會減少到足 夠低的程度’使得中止被拒絕之請求(而非自動重試被拒 絕之請求)將為一可接受之選項。 端點中之降低ExChangeCapabilities(交換能力)傳回碼發 生之可能性的另一實施(選項2)為在每一請求之前將 EXChangeCapabilities(交換能力)請求發布至端點。藉由此 方法’不必快取能力。僅在需要時即時查詢能力。如同選 項1,此方法不完全消除得到ExchangeCapabimies(交換能 力)傳回碼之可能性,但其將機會減少到足夠低的程度以 至於中止被拒絕之請求(而非自動重試)將為可接受的。應 注意,此方法可能顯著增加端點間之流之數目。在最簡單 的形式中,其將使流之數目加倍4存在涉及至—特定端 點之多個流來完成任務的特定使用者任務,則起始端點可 每一使用者任務一次(而非每一請求流一次)地將 ExchangeCapabilities(交換能力)發布至目標端點,由此減 小ExchangeCapabilities(交換能力)流之總數目。 對於上文所描述之兩個選項,每一端點必須追蹤已與之 交換能力的其他端點,且針對其為目標的並非交換能力之 每一流而驗證其已與彼流之源交換其當前能力。若一改變 146527.doc -57- 201102805 或可能已改變其能力之事件發生,則其必須再次與其他端 點交換其新能力或潛在新能力。 每當一可能改變HYP之能力的事件(亦即,同時韌體更 新)發生時,HYP將在所有作用中的通用超管理器管道 HMC會期及LP事件會期上起始一 ExchangeCapabilities(交 換能力)請求。類似地,每當一可能同時改變端點之能力 的事件(亦即,同時程式碼更新)發生時,端點應起始對 HYP之ExchangeCapabilities(交換能力)請求。因此,端點 必須作為源(亦即,當端點經受同時程式碼更新時)與目標 (亦即,當HYP經受同時韌體更新時)兩者支援與HYP之 ExchangeCapabilities(交換能力)流。經由開啟會期命令在 HMC或分割區通電時交換能力。 關於共用記憶體分割區資料處理系統之其他細節提供於 以下共同申請之專利申請案中,該等專利申請案中之每一 者的全部内容特此以引用的方式併入本文中:美國第 _號(代理人案號 ROC 920080415US1), 「Hypervisor Page Fault Processing in a Shared MemoryBusy Return Code Handling The recommendation to busy return code from HYP in the endpoint is to retry the turbulence after a short delay and repeat until a reasonable amount of time has elapsed or the turbulence last succeeded. A reasonable amount can be based on the command or request being sent. 0 Exchange Capability Disposition The recommendation to return the ExchangeCapabiUties (replacement capability) code from HYP or other endpoints is to issue ExchangeCapabUities (Exchange Capability). Request for the month and then retry the rejected request. Alternatively, the endpoint may choose to issue an Exchange Capabilities request and abort rather than retry the rejected request. The user will then have to manually retry the failed request. This alternative can be attractive in situations where it is difficult to automatically retry a failed request. If this alternative is implemented, it should be known that although the capability change is rare, the split/power down/power-on can occur relatively frequently and will cause the partition to return the EXChangeCapabilities (return capability) return code (until again with the source of the request) /Starting party exchange ability). (d), measures can be taken to reduce the possibility of ExchangeCapabilities (return capability) return codes. One of the endpoint implementations (Option D is to publish ExchangeCapabUities requests to all roles whenever an event that may have changed capabilities (such as 'simultaneous code update (or removal) or endpoint power up) occurs 146527.doc •56· 201102805 endpoint. By doing so, the endpoint caches all other possible endpoints and notifies the other endpoints by something that may change the capabilities of the other endpoints. To maintain the ability of the cache to be current. This method does not completely prevent ExChangeCapabilities (transfer capability) transmission (four) because in the exchange capacity _, the stream may be in the process of transmission, but it will indeed return ExchangeCapabiHties (exchange capability) The chance of the code is reduced to a low enough level to make the request to abort the rejection (rather than the automatic retry of the rejected request) will be an acceptable option. The reduced ExChangeCapabilities in the endpoint will occur. Another implementation of the possibility (option 2) is to issue an EXChangeCapabilities request to the endpoint before each request. 'Do not have the ability to cache. Just query the ability when needed. As with option 1, this method does not completely eliminate the possibility of getting the ExchangeCapabimies (switching ability) to return the code, but it reduces the chance to a low enough level to stop being Requests for rejection (rather than automatic retry) will be acceptable. It should be noted that this method may significantly increase the number of flows between endpoints. In the simplest form, it will double the number of streams to 4 - a specific user task for a particular endpoint to complete a task, the originating endpoint may publish ExchangeCapabilities to the target endpoint once per user task (rather than once per request stream). This reduces the total number of ExchangeCapabilities flows. For the two options described above, each endpoint must track the other endpoints with which it has exchanged capabilities, and for which it is not the ability to exchange First-class and verify that it has exchanged its current capabilities with the source of the stream. If a change is made 146527.doc -57- 201102805 or an event that may have changed its capabilities If it is born, it must exchange its new capabilities or potential new capabilities with other endpoints. Whenever an event that may change the capabilities of HYP (ie, simultaneous firmware update) occurs, HYP will be generic in all roles. The Manager Pipeline HMC Session and the LP Event Session start an ExchangeCapabilities request. Similarly, whenever an event that may change the capabilities of the endpoint (ie, simultaneous code update) occurs, the endpoint The point should initiate an Exchange Capabilities request for HYP. Therefore, the endpoint must support the Exchange Capabilities flow with HYP as both the source (i.e., when the endpoint is subject to simultaneous code update) and the target (i.e., when the HYP is undergoing simultaneous firmware updates). The ability to exchange when the HMC or partition is powered on via the open session command. Further details regarding the shared memory partition data processing system are provided in the following co-pending patent application, the entire contents of each of which are hereby incorporated by reference herein in (Agent Case No. ROC 920080415US1), "Hypervisor Page Fault Processing in a Shared Memory

Partition Data Processing System」;美國第_ 號(代理人案號 ROC 920080416US1),「Managing Assignment of Partition Services to Virtual Input/OutputPartition Data Processing System; US No. _ (agent case number ROC 920080416US1), "Managing Assignment of Partition Services to Virtual Input/Output

Adapters」;美國第—_號(代理人案號ROC 920080417US1),「Automated Paging Device Management in a Shared Memory Partition Data Processing System」;美 國第_號(代理人案號ROC 920080418US1), 146527.doc -58- 201102805 「.Dynamic Control of Partition Memory Affinity in a Shared Memory Partition Data Processing System」;美國第 _號(代理人案號 ROC 920080419USI), 「Transparent Hypervisor Pinning of Critical Memory Areas in a Shared Memory Partition Data Processing System」;美 國第__號(代理人案號ROC 920080420US1), 「Shared Memory Partition Data Processing System withAdapters; US No. - (No. ROC 920080417US1), "Automated Paging Device Management in a Shared Memory Partition Data Processing System"; US No. (Agent Case No. ROC 920080418US1), 146527.doc -58 - 201102805 ".Dynamic Control of Partition Memory Affinity in a Shared Memory Partition Data Processing System"; US No. (Agent No. ROC 920080419USI), "Transparent Hypervisor Pinning of Critical Memory Areas in a Shared Memory Partition Data Processing System" ; US __ (agent case number ROC 920080420US1), "Shared Memory Partition Data Processing System with

Hypervisor Managed Paging」;美國第_號(代 理人案號ROC 920080421US1),「Controlled Shut-Down of Partitions Within a Shared Memory Partition DataHypervisor Managed Paging"; US No. _ (Agent Case No. ROC 920080421US1), "Controlled Shut-Down of Partitions Within a Shared Memory Partition Data

Processing System」;及美國第_號(代理人案 號 ROC 920080422US1),「Managing Migration of a SharedProcessing System"; and US No. _ (agent case number ROC 920080422US1), "Managing Migration of a Shared

Memory Logical Partition From a Source System to a Target System」° 本發明之一或多個態樣可包括於一具有(例如)電腦可用 媒體之製品(例如,一或多個電腦程式產品)中。媒體在其 中具有(例如)電腦可讀程式碼構件或邏輯(例如,指令、程 式碼、命令等)以提供並促進本發明之能力◊製品可被包 括作為電腦系統之一部分或單獨銷售。 參看圖14描述併有本發明之一或多個態樣之製品或電腦 程式產品之一實例。電腦程式產品1400包括(例如)一或多 個電腦可讀媒體1410,該一或多個電腦可讀媒體1410用以 將電腦可讀程式碼構件或邏輯1420儲存於其上以提供並促 進本發明之一或多個態樣。媒體可為電子、磁性、光學、 146527.doc -59- 201102805 電磁、紅外線或半導體系統(或裝置或器件)或者傳播媒 體。電腦可讀媒體之實例包括半導體或固態記憶體、磁 帶、抽取式電腦磁片、隨機存取記憶體(ram)、唯讀記憶 體(職)、硬磁碟及光碟。光碟之實例包括緊密光碟唯讀 記憶體(CD-ROM)、緊密光碟讀/寫(CD R/W)& dvd。 由-或多個電腦可讀程式㈣件或邏輯定義之程式指令 序列或-或多個相關模組之邏輯套組指引本發明之 個態樣的執行。 儘官上文描述各種實施例,但此等實施例僅為實例。 此環境可包括模擬器(例>,軟體或其他模擬相 =其\模擬-特定架構或其子集。在此環境中,模掏 使^個模擬功能可實施本發明之-或多個態樣,即 執仃該模擬器之電腦可 架構亦如此。作為…,;與㈣之能力不同的 ^ ^ 4T 模擬模式中,對正模擬之特 疋心7或操作進行解碼, 別指令或操#。 建置-適當模擬功能以實施個 在-模擬環境中,主機電腦 用以健存指令及資料;_指令提^體’其 提取指令 ^取早兀,其用以自記憶體 …Μ 以提供對經提取指令之本端緩衝. 心令解碼單元,其 心衝’ 提取之指令的類型;及-指令::單之指令:用以判定已 令。執行可包括:將資料自心其广執行指 料"存器储存回至記憶雜;載=存…將資 *-類型的算術或邏輯運算。在解'單元判定之 實例中,以軟體來實施 146527.doc 201102805 每一單元。舉例而言,將由該等單元執行之操作實 擬裔軟體内之一或多個副常式。 、 另卜可使用適用於健存及/或執行程式碼之資料處理 系統,其包括直接或經由系統匯流排間接輕接至記憶體元 件的至少—處理器。記憶體元件包括(例如)在程式碼之實 際執行期間使用之本端記憶體、大容量儲存器,及快取^ 憶體,快取記憶體提供對至少某—程式碼之暫時儲存以便 減少在執行期間必須自大容量儲存器榻取程式碼的次數。 輸入/輪出或I/O器件(包括(但*限於)鍵盤 標器件、⑽D、磁帶、CD、DVD、隨身碟及其他記^ 媒體等)可直接或經由介入y 、 耦接至该系統。網路 -可耦接至系統以使得資料處理系統能夠經由介入 公用網路而輕接至其他資料處理系統或遠端印表 機或儲存益件。數據機、 用類型之網路配接器中的少數幾種。 卞僅為了 本發明之-或多個態 其某-組合來以〜力7以軋體,體、硬體或 存器件,其具體化可由 機^取之至少一程式儲 指令的至少-程式亥機讀订以執行本發明之能力之 本文中所描繪之流程圖僅 神的情況下,可存在例。在不脱離本發明之精 操作)之許多變化。舉例而言,可按不it圖或步驟(或 m , , 了按不同次序執行該等步 所主張之本:明、删除或修改步驟。將所有此等變化視為 所王拫之本發明的—部分。 146527.doc -61 - 201102805 儘管已在本文令詳細地描 ^田繪並锸述了實施例,但對於孰 各相關技術者將顯而层g ' “’可在不脫離本發明之精神的情 /兄下作出各種修改、、灭Ά J' 替代及其類似者,且因此,將 此等修改、添加、替抑乃1 #5 y、&、 、頰似者視為在如以下申請專利 範圍中所定義之本發明的範疇内。 【圖式簡單說明】 圖為根據本發明之態樣的經邏輯分割的資料處理系統 之各種硬體組件的高層級方塊圖; 圖2為根據本發明之態樣的在資料處理系統中的不同硬 體及軟體抽象層級處之邏輯分割區的概念說明; 圖3為根據本發明之態樣的具有以超管理器為基礎之通 信設施的經邏輯分割的資料處理系統之-實體實施例的方 塊圖說明; 圖4為根據本發明之態樣的圖3之資料處理系統及通信設 施二較低層級邏輯視圖,其中通信會期建立於HMC與超管 理盗管道之間’及邏輯分割區與超管理器管道之間; 為根據本發明之態樣的圖3及圖4之以超管理器為基 礎之通信設施的較高層級邏輯描繪; 圖6為根據本發明之態樣的用於處理自經邏輯分割的資 料處理系統之源端點接收到之通信的超管理器邏輯的一實 施例的流程圖; 圖7為根據本發明之態樣的用於在超管理器處處理泛用 傳达基元(Hypepipe)的邏輯的一實施例的流程圖; 圖8為根據本發明之態樣的用於處理自經邏輯分割的資 146527.doc •62- 201102805 料處理系統之一端點接收到之應答(ACK)的超管理器邏輯 的一實施例的流程圖; 圖9為根據本發明之態樣的用於建置一 exchange capabimies (交換能力)請求並將其發送至經邏輯分割的資料處理系統 之目標端點的超管理器邏輯的一實施例的流程圖; 圖10為根據本發明之態樣的在一可引起超管理器能力改 變之事件在經邏輯分割的資料處理系統内發生時執行的超 管理器邏輯的一實施例的流程圖; 圖11為根據本發明之態樣的用於在經邏輯分割的資料處 理系統之目標端點處處理目標傳送基元的端點邏輯的一實 施例的流程圖; 圖12為根據本發明之態樣的用於主動地自端點起始一 exchange capabilities(交換能力)請求的端點邏輯的一實施 例的流程圖; 圖13為根據本發明之態樣的用於回應於在通信設施之端 點處的能力改變事件而被動地執行能力交換的端點邏輯的 一實施例的流程圖;及 圖14描繪併有本發明之一或多個態樣之電腦程式產品的 * 一實施例。 【主要元件符號說明】 100 101 102 103 可經邏輯分割的電腦系統 中央處理單元(CPU) 主記憶體 服務處理器 146527.doc -63- 201102805 105 匯流排 106 終端機介面 107 儲存器介面 108 其他I/O器件介面 109 通信/網路介面 111Α、111Β、 111C ' 111D 電路卡 112A、112B、 112C、112D、 112E 、 112F 電路卡 113 電路卡 114 硬體管理控制台(HMC) 115A、115B、 115C 電路卡 116A、116B、 116C 電路卡 117A、117B 電路卡 118A、118B 電路卡 119A、119B 電路卡 121A、121B、 121C 使用者終端機 122A、122B、 122C 資料儲存器件 123 印表機 ·64· 146527.doc 201102805 124 傳真機 126 網路 201 硬體層級 202 不可分派超管理器 203 可分派超管理器 204 邏輯分割區 205 邏輯分割區 206 邏輯分割區 207 邏輯分割區 211 OS核心 212 OS核心 213 OS核心 214 OS核心 221 視覺指示器 222 分配至分割區之可分割實體 223 可分割實體位置 300 經邏輯分割的資料處理系統 310 邏輯分割區 320 超管理器 325 超管理器管道 330 靈活服務處理器 340 硬體管理控制台(HMC) 400 邏輯HMC命令會期/HMC對超管理器 會期 146527.doc •65- 201102805 405 訊息缓衝器集區 410 邏輯分割事件會期/邏輯分割對超管 理器會期 415 訊息緩衝器集區 500 邏輯HMC對分割區通信會期 1400 電腦程式產品 1410 電腦可讀媒體 1420 電腦可讀程式碼構件或邏輯 146527.doc -66-One or more aspects of the present invention can be included in an article (e.g., one or more computer program products) having, for example, computer usable media. The media has, among other things, computer readable code components or logic (e.g., instructions, program code, commands, etc.) to provide and facilitate the capabilities of the present invention. The article can be included as part of a computer system or sold separately. An example of an article or computer program product having one or more aspects of the present invention is described with reference to FIG. The computer program product 1400 includes, for example, one or more computer readable media 1410 for storing computer readable code components or logic 1420 thereon to provide and facilitate the present invention. One or more aspects. The media can be electronic, magnetic, optical, 146527.doc -59- 201102805 electromagnetic, infrared or semiconductor systems (or devices or devices) or propagating media. Examples of computer readable media include semiconductor or solid state memory, magnetic tape, removable computer magnetic disks, random access memory (ram), read only memory (job), hard disk and optical disk. Examples of optical disks include Compact Disc Read Only Memory (CD-ROM), Compact Disc Read/Write (CD R/W) & dvd. The execution of aspects of the present invention is directed to a sequence of program instructions or sequences of logic modules defined by - or a plurality of computer readable programs (s) or logic. Various embodiments are described above, but such embodiments are merely examples. This environment may include a simulator (example >, software or other analog phase = its \ analog - specific architecture or a subset thereof. In this environment, the simulation enables the simulation function to implement the invention - or multiple states The same is true, that is, the computer architecture of the emulator is also the same. As the ^^4T emulation mode with different capabilities than (4), the special emulation 7 or operation of the simulation is decoded, and the instruction or operation# Build-appropriate simulation function to implement an in-simulation environment, the host computer is used to store instructions and data; _ command to raise the body's fetch instruction ^ early, which is used to self-memory...Μ to provide The local buffer of the extracted instruction. The heart decoding unit, the type of the instruction that is extracted by the heart; and the instruction: the instruction of the single: used to determine the order. The execution may include: performing the data self-discipline Refer to the storage " store to return to the memory; load = save ... will be *-type arithmetic or logical operation. In the solution of the unit determination, software 146527.doc 201102805 each unit. For example The operations that will be performed by these units One or more secondary routines. Alternatively, a data processing system suitable for storing and/or executing code may be used, including at least a processor that is indirectly connected to the memory component either directly or via a system bus. The memory component includes, for example, a local memory, a large-capacity memory, and a cache memory used during actual execution of the code, and the cache memory provides temporary storage of at least some code to reduce The number of times the code must be retrieved from the mass storage device during execution. Input/rounding or I/O devices (including (but limited to) keyboard devices, (10)D, tape, CD, DVD, flash drive, and other media The system can be coupled to the system either directly or via intervention y. The network can be coupled to the system to enable the data processing system to be tapped to other data processing systems or remote printers or via the intervening public network. A few of the data adapters and types of network adapters. 卞 Only for the present invention - or a plurality of states - a combination of ~ force 7 for rolling, body, hardware or storage Device, its materialization In the case where at least one of the program storage instructions can be read by the machine to perform the present invention, the flowchart depicted in this document can be used only in the case of God. Without departing from the essence of the present invention Many changes in operation). For example, the steps claimed in the steps may be performed in a different order, such as a clear, delete, or modification step. All such changes are considered to be part of the invention of the king. 146527.doc -61 - 201102805 Although the embodiments have been described in detail and described in detail herein, it will be apparent to those skilled in the art that Emotions/brothers make various modifications, annihilation, J' substitutions and the like, and therefore, these modifications, additions, and substitutions are considered to be 1 #5 y, & The scope of the invention as defined in the scope of the patent. [Brief Description] FIG. 2 is a high level block diagram of various hardware components of a logically segmented data processing system in accordance with an aspect of the present invention; FIG. Conceptual illustration of logical partitions at different hardware and software abstraction levels in a data processing system; FIG. 3 is a logic diagram of a hypervisor-based communication facility in accordance with aspects of the present invention; Segmented data processing system - block diagram illustration of an entity embodiment; FIG. 4 is a lower level logical view of the data processing system and communication facility of FIG. 3 in accordance with an aspect of the present invention, wherein the communication session is established between the HMC and the super-managed pipe. And between the logical partition and the hypervisor pipeline; a higher level logical depiction of the hypervisor-based communication facility of FIGS. 3 and 4 in accordance with aspects of the present invention; FIG. 6 is a diagram in accordance with the present invention. A flowchart of an embodiment of hypervisor logic for processing communications received by a source endpoint of a logically segmented data processing system; FIG. 7 is a diagram of a hypervisor in accordance with aspects of the present invention. A flowchart of an embodiment of a logic for processing a general-purpose communication primitive (Hypepipe); FIG. 8 is a diagram of a material processing system for processing logical segmentation according to aspects of the present invention; 146527.doc • 62-201102805 A flowchart of an embodiment of a hypervisor logic for an acknowledgement (ACK) received by an endpoint; FIG. 9 is a diagram for constructing an exchange capabimies request and transmitting it to the aspect of the present invention Logical division A flowchart of an embodiment of hypervisor logic for a target endpoint of a cut data processing system; FIG. 10 is a logically partitioned data processing in response to an event that may cause a hypervisor capability change in accordance with an aspect of the present invention; A flowchart of an embodiment of hypervisor logic executing when occurring within a system; FIG. 11 is a diagram of processing a target transport primitive at a target endpoint of a logically partitioned data processing system in accordance with an aspect of the present invention; A flowchart of an embodiment of point logic; FIG. 12 is a flow diagram of an embodiment of endpoint logic for actively initiating an exchange capabilities request from an endpoint in accordance with an aspect of the present invention; 13 is a flow diagram of an embodiment of endpoint logic for passively performing capability exchange in response to a capability change event at an endpoint of a communication facility in accordance with aspects of the present invention; and FIG. 14 depicts and incorporates the present invention An embodiment of one or more aspects of a computer program product. [Main component symbol description] 100 101 102 103 Computer system central processing unit (CPU) which can be logically divided Main memory service processor 146527.doc -63- 201102805 105 Bus 86 Terminal interface 107 Memory interface 108 Other I /O device interface 109 communication / network interface 111 Α, 111 Β, 111C ' 111D circuit card 112A, 112B, 112C, 112D, 112E, 112F circuit card 113 circuit card 114 hardware management console (HMC) 115A, 115B, 115C circuit Cards 116A, 116B, 116C Circuit cards 117A, 117B Circuit cards 118A, 118B Circuit cards 119A, 119B Circuit cards 121A, 121B, 121C User terminals 122A, 122B, 122C Data storage device 123 Printer · 64 · 146527.doc 201102805 124 Fax machine 126 Network 201 Hardware level 202 Non-dispatchable hypervisor 203 Distributable hypervisor 204 Logical partition 205 Logical partition 206 Logical partition 207 Logical partition 211 OS core 212 OS core 213 OS core 214 OS Core 221 visual indicator 222 assignable to partitionable entity 223 separable physical location 300 logically partitioned data processing system 310 logical partition 320 hypervisor 325 hypervisor pipeline 330 flexible service processor 340 hardware management console (HMC) 400 logical HMC command session / HMC to hypervisor session 146527 .doc •65- 201102805 405 Message Buffer Pool 410 Logical Split Event Session/Logical Split to HyperManager Session 415 Message Buffer Set 500 Logical HMC Pair Partition Communication Session 1400 Computer Program Product 1410 Computer Available Read media 1420 computer readable code component or logic 146527.doc -66-

Claims (1)

201102805 七、申請專利範圍: 1 · 一種在一經邏輯分割的資料處理系統之一硬體管理控制 台與一邏輯分割區之間通信的方法,該方法包含: 由一源端點將該源端點之一請求或一回應作為貨物封 裝於一泛用傳送基元中,該源端點為該資料處理系統之 一硬體管理控制台或一邏輯分割區中之一者,其中該硬 體管理控制台為用於分割管理之一使用者介面;及 經由該資料處理系統之一超管理器將該泛用傳送基元 自該源端點轉遞至一目標端點,其中該超管理器接收在 該源端點處經封裝之該泛用傳送基元且將該泛用傳送基 元之該貨物轉遞至該目標端點,該貨物包含該源端點之 該請求或该回應,且其中在由該超管理器進行之該接收 及該轉遞中未由該超管理器對該貨物進行檢驗或剖析, 且该目標端點為該資料處理系統之該邏輯分割區或該硬 體管理控制台中之另一者。 2.如請求項1之方法,其中該接收到之泛用傳送基元為一 源傳送基元’且其中該轉遞包含由該超管理器建置一目 標傳送基元,該建置包含將該源傳送基元之該貨物複製 至该目標傳送基元中,且將該目標傳送基元自該超管理 器轉遞至該目標端點。 3 ·如凊求項2之方法,其中該接收包含在該超管理器處將 該源傳送基元接收至一接收緩衝器中,且該建置包含在 該超#理器處在一目標訊息緩衝器中建置該目標傳送基 元’该目標訊息緩衝器係來自在該超管理器處之與該目 146527.doc 201102805 標端點相關聯的訊息緩衝器之一集區(p〇〇l)。 “如請求項3之方法,冑中該建置進一步包含判定與該目 标端點相關聯之訊息緩衝器之該集區中是否有一目標訊 息緩衝器可用,且若是,則自與該目標端點相關聯之訊 息緩衝器之該集區獲得該目標訊息緩衝器且在該所獲得 之目標訊息緩衝器中建置該目標傳送基元,該建置包含 由該超管理器將包含該源端點之該請求或該回應的該貨 物自該接收緩衝器複製至該目標訊息缓衝器中而不對該 貨物進行檢驗或剖析。 5·如4求項2之方法’其中將該目標傳送基元轉遞至該目 標端點包含將該目標傳送基元非同步地轉遞至該目標端 點。 6·如請求項2之方法,其中該建置包含由該超管理器驗證 該源傳送基元中之該貨物之大小在一可接受大小範圍 内,且驗證隨該源傳送基元提供之一目標端點識別 效的。 ” 7.如叫求項2之方法,其中該目標端點在接收到該目標傳 送基元之後判定其貨物是否含有一有效請求,且若是, 則處理該請求且經由該超管理器將一回應作為另一泛用 傳送基元内之貨物傳回至該源端點,而若隨該目標傳送 基元—起接收到之該貨物並非—有效請求,則該 點判定該貨物是否含有一有效回應,且若是,則處:該 回應。 8·如请求们之方法,其進一步包含開啟該硬體管理控制 146527.doc 201102805 台與該超管理器之間的一通信會期,且開啟該邏輯分割 區與該超管理器之間的一通信會期,其中該開啟該等通 信會期發生在對該泛用傳送基元之該封裝及該轉遞之 前。 9. 如睛求項8之方法’其進一步包含回應於改變在該經邏 輯分割的資料處理系統之該超管理器或一端點處之能力 的一事件而在該邏輯分割區、該超管理器及該硬體管理 控制台之間自動交換通信能力,該端點為該邏輯分割區 或該硬體管理控制台中之一者,其中在該邏輯分割區與 該硬體管理控制台之間交換能力包含以下各項中之L 者: (1)在該邏輯分割區處冑—交換能力冑求作$貨物封裝 於泛用傳送基兀中,且經由該超管理器將該泛用傳送 基,自該邏輯分割區轉遞至該硬體管理控制台而不由該 超官理器對該貨物進行檢驗或剖析;或 ⑼由該超管理器起始—交換能力請求,且將該交· 力請求自該超管理器轉遞至該邏輯分割區。 、月b 10. —種經邏輯分割的資料處理系統其包含: 至少-處理器’其包含至少—邏輯分割區; 至;-硬體官理控制台,每一硬體管理控制台 分割管理之一使用者介面;及 ; 超官理器,其使該至少一硬體管理控制台與嗜至丨 一邏輯分㈣介面連接,且包含用於經由該超管理W 該至少-硬體管理控制台與該至少_邏輯分割區之間通 146527.doc 201102805 信的一通信設施,該通信包括: 由一源端點將該源端點之一請求或一回應作為貨物 封裝於一泛用傳送基元中,該源端點為該至少一硬體 管理控制台中之一硬體管理控制台或該至少一邏輯分 割區中之一邏輯分割區;及 經由該超管理器將該泛用傳送基元自該源端點轉遞 至一目標端點,其中該超管理器接收在該源端點處經 封裝之該泛用傳送基元且將該泛用傳送基元之該貨物 轉遞至該目標端點,該貨物包含該請求或該回應,且 其中在由該超管理器進行之該接收及該轉遞中未由該 超管理器對該貨物進行檢驗或剖析,且該目標端點為 該至少一邏輯分割區中之該邏輯分割區或該至少一硬 體管理控制台中之該硬體管理控制台中的另一者。 11·如請求項10之經邏輯分割的資料處理系統,其中該接收 到之泛用傳送基元為一源傳送基元,且其中該轉遞包含 由該超管理器建置一目標傳送基元,該建置包含將該源 傳送基元之該貨物複製至該目標傳送基元中,且將該目 標傳送基元自該超管理器轉遞至該目標端點。 12.如請求項1!之經邏輯分割的資料處理系統,其中該超管 理器進一步包括與該目標端點相關聯之訊息緩衝器之一 集區,且其中該超管理器將該泛用傳送基元接收至一接 收緩衝器中且在一目標訊息緩衝器中建置該目標傳送基 兀,忒目標訊息緩衝|§為自與該目標端點相關聯之訊息 緩衝器之s亥集區操取到之一訊息緩衝器。 146527.doc 201102805 13. 如請求項12之經遜輯分割的資料處理系統,其中該建置 該目標傳送基元進一步包括判定與該目標端點相關聯之 訊息緩衝器之該集區中是否有一目標訊息緩衝器可用, 且若是,則自與該目標端點相關聯之訊息緩衝器之該集 區獲得該目標訊息緩衝器且在該所獲得之目標訊息緩衝 器中建置該目標傳送基元,該建置包括由該超管理器將 包含邊源端點之該請求或該回應的該貨物自該接收緩衝 器複製至該目標訊息緩衝器中而不對該貨物進行檢驗或 剖析。 14. 如請求項11之經邏輯分割的資料處理系統其中將該目 標傳送基兀轉遞至該目標端點包含由該超管理器將該目 標傳送基元非同步地轉遞至該目標端點。 15. 如請求項1〇之經邏輯分割的資料處理系統,其中該至少 一處理器包含多個邏輯分割區,且其中該至少一硬體管 理控制台跨越耦接至該超管理器之—靈活服務處理器而 連接至該經邏輯分割的資料處理系統之該超管理器,且 該通信設施進一步包含在該至少一硬體管理控制台中之 每一硬體管理控制台與該超管理器之間的一各別開啟通 k會期’及在該多個邏輯分割區中之每一邏輯分割區與 該超管理器之間的一各別開啟通信會期。 16 ·如请求項15之經邏輯分割的資料處理系統,其中該通信 δ史施經組態以回應於改變在該經邏輯分割的資料處理系 統之該超官理器或一端點處之能力的一事件而在該等邏 輯分割區、該超管理器及該至少一硬體管理控制台之間 146527.doc 201102805 區中之一邏輯 硬體管理控制 自動交換能力,該端點為該多個邏輯分割 分割區或該至少一硬體管理控制台中之— 台。 17. 一具有用以促進一經邏輯分割的資料處理糸 示、、死之一 管理控制台與一邏輯分割區之間之通 乐·品,其包 至少-電腦可讀媒體,其具有用以促進—經邏 的資料處理系統之一硬體管理控制台與—邏輯八割2 ° 間之通信的電腦可讀程式碼邏輯,當在—處 Υ區之 如理盗上執行 時’該電腦可讀程式碼邏輯執行以下動作: 由-源端點將該源端點之—請求或1應作為貨物 封裝於一泛用傳送基元中,該源端點為該資料處理系 統之一硬體管理控制台或一邏輯分割區中之—者,其 中該硬體管理控制台為用於分割管理之一使用者介 面;及 經由該資料處理系統之一超管理器將該泛用傳送基 元自該源J而點轉遞至一目標端點,其中該超管理器接 收在該源端點處經封裝之該泛用傳送基元且將該泛用 傳送基元之該貨物轉遞至該目標端點,該貨物包含該 源端點之該請求或該回應,且其中在由該超管理器進 行之該接收及該轉遞中未由該超管理器對該貨物進行 檢驗或剖析’且該目標端點為該資料處理系統之該邏 輯分割區或該硬體管理控制台中之另一者。 18.如晴求項17之製品’其中該接收到之泛用傳送基元為一 146527.doc 201102805 源傳送基元,且其中該轉遞包含由該超管理器建置一目 標傳送基元,該建置包含將該源傳送基元之該貨物複製 至該目標傳送基元中,且將該目標傳送基元自該超管理 器轉遞至該目標端點。 19. 20. 如請求項18之製品’其中該接收包含在該超管理器處將 該源傳送基元接收至一接收緩衝器中,且該建置包含在 "亥超f理器處在一目標訊息緩衝器中建置該目標傳送基 70,该目標訊息緩衝器來自在該超管理器處之與該目標 端點相關聯的訊息緩衝器之一集.區。 如π求項19之製品,其中該建置進一步包含判定與該目 標端點相關聯之訊息緩衝器之該集區中是否有一目標訊 …緩衝器可用’且若是’則自與該目標端點相關聯之訊 息緩衝器之該集㈣得該目標訊息緩衝器且在該所獲得 之目標訊息緩衝器中建置該目標傳送基元,該建置包含 由該超管理器將包含該源端點之該請求或該回應的該貨 ^自該接收緩衝器複製至該目襟訊息緩衝器中而不對該 貝物進行檢驗或剖析。 146527.doc201102805 VII. Patent application scope: 1 · A method for communication between a hardware management console and a logical partition of a logically divided data processing system, the method comprising: the source endpoint being A request or a response is packaged as a commodity in a general-purpose transport primitive, the source endpoint being one of a hardware management console or a logical partition of the data processing system, wherein the hardware management console a user interface for partition management; and forwarding the generic transport primitive from the source endpoint to a target endpoint via a hypervisor of the data processing system, wherein the hypervisor receives The generalized transport primitive encapsulated by the source endpoint and the goods of the generic transport primitive are forwarded to the target endpoint, the shipment containing the request or the response of the source endpoint, and wherein The super-manager performs the inspection and parsing of the goods by the hyper-manager in the receiving and the transfer, and the target endpoint is the logical partition of the data processing system or the hardware management console The other one. 2. The method of claim 1, wherein the received generalized transport primitive is a source transport primitive 'and wherein the forwarding comprises constructing a target transport primitive by the hypervisor, the The goods of the source transport primitive are copied into the target transport primitive and the target transport primitive is forwarded from the hypervisor to the target endpoint. 3. The method of claim 2, wherein the receiving comprises receiving, at the hypervisor, the source transport primitive into a receive buffer, and the constructing comprises at the super processor a target message The target transport primitive is built in the buffer. The target message buffer is from one of the message buffers associated with the endpoint of the destination 146527.doc 201102805 at the hypervisor (p〇〇l ). "A method of claim 3, wherein the constructing further comprises determining whether a target message buffer is available in the set of message buffers associated with the target endpoint, and if so, from the target endpoint The pool of associated message buffers obtains the target message buffer and constructs the target transport primitive in the obtained target message buffer, the build containing the source endpoint to be included by the hypervisor The request or the response of the goods is copied from the receiving buffer to the target message buffer without checking or dissecting the goods. 5. The method of claim 2, wherein the target transfer primitive is transferred Delivering to the target endpoint includes asynchronously forwarding the target transport primitive to the target endpoint. 6. The method of claim 2, wherein the constructing comprises verifying, by the hypervisor, the source transport primitive The size of the goods is within an acceptable size range, and verification is determined by the target endpoint provided by the source transport primitive. 7. The method of claim 2, wherein the target endpoint is receiving The head Determining whether the shipment contains a valid request after the primitive is transmitted, and if so, processing the request and transmitting a response via the hypervisor to the source endpoint as another generalized transport primitive, and if With the target transmitting primitive - the received payload is not a valid request, the point determines whether the shipment contains a valid response, and if so, the response. 8. The method of requesting, further comprising: enabling a communication session between the hardware management control 146527.doc 201102805 and the hypervisor, and enabling a connection between the logical partition and the hypervisor The communication session, wherein the opening of the communication session occurs before the encapsulation of the general purpose transport primitive and the transfer. 9. The method of claim 8, wherein the further includes an event in response to changing the capability at the hypervisor or an endpoint of the logically segmented data processing system in the logical partition, the hypervisor And automatically exchange communication capability between the hardware management console, the endpoint is one of the logical partition or the hardware management console, wherein the logical partition and the hardware management console exchange capability Including the L of the following: (1) At the logical partition, the exchange capability is requested to be packaged in a general-purpose transport base, and the general-purpose transport base is self-managed via the hypervisor. The logical partition is forwarded to the hardware management console without the inspection or profiling of the goods by the hyper-manufacturer; or (9) initiated by the hyper-manager-exchange capability request, and the communication request is from The hypervisor is forwarded to the logical partition. , month b 10. A logically segmented data processing system comprising: at least a processor 'which contains at least a logical partition; to; a hardware administrative console, each hardware management console partition management a user interface; and; a super-orientation device that connects the at least one hardware management console to the hobby-to-one (four) interface, and includes the at least-hardware management console via the hyper-management And a communication facility between the at least _ logical partition 146527.doc 201102805, the communication comprising: including, by a source endpoint, the request or a response of the source endpoint as a commodity in a general-purpose transport primitive The source endpoint is a hardware management console of the at least one hardware management console or one of the at least one logical partition; and the generic transport primitive is from the hypervisor via the hypervisor The source endpoint forwards to a target endpoint, wherein the hypervisor receives the generalized transport primitive encapsulated at the source endpoint and forwards the shipment of the generic transport primitive to the target endpoint , the goods contain the request Or the response, and wherein the shipment is not verified or parsed by the hypervisor in the receiving and the transfer by the hypervisor, and the target endpoint is the one of the at least one logical partition The logical partition or the other of the hardware management consoles in the at least one hardware management console. 11. The data processing system of claim 10, wherein the received generalized transport primitive is a source transport primitive, and wherein the transferring includes constructing a target transport primitive by the hypervisor The constructing includes copying the goods of the source transport primitive into the target transport primitive and forwarding the target transport primitive from the hypervisor to the target endpoint. 12. The logically segmented data processing system of claim 1; wherein the hypervisor further comprises a pool of message buffers associated with the target endpoint, and wherein the hypervisor transmits the generic The primitive is received into a receive buffer and the target transport base is built in a target message buffer, and the target message buffer is configured as a message buffer from the target buffer associated with the target endpoint. Get one of the message buffers. 146527.doc 201102805 13. The data processing system of claim 12, wherein the constructing the target transport primitive further comprises determining whether there is one of the regions of the message buffer associated with the target endpoint a target message buffer is available, and if so, the target message buffer is obtained from the pool of message buffers associated with the target endpoint and the target transport primitive is built in the obtained target message buffer The constructing includes copying, by the hypervisor, the request including the edge source endpoint or the response to the goods from the receiving buffer to the target message buffer without inspecting or parsing the goods. 14. The data processing system of claim 11, wherein the forwarding of the target transport base to the target endpoint comprises asynchronously forwarding the target transport primitive to the target endpoint by the hypervisor . 15. The data processing system of claim 1, wherein the at least one processor comprises a plurality of logical partitions, and wherein the at least one hardware management console is coupled to the hypervisor - flexible The service processor is coupled to the hypervisor of the logically partitioned data processing system, and the communication facility is further included between each hardware management console in the at least one hardware management console and the hypervisor Each of the plurality of logical partitions and each of the plurality of logical partitions and the hypervisor open a communication session. 16. The logically segmented data processing system of claim 15, wherein the communication delta history is configured to respond to changing capabilities at the super-orientator or an endpoint of the logically-divided data processing system. An event in one of the logical partitions, the hypervisor, and the at least one hardware management console 146527.doc 201102805 zone one of the logical hardware management controls the automatic switching capability, the endpoint is the multiple logic Split the partition or the at least one of the hardware management consoles. 17. A device having at least a computer readable medium having a data processing means for facilitating logical segmentation, and a management console and a logical partition, the package having at least a computer readable medium - Computer-readable code logic for communication between a hardware management console and a logical communication system, which is readable by the computer. The code logic performs the following actions: The source endpoint's request or 1 should be encapsulated in a generic transport primitive by the source endpoint, which is one of the data processing system hardware management consoles. Or a logical partition, wherein the hardware management console is a user interface for split management; and the general transfer primitive is from the source J via one of the data processing systems. And the point is forwarded to a target endpoint, wherein the hypervisor receives the generalized transport primitive encapsulated at the source endpoint and forwards the goods of the generic transport primitive to the target endpoint, The goods contain the source endpoint The request or the response, and wherein the shipment is not verified or parsed by the hypervisor in the receipt and the transfer by the hypervisor and the target endpoint is the logic of the data processing system The partition or the other of the hardware management consoles. 18. The article of claim 17, wherein the received general-purpose transport primitive is a 146527.doc 201102805 source transport primitive, and wherein the transfer includes constructing a target transport primitive by the hypervisor, The constructing includes copying the shipment of the source transport primitive into the target transport primitive and forwarding the target transport primitive from the hypervisor to the target endpoint. 19. 20. The article of claim 18, wherein the receiving comprises receiving the source transport primitive into a receive buffer at the hypervisor, and the constructing is included in the "Haifu processor The target transport base 70 is built into a target message buffer from a set of regions of the message buffer associated with the target endpoint at the hypervisor. An article of claim 19, wherein the constructing further comprises determining whether a target buffer is available in the region of the message buffer associated with the target endpoint, and if yes, then the target endpoint The set of associated message buffers (4) obtains the target message buffer and constructs the target transport primitive in the obtained target message buffer, the build containing the source endpoint to be included by the hypervisor The request or the item of the response is copied from the receiving buffer to the directory message buffer without checking or parsing the item. 146527.doc
TW099106799A 2009-03-13 2010-03-09 Hypervisor-based facility for communicating between a hardware management console and a logical partition TWI463304B (en)

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US12/403,402 US8230077B2 (en) 2008-06-06 2009-03-13 Hypervisor-based facility for communicating between a hardware management console and a logical partition

Publications (2)

Publication Number Publication Date
TW201102805A true TW201102805A (en) 2011-01-16
TWI463304B TWI463304B (en) 2014-12-01

Family

ID=44838280

Family Applications (1)

Application Number Title Priority Date Filing Date
TW099106799A TWI463304B (en) 2009-03-13 2010-03-09 Hypervisor-based facility for communicating between a hardware management console and a logical partition

Country Status (1)

Country Link
TW (1) TWI463304B (en)

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7502842B2 (en) * 2003-09-25 2009-03-10 International Business Machines Corporation Auto-configuration of an internal VLAN network interface
US8024726B2 (en) * 2004-05-28 2011-09-20 International Business Machines Corporation System for correct distribution of hypervisor work
US20060123217A1 (en) * 2004-12-07 2006-06-08 International Business Machines Corporation Utilization zones for automated resource management
US7693811B2 (en) * 2006-02-28 2010-04-06 International Business Machines Corporation Generating unique identifiers for logical partitions

Also Published As

Publication number Publication date
TWI463304B (en) 2014-12-01

Similar Documents

Publication Publication Date Title
US8230077B2 (en) Hypervisor-based facility for communicating between a hardware management console and a logical partition
CN104714905B (en) Method and system for performing failover operations
US20180375782A1 (en) Data buffering
CN1881944B (en) Improved distributed kernel operating system
TWI337823B (en) Message delivery with configurable assurances and features between two endpoints
KR100992050B1 (en) Method and system for protocol offload and direct i/o with i/o sharing in a virtualized network environment
JP4636629B2 (en) A system that can provide remote recovery of a remote server
TWI306196B (en) Apparatus for managing read and write data congestion, method and host bus adapter for avoiding data congestion, host computer, storage area network and computer readable recording medium with instructions for avoiding data congestion
TW477950B (en) Event-driven communications interface for logically-partitioned computer
TWI252651B (en) System, method, and product for managing data transfers in a network
CN1881945A (en) Improved distributed kernel operating system
JP2006236391A (en) System and method for positive response of message reception on communication network of packet base
WO2017049927A1 (en) Message dispatching method, device and system
CN105144105A (en) System and method for a scalable crash-consistent snapshot operation
US9509529B1 (en) Assured messaging system with differentiated real time traffic
CN107179879A (en) Method and apparatus for the Data Migration of storage device
CN1650582A (en) Methods and apparatus for implementing a high availability fibre channel switch
CN107210967A (en) System and method for optimizing network transmission
JP2003515813A (en) Quorum resource arbiter in storage network
JP2015532073A (en) System and method for small batch processing of usage requests
JP6336602B2 (en) Packet flow control method, related apparatus, and computing node
US8028017B2 (en) Virtual controllers with a large data center
US10459791B2 (en) Storage device having error communication logical ports
JP4071098B2 (en) Architecture and runtime environment for network filter drivers
CN1508714A (en) Method and system for determining activity of high availability mass