/*********************************************************************/ /* Copyright 2009, 2010 The University of Texas at Austin. */ /* All rights reserved. */ /* */ /* Redistribution and use in source and binary forms, with or */ /* without modification, are permitted provided that the following */ /* conditions are met: */ /* */ /* 1. Redistributions of source code must retain the above */ /* copyright notice, this list of conditions and the following */ /* disclaimer. */ /* */ /* 2. Redistributions in binary form must reproduce the above */ /* copyright notice, this list of conditions and the following */ /* disclaimer in the documentation and/or other materials */ /* provided with the distribution. */ /* */ /* THIS SOFTWARE IS PROVIDED BY THE UNIVERSITY OF TEXAS AT */ /* AUSTIN ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, */ /* INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF */ /* MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE */ /* DISCLAIMED. IN NO EVENT SHALL THE UNIVERSITY OF TEXAS AT */ /* AUSTIN OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, */ /* INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES */ /* (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE */ /* GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR */ /* BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF */ /* LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT */ /* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT */ /* OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE */ /* POSSIBILITY OF SUCH DAMAGE. */ /* */ /* The views and conclusions contained in the software and */ /* documentation are those of the authors and should not be */ /* interpreted as representing official policies, either expressed */ /* or implied, of The University of Texas at Austin. */ /*********************************************************************/ #define ASSEMBLER #include "common.h" #ifndef __64BIT__ #define LOAD lwz #else #define LOAD ld #endif #ifdef __64BIT__ #define STACKSIZE 360 #else #define STACKSIZE 272 #endif #define ALIGN_SIZE 0xffff #define SWAP 0 #define NEG 16 #define ALPHA_R 32 #define ALPHA_I 48 #define FZERO 64 #define M r3 #define N r4 #define K r5 #ifdef linux #ifndef __64BIT__ #define A r6 #define B r7 #define C r8 #define LDC r9 #else #define A r8 #define B r9 #define C r10 #define LDC r6 #endif #endif #if defined(_AIX) || defined(__APPLE__) #if !defined(__64BIT__) && defined(DOUBLE) #define A r10 #define B r6 #define C r7 #define LDC r8 #else #define A r8 #define B r9 #define C r10 #define LDC r6 #endif #endif #define STACK r11 #define I r21 #define J r22 #define AO r23 #define BO r24 #define CO1 r25 #define CO2 r26 #define PREA r29 #define PREB r29 #define PREC r30 #define VREG r31 #define LOAD_A lvx #define LOAD_B lvx #define OFFSET_0 0 #define OFFSET_1 r14 #define OFFSET_2 r15 #define OFFSET_3 r16 #define OFFSET_4 r17 #define OFFSET_5 r18 #define OFFSET_6 r19 #define OFFSET_7 r20 #define c01 v0 #define c02 v1 #define c03 v2 #define c04 v3 #define c05 v4 #define c06 v5 #define c07 v6 #define c08 v7 #define c09 v8 #define c10 v9 #define c11 v10 #define c12 v11 #define c13 v12 #define c14 v13 #define c15 v14 #define c16 v15 #define a1 v16 #define a2 v17 #define a3 v18 #define a4 v19 #define a5 v20 #define a6 v21 #define a7 v22 #define a8 v23 #define b1 v24 #define b2 v25 #define bp1 v26 #define bp2 v27 #define C1 v16 #define C2 v17 #define C3 v18 #define C4 v19 #define C5 v20 #define c00 v24 #define VZERO v25 #define PERMRSHIFT1 v26 #define PERMRSHIFT2 v27 #define swap v28 #define neg v29 #define alpha_r v30 #define alpha_i v31 #ifndef NEEDPARAM PROLOGUE PROFCODE addi SP, SP, -STACKSIZE mr STACK, SP li r0, 0 * 16 stvx v20, SP, r0 li r0, 1 * 16 stvx v21, SP, r0 li r0, 2 * 16 stvx v22, SP, r0 li r0, 3 * 16 stvx v23, SP, r0 li r0, 4 * 16 stvx v24, SP, r0 li r0, 5 * 16 stvx v25, SP, r0 li r0, 6 * 16 stvx v26, SP, r0 li r0, 7 * 16 stvx v27, SP, r0 li r0, 8 * 16 stvx v28, SP, r0 li r0, 9 * 16 stvx v29, SP, r0 li r0, 10 * 16 stvx v30, SP, r0 li r0, 11 * 16 stvx v31, SP, r0 #ifdef __64BIT__ std r31, 192(SP) std r30, 200(SP) std r29, 208(SP) std r28, 216(SP) std r27, 224(SP) std r26, 232(SP) std r25, 240(SP) std r24, 248(SP) std r23, 256(SP) std r22, 264(SP) std r21, 272(SP) std r20, 280(SP) std r19, 288(SP) std r18, 296(SP) std r17, 304(SP) std r16, 312(SP) std r15, 320(SP) std r14, 328(SP) #else stw r31, 192(SP) stw r30, 196(SP) stw r29, 200(SP) stw r28, 204(SP) stw r27, 208(SP) stw r26, 212(SP) stw r25, 216(SP) stw r24, 220(SP) stw r23, 224(SP) stw r22, 228(SP) stw r21, 232(SP) stw r20, 236(SP) stw r19, 240(SP) stw r18, 244(SP) stw r17, 248(SP) stw r16, 252(SP) stw r15, 256(SP) stw r14, 260(SP) #endif #ifdef linux #ifdef __64BIT__ ld LDC, FRAMESLOT(0) + STACKSIZE(SP) #endif #endif #if defined(_AIX) || defined(__APPLE__) #ifdef __64BIT__ ld LDC, FRAMESLOT(0) + STACKSIZE(SP) #else #ifdef DOUBLE lwz B, FRAMESLOT(0) + STACKSIZE(SP) lwz C, FRAMESLOT(1) + STACKSIZE(SP) lwz LDC, FRAMESLOT(2) + STACKSIZE(SP) #else lwz LDC, FRAMESLOT(0) + STACKSIZE(SP) #endif #endif #endif li r0, -1 mfspr VREG, VRsave mtspr VRsave, r0 addi SP, SP, -128 li r0, -8192 and SP, SP, r0 fneg f3, f1 fneg f4, f2 #if defined(NN) || defined(NT) || defined(TN) || defined(TT) || \ defined(NC) || defined(TC) || defined(NR) || defined(TR) stfs f1, ALPHA_R + 0(SP) stfs f1, ALPHA_R + 4(SP) stfs f1, ALPHA_R + 8(SP) stfs f1, ALPHA_R + 12(SP) stfs f4, ALPHA_I + 0(SP) stfs f2, ALPHA_I + 4(SP) stfs f4, ALPHA_I + 8(SP) stfs f2, ALPHA_I + 12(SP) #else stfs f1, ALPHA_R + 0(SP) stfs f3, ALPHA_R + 4(SP) stfs f1, ALPHA_R + 8(SP) stfs f3, ALPHA_R + 12(SP) stfs f2, ALPHA_I + 0(SP) stfs f2, ALPHA_I + 4(SP) stfs f2, ALPHA_I + 8(SP) stfs f2, ALPHA_I + 12(SP) #endif li I, Address_L(0x04050607) addis I, I, Address_H(0x04050607) stw I, SWAP + 0(SP) li I, Address_L(0x00010203) addis I, I, Address_H(0x00010203) stw I, SWAP + 4(SP) li I, Address_L(0x0c0d0e0f) addis I, I, Address_H(0x0c0d0e0f) stw I, SWAP + 8(SP) li I, Address_L(0x08090a0b) addis I, I, Address_H(0x08090a0b) stw I, SWAP + 12(SP) #if defined(NN) || defined(NT) || defined(TN) || defined(TT) || \ defined(RR) || defined(RC) || defined(CR) || defined(CC) lis I, 0x8000 stw I, NEG + 0(SP) stw I, NEG + 8(SP) li I, 0 stw I, NEG + 4(SP) stw I, NEG + 12(SP) #else li I, 0 stw I, NEG + 0(SP) stw I, NEG + 8(SP) lis I, 0x8000 stw I, NEG + 4(SP) stw I, NEG + 12(SP) #endif li r0, 0 stw r0, FZERO(SP) slwi LDC, LDC, ZBASE_SHIFT li PREC, (15 * SIZE) li PREB, (25 * 8 * SIZE) li OFFSET_1, 4 * SIZE li OFFSET_2, 8 * SIZE li OFFSET_3, 12 * SIZE li OFFSET_4, 16 * SIZE li OFFSET_5, 20 * SIZE li OFFSET_6, 24 * SIZE li OFFSET_7, 28 * SIZE cmpwi cr0, M, 0 ble LL(999) cmpwi cr0, N, 0 ble LL(999) cmpwi cr0, K, 0 ble LL(999) srawi. J, N, 1 ble LL(50) .align 4 LL(01): mr CO1, C add CO2, C, LDC add C, CO2, LDC mr AO, A srawi. I, M, 3 ble LL(20) .align 4 LL(11): vxor c01, c01, c01 LOAD_B b1, OFFSET_0, B vxor c02, c02, c02 LOAD_A a1, OFFSET_0, AO vxor c03, c03, c03 LOAD_A a2, OFFSET_1, AO vxor c04, c04, c04 LOAD_A a3, OFFSET_2, AO vxor c05, c05, c05 LOAD_A a4, OFFSET_3, AO vxor c06, c06, c06 LOAD_B b2, OFFSET_2, B vxor c07, c07, c07 LOAD_A a5, OFFSET_4, AO vxor c08, c08, c08 LOAD_A a6, OFFSET_5, AO vxor c09, c09, c09 dcbtst CO1, PREC vxor c10, c10, c10 dcbtst CO2, PREC vxor c11, c11, c11 vxor c12, c12, c12 vxor c13, c13, c13 mr BO, B vxor c14, c14, c14 srawi. r0, K, 2 vxor c15, c15, c15 mtspr CTR, r0 vxor c16, c16, c16 vspltw bp1, b1, 0 ble LL(15) .align 4 LL(12): /* 1 */ vmaddfp c01, a1, bp1, c01 vspltw bp2, b1, 1 vmaddfp c02, a2, bp1, c02 addi AO, AO, 8 * SIZE vmaddfp c03, a3, bp1, c03 LOAD_A a7, OFFSET_4, AO vmaddfp c04, a4, bp1, c04 LOAD_A a8, OFFSET_5, AO /* 2 */ vmaddfp c05, a1, bp2, c05 vspltw bp1, b1, 2 vmaddfp c06, a2, bp2, c06 dcbt BO, PREB vmaddfp c07, a3, bp2, c07 dcbt AO, PREB vmaddfp c08, a4, bp2, c08 addi AO, AO, 8 * SIZE /* 3 */ vmaddfp c09, a1, bp1, c09 vspltw bp2, b1, 3 vmaddfp c10, a2, bp1, c10 LOAD_B b1, OFFSET_1, BO vmaddfp c11, a3, bp1, c11 dcbt AO, PREB vmaddfp c12, a4, bp1, c12 addi AO, AO, 8 * SIZE /* 4 */ vmaddfp c13, a1, bp2, c13 vspltw bp1, b1, 0 vmaddfp c14, a2, bp2, c14 LOAD_A a1, OFFSET_2, AO vmaddfp c15, a3, bp2, c15 dcbt AO, PREB vmaddfp c16, a4, bp2, c16 addi AO, AO, 8 * SIZE /* 5 */ vmaddfp c01, a5, bp1, c01 vspltw bp2, b1, 1 vmaddfp c02, a6, bp1, c02 LOAD_A a2, OFFSET_1, AO vmaddfp c03, a7, bp1, c03 LOAD_A a3, OFFSET_2, AO vmaddfp c04, a8, bp1, c04 LOAD_A a4, OFFSET_3, AO /* 6 */ vmaddfp c05, a5, bp2, c05 vspltw bp1, b1, 2 vmaddfp c06, a6, bp2, c06 nop vmaddfp c07, a7, bp2, c07 dcbt AO, PREA vmaddfp c08, a8, bp2, c08 addi AO, AO, 8 * SIZE /* 7 */ vmaddfp c09, a5, bp1, c09 vspltw bp2, b1, 3 vmaddfp c10, a6, bp1, c10 LOAD_B b1, OFFSET_4, BO vmaddfp c11, a7, bp1, c11 nop vmaddfp c12, a8, bp1, c12 nop /* 8 */ vmaddfp c13, a5, bp2, c13 vspltw bp1, b2, 0 vmaddfp c14, a6, bp2, c14 LOAD_A a5, OFFSET_2, AO vmaddfp c15, a7, bp2, c15 LOAD_A a6, OFFSET_3, AO vmaddfp c16, a8, bp2, c16 LOAD_A a7, OFFSET_4, AO /* 9 */ vmaddfp c01, a1, bp1, c01 vspltw bp2, b2, 1 vmaddfp c02, a2, bp1, c02 LOAD_A a8, OFFSET_5, AO vmaddfp c03, a3, bp1, c03 addi BO, BO, 8 * SIZE vmaddfp c04, a4, bp1, c04 nop /* 10 */ vmaddfp c05, a1, bp2, c05 vspltw bp1, b2, 2 vmaddfp c06, a2, bp2, c06 nop vmaddfp c07, a3, bp2, c07 nop vmaddfp c08, a4, bp2, c08 nop /* 11 */ vmaddfp c09, a1, bp1, c09 vspltw bp2, b2, 3 vmaddfp c10, a2, bp1, c10 LOAD_B b2, OFFSET_1, BO vmaddfp c11, a3, bp1, c11 dcbt AO, PREA vmaddfp c12, a4, bp1, c12 addi AO, AO, 8 * SIZE /* 12 */ vmaddfp c13, a1, bp2, c13 vspltw bp1, b2, 0 vmaddfp c14, a2, bp2, c14 LOAD_A a1, OFFSET_4, AO vmaddfp c15, a3, bp2, c15 LOAD_A a2, OFFSET_5, AO vmaddfp c16, a4, bp2, c16 LOAD_A a3, OFFSET_6, AO /* 13 */ vmaddfp c01, a5, bp1, c01 vspltw bp2, b2, 1 vmaddfp c02, a6, bp1, c02 LOAD_A a4, OFFSET_7, AO vmaddfp c03, a7, bp1, c03 dcbt AO, PREA vmaddfp c04, a8, bp1, c04 addi AO, AO, 8 * SIZE /* 14 */ vmaddfp c05, a5, bp2, c05 vspltw bp1, b2, 2 vmaddfp c06, a6, bp2, c06 nop vmaddfp c07, a7, bp2, c07 dcbt AO, PREA vmaddfp c08, a8, bp2, c08 addi AO, AO, 8 * SIZE /* 15 */ vmaddfp c09, a5, bp1, c09 vspltw bp2, b2, 3 vmaddfp c10, a6, bp1, c10 LOAD_B b2, OFFSET_4, BO vmaddfp c11, a7, bp1, c11 dcbt AO, PREA vmaddfp c12, a8, bp1, c12 addi BO, BO, 8 * SIZE /* 16 */ vmaddfp c13, a5, bp2, c13 vspltw bp1, b1, 0 vmaddfp c14, a6, bp2, c14 LOAD_A a5, OFFSET_4, AO vmaddfp c15, a7, bp2, c15 LOAD_A a6, OFFSET_5, AO vmaddfp c16, a8, bp2, c16 bdnz+ LL(12) .align 4 LL(15): lvx swap, OFFSET_0, SP lvx neg, OFFSET_1, SP lvx alpha_r, OFFSET_2, SP lvx alpha_i, OFFSET_3, SP andi. r0, K, 3 mtspr CTR, r0 ble+ LL(18) .align 4 LL(16): vmaddfp c01, a1, bp1, c01 vspltw bp2, b1, 1 vmaddfp c02, a2, bp1, c02 nop vmaddfp c03, a3, bp1, c03 nop vmaddfp c04, a4, bp1, c04 nop vmaddfp c05, a1, bp2, c05 vspltw bp1, b1, 2 vmaddfp c06, a2, bp2, c06 nop vmaddfp c07, a3, bp2, c07 nop vmaddfp c08, a4, bp2, c08 nop vmaddfp c09, a1, bp1, c09 vspltw bp2, b1, 3 vmaddfp c10, a2, bp1, c10 LOAD_B b1, OFFSET_1, BO vmaddfp c11, a3, bp1, c11 addi AO, AO, 16 * SIZE vmaddfp c12, a4, bp1, c12 addi BO, BO, 4 * SIZE vmaddfp c13, a1, bp2, c13 vspltw bp1, b1, 0 vmaddfp c14, a2, bp2, c14 LOAD_A a1, OFFSET_0, AO vmaddfp c15, a3, bp2, c15 LOAD_A a2, OFFSET_1, AO vmaddfp c16, a4, bp2, c16 LOAD_A a3, OFFSET_2, AO LOAD_A a4, OFFSET_3, AO bdnz+ LL(16) .align 4 LL(18): vxor VZERO, VZERO, VZERO vperm c05, c05, c05, swap vperm c06, c06, c06, swap vperm c07, c07, c07, swap vperm c08, c08, c08, swap vperm c13, c13, c13, swap vperm c14, c14, c14, swap vperm c15, c15, c15, swap vperm c16, c16, c16, swap vxor c05, c05, neg vxor c06, c06, neg vxor c07, c07, neg vxor c08, c08, neg vxor c13, c13, neg vxor c14, c14, neg vxor c15, c15, neg vxor c16, c16, neg vaddfp c01, c01, c05 vaddfp c02, c02, c06 vaddfp c03, c03, c07 vaddfp c04, c04, c08 vaddfp c09, c09, c13 vaddfp c10, c10, c14 vaddfp c11, c11, c15 vaddfp c12, c12, c16 vperm c05, c01, c01, swap vperm c06, c02, c02, swap vperm c07, c03, c03, swap vperm c08, c04, c04, swap vperm c13, c09, c09, swap vperm c14, c10, c10, swap vperm c15, c11, c11, swap vperm c16, c12, c12, swap vmaddfp c01, alpha_r, c01, VZERO vmaddfp c02, alpha_r, c02, VZERO vmaddfp c03, alpha_r, c03, VZERO vmaddfp c04, alpha_r, c04, VZERO vmaddfp c01, alpha_i, c05, c01 vmaddfp c02, alpha_i, c06, c02 vmaddfp c03, alpha_i, c07, c03 vmaddfp c04, alpha_i, c08, c04 vmaddfp c09, alpha_r, c09, VZERO vmaddfp c10, alpha_r, c10, VZERO vmaddfp c11, alpha_r, c11, VZERO vmaddfp c12, alpha_r, c12, VZERO vmaddfp c09, alpha_i, c13, c09 vmaddfp c10, alpha_i, c14, c10 vmaddfp c11, alpha_i, c15, c11 vmaddfp c12, alpha_i, c16, c12 lvx C1, OFFSET_0, CO1 lvx C2, OFFSET_1, CO1 lvx C3, OFFSET_2, CO1 lvx C4, OFFSET_3, CO1 lvx C5, OFFSET_4, CO1 lvsr PERMRSHIFT1, 0, CO1 lvsr PERMRSHIFT2, 0, CO2 vperm c00, VZERO, c01, PERMRSHIFT1 vperm c01, c01, c02, PERMRSHIFT1 vperm c02, c02, c03, PERMRSHIFT1 vperm c03, c03, c04, PERMRSHIFT1 vperm c04, c04, VZERO, PERMRSHIFT1 vaddfp c00, c00, C1 vaddfp c01, c01, C2 vaddfp c02, c02, C3 vaddfp c03, c03, C4 vaddfp c04, c04, C5 stvx c00, OFFSET_0, CO1 stvx c01, OFFSET_1, CO1 stvx c02, OFFSET_2, CO1 stvx c03, OFFSET_3, CO1 stvx c04, OFFSET_4, CO1 lvx C1, OFFSET_0, CO2 lvx C2, OFFSET_1, CO2 lvx C3, OFFSET_2, CO2 lvx C4, OFFSET_3, CO2 lvx C5, OFFSET_4, CO2 vperm c00, VZERO, c09, PERMRSHIFT2 vperm c09, c09, c10, PERMRSHIFT2 vperm c10, c10, c11, PERMRSHIFT2 vperm c11, c11, c12, PERMRSHIFT2 vperm c12, c12, VZERO, PERMRSHIFT2 vaddfp c00, c00, C1 vaddfp c09, c09, C2 vaddfp c10, c10, C3 vaddfp c11, c11, C4 vaddfp c12, c12, C5 stvx c00, OFFSET_0, CO2 stvx c09, OFFSET_1, CO2 stvx c10, OFFSET_2, CO2 stvx c11, OFFSET_3, CO2 stvx c12, OFFSET_4, CO2 addi CO1, CO1, 16 * SIZE addi CO2, CO2, 16 * SIZE addic. I, I, -1 bgt+ LL(11) .align 4 LL(20): andi. I, M, 4 ble LL(30) vxor c01, c01, c01 LOAD_A a1, OFFSET_0, AO vxor c02, c02, c02 LOAD_A a2, OFFSET_1, AO vxor c05, c05, c05 LOAD_A a3, OFFSET_2, AO vxor c06, c06, c06 LOAD_A a4, OFFSET_3, AO vxor c09, c09, c09 LOAD_B b1, OFFSET_0, B vxor c10, c10, c10 LOAD_B b2, OFFSET_1, B vxor c13, c13, c13 vxor c14, c14, c14 mr BO, B vspltw bp1, b1, 0 srawi. r0, K, 1 mtspr CTR, r0 ble LL(25) .align 4 LL(22): vmaddfp c01, a1, bp1, c01 vspltw bp2, b1, 1 addi AO, AO, 16 * SIZE vmaddfp c02, a2, bp1, c02 addi BO, BO, 8 * SIZE vmaddfp c05, a1, bp2, c05 vspltw bp1, b1, 2 vmaddfp c06, a2, bp2, c06 vmaddfp c09, a1, bp1, c09 vspltw bp2, b1, 3 LOAD_B b1, OFFSET_0, BO vmaddfp c10, a2, bp1, c10 vmaddfp c13, a1, bp2, c13 LOAD_A a1, OFFSET_0, AO vspltw bp1, b2, 0 vmaddfp c14, a2, bp2, c14 LOAD_A a2, OFFSET_1, AO vmaddfp c01, a3, bp1, c01 vspltw bp2, b2, 1 vmaddfp c02, a4, bp1, c02 vmaddfp c05, a3, bp2, c05 vspltw bp1, b2, 2 vmaddfp c06, a4, bp2, c06 vmaddfp c09, a3, bp1, c09 vspltw bp2, b2, 3 LOAD_B b2, OFFSET_1, BO vmaddfp c10, a4, bp1, c10 vmaddfp c13, a3, bp2, c13 LOAD_A a3, OFFSET_2, AO vmaddfp c14, a4, bp2, c14 LOAD_A a4, OFFSET_3, AO vspltw bp1, b1, 0 bdnz LL(22) .align 4 LL(25): andi. r0, K, 1 ble+ LL(28) .align 4 LL(26): vmaddfp c01, a1, bp1, c01 vspltw bp2, b1, 1 vmaddfp c02, a2, bp1, c02 nop vmaddfp c05, a1, bp2, c05 vspltw bp1, b1, 2 vmaddfp c06, a2, bp2, c06 nop vmaddfp c09, a1, bp1, c09 vspltw bp2, b1, 3 vmaddfp c10, a2, bp1, c10 addi AO, AO, 8 * SIZE vmaddfp c13, a1, bp2, c13 addi BO, BO, 4 * SIZE vmaddfp c14, a2, bp2, c14 nop .align 4 LL(28): vxor VZERO, VZERO, VZERO lvx swap, OFFSET_0, SP lvx neg, OFFSET_1, SP lvx alpha_r, OFFSET_2, SP lvx alpha_i, OFFSET_3, SP vperm c05, c05, c05, swap vperm c06, c06, c06, swap vperm c13, c13, c13, swap vperm c14, c14, c14, swap vxor c05, c05, neg vxor c06, c06, neg vxor c13, c13, neg vxor c14, c14, neg vaddfp c01, c01, c05 vaddfp c02, c02, c06 vaddfp c09, c09, c13 vaddfp c10, c10, c14 vperm c05, c01, c01, swap vperm c06, c02, c02, swap vperm c13, c09, c09, swap vperm c14, c10, c10, swap vmaddfp c01, alpha_r, c01, VZERO vmaddfp c02, alpha_r, c02, VZERO vmaddfp c01, alpha_i, c05, c01 vmaddfp c02, alpha_i, c06, c02 vmaddfp c09, alpha_r, c09, VZERO vmaddfp c10, alpha_r, c10, VZERO vmaddfp c09, alpha_i, c13, c09 vmaddfp c10, alpha_i, c14, c10 lvx C1, OFFSET_0, CO1 lvx C2, OFFSET_1, CO1 lvx C3, OFFSET_2, CO1 lvsr PERMRSHIFT1, 0, CO1 lvsr PERMRSHIFT2, 0, CO2 vperm c00, VZERO, c01, PERMRSHIFT1 vperm c01, c01, c02, PERMRSHIFT1 vperm c02, c02, VZERO, PERMRSHIFT1 vaddfp c00, c00, C1 vaddfp c01, c01, C2 vaddfp c02, c02, C3 stvx c00, OFFSET_0, CO1 stvx c01, OFFSET_1, CO1 stvx c02, OFFSET_2, CO1 lvx C1, OFFSET_0, CO2 lvx C2, OFFSET_1, CO2 lvx C3, OFFSET_2, CO2 vperm c00, VZERO, c09, PERMRSHIFT2 vperm c09, c09, c10, PERMRSHIFT2 vperm c10, c10, VZERO, PERMRSHIFT2 vaddfp c00, c00, C1 vaddfp c09, c09, C2 vaddfp c10, c10, C3 stvx c00, OFFSET_0, CO2 stvx c09, OFFSET_1, CO2 stvx c10, OFFSET_2, CO2 addi CO1, CO1, 8 * SIZE addi CO2, CO2, 8 * SIZE .align 4 LL(30): andi. I, M, 2 ble LL(40) vxor c01, c01, c01 LOAD_A a1, OFFSET_0, AO vxor c02, c02, c02 LOAD_A a2, OFFSET_1, AO vxor c05, c05, c05 LOAD_B b1, OFFSET_0, B vxor c06, c06, c06 LOAD_B b2, OFFSET_1, B vxor c09, c09, c09 vxor c10, c10, c10 vxor c13, c13, c13 vxor c14, c14, c14 vspltw bp1, b1, 0 mr BO, B srawi. r0, K, 1 mtspr CTR, r0 ble LL(35) .align 4 LL(32): vmaddfp c01, a1, bp1, c01 addi AO, AO, 8 * SIZE vspltw bp2, b1, 1 vmaddfp c05, a1, bp2, c05 addi BO, BO, 8 * SIZE vspltw bp1, b1, 2 vmaddfp c09, a1, bp1, c09 vspltw bp2, b1, 3 vmaddfp c13, a1, bp2, c13 LOAD_A a1, OFFSET_0, AO vspltw bp1, b2, 0 LOAD_B b1, OFFSET_0, BO vmaddfp c02, a2, bp1, c02 vspltw bp2, b2, 1 vmaddfp c06, a2, bp2, c06 vspltw bp1, b2, 2 vmaddfp c10, a2, bp1, c10 vspltw bp2, b2, 3 LOAD_B b2, OFFSET_1, BO vmaddfp c14, a2, bp2, c14 LOAD_A a2, OFFSET_1, AO vspltw bp1, b1, 0 bdnz LL(32) .align 4 LL(35): andi. r0, K, 1 ble+ LL(38) .align 4 LL(36): vmaddfp c01, a1, bp1, c01 vspltw bp2, b1, 1 vmaddfp c05, a1, bp2, c05 vspltw bp1, b1, 2 vmaddfp c09, a1, bp1, c09 vspltw bp2, b1, 3 vmaddfp c13, a1, bp2, c13 addi AO, AO, 4 * SIZE addi BO, BO, 4 * SIZE .align 4 LL(38): vaddfp c01, c01, c02 vaddfp c05, c05, c06 vaddfp c09, c09, c10 vaddfp c13, c13, c14 vxor VZERO, VZERO, VZERO lvx swap, OFFSET_0, SP lvx neg, OFFSET_1, SP lvx alpha_r, OFFSET_2, SP lvx alpha_i, OFFSET_3, SP vperm c05, c05, c05, swap vperm c13, c13, c13, swap vxor c05, c05, neg vxor c13, c13, neg vaddfp c01, c01, c05 vaddfp c09, c09, c13 vperm c05, c01, c01, swap vperm c13, c09, c09, swap vmaddfp c01, alpha_r, c01, VZERO vmaddfp c01, alpha_i, c05, c01 vmaddfp c09, alpha_r, c09, VZERO vmaddfp c09, alpha_i, c13, c09 lvx C1, OFFSET_0, CO1 lvx C2, OFFSET_1, CO1 lvsr PERMRSHIFT1, 0, CO1 lvsr PERMRSHIFT2, 0, CO2 vperm c00, VZERO, c01, PERMRSHIFT1 vperm c01, c01, VZERO, PERMRSHIFT1 vaddfp c00, c00, C1 vaddfp c01, c01, C2 stvx c00, OFFSET_0, CO1 stvx c01, OFFSET_1, CO1 lvx C1, OFFSET_0, CO2 lvx C2, OFFSET_1, CO2 vperm c00, VZERO, c09, PERMRSHIFT2 vperm c09, c09, VZERO, PERMRSHIFT2 vaddfp c00, c00, C1 vaddfp c09, c09, C2 stvx c00, OFFSET_0, CO2 stvx c09, OFFSET_1, CO2 addi CO1, CO1, 4 * SIZE addi CO2, CO2, 4 * SIZE .align 4 LL(40): andi. I, M, 1 ble LL(49) mr BO, B LFD f8, 0 * SIZE(AO) LFD f9, 1 * SIZE(AO) LFD f10, 0 * SIZE(BO) LFD f11, 1 * SIZE(BO) LFD f12, 2 * SIZE(BO) LFD f13, 3 * SIZE(BO) lfs f0, FZERO(SP) fmr f1, f0 fmr f2, f0 fmr f3, f0 fmr f4, f0 fmr f5, f0 fmr f6, f0 fmr f7, f0 srawi. r0, K, 1 mtspr CTR, r0 ble LL(45) .align 4 LL(42): fmadd f0, f8, f10, f0 fmadd f2, f8, f11, f2 fmadd f4, f8, f12, f4 fmadd f6, f8, f13, f6 fmadd f1, f9, f10, f1 fmadd f3, f9, f11, f3 fmadd f5, f9, f12, f5 fmadd f7, f9, f13, f7 LFD f8, 2 * SIZE(AO) LFD f9, 3 * SIZE(AO) LFD f10, 4 * SIZE(BO) LFD f11, 5 * SIZE(BO) LFD f12, 6 * SIZE(BO) LFD f13, 7 * SIZE(BO) fmadd f0, f8, f10, f0 fmadd f2, f8, f11, f2 fmadd f4, f8, f12, f4 fmadd f6, f8, f13, f6 fmadd f1, f9, f10, f1 fmadd f3, f9, f11, f3 fmadd f5, f9, f12, f5 fmadd f7, f9, f13, f7 LFD f8, 4 * SIZE(AO) LFD f9, 5 * SIZE(AO) LFD f10, 8 * SIZE(BO) LFD f11, 9 * SIZE(BO) LFD f12, 10 * SIZE(BO) LFD f13, 11 * SIZE(BO) addi AO, AO, 4 * SIZE addi BO, BO, 8 * SIZE bdnz LL(42) .align 4 LL(45): andi. r0, K, 1 ble LL(48) .align 4 LL(46): fmadd f0, f8, f10, f0 fmadd f2, f8, f11, f2 fmadd f4, f8, f12, f4 fmadd f6, f8, f13, f6 fmadd f1, f9, f10, f1 fmadd f3, f9, f11, f3 fmadd f5, f9, f12, f5 fmadd f7, f9, f13, f7 addi AO, AO, 2 * SIZE addi BO, BO, 4 * SIZE .align 4 LL(48): #if defined(NN) || defined(NT) || defined(TN) || defined(TT) fsub f0, f0, f3 fadd f1, f1, f2 fsub f4, f4, f7 fadd f5, f5, f6 #elif defined(NR) || defined(NC) || defined(TR) || defined(TC) fadd f0, f0, f3 fsub f1, f1, f2 fadd f4, f4, f7 fsub f5, f5, f6 #elif defined(RN) || defined(RT) || defined(CN) || defined(CT) fadd f0, f0, f3 fsub f1, f2, f1 fadd f4, f4, f7 fsub f5, f6, f5 #else /* RR, RC, CR, CC */ fsub f0, f0, f3 fadd f1, f1, f2 fsub f4, f4, f7 fadd f5, f5, f6 #endif LFD f8, 0 * SIZE(CO1) LFD f9, 1 * SIZE(CO1) LFD f10, 0 * SIZE(CO2) LFD f11, 1 * SIZE(CO2) lfs f12, ALPHA_R + 0(SP) lfs f13, ALPHA_I + 4(SP) #if defined(RR) || defined(RC) || defined(CR) || defined(CC) fmadd f8, f12, f0, f8 fnmsub f9, f12, f1, f9 fmadd f10, f12, f4, f10 fnmsub f11, f12, f5, f11 fmadd f8, f13, f1, f8 fmadd f9, f13, f0, f9 fmadd f10, f13, f5, f10 fmadd f11, f13, f4, f11 #else fmadd f8, f12, f0, f8 fmadd f9, f12, f1, f9 fmadd f10, f12, f4, f10 fmadd f11, f12, f5, f11 fnmsub f8, f13, f1, f8 fmadd f9, f13, f0, f9 fnmsub f10, f13, f5, f10 fmadd f11, f13, f4, f11 #endif STFD f8, 0 * SIZE(CO1) STFD f9, 1 * SIZE(CO1) STFD f10, 0 * SIZE(CO2) STFD f11, 1 * SIZE(CO2) LL(49): mr B, BO addic. J, J, -1 bgt LL(01) .align 4 LL(50): andi. J, N, 1 ble LL(999) mr CO1, C mr AO, A srawi. I, M, 3 ble LL(70) .align 4 LL(61): vxor c01, c01, c01 LOAD_B b1, OFFSET_0, B vxor c02, c02, c02 vxor c03, c03, c03 LOAD_A a1, OFFSET_0, AO vxor c04, c04, c04 LOAD_A a2, OFFSET_1, AO vxor c05, c05, c05 LOAD_A a3, OFFSET_2, AO vxor c06, c06, c06 LOAD_A a4, OFFSET_3, AO vxor c07, c07, c07 vxor c08, c08, c08 mr BO, B dcbtst CO1, PREC dcbtst CO2, PREC vspltw bp1, b1, 0 srawi. r0, K, 1 mtspr CTR, r0 ble LL(65) .align 4 LL(62): LOAD_A a5, OFFSET_4, AO LOAD_A a6, OFFSET_5, AO LOAD_A a7, OFFSET_6, AO LOAD_A a8, OFFSET_7, AO vmaddfp c01, a1, bp1, c01 vspltw bp2, b1, 1 vmaddfp c02, a2, bp1, c02 vmaddfp c03, a3, bp1, c03 vmaddfp c04, a4, bp1, c04 vmaddfp c05, a1, bp2, c05 vspltw bp1, b1, 2 vmaddfp c06, a2, bp2, c06 vmaddfp c07, a3, bp2, c07 vmaddfp c08, a4, bp2, c08 vmaddfp c01, a5, bp1, c01 vspltw bp2, b1, 3 vmaddfp c02, a6, bp1, c02 vmaddfp c03, a7, bp1, c03 vmaddfp c04, a8, bp1, c04 LOAD_B b1, OFFSET_1, BO vspltw bp1, b1, 0 vmaddfp c05, a5, bp2, c05 vmaddfp c06, a6, bp2, c06 vmaddfp c07, a7, bp2, c07 vmaddfp c08, a8, bp2, c08 addi AO, AO, 32 * SIZE addi BO, BO, 4 * SIZE LOAD_A a1, OFFSET_0, AO LOAD_A a2, OFFSET_1, AO LOAD_A a3, OFFSET_2, AO LOAD_A a4, OFFSET_3, AO bdnz LL(62) .align 4 LL(65): andi. r0, K, 1 ble+ LL(68) .align 4 LL(66): vmaddfp c01, a1, bp1, c01 vspltw bp2, b1, 1 vmaddfp c02, a2, bp1, c02 addi AO, AO, 16 * SIZE vmaddfp c03, a3, bp1, c03 addi BO, BO, 2 * SIZE vmaddfp c04, a4, bp1, c04 nop vmaddfp c05, a1, bp2, c05 vmaddfp c06, a2, bp2, c06 vmaddfp c07, a3, bp2, c07 vmaddfp c08, a4, bp2, c08 .align 4 LL(68): vxor VZERO, VZERO, VZERO lvx swap, OFFSET_0, SP lvx neg, OFFSET_1, SP lvx alpha_r, OFFSET_2, SP lvx alpha_i, OFFSET_3, SP vperm c05, c05, c05, swap vperm c06, c06, c06, swap vperm c07, c07, c07, swap vperm c08, c08, c08, swap vxor c05, c05, neg vxor c06, c06, neg vxor c07, c07, neg vxor c08, c08, neg vaddfp c01, c01, c05 vaddfp c02, c02, c06 vaddfp c03, c03, c07 vaddfp c04, c04, c08 vperm c05, c01, c01, swap vperm c06, c02, c02, swap vperm c07, c03, c03, swap vperm c08, c04, c04, swap vmaddfp c01, alpha_r, c01, VZERO vmaddfp c02, alpha_r, c02, VZERO vmaddfp c03, alpha_r, c03, VZERO vmaddfp c04, alpha_r, c04, VZERO vmaddfp c01, alpha_i, c05, c01 vmaddfp c02, alpha_i, c06, c02 vmaddfp c03, alpha_i, c07, c03 vmaddfp c04, alpha_i, c08, c04 lvx C1, OFFSET_0, CO1 lvx C2, OFFSET_1, CO1 lvx C3, OFFSET_2, CO1 lvx C4, OFFSET_3, CO1 lvx C5, OFFSET_4, CO1 lvsr PERMRSHIFT1, 0, CO1 vperm c00, VZERO, c01, PERMRSHIFT1 vperm c01, c01, c02, PERMRSHIFT1 vperm c02, c02, c03, PERMRSHIFT1 vperm c03, c03, c04, PERMRSHIFT1 vperm c04, c04, VZERO, PERMRSHIFT1 vaddfp c00, c00, C1 vaddfp c01, c01, C2 vaddfp c02, c02, C3 vaddfp c03, c03, C4 vaddfp c04, c04, C5 stvx c00, OFFSET_0, CO1 stvx c01, OFFSET_1, CO1 stvx c02, OFFSET_2, CO1 stvx c03, OFFSET_3, CO1 stvx c04, OFFSET_4, CO1 addi CO1, CO1, 16 * SIZE addic. I, I, -1 bgt+ LL(61) .align 4 LL(70): andi. I, M, 4 ble LL(80) vxor c01, c01, c01 LOAD_B b1, OFFSET_0, B vxor c02, c02, c02 vxor c03, c03, c03 LOAD_A a1, OFFSET_0, AO vxor c04, c04, c04 LOAD_A a2, OFFSET_1, AO vxor c05, c05, c05 LOAD_A a3, OFFSET_2, AO vxor c06, c06, c06 LOAD_A a4, OFFSET_3, AO vxor c07, c07, c07 vxor c08, c08, c08 mr BO, B vspltw bp1, b1, 0 srawi. r0, K, 1 mtspr CTR, r0 ble LL(75) .align 4 LL(72): vmaddfp c01, a1, bp1, c01 vspltw bp2, b1, 1 vmaddfp c02, a2, bp1, c02 vmaddfp c05, a1, bp2, c05 vspltw bp1, b1, 2 vmaddfp c06, a2, bp2, c06 vmaddfp c03, a3, bp1, c03 vspltw bp2, b1, 3 vmaddfp c04, a4, bp1, c04 LOAD_B b1, OFFSET_1, BO vspltw bp1, b1, 0 vmaddfp c07, a3, bp2, c07 vmaddfp c08, a4, bp2, c08 addi AO, AO, 16 * SIZE addi BO, BO, 4 * SIZE LOAD_A a1, OFFSET_0, AO LOAD_A a2, OFFSET_1, AO LOAD_A a3, OFFSET_2, AO LOAD_A a4, OFFSET_3, AO bdnz LL(72) .align 4 LL(75): andi. r0, K, 1 ble+ LL(78) .align 4 LL(76): vmaddfp c01, a1, bp1, c01 vspltw bp2, b1, 1 vmaddfp c02, a2, bp1, c02 addi AO, AO, 8 * SIZE vmaddfp c05, a1, bp2, c05 addi BO, BO, 2 * SIZE vmaddfp c06, a2, bp2, c06 .align 4 LL(78): vaddfp c01, c01, c03 vaddfp c02, c02, c04 vaddfp c05, c05, c07 vaddfp c06, c06, c08 vxor VZERO, VZERO, VZERO lvx swap, OFFSET_0, SP lvx neg, OFFSET_1, SP lvx alpha_r, OFFSET_2, SP lvx alpha_i, OFFSET_3, SP vperm c05, c05, c05, swap vperm c06, c06, c06, swap vxor c05, c05, neg vxor c06, c06, neg vaddfp c01, c01, c05 vaddfp c02, c02, c06 vperm c05, c01, c01, swap vperm c06, c02, c02, swap vmaddfp c01, alpha_r, c01, VZERO vmaddfp c02, alpha_r, c02, VZERO vmaddfp c01, alpha_i, c05, c01 vmaddfp c02, alpha_i, c06, c02 lvx C1, OFFSET_0, CO1 lvx C2, OFFSET_1, CO1 lvx C3, OFFSET_2, CO1 lvsr PERMRSHIFT1, 0, CO1 vperm c00, VZERO, c01, PERMRSHIFT1 vperm c01, c01, c02, PERMRSHIFT1 vperm c02, c02, VZERO, PERMRSHIFT1 vaddfp c00, c00, C1 vaddfp c01, c01, C2 vaddfp c02, c02, C3 stvx c00, OFFSET_0, CO1 stvx c01, OFFSET_1, CO1 stvx c02, OFFSET_2, CO1 addi CO1, CO1, 8 * SIZE .align 4 LL(80): andi. I, M, 2 ble LL(90) vxor c01, c01, c01 LOAD_B b1, OFFSET_0, B vxor c02, c02, c02 LOAD_A a1, OFFSET_0, AO LOAD_A a2, OFFSET_1, AO vxor c05, c05, c05 vxor c06, c06, c06 mr BO, B vspltw bp1, b1, 0 srawi. r0, K, 1 mtspr CTR, r0 ble LL(85) .align 4 LL(82): vmaddfp c01, a1, bp1, c01 vspltw bp2, b1, 1 vmaddfp c05, a1, bp2, c05 vspltw bp1, b1, 2 vmaddfp c02, a2, bp1, c02 vspltw bp2, b1, 3 LOAD_B b1, OFFSET_1, BO vspltw bp1, b1, 0 vmaddfp c06, a2, bp2, c06 addi AO, AO, 8 * SIZE addi BO, BO, 4 * SIZE LOAD_A a1, OFFSET_0, AO LOAD_A a2, OFFSET_1, AO bdnz LL(82) .align 4 LL(85): andi. r0, K, 1 ble+ LL(88) .align 4 LL(86): vspltw bp2, b1, 1 vmaddfp c01, a1, bp1, c01 vmaddfp c05, a1, bp2, c05 addi AO, AO, 4 * SIZE addi BO, BO, 2 * SIZE .align 4 LL(88): vaddfp c01, c01, c02 vaddfp c05, c05, c06 vaddfp c09, c09, c10 vaddfp c13, c13, c14 vxor VZERO, VZERO, VZERO lvx swap, OFFSET_0, SP lvx neg, OFFSET_1, SP lvx alpha_r, OFFSET_2, SP lvx alpha_i, OFFSET_3, SP vperm c05, c05, c05, swap vxor c05, c05, neg vaddfp c01, c01, c05 vperm c05, c01, c01, swap vmaddfp c01, alpha_r, c01, VZERO vmaddfp c01, alpha_i, c05, c01 lvx C1, OFFSET_0, CO1 lvx C2, OFFSET_1, CO1 lvsr PERMRSHIFT1, 0, CO1 vperm c00, VZERO, c01, PERMRSHIFT1 vperm c01, c01, VZERO, PERMRSHIFT1 vaddfp c00, c00, C1 vaddfp c01, c01, C2 stvx c00, OFFSET_0, CO1 stvx c01, OFFSET_1, CO1 addi CO1, CO1, 4 * SIZE .align 4 LL(90): andi. I, M, 1 ble LL(999) mr BO, B LFD f8, 0 * SIZE(AO) LFD f9, 1 * SIZE(AO) LFD f10, 0 * SIZE(BO) LFD f11, 1 * SIZE(BO) LFD f12, 2 * SIZE(BO) LFD f13, 3 * SIZE(BO) lfs f0, FZERO(SP) fmr f1, f0 fmr f2, f0 fmr f3, f0 srawi. r0, K, 1 mtspr CTR, r0 ble LL(95) .align 4 LL(92): fmadd f0, f8, f10, f0 fmadd f2, f8, f11, f2 fmadd f1, f9, f10, f1 fmadd f3, f9, f11, f3 LFD f8, 2 * SIZE(AO) LFD f9, 3 * SIZE(AO) LFD f10, 4 * SIZE(BO) LFD f11, 5 * SIZE(BO) fmadd f0, f8, f12, f0 fmadd f2, f8, f13, f2 fmadd f1, f9, f12, f1 fmadd f3, f9, f13, f3 LFD f8, 4 * SIZE(AO) LFD f9, 5 * SIZE(AO) LFD f12, 6 * SIZE(BO) LFD f13, 7 * SIZE(BO) addi AO, AO, 4 * SIZE addi BO, BO, 4 * SIZE bdnz LL(92) .align 4 LL(95): andi. r0, K, 1 ble LL(98) .align 4 LL(96): fmadd f0, f8, f10, f0 fmadd f2, f8, f11, f2 fmadd f1, f9, f10, f1 fmadd f3, f9, f11, f3 .align 4 LL(98): #if defined(NN) || defined(NT) || defined(TN) || defined(TT) fsub f0, f0, f3 fadd f1, f1, f2 #elif defined(NR) || defined(NC) || defined(TR) || defined(TC) fadd f0, f0, f3 fsub f1, f1, f2 #elif defined(RN) || defined(RT) || defined(CN) || defined(CT) fadd f0, f0, f3 fsub f1, f2, f1 #else /* RR, RC, CR, CC */ fsub f0, f0, f3 fadd f1, f1, f2 #endif LFD f8, 0 * SIZE(CO1) LFD f9, 1 * SIZE(CO1) lfs f12, ALPHA_R + 0(SP) lfs f13, ALPHA_I + 4(SP) #if defined(RR) || defined(RC) || defined(CR) || defined(CC) fmadd f8, f12, f0, f8 fnmsub f9, f12, f1, f9 fmadd f8, f13, f1, f8 fmadd f9, f13, f0, f9 #else fmadd f8, f12, f0, f8 fmadd f9, f12, f1, f9 fnmsub f8, f13, f1, f8 fmadd f9, f13, f0, f9 #endif STFD f8, 0 * SIZE(CO1) STFD f9, 1 * SIZE(CO1) .align 4 LL(999): mr SP, STACK li r0, 0 * 16 lvx v20, SP, r0 li r0, 1 * 16 lvx v21, SP, r0 li r0, 2 * 16 lvx v22, SP, r0 li r0, 3 * 16 lvx v23, SP, r0 li r0, 4 * 16 lvx v24, SP, r0 li r0, 5 * 16 lvx v25, SP, r0 li r0, 6 * 16 lvx v26, SP, r0 li r0, 7 * 16 lvx v27, SP, r0 li r0, 8 * 16 lvx v28, SP, r0 li r0, 9 * 16 lvx v29, SP, r0 li r0, 10 * 16 lvx v30, SP, r0 li r0, 11 * 16 lvx v31, SP, r0 mtspr VRsave, VREG #ifdef __64BIT__ ld r31, 192(SP) ld r30, 200(SP) ld r29, 208(SP) ld r28, 216(SP) ld r27, 224(SP) ld r26, 232(SP) ld r25, 240(SP) ld r24, 248(SP) ld r23, 256(SP) ld r22, 264(SP) ld r21, 272(SP) ld r20, 280(SP) ld r19, 288(SP) ld r18, 296(SP) ld r17, 304(SP) ld r16, 312(SP) ld r15, 320(SP) ld r14, 328(SP) #else lwz r31, 192(SP) lwz r30, 196(SP) lwz r29, 200(SP) lwz r28, 204(SP) lwz r27, 208(SP) lwz r26, 212(SP) lwz r25, 216(SP) lwz r24, 220(SP) lwz r23, 224(SP) lwz r22, 228(SP) lwz r21, 232(SP) lwz r20, 236(SP) lwz r19, 240(SP) lwz r18, 244(SP) lwz r17, 248(SP) lwz r16, 252(SP) lwz r15, 256(SP) lwz r14, 260(SP) #endif addi SP, SP, STACKSIZE blr EPILOGUE #endif