/*---------------------------------------------------------------*/ /*--- begin guest_ppc_helpers.c ---*/ /*---------------------------------------------------------------*/ /* This file is part of Valgrind, a dynamic binary instrumentation framework. Copyright (C) 2004-2017 OpenWorks LLP info@open-works.net This program is free software; you can redistribute it and/or modify it under the terms of the GNU General Public License as published by the Free Software Foundation; either version 2 of the License, or (at your option) any later version. This program is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more details. You should have received a copy of the GNU General Public License along with this program; if not, see . The GNU General Public License is contained in the file COPYING. Neither the names of the U.S. Department of Energy nor the University of California nor the names of its contributors may be used to endorse or promote products derived from this software without prior written permission. */ #include "libvex_basictypes.h" #include "libvex_emnote.h" #include "libvex_guest_ppc32.h" #include "libvex_guest_ppc64.h" #include "libvex_ir.h" #include "libvex.h" #include "main_util.h" #include "main_globals.h" #include "guest_generic_bb_to_IR.h" #include "guest_ppc_defs.h" /* This file contains helper functions for ppc32 and ppc64 guest code. Calls to these functions are generated by the back end. These calls are of course in the host machine code and this file will be compiled to host machine code, so that all makes sense. Only change the signatures of these helper functions very carefully. If you change the signature here, you'll have to change the parameters passed to it in the IR calls constructed by guest-ppc/toIR.c. */ /*---------------------------------------------------------------*/ /*--- Misc integer helpers. ---*/ /*---------------------------------------------------------------*/ /* CALLED FROM GENERATED CODE */ /* DIRTY HELPER (non-referentially-transparent) */ /* Horrible hack. On non-ppc platforms, return 1. */ /* Reads a complete, consistent 64-bit TB value. */ ULong ppcg_dirtyhelper_MFTB ( void ) { # if defined(__powerpc__) ULong res; UInt lo, hi1, hi2; while (1) { __asm__ __volatile__ ("\n" "\tmftbu %0\n" "\tmftb %1\n" "\tmftbu %2\n" : "=r" (hi1), "=r" (lo), "=r" (hi2) ); if (hi1 == hi2) break; } res = ((ULong)hi1) << 32; res |= (ULong)lo; return res; # else return 1ULL; # endif } /* CALLED FROM GENERATED CODE */ /* DIRTY HELPER (non-referentially transparent) */ UInt ppc32g_dirtyhelper_MFSPR_268_269 ( UInt r269 ) { # if defined(__powerpc__) UInt spr; if (r269) { __asm__ __volatile__("mfspr %0,269" : "=b"(spr)); } else { __asm__ __volatile__("mfspr %0,268" : "=b"(spr)); } return spr; # else return 0; # endif } /* CALLED FROM GENERATED CODE */ /* DIRTY HELPER (I'm not really sure what the side effects are) */ UInt ppc32g_dirtyhelper_MFSPR_287 ( void ) { # if defined(__powerpc__) UInt spr; __asm__ __volatile__("mfspr %0,287" : "=b"(spr)); return spr; # else return 0; # endif } /* CALLED FROM GENERATED CODE */ /* DIRTY HELPER (reads guest state, writes guest mem) */ void ppc32g_dirtyhelper_LVS ( VexGuestPPC32State* gst, UInt vD_off, UInt sh, UInt shift_right ) { static UChar ref[32] = { 0x00, 0x01, 0x02, 0x03, 0x04, 0x05, 0x06, 0x07, 0x08, 0x09, 0x0A, 0x0B, 0x0C, 0x0D, 0x0E, 0x0F, 0x10, 0x11, 0x12, 0x13, 0x14, 0x15, 0x16, 0x17, 0x18, 0x19, 0x1A, 0x1B, 0x1C, 0x1D, 0x1E, 0x1F }; U128* pU128_src; U128* pU128_dst; vassert( vD_off <= sizeof(VexGuestPPC32State)-8 ); vassert( sh <= 15 ); vassert( shift_right <= 1 ); if (shift_right) sh = 16-sh; /* else shift left */ pU128_src = (U128*)&ref[sh]; pU128_dst = (U128*)( ((UChar*)gst) + vD_off ); (*pU128_dst)[0] = (*pU128_src)[0]; (*pU128_dst)[1] = (*pU128_src)[1]; (*pU128_dst)[2] = (*pU128_src)[2]; (*pU128_dst)[3] = (*pU128_src)[3]; } /* CALLED FROM GENERATED CODE */ /* DIRTY HELPER (reads guest state, writes guest mem) */ void ppc64g_dirtyhelper_LVS ( VexGuestPPC64State* gst, UInt vD_off, UInt sh, UInt shift_right, UInt endness ) { UChar ref[32]; ULong i; Int k; /* ref[] used to be a static const array, but this doesn't work on ppc64 because VEX doesn't load the TOC pointer for the call here, and so we wind up picking up some totally random other data. (It's a wonder we don't segfault.) So, just to be clear, this "fix" (vex r2073) is really a kludgearound for the fact that VEX's 64-bit ppc code generation doesn't provide a valid TOC pointer for helper function calls. Ick. (Bug 250038) */ for (i = 0; i < 32; i++) ref[i] = i; U128* pU128_src; U128* pU128_dst; vassert( vD_off <= sizeof(VexGuestPPC64State)-8 ); vassert( sh <= 15 ); vassert( shift_right <= 1 ); if (shift_right) sh = 16-sh; /* else shift left */ pU128_src = (U128*)&ref[sh]; pU128_dst = (U128*)( ((UChar*)gst) + vD_off ); if ((0x1 & endness) == 0x0) { /* Little endian */ unsigned char *srcp, *dstp; srcp = (unsigned char *)pU128_src; dstp = (unsigned char *)pU128_dst; for (k = 15; k >= 0; k--, srcp++) dstp[k] = *srcp; } else { (*pU128_dst)[0] = (*pU128_src)[0]; (*pU128_dst)[1] = (*pU128_src)[1]; (*pU128_dst)[2] = (*pU128_src)[2]; (*pU128_dst)[3] = (*pU128_src)[3]; } } /* Helper-function specialiser. */ IRExpr* guest_ppc32_spechelper ( const HChar* function_name, IRExpr** args, IRStmt** precedingStmts, Int n_precedingStmts ) { return NULL; } IRExpr* guest_ppc64_spechelper ( const HChar* function_name, IRExpr** args, IRStmt** precedingStmts, Int n_precedingStmts ) { return NULL; } /* 16-bit floating point number is stored in the lower 16-bits of 32-bit value */ #define I16_EXP_MASK 0x7C00 #define I16_FRACTION_MASK 0x03FF #define I32_EXP_MASK 0x7F800000 #define I32_FRACTION_MASK 0x007FFFFF #define I64_EXP_MASK 0x7FF0000000000000ULL #define I64_FRACTION_MASK 0x000FFFFFFFFFFFFFULL #define V128_EXP_MASK 0x7FFF000000000000ULL #define V128_FRACTION_MASK 0x0000FFFFFFFFFFFFULL /* upper 64-bit fractional mask */ ULong generate_C_FPCC_helper( ULong irType, ULong src_hi, ULong src ) { UInt NaN, inf, zero, norm, dnorm, pos; UInt bit0, bit1, bit2, bit3; UInt sign_bit = 0; ULong exp_mask = 0, exp_part = 0, frac_part = 0; ULong fpcc, c; if ( irType == Ity_I16 ) { frac_part = I16_FRACTION_MASK & src; exp_mask = I16_EXP_MASK; exp_part = exp_mask & src; sign_bit = src >> 15; } else if ( irType == Ity_I32 ) { frac_part = I32_FRACTION_MASK & src; exp_mask = I32_EXP_MASK; exp_part = exp_mask & src; sign_bit = src >> 31; } else if ( irType == Ity_I64 ) { frac_part = I64_FRACTION_MASK & src; exp_mask = I64_EXP_MASK; exp_part = exp_mask & src; sign_bit = src >> 63; } else if ( irType == Ity_F128 ) { /* only care if the frac part is zero or non-zero */ frac_part = (V128_FRACTION_MASK & src_hi) | src; exp_mask = V128_EXP_MASK; exp_part = exp_mask & src_hi; sign_bit = src_hi >> 63; } else { vassert(0); // Unknown value of irType } /* NaN: exponene is all ones, fractional part not zero */ if ((exp_part == exp_mask) && (frac_part != 0)) NaN = 1; else NaN = 0; /* inf: exponent all 1's, fraction part is zero */ if ((exp_part == exp_mask) && (frac_part == 0)) inf = 1; else inf = 0; /* zero: exponent is 0, fraction part is zero */ if ((exp_part == 0) && (frac_part == 0)) zero = 1; else zero = 0; /* norm: exponent is not 0, exponent is not all 1's */ if ((exp_part != 0) && (exp_part != exp_mask)) norm = 1; else norm = 0; /* dnorm: exponent is all 0's, fraction is not 0 */ if ((exp_part == 0) && (frac_part != 0)) dnorm = 1; else dnorm = 0; /* pos: MSB is 1 */ if (sign_bit == 0) pos = 1; else pos = 0; /* calculate FPCC */ /* If the result is NaN then must force bits 1, 2 and 3 to zero * to get correct result. */ bit0 = NaN | inf; bit1 = (!NaN) & zero; bit2 = (!NaN) & ((pos & dnorm) | (pos & norm) | (pos & inf)) & ((!zero) & (!NaN)); bit3 = (!NaN) & (((!pos) & dnorm) |((!pos) & norm) | ((!pos) & inf)) & ((!zero) & (!NaN)); fpcc = (bit3 << 3) | (bit2 << 2) | (bit1 << 1) | bit0; /* calculate C */ c = NaN | ((!pos) & dnorm) | ((!pos) & zero) | (pos & dnorm); /* return C in the upper 32-bits and FPCC in the lower 32 bits */ return (c <<32) | fpcc; } /*---------------------------------------------------------------*/ /*--- Misc BCD clean helpers. ---*/ /*---------------------------------------------------------------*/ /* NOTE, the clean and dirty helpers need to called using the * fnptr_to_fnentry() function wrapper to handle the Big Endian * pointer-to-function ABI and the Little Endian ABI. */ /* This C-helper takes a 128-bit BCD value as two 64-bit pieces. * It checks the string to see if it is a valid 128-bit BCD value. * A valid BCD value has a sign value in bits [3:0] between 0xA * and 0xF inclusive. each of the BCD digits represented as a 4-bit * hex number in bits BCD value[128:4] mut be between 0 and 9 * inclusive. Returns an unsigned 64-bit value if valid. */ ULong is_BCDstring128_helper( ULong Signed, ULong bcd_string_hi, ULong bcd_string_low ) { Int i; ULong valid_bcd, sign_valid = False; ULong digit; UInt sign; if ( Signed == True ) { sign = bcd_string_low & 0xF; if( ( sign >= 0xA ) && ( sign <= 0xF ) ) sign_valid = True; /* Change the sign digit to a zero * so the for loop below works the same * for signed and unsigned BCD stings */ bcd_string_low &= 0xFFFFFFFFFFFFFFF0ULL; } else { sign_valid = True; /* set sign to True so result is only based on the validity of the digits */ } valid_bcd = True; // Assume true to start for( i = 0; i < 32; i++ ) { /* check high and low 64-bit strings in parallel */ digit = bcd_string_low & 0xF; if ( digit > 0x9 ) valid_bcd = False; bcd_string_low = bcd_string_low >> 4; digit = bcd_string_hi & 0xF; if ( digit > 0x9 ) valid_bcd = False; bcd_string_hi = bcd_string_hi >> 4; } return valid_bcd & sign_valid; } /* This clean helper takes a signed 32-bit BCD value and a carry in * and adds 1 to the value of the BCD value. The BCD value is passed * in as a single 64-bit value. The incremented value is returned in * the lower 32 bits of the result. If the input was signed the sign of * the result is the same as the input. The carry out is returned in * bits [35:32] of the result. */ ULong increment_BCDstring32_helper( ULong Signed, ULong bcd_string, ULong carry_in ) { UInt i, num_digits = 8; ULong bcd_value, result = 0; ULong carry, digit, new_digit; carry = carry_in; if ( Signed == True ) { bcd_value = bcd_string >> 4; /* remove sign */ num_digits = num_digits - 1; } else { bcd_value = bcd_string; } for( i = 0; i < num_digits; i++ ) { digit = bcd_value & 0xF; bcd_value = bcd_value >> 4; new_digit = digit + carry; if ( new_digit > 10 ) { carry = 1; new_digit = new_digit - 10; } else { carry = 0; } result = result | (new_digit << (i*4) ); } if ( Signed == True ) { result = ( carry << 32) | ( result << 4 ) | ( bcd_string & 0xF ); } else { result = ( carry << 32) | result; } return result; } /*---------------------------------------------------------------*/ /*--- Misc packed decimal clean helpers. ---*/ /*---------------------------------------------------------------*/ /* This C-helper takes a 64-bit packed decimal value stored in a * 64-bit value. It converts the zoned decimal format. The lower * byte may contain a sign value, set it to zero. If return_upper * is zero, return lower 64 bits of result, otherwise return upper * 64 bits of the result. */ ULong convert_to_zoned_helper( ULong src_hi, ULong src_low, ULong upper_byte, ULong return_upper ) { UInt i, sh; ULong tmp = 0, new_value; /* Remove the sign from the source. Put in the upper byte of result. * Sign inserted later. */ if ( return_upper == 0 ) { /* return lower 64-bit result */ for(i = 0; i < 7; i++) { sh = ( 8 - i ) * 4; new_value = ( ( src_low >> sh ) & 0xf ) | upper_byte; tmp = tmp | ( new_value << ( ( 7 - i ) * 8 ) ); } } else { /* Byte for i=0 is in upper 64-bit of the source, do it separately */ new_value = ( src_hi & 0xf ) | upper_byte; tmp = tmp | new_value << 56; for( i = 1; i < 8; i++ ) { sh = ( 16 - i ) * 4; new_value = ( ( src_low >> sh ) & 0xf ) | upper_byte; tmp = tmp | ( new_value << ( ( 7 - i ) * 8 ) ); } } return tmp; } /* This C-helper takes the lower 64-bits of the 128-bit packed decimal * src value. It converts the src value to a 128-bit national format. * If return_upper is zero, the helper returns lower 64 bits of result, * otherwise it returns the upper 64-bits of the result. */ ULong convert_to_national_helper( ULong src, ULong return_upper ) { UInt i; UInt sh = 3, max = 4, min = 0; /* initialize max, min for return upper */ ULong tmp = 0, new_value; if ( return_upper == 0 ) { /* return lower 64-bit result */ min = 4; max = 7; sh = 7; } for( i = min; i < max; i++ ) { new_value = ( ( src >> ( ( 7 - i ) * 4 ) ) & 0xf ) | 0x0030; tmp = tmp | ( new_value << ( ( sh - i ) * 16 ) ); } return tmp; } /* This C-helper takes a 128-bit zoned value stored in a 128-bit * value. It converts it to the packed 64-bit decimal format without a * a sign value. The sign is supposed to be in bits [3:0] and the packed * value in bits [67:4]. This helper leaves it to the caller to put the * result into a V128 and shift the returned value over and put the sign * in. */ ULong convert_from_zoned_helper( ULong src_hi, ULong src_low ) { UInt i; ULong tmp = 0, nibble; /* Unroll the i = 0 iteration so the sizes of the loop for the upper * and lower extraction match. Skip sign in lease significant byte. */ nibble = ( src_hi >> 56 ) & 0xF; tmp = tmp | ( nibble << 60 ); for( i = 1; i < 8; i++ ) { /* get the high nibbles, put into result */ nibble = ( src_hi >> ( ( 7 - i ) * 8 ) ) & 0xF; tmp = tmp | ( nibble << ( ( 15 - i ) * 4 ) ); /* get the low nibbles, put into result */ nibble = ( src_low >> ( ( 8 - i ) * 8 ) ) & 0xF; tmp = tmp | ( nibble << ( ( 8 - i ) * 4 ) ); } return tmp; } /* This C-helper takes a 128-bit national value stored in a 128-bit * value. It converts it to a signless packed 64-bit decimal format. */ ULong convert_from_national_helper( ULong src_hi, ULong src_low ) { UInt i; ULong tmp = 0, hword; src_low = src_low & 0xFFFFFFFFFFFFFFF0ULL; /* remove the sign */ for( i = 0; i < 4; i++ ) { /* get the high half-word, put into result */ hword = ( src_hi >> ( ( 3 - i ) * 16 ) ) & 0xF; tmp = tmp | ( hword << ( ( 7 - i ) * 4 ) ); /* get the low half-word, put into result */ hword = ( src_low >> ( ( 3 - i ) * 16 ) ) & 0xF; tmp = tmp | ( hword << ( ( 3 - i ) * 4 ) ); } return tmp; } /*------------------------------------------------*/ /*--- Population count ---------------------------*/ /*------------------------------------------------*/ ULong population_count64_helper( ULong src ) { /* Fast population count based on algorithm in the "Hacker's Delight" by Henery S. Warren. */ src = (src & 0x5555555555555555) + ((src >> 1) & 0x5555555555555555); src = (src & 0x3333333333333333) + ((src >> 2) & 0x3333333333333333); src = (src & 0x0F0F0F0F0F0F0F0F) + ((src >> 4) & 0x0F0F0F0F0F0F0F0F); src = (src & 0x00FF00FF00FF00FF) + ((src >> 8) & 0x00FF00FF00FF00FF); src = (src & 0x0000FFFF0000FFFF) + ((src >> 16) & 0x0000FFFF0000FFFF); src = (src & 0x00000000FFFFFFFF) + ((src >> 32) & 0x00000000FFFFFFFF); return src & 0x3F; } /*------------------------------------------------*/ /*---- Extract/Deposit bits under mask helpers ---*/ /*------------------------------------------------*/ ULong extract_bits_under_mask_helper( ULong src, ULong mask, UInt flag ) { UInt i; ULong ones, zeros, mask_bit, bit_src; zeros = 0; ones = 0; for (i=0; i<64; i++){ mask_bit = 0x1 & (mask >> (63-i)); bit_src = 0x1 & (src >> (63-i)); ones = ones << mask_bit; ones = ones | (mask_bit & bit_src); zeros = zeros << (1^mask_bit); zeros = zeros | ((1^mask_bit) & bit_src); } if (flag == 1) return ones; else return zeros; } UInt count_bits_under_mask_helper( ULong src, ULong mask, UInt flag ) { UInt i, count_extracted_1, count_extracted_0;; ULong mask_bit; count_extracted_1 = 0; count_extracted_0 = 0; for (i=0; i<64; i++){ mask_bit = 0x1 & (mask >> (63-i)); if (mask_bit == 1) count_extracted_1++; if ((1^mask_bit) == 1) count_extracted_0++; } if (flag == 1) return count_extracted_1; else return count_extracted_0; } ULong deposit_bits_under_mask_helper( ULong src, ULong mask ) { UInt i, src_bit_pos; ULong result, mask_bit, bit_src; result = 0; src_bit_pos = 0; for (i=0; i<64; i++){ mask_bit = 0x1 & (mask >> i); if (mask_bit == 1) { bit_src = 0x1 & (src >> src_bit_pos); result = result | (bit_src << i); src_bit_pos++; } } return result; } /*----------------------------------------------*/ /*--- Vector Evaluate Inst helper --------------*/ /*----------------------------------------------*/ /* This is a 64-bit version of the VXS Vector Evaluate instruction xxeval. */ ULong vector_evaluate64_helper( ULong srcA, ULong srcB, ULong srcC, ULong IMM ) { #define MAX_BITS 64 #define MAX_IMM_BITS 8 UInt i, select; ULong bitA, bitB, bitC, result; ULong bitIMM; result = 0; for (i=0; i> i); bitB = 0x1 & (srcB >> i); bitC = 0x1 & (srcC >> i); /* The value of select is IBM numbering based, i.e. MSB is bit 0 */ select = (bitA << 2) | (bitB << 1) | bitC; bitIMM = (IMM >> (MAX_IMM_BITS - 1 - select)) & 0x1; result = result | (bitIMM << i); } return result; #undef MAX_BITS #undef MAX_IMM_BITS } /*------------------------------------------------*/ /*---- VSX Matrix signed integer GER functions ---*/ /*------------------------------------------------*/ static UInt exts4( UInt src) { /* Input is an 4-bit value. Extend bit 3 to bits [31:4] */ if (( src >> 3 ) & 0x1) return src | 0xFFFFFFF0; /* sign bit is a 1, extend */ else return src & 0xF; /* make sure high order bits are zero */ } static UInt exts8( UInt src) { /* Input is an 8-bit value. Extend bit 7 to bits [31:8] */ if (( src >> 7 ) & 0x1) return src | 0xFFFFFF00; /* sign bit is a 1, extend */ else return src & 0xFF; /* make sure high order bits are zero */ } static UInt extz8( UInt src) { /* Input is an 8-bit value. Extend src on the left with zeros. */ return src & 0xFF; /* make sure high order bits are zero */ } static ULong exts16to64( UInt src) { /* Input is an 16-bit value. Extend bit 15 to bits [63:16] */ if (( src >> 15 ) & 0x1) return ((ULong) src) | 0xFFFFFFFFFFFF0000ULL; /* sign is 1, extend */ else /* make sure high order bits are zero */ return ((ULong) src) & 0xFFFFULL; } static UInt chop64to32( Long src ) { /* Take a 64-bit input, return the lower 32-bits */ return (UInt)(0xFFFFFFFF & src); } static UInt clampS64toS32( Long src ) { /* Take a 64-bit signed input, clamp positive values to 2^31, clamp negative values at -2^31. Return the result in an unsigned 32-bit value. */ Long max_val = 2147483647; // 2^31-1 if ( src > max_val) return (UInt)max_val; if (src < -max_val) return (UInt)-max_val; return (UInt)src; } void write_ACC_entry (VexGuestPPC64State* gst, UInt offset, UInt acc, UInt reg, UInt *acc_word) { U128* pU128_dst; vassert( (acc >= 0) && (acc < 8) ); vassert( (reg >= 0) && (reg < 4) ); pU128_dst = (U128*) (((UChar*)gst) + offset + acc*4*sizeof(U128) + reg*sizeof(U128)); /* The U128 type is defined as an array of unsigned intetgers. */ (*pU128_dst)[0] = acc_word[0]; (*pU128_dst)[1] = acc_word[1]; (*pU128_dst)[2] = acc_word[2]; (*pU128_dst)[3] = acc_word[3]; return; } void get_ACC_entry (VexGuestPPC64State* gst, UInt offset, UInt acc, UInt reg, UInt *acc_word) { U128* pU128_src; acc_word[3] = 0xDEAD; acc_word[2] = 0xBEEF; acc_word[1] = 0xBAD; acc_word[0] = 0xBEEF; vassert( (acc >= 0) && (acc < 8) ); vassert( (reg >= 0) && (reg < 4) ); pU128_src = (U128*) (((UChar*)gst) + offset + acc*4*sizeof(U128) + reg*sizeof(U128)); /* The U128 type is defined as an array of unsigned intetgers. */ acc_word[0] = (*pU128_src)[0]; acc_word[1] = (*pU128_src)[1]; acc_word[2] = (*pU128_src)[2]; acc_word[3] = (*pU128_src)[3]; return; } void vsx_matrix_4bit_ger_dirty_helper ( VexGuestPPC64State* gst, UInt offset_ACC, ULong srcA_hi, ULong srcA_lo, ULong srcB_hi, ULong srcB_lo, UInt masks_inst ) { /* This helper calculates the result for one of the four ACC entires. It is called twice, to get the hi and then the low 64-bit of the 128-bit result. */ UInt i, j, mask, sum, inst, acc_entry, prefix_inst; UInt srcA_nibbles[4][8]; /* word, nibble */ UInt srcB_nibbles[4][8]; /* word, nibble */ UInt acc_word[4]; UInt prod0, prod1, prod2, prod3, prod4, prod5, prod6, prod7; UInt result[4]; UInt pmsk = 0; UInt xmsk = 0; UInt ymsk = 0; mask = 0xF; inst = (masks_inst >> 5) & 0xFF; prefix_inst = (masks_inst >> 13) & 0x1; acc_entry = masks_inst & 0xF; /* LE word numbering */ if ( prefix_inst == 0 ) { /* Set the masks for non-prefix instructions */ pmsk = 0b11111111; xmsk = 0b1111; ymsk = 0b1111; } else { pmsk = (masks_inst >> 22) & 0xFF; xmsk = (masks_inst >> 18) & 0xF; ymsk = (masks_inst >> 14) & 0xF; } /* Address nibbles using IBM numbering */ for( i = 0; i < 4; i++) { /* Get the ACC contents directly from the PPC64 state */ get_ACC_entry (gst, offset_ACC, acc_entry, 3-i, acc_word); // input is in double words for( j = 0; j< 8; j++) { srcA_nibbles[3][j] = (srcA_hi >> (60-4*j)) & mask; // hi bits [63:32] srcA_nibbles[2][j] = (srcA_hi >> (28-4*j)) & mask; // hi bits [31:0] srcA_nibbles[1][j] = (srcA_lo >> (60-4*j)) & mask; // lo bits [63:32] srcA_nibbles[0][j] = (srcA_lo >> (28-4*j)) & mask; // lo bits [31:0] srcB_nibbles[3][j] = (srcB_hi >> (60-4*j)) & mask; srcB_nibbles[2][j] = (srcB_hi >> (28-4*j)) & mask; srcB_nibbles[1][j] = (srcB_lo >> (60-4*j)) & mask; srcB_nibbles[0][j] = (srcB_lo >> (28-4*j)) & mask; } for( j = 0; j < 4; j++) { if (((xmsk >> i) & 0x1) & ((ymsk >> j) & 0x1)) { if (((pmsk >> 7) & 0x1) == 0) prod0 = 0; else prod0 = exts4( srcA_nibbles[i][0] ) * exts4( srcB_nibbles[j][0] ); if (((pmsk >> 6) & 0x1) == 0) prod1 = 0; else prod1 = exts4( srcA_nibbles[i][1] ) * exts4( srcB_nibbles[j][1] ); if (((pmsk >> 5) & 0x1) == 0) prod2 = 0; else prod2 = exts4( srcA_nibbles[i][2] ) * exts4( srcB_nibbles[j][2] ); if (((pmsk >> 4) & 0x1) == 0) prod3 = 0; else prod3 = exts4( srcA_nibbles[i][3] ) * exts4( srcB_nibbles[j][3] ); if (((pmsk >> 3) & 0x1) == 0) prod4 = 0; else prod4 = exts4( srcA_nibbles[i][4] ) * exts4( srcB_nibbles[j][4] ); if (((pmsk >> 2) & 0x1) == 0) prod5 = 0; else prod5 = exts4( srcA_nibbles[i][5] ) * exts4( srcB_nibbles[j][5] ); if (((pmsk >> 1) & 0x1) == 0) prod6 = 0; else prod6 = exts4( srcA_nibbles[i][6] ) * exts4( srcB_nibbles[j][6] ); if ((pmsk & 0x1) == 0) prod7 = 0; else prod7 = exts4( srcA_nibbles[i][7] ) * exts4( srcB_nibbles[j][7] ); /* sum is UInt so the result is choped to 32-bits */ sum = prod0 + prod1 + prod2 + prod3 + prod4 + prod5 + prod6 + prod7; if ( inst == XVI4GER8 ) result[j] = sum; else if ( inst == XVI4GER8PP ) result[j] = sum + acc_word[j]; } else { result[j] = 0; } } write_ACC_entry (gst, offset_ACC, acc_entry, 3-i, result); } } void vsx_matrix_8bit_ger_dirty_helper( VexGuestPPC64State* gst, UInt offset_ACC, ULong srcA_hi, ULong srcA_lo, ULong srcB_hi, ULong srcB_lo, UInt masks_inst ) { UInt i, j, mask, sum, inst, acc_entry, prefix_inst; UInt srcA_bytes[4][4]; /* word, byte */ UInt srcB_bytes[4][4]; /* word, byte */ UInt acc_word[4]; UInt prod0, prod1, prod2, prod3; UInt result[4]; UInt pmsk = 0; UInt xmsk = 0; UInt ymsk = 0; mask = 0xFF; inst = (masks_inst >> 5) & 0xFF; prefix_inst = (masks_inst >> 13) & 0x1; acc_entry = masks_inst & 0xF; /* LE word numbering */ if ( prefix_inst == 0 ) { /* Set the masks */ pmsk = 0b1111; xmsk = 0b1111; ymsk = 0b1111; } else { pmsk = (masks_inst >> 26) & 0xF; xmsk = (masks_inst >> 18) & 0xF; ymsk = (masks_inst >> 14) & 0xF; } /* Address byes using IBM numbering */ for( i = 0; i < 4; i++) { /* Get the ACC contents directly from the PPC64 state */ get_ACC_entry (gst, offset_ACC, acc_entry, 3-i, acc_word); for( j = 0; j< 4; j++) { srcA_bytes[3][j] = (srcA_hi >> (56-8*j)) & mask; srcA_bytes[2][j] = (srcA_hi >> (24-8*j)) & mask; srcA_bytes[1][j] = (srcA_lo >> (56-8*j)) & mask; srcA_bytes[0][j] = (srcA_lo >> (24-8*j)) & mask; srcB_bytes[3][j] = (srcB_hi >> (56-8*j)) & mask; srcB_bytes[2][j] = (srcB_hi >> (24-8*j)) & mask; srcB_bytes[1][j] = (srcB_lo >> (56-8*j)) & mask; srcB_bytes[0][j] = (srcB_lo >> (24-8*j)) & mask; } for( j = 0; j < 4; j++) { if (((xmsk >> i) & 0x1) & ((ymsk >> j) & 0x1)) { if (((pmsk >> 3) & 0x1) == 0) prod0 = 0; else prod0 = exts8( srcA_bytes[i][0] ) * extz8( srcB_bytes[j][0] ); if (((pmsk >> 2) & 0x1) == 0) prod1 = 0; else prod1 = exts8( srcA_bytes[i][1] ) * extz8( srcB_bytes[j][1] ); if (((pmsk >> 1) & 0x1) == 0) prod2 = 0; else prod2 = exts8( srcA_bytes[i][2] ) * extz8( srcB_bytes[j][2] ); if (((pmsk >> 0) & 0x1) == 0) prod3 = 0; else prod3 = exts8( srcA_bytes[i][3] ) * extz8( srcB_bytes[j][3] ); /* sum is UInt so the result is choped to 32-bits */ sum = prod0 + prod1 + prod2 + prod3; if ( inst == XVI8GER4 ) result[j] = sum; else if ( inst == XVI8GER4PP ) result[j] = sum + acc_word[j]; } else { result[j] = 0; } } write_ACC_entry (gst, offset_ACC, acc_entry, 3-i, result); } } void vsx_matrix_16bit_ger_dirty_helper( VexGuestPPC64State* gst, UInt offset_ACC, ULong srcA_hi, ULong srcA_lo, ULong srcB_hi, ULong srcB_lo, UInt masks_inst ) { UInt i, j, mask, inst, acc_entry, prefix_inst; ULong sum; UInt srcA_word[4][2]; /* word, hword */ UInt srcB_word[4][2]; /* word, hword */ UInt acc_word[4]; ULong prod0, prod1; UInt result[4]; UInt pmsk = 0; UInt xmsk = 0; UInt ymsk = 0; mask = 0xFFFF; inst = (masks_inst >> 5) & 0xFF; prefix_inst = (masks_inst >> 13) & 0x1; acc_entry = masks_inst & 0xF; /* LE word numbering */ if ( prefix_inst == 0 ) { /* Set the masks for non prefix instructions */ pmsk = 0b11; xmsk = 0b1111; ymsk = 0b1111; } else { pmsk = (masks_inst >> 28) & 0x3; xmsk = (masks_inst >> 18) & 0xF; ymsk = (masks_inst >> 14) & 0xF; } /* Address half-words using IBM numbering */ for( i = 0; i < 4; i++) { /* Get the ACC contents directly from the PPC64 state */ get_ACC_entry (gst, offset_ACC, acc_entry, 3-i, acc_word); for( j = 0; j< 2; j++) { srcA_word[3][j] = (srcA_hi >> (48-16*j)) & mask; srcA_word[2][j] = (srcA_hi >> (16-16*j)) & mask; srcA_word[1][j] = (srcA_lo >> (48-16*j)) & mask; srcA_word[0][j] = (srcA_lo >> (16-16*j)) & mask; srcB_word[3][j] = (srcB_hi >> (48-16*j)) & mask; srcB_word[2][j] = (srcB_hi >> (16-16*j)) & mask; srcB_word[1][j] = (srcB_lo >> (48-16*j)) & mask; srcB_word[0][j] = (srcB_lo >> (16-16*j)) & mask; } for( j = 0; j < 4; j++) { if (((xmsk >> i) & 0x1) & ((ymsk >> j) & 0x1)) { if (((pmsk >> 1) & 0x1) == 0) prod0 = 0; else prod0 = exts16to64( srcA_word[i][0] ) * exts16to64( srcB_word[j][0] ); if (((pmsk >> 0) & 0x1) == 0) prod1 = 0; else prod1 = exts16to64( srcA_word[i][1] ) * exts16to64( srcB_word[j][1] ); /* sum is UInt so the result is choped to 32-bits */ sum = prod0 + prod1; if ( inst == XVI16GER2 ) result[j] = chop64to32( sum ); else if ( inst == XVI16GER2S ) result[j] = clampS64toS32( sum ); else if ( inst == XVI16GER2PP ) { result[j] = chop64to32( sum + acc_word[j] ); } else if ( inst == XVI16GER2SPP ) { result[j] = clampS64toS32( sum + acc_word[j] ); } } else { result[j] = 0; } } write_ACC_entry (gst, offset_ACC, acc_entry, 3-i, result); } } //matrix 16 float stuff union convert_t { UInt u32; ULong u64; Float f; Double d; }; static Float reinterpret_int_as_float( UInt input ) { /* Reinterpret the bit pattern of an int as a float. */ __attribute__ ((aligned (128))) union convert_t conv; conv.u32 = input; return conv.f; } static UInt reinterpret_float_as_int( Float input ) { /* Reinterpret the bit pattern of an int as a float. */ __attribute__ ((aligned (128))) union convert_t conv; conv.f = input; return conv.u32; } static Double reinterpret_long_as_double( ULong input ) { /* Reinterpret the bit pattern of an int as a float. */ __attribute__ ((aligned (128))) union convert_t conv; conv.u64 = input; return conv.d; } static ULong reinterpret_double_as_long( Double input ) { /* Reinterpret the bit pattern of an int as a float. */ __attribute__ ((aligned (128))) union convert_t conv; conv.d = input; return conv.u64; } static Double conv_f16_to_double( ULong input ) { # if defined (HAS_XSCVHPDP) // This all seems to be very alignment sensitive?? __attribute__ ((aligned (64))) ULong src; __attribute__ ((aligned (64))) Double result; src = input; __asm__ __volatile__ ("xscvhpdp %x0,%x1" : "=wa" (result) : "wa" (src)); return result; # else return 0.0; # endif } static Float conv_double_to_float( Double src ) { return (float) src ; } static Double negate_double( Double input ) { /* Don't negate a NaN value. A NaN has an exponet of all 1's, non zero fraction. */ __attribute__ ((aligned (128))) union convert_t conv; conv.d = input; if ( ( ( conv.u64 & I64_EXP_MASK) == I64_EXP_MASK ) && ( ( conv.u64 & I64_FRACTION_MASK ) != 0 ) ) return input; else return -input; } static Float negate_float( Float input ) { /* Don't negate a NaN value. A NaN has an exponet of all 1's, non zero fraction. */ __attribute__ ((aligned (128))) union convert_t conv; conv.f = input; if ( ( ( conv.u32 & I32_EXP_MASK) == I32_EXP_MASK ) && ( ( conv.u32 & I32_FRACTION_MASK ) != 0 ) ) return input; else return -input; } void vsx_matrix_16bit_float_ger_dirty_helper( VexGuestPPC64State* gst, UInt offset_ACC, ULong srcA_hi, ULong srcA_lo, ULong srcB_hi, ULong srcB_lo, UInt masks_inst ) { UInt i, j, mask, inst, acc_entry, prefix_inst; UInt srcA_word[4][2]; /* word, hword */ UInt srcB_word[4][2]; /* word, hword */ Double src10, src11, src20, src21; UInt acc_word_input[4]; Float acc_word[4]; Double prod; Double msum; UInt result[4]; UInt pmsk = 0; UInt xmsk = 0; UInt ymsk = 0; mask = 0xFFFF; inst = (masks_inst >> 5) & 0xFF; prefix_inst = (masks_inst >> 13) & 0x1; acc_entry = masks_inst & 0xF; if ( prefix_inst == 0 ) { /* Set the masks for non-prefix instructions */ pmsk = 0b11; xmsk = 0b1111; ymsk = 0b1111; } else { /* Use mask supplied with prefix inst */ pmsk = (masks_inst >> 28) & 0x3; xmsk = (masks_inst >> 18) & 0xF; ymsk = (masks_inst >> 14) & 0xF; } /* Address half-words using IBM numbering */ for( i = 0; i < 4; i++) { /* Get the ACC contents directly from the PPC64 state */ get_ACC_entry (gst, offset_ACC, acc_entry, 3-i, acc_word_input); acc_word[3] = reinterpret_int_as_float( acc_word_input[3] ); acc_word[2] = reinterpret_int_as_float( acc_word_input[2] ); acc_word[1] = reinterpret_int_as_float( acc_word_input[1] ); acc_word[0] = reinterpret_int_as_float( acc_word_input[0] ); for( j = 0; j < 2; j++) { // input is in double words srcA_word[3][j] = (UInt)((srcA_hi >> (48-16*j)) & mask); srcA_word[2][j] = (UInt)((srcA_hi >> (16-16*j)) & mask); srcA_word[1][j] = (UInt)((srcA_lo >> (48-16*j)) & mask); srcA_word[0][j] = (UInt)((srcA_lo >> (16-16*j)) & mask); srcB_word[3][j] = (UInt)((srcB_hi >> (48-16*j)) & mask); srcB_word[2][j] = (UInt)((srcB_hi >> (16-16*j)) & mask); srcB_word[1][j] = (UInt)((srcB_lo >> (48-16*j)) & mask); srcB_word[0][j] = (UInt)((srcB_lo >> (16-16*j)) & mask); } for( j = 0; j < 4; j++) { if (((pmsk >> 1) & 0x1) == 0) { src10 = 0; src20 = 0; } else { src10 = conv_f16_to_double((ULong)srcA_word[i][0]); src20 = conv_f16_to_double((ULong)srcB_word[j][0]); } if ((pmsk & 0x1) == 0) { src11 = 0; src21 = 0; } else { src11 = conv_f16_to_double((ULong)srcA_word[i][1]); src21 = conv_f16_to_double((ULong)srcB_word[j][1]); } prod = src10 * src20; msum = prod + src11 * src21; if (((xmsk >> i) & 0x1) & ((ymsk >> j) & 0x1)) { /* Note, we do not track the exception handling bits ox, ux, xx, si, mz, vxsnan and vximz in the FPSCR. */ if ( inst == XVF16GER2 ) result[j] = reinterpret_float_as_int( conv_double_to_float(msum) ); else if ( inst == XVF16GER2PP ) result[j] = reinterpret_float_as_int( conv_double_to_float(msum) + acc_word[j] ); else if ( inst == XVF16GER2PN ) result[j] = reinterpret_float_as_int( conv_double_to_float(msum) + negate_float( acc_word[j] ) ); else if ( inst == XVF16GER2NP ) result[j] = reinterpret_float_as_int( conv_double_to_float( negate_double( msum ) ) + acc_word[j] ); else if ( inst == XVF16GER2NN ) result[j] = reinterpret_float_as_int( conv_double_to_float( negate_double( msum ) ) + negate_float( acc_word[j] ) ); } else { result[j] = 0; } } write_ACC_entry (gst, offset_ACC, acc_entry, 3-i, result); } } void vsx_matrix_32bit_float_ger_dirty_helper( VexGuestPPC64State* gst, UInt offset_ACC, ULong srcA_hi, ULong srcA_lo, ULong srcB_hi, ULong srcB_lo, UInt masks_inst ) { UInt i, j, mask, inst, acc_entry, prefix_inst; Float srcA_word[4]; Float srcB_word[4]; UInt acc_word_input[4]; Float acc_word[4]; UInt result[4]; UInt xmsk = 0; UInt ymsk = 0; Float src1, src2, acc; mask = 0xFFFFFFFF; inst = (masks_inst >> 5) & 0xFF; prefix_inst = (masks_inst >> 13) & 0x1; acc_entry = masks_inst & 0xF; if ( prefix_inst == 0 ) { /* Set the masks for non-prefix instructions */ xmsk = 0b1111; ymsk = 0b1111; } else { xmsk = (masks_inst >> 18) & 0xF; ymsk = (masks_inst >> 14) & 0xF; } srcA_word[3] = reinterpret_int_as_float( (srcA_hi >> 32) & mask ); srcA_word[2] = reinterpret_int_as_float( srcA_hi & mask ); srcA_word[1] = reinterpret_int_as_float( (srcA_lo >> 32) & mask ); srcA_word[0] = reinterpret_int_as_float( srcA_lo & mask ); srcB_word[3] = reinterpret_int_as_float( (srcB_hi >> 32) & mask ); srcB_word[2] = reinterpret_int_as_float( srcB_hi & mask ); srcB_word[1] = reinterpret_int_as_float( (srcB_lo >> 32) & mask ); srcB_word[0] = reinterpret_int_as_float( srcB_lo & mask ); /* Address byes using IBM numbering */ for( i = 0; i < 4; i++) { /* Get the ACC contents directly from the PPC64 state */ get_ACC_entry (gst, offset_ACC, acc_entry, 3-i, acc_word_input); acc_word[3] = reinterpret_int_as_float( acc_word_input[3] ); acc_word[2] = reinterpret_int_as_float( acc_word_input[2] ); acc_word[1] = reinterpret_int_as_float( acc_word_input[1] ); acc_word[0] = reinterpret_int_as_float( acc_word_input[0] ); for( j = 0; j < 4; j++) { if ((((xmsk >> i) & 0x1) & ((ymsk >> j) & 0x1)) == 0x1) { /* Note, we do not track the exception handling bits ox, ux, xx, si, mz, vxsnan and vximz in the FPSCR. */ src1 = srcA_word[i]; src2 = srcB_word[j]; acc = acc_word[j]; if ( inst == XVF32GER ) result[j] = reinterpret_float_as_int( src1 * src2 ); else if ( inst == XVF32GERPP ) result[j] = reinterpret_float_as_int( ( src1 * src2 ) + acc ); else if ( inst == XVF32GERPN ) result[j] = reinterpret_float_as_int( ( src1 * src2 ) + negate_float( acc ) ); else if ( inst == XVF32GERNP ) result[j] = reinterpret_float_as_int( negate_float( src1 * src2 ) + acc ); else if ( inst == XVF32GERNN ) result[j] = reinterpret_float_as_int( negate_float( src1 * src2 ) + negate_float( acc ) ); } else { result[j] = 0; } } write_ACC_entry (gst, offset_ACC, acc_entry, 3-i, result); } } void vsx_matrix_64bit_float_ger_dirty_helper( VexGuestPPC64State* gst, UInt offset_ACC, ULong srcX_hi, ULong srcX_lo, ULong srcY_hi, ULong srcY_lo, UInt masks_inst ) { /* This function just computes the result for one entry in the ACC. */ UInt i, j, inst, acc_entry, prefix_inst; Double srcX_dword[4]; Double srcY_dword[2]; Double result[2]; UInt result_uint[4]; ULong result_ulong[2]; Double acc_dword[4]; ULong acc_word_ulong[2]; UInt acc_word_input[4]; UInt xmsk = 0; UInt ymsk = 0; UInt start_i; Double src1, src2, acc; inst = (masks_inst >> 8) & 0xFF; prefix_inst = (masks_inst >> 16) & 0x1; start_i = (masks_inst >> 4) & 0xF; acc_entry = masks_inst & 0xF; if ( prefix_inst == 0 ) { /* Set the masks for non-prefix instructions */ xmsk = 0b1111; ymsk = 0b11; } else { xmsk = (masks_inst >> 21) & 0xF; ymsk = (masks_inst >> 19) & 0x3; } /* Need to store the srcX_dword in the correct index for the following for loop. */ srcX_dword[1+start_i] = reinterpret_long_as_double( srcX_lo); srcX_dword[0+start_i] = reinterpret_long_as_double( srcX_hi ); srcY_dword[1] = reinterpret_long_as_double( srcY_lo ); srcY_dword[0] = reinterpret_long_as_double( srcY_hi ); for( i = start_i; i < start_i+2; i++) { /* Get the ACC contents directly from the PPC64 state */ get_ACC_entry (gst, offset_ACC, acc_entry, 3 - i, acc_word_input); acc_word_ulong[1] = acc_word_input[3]; acc_word_ulong[1] = (acc_word_ulong[1] << 32) | acc_word_input[2]; acc_word_ulong[0] = acc_word_input[1]; acc_word_ulong[0] = (acc_word_ulong[0] << 32) | acc_word_input[0]; acc_dword[0] = reinterpret_long_as_double( acc_word_ulong[0] ); acc_dword[1] = reinterpret_long_as_double( acc_word_ulong[1]); for( j = 0; j < 2; j++) { if (((xmsk >> i) & 0x1) & ((ymsk >> j) & 0x1)) { /* Note, we do not track the exception handling bits ox, ux, xx, si, mz, vxsnan and vximz in the FPSCR. */ src1 = srcX_dword[i]; src2 = srcY_dword[j]; acc = acc_dword[j]; if ( inst == XVF64GER ) result[j] = src1 * src2; else if ( inst == XVF64GERPP ) result[j] = ( src1 * src2 ) + acc; else if ( inst == XVF64GERPN ) result[j] = ( src1 * src2 ) + negate_double( acc ); else if ( inst == XVF64GERNP ) result[j] = negate_double( src1 * src2 ) + acc; else if ( inst == XVF64GERNN ) result[j] = negate_double( src1 * src2 ) + negate_double( acc ); } else { result[j] = 0; } } /* Need to store the two double float values as two unsigned ints in order to store them to the ACC. */ result_ulong[0] = reinterpret_double_as_long ( result[0] ); result_ulong[1] = reinterpret_double_as_long ( result[1] ); result_uint[0] = result_ulong[0] & 0xFFFFFFFF; result_uint[1] = (result_ulong[0] >> 32) & 0xFFFFFFFF; result_uint[2] = result_ulong[1] & 0xFFFFFFFF; result_uint[3] = (result_ulong[1] >> 32) & 0xFFFFFFFF; write_ACC_entry (gst, offset_ACC, acc_entry, 3 - i, result_uint); } } /*----------------------------------------------*/ /*--- The exported fns .. ---*/ /*----------------------------------------------*/ /* VISIBLE TO LIBVEX CLIENT */ UInt LibVEX_GuestPPC32_get_CR ( /*IN*/const VexGuestPPC32State* vex_state ) { # define FIELD(_n) \ ( ( (UInt) \ ( (vex_state->guest_CR##_n##_321 & (7<<1)) \ | (vex_state->guest_CR##_n##_0 & 1) \ ) \ ) \ << (4 * (7-(_n))) \ ) return FIELD(0) | FIELD(1) | FIELD(2) | FIELD(3) | FIELD(4) | FIELD(5) | FIELD(6) | FIELD(7); # undef FIELD } /* VISIBLE TO LIBVEX CLIENT */ /* Note: %CR is 32 bits even for ppc64 */ UInt LibVEX_GuestPPC64_get_CR ( /*IN*/const VexGuestPPC64State* vex_state ) { # define FIELD(_n) \ ( ( (UInt) \ ( (vex_state->guest_CR##_n##_321 & (7<<1)) \ | (vex_state->guest_CR##_n##_0 & 1) \ ) \ ) \ << (4 * (7-(_n))) \ ) return FIELD(0) | FIELD(1) | FIELD(2) | FIELD(3) | FIELD(4) | FIELD(5) | FIELD(6) | FIELD(7); # undef FIELD } /* VISIBLE TO LIBVEX CLIENT */ void LibVEX_GuestPPC32_put_CR ( UInt cr_native, /*OUT*/VexGuestPPC32State* vex_state ) { UInt t; # define FIELD(_n) \ do { \ t = cr_native >> (4*(7-(_n))); \ vex_state->guest_CR##_n##_0 = toUChar(t & 1); \ vex_state->guest_CR##_n##_321 = toUChar(t & (7<<1)); \ } while (0) FIELD(0); FIELD(1); FIELD(2); FIELD(3); FIELD(4); FIELD(5); FIELD(6); FIELD(7); # undef FIELD } /* VISIBLE TO LIBVEX CLIENT */ /* Note: %CR is 32 bits even for ppc64 */ void LibVEX_GuestPPC64_put_CR ( UInt cr_native, /*OUT*/VexGuestPPC64State* vex_state ) { UInt t; # define FIELD(_n) \ do { \ t = cr_native >> (4*(7-(_n))); \ vex_state->guest_CR##_n##_0 = toUChar(t & 1); \ vex_state->guest_CR##_n##_321 = toUChar(t & (7<<1)); \ } while (0) FIELD(0); FIELD(1); FIELD(2); FIELD(3); FIELD(4); FIELD(5); FIELD(6); FIELD(7); # undef FIELD } /* VISIBLE TO LIBVEX CLIENT */ UInt LibVEX_GuestPPC32_get_XER ( /*IN*/const VexGuestPPC32State* vex_state ) { UInt w = 0; w |= ( ((UInt)vex_state->guest_XER_BC) & 0xFF ); w |= ( (((UInt)vex_state->guest_XER_SO) & 0x1) << 31 ); w |= ( (((UInt)vex_state->guest_XER_OV) & 0x1) << 30 ); w |= ( (((UInt)vex_state->guest_XER_CA) & 0x1) << 29 ); w |= ( (((UInt)vex_state->guest_XER_OV32) & 0x1) << 19 ); w |= ( (((UInt)vex_state->guest_XER_CA32) & 0x1) << 18 ); return w; } /* VISIBLE TO LIBVEX CLIENT */ /* Note: %XER is 32 bits even for ppc64 */ UInt LibVEX_GuestPPC64_get_XER ( /*IN*/const VexGuestPPC64State* vex_state ) { UInt w = 0; w |= ( ((UInt)vex_state->guest_XER_BC) & 0xFF ); w |= ( (((UInt)vex_state->guest_XER_SO) & 0x1) << 31 ); w |= ( (((UInt)vex_state->guest_XER_OV) & 0x1) << 30 ); w |= ( (((UInt)vex_state->guest_XER_CA) & 0x1) << 29 ); w |= ( (((UInt)vex_state->guest_XER_OV32) & 0x1) << 19 ); w |= ( (((UInt)vex_state->guest_XER_CA32) & 0x1) << 18 ); return w; } /* VISIBLE TO LIBVEX CLIENT */ void LibVEX_GuestPPC32_put_XER ( UInt xer_native, /*OUT*/VexGuestPPC32State* vex_state ) { vex_state->guest_XER_BC = toUChar(xer_native & 0xFF); vex_state->guest_XER_SO = toUChar((xer_native >> 31) & 0x1); vex_state->guest_XER_OV = toUChar((xer_native >> 30) & 0x1); vex_state->guest_XER_CA = toUChar((xer_native >> 29) & 0x1); vex_state->guest_XER_OV32 = toUChar((xer_native >> 19) & 0x1); vex_state->guest_XER_CA32 = toUChar((xer_native >> 18) & 0x1); } /* VISIBLE TO LIBVEX CLIENT */ /* Note: %XER is 32 bits even for ppc64 */ void LibVEX_GuestPPC64_put_XER ( UInt xer_native, /*OUT*/VexGuestPPC64State* vex_state ) { vex_state->guest_XER_BC = toUChar(xer_native & 0xFF); vex_state->guest_XER_SO = toUChar((xer_native >> 31) & 0x1); vex_state->guest_XER_OV = toUChar((xer_native >> 30) & 0x1); vex_state->guest_XER_CA = toUChar((xer_native >> 29) & 0x1); vex_state->guest_XER_OV32 = toUChar((xer_native >> 19) & 0x1); vex_state->guest_XER_CA32 = toUChar((xer_native >> 18) & 0x1); } /* VISIBLE TO LIBVEX CLIENT */ void LibVEX_GuestPPC32_initialise ( /*OUT*/VexGuestPPC32State* vex_state ) { Int i; vex_state->host_EvC_FAILADDR = 0; vex_state->host_EvC_COUNTER = 0; vex_state->pad3 = 0; vex_state->pad4 = 0; vex_state->guest_GPR0 = 0; vex_state->guest_GPR1 = 0; vex_state->guest_GPR2 = 0; vex_state->guest_GPR3 = 0; vex_state->guest_GPR4 = 0; vex_state->guest_GPR5 = 0; vex_state->guest_GPR6 = 0; vex_state->guest_GPR7 = 0; vex_state->guest_GPR8 = 0; vex_state->guest_GPR9 = 0; vex_state->guest_GPR10 = 0; vex_state->guest_GPR11 = 0; vex_state->guest_GPR12 = 0; vex_state->guest_GPR13 = 0; vex_state->guest_GPR14 = 0; vex_state->guest_GPR15 = 0; vex_state->guest_GPR16 = 0; vex_state->guest_GPR17 = 0; vex_state->guest_GPR18 = 0; vex_state->guest_GPR19 = 0; vex_state->guest_GPR20 = 0; vex_state->guest_GPR21 = 0; vex_state->guest_GPR22 = 0; vex_state->guest_GPR23 = 0; vex_state->guest_GPR24 = 0; vex_state->guest_GPR25 = 0; vex_state->guest_GPR26 = 0; vex_state->guest_GPR27 = 0; vex_state->guest_GPR28 = 0; vex_state->guest_GPR29 = 0; vex_state->guest_GPR30 = 0; vex_state->guest_GPR31 = 0; /* Initialise the vector state. */ # define VECZERO(_vr) _vr[0]=_vr[1]=_vr[2]=_vr[3] = 0; VECZERO(vex_state->guest_VSR0 ); VECZERO(vex_state->guest_VSR1 ); VECZERO(vex_state->guest_VSR2 ); VECZERO(vex_state->guest_VSR3 ); VECZERO(vex_state->guest_VSR4 ); VECZERO(vex_state->guest_VSR5 ); VECZERO(vex_state->guest_VSR6 ); VECZERO(vex_state->guest_VSR7 ); VECZERO(vex_state->guest_VSR8 ); VECZERO(vex_state->guest_VSR9 ); VECZERO(vex_state->guest_VSR10); VECZERO(vex_state->guest_VSR11); VECZERO(vex_state->guest_VSR12); VECZERO(vex_state->guest_VSR13); VECZERO(vex_state->guest_VSR14); VECZERO(vex_state->guest_VSR15); VECZERO(vex_state->guest_VSR16); VECZERO(vex_state->guest_VSR17); VECZERO(vex_state->guest_VSR18); VECZERO(vex_state->guest_VSR19); VECZERO(vex_state->guest_VSR20); VECZERO(vex_state->guest_VSR21); VECZERO(vex_state->guest_VSR22); VECZERO(vex_state->guest_VSR23); VECZERO(vex_state->guest_VSR24); VECZERO(vex_state->guest_VSR25); VECZERO(vex_state->guest_VSR26); VECZERO(vex_state->guest_VSR27); VECZERO(vex_state->guest_VSR28); VECZERO(vex_state->guest_VSR29); VECZERO(vex_state->guest_VSR30); VECZERO(vex_state->guest_VSR31); VECZERO(vex_state->guest_VSR32); VECZERO(vex_state->guest_VSR33); VECZERO(vex_state->guest_VSR34); VECZERO(vex_state->guest_VSR35); VECZERO(vex_state->guest_VSR36); VECZERO(vex_state->guest_VSR37); VECZERO(vex_state->guest_VSR38); VECZERO(vex_state->guest_VSR39); VECZERO(vex_state->guest_VSR40); VECZERO(vex_state->guest_VSR41); VECZERO(vex_state->guest_VSR42); VECZERO(vex_state->guest_VSR43); VECZERO(vex_state->guest_VSR44); VECZERO(vex_state->guest_VSR45); VECZERO(vex_state->guest_VSR46); VECZERO(vex_state->guest_VSR47); VECZERO(vex_state->guest_VSR48); VECZERO(vex_state->guest_VSR49); VECZERO(vex_state->guest_VSR50); VECZERO(vex_state->guest_VSR51); VECZERO(vex_state->guest_VSR52); VECZERO(vex_state->guest_VSR53); VECZERO(vex_state->guest_VSR54); VECZERO(vex_state->guest_VSR55); VECZERO(vex_state->guest_VSR56); VECZERO(vex_state->guest_VSR57); VECZERO(vex_state->guest_VSR58); VECZERO(vex_state->guest_VSR59); VECZERO(vex_state->guest_VSR60); VECZERO(vex_state->guest_VSR61); VECZERO(vex_state->guest_VSR62); VECZERO(vex_state->guest_VSR63); VECZERO( vex_state->guest_ACC_0_r0 ); VECZERO( vex_state->guest_ACC_0_r1 ); VECZERO( vex_state->guest_ACC_0_r2 ); VECZERO( vex_state->guest_ACC_0_r3 ); VECZERO( vex_state->guest_ACC_1_r0 ); VECZERO( vex_state->guest_ACC_1_r1 ); VECZERO( vex_state->guest_ACC_1_r2 ); VECZERO( vex_state->guest_ACC_1_r3 ); VECZERO( vex_state->guest_ACC_2_r0 ); VECZERO( vex_state->guest_ACC_2_r1 ); VECZERO( vex_state->guest_ACC_2_r2 ); VECZERO( vex_state->guest_ACC_2_r3 ); VECZERO( vex_state->guest_ACC_3_r0 ); VECZERO( vex_state->guest_ACC_3_r1 ); VECZERO( vex_state->guest_ACC_3_r2 ); VECZERO( vex_state->guest_ACC_3_r3 ); VECZERO( vex_state->guest_ACC_4_r0 ); VECZERO( vex_state->guest_ACC_4_r1 ); VECZERO( vex_state->guest_ACC_4_r2 ); VECZERO( vex_state->guest_ACC_4_r3 ); VECZERO( vex_state->guest_ACC_5_r0 ); VECZERO( vex_state->guest_ACC_5_r1 ); VECZERO( vex_state->guest_ACC_5_r2 ); VECZERO( vex_state->guest_ACC_5_r3 ); VECZERO( vex_state->guest_ACC_6_r0 ); VECZERO( vex_state->guest_ACC_6_r1 ); VECZERO( vex_state->guest_ACC_6_r2 ); VECZERO( vex_state->guest_ACC_6_r3 ); VECZERO( vex_state->guest_ACC_7_r0 ); VECZERO( vex_state->guest_ACC_7_r1 ); VECZERO( vex_state->guest_ACC_7_r2 ); VECZERO( vex_state->guest_ACC_7_r3 ); # undef VECZERO vex_state->guest_CIA = 0; vex_state->guest_LR = 0; vex_state->guest_CTR = 0; vex_state->guest_XER_SO = 0; vex_state->guest_XER_OV = 0; vex_state->guest_XER_CA = 0; vex_state->guest_XER_BC = 0; vex_state->guest_XER_OV32 = 0; vex_state->guest_XER_CA32 = 0; vex_state->guest_CR0_321 = 0; vex_state->guest_CR0_0 = 0; vex_state->guest_CR1_321 = 0; vex_state->guest_CR1_0 = 0; vex_state->guest_CR2_321 = 0; vex_state->guest_CR2_0 = 0; vex_state->guest_CR3_321 = 0; vex_state->guest_CR3_0 = 0; vex_state->guest_CR4_321 = 0; vex_state->guest_CR4_0 = 0; vex_state->guest_CR5_321 = 0; vex_state->guest_CR5_0 = 0; vex_state->guest_CR6_321 = 0; vex_state->guest_CR6_0 = 0; vex_state->guest_CR7_321 = 0; vex_state->guest_CR7_0 = 0; vex_state->guest_FPROUND = PPCrm_NEAREST; vex_state->guest_DFPROUND = PPCrm_NEAREST; vex_state->guest_C_FPCC = 0; vex_state->pad2 = 0; vex_state->guest_VRSAVE = 0; # if defined(VGP_ppc64be_linux) /* By default, the HW for BE sets the VSCR[NJ] bit to 1. VSR is a 128-bit register, NJ bit is bit 111 (IBM numbering). However, VSCR is modeled as a 64-bit register. */ vex_state->guest_VSCR = 0x1 << (127 - 111); # else /* LE API requires NJ be set to 0. */ vex_state->guest_VSCR = 0x0; #endif vex_state->guest_EMNOTE = EmNote_NONE; vex_state->guest_CMSTART = 0; vex_state->guest_CMLEN = 0; vex_state->guest_NRADDR = 0; vex_state->guest_NRADDR_GPR2 = 0; vex_state->guest_REDIR_SP = -1; for (i = 0; i < VEX_GUEST_PPC32_REDIR_STACK_SIZE; i++) vex_state->guest_REDIR_STACK[i] = 0; vex_state->guest_IP_AT_SYSCALL = 0; vex_state->guest_SPRG3_RO = 0; vex_state->guest_PPR = 0x4ULL << 50; // medium priority vex_state->guest_PSPB = 0x100; // an arbitrary non-zero value to start with vex_state->padding1 = 0; /* vex_state->padding2 = 0; currently not used */ } /* VISIBLE TO LIBVEX CLIENT */ void LibVEX_GuestPPC64_initialise ( /*OUT*/VexGuestPPC64State* vex_state ) { Int i; vex_state->host_EvC_FAILADDR = 0; vex_state->host_EvC_COUNTER = 0; vex_state->pad0 = 0; vex_state->guest_GPR0 = 0; vex_state->guest_GPR1 = 0; vex_state->guest_GPR2 = 0; vex_state->guest_GPR3 = 0; vex_state->guest_GPR4 = 0; vex_state->guest_GPR5 = 0; vex_state->guest_GPR6 = 0; vex_state->guest_GPR7 = 0; vex_state->guest_GPR8 = 0; vex_state->guest_GPR9 = 0; vex_state->guest_GPR10 = 0; vex_state->guest_GPR11 = 0; vex_state->guest_GPR12 = 0; vex_state->guest_GPR13 = 0; vex_state->guest_GPR14 = 0; vex_state->guest_GPR15 = 0; vex_state->guest_GPR16 = 0; vex_state->guest_GPR17 = 0; vex_state->guest_GPR18 = 0; vex_state->guest_GPR19 = 0; vex_state->guest_GPR20 = 0; vex_state->guest_GPR21 = 0; vex_state->guest_GPR22 = 0; vex_state->guest_GPR23 = 0; vex_state->guest_GPR24 = 0; vex_state->guest_GPR25 = 0; vex_state->guest_GPR26 = 0; vex_state->guest_GPR27 = 0; vex_state->guest_GPR28 = 0; vex_state->guest_GPR29 = 0; vex_state->guest_GPR30 = 0; vex_state->guest_GPR31 = 0; /* Initialise the vector state. */ # define VECZERO(_vr) _vr[0]=_vr[1]=_vr[2]=_vr[3] = 0; VECZERO(vex_state->guest_VSR0 ); VECZERO(vex_state->guest_VSR1 ); VECZERO(vex_state->guest_VSR2 ); VECZERO(vex_state->guest_VSR3 ); VECZERO(vex_state->guest_VSR4 ); VECZERO(vex_state->guest_VSR5 ); VECZERO(vex_state->guest_VSR6 ); VECZERO(vex_state->guest_VSR7 ); VECZERO(vex_state->guest_VSR8 ); VECZERO(vex_state->guest_VSR9 ); VECZERO(vex_state->guest_VSR10); VECZERO(vex_state->guest_VSR11); VECZERO(vex_state->guest_VSR12); VECZERO(vex_state->guest_VSR13); VECZERO(vex_state->guest_VSR14); VECZERO(vex_state->guest_VSR15); VECZERO(vex_state->guest_VSR16); VECZERO(vex_state->guest_VSR17); VECZERO(vex_state->guest_VSR18); VECZERO(vex_state->guest_VSR19); VECZERO(vex_state->guest_VSR20); VECZERO(vex_state->guest_VSR21); VECZERO(vex_state->guest_VSR22); VECZERO(vex_state->guest_VSR23); VECZERO(vex_state->guest_VSR24); VECZERO(vex_state->guest_VSR25); VECZERO(vex_state->guest_VSR26); VECZERO(vex_state->guest_VSR27); VECZERO(vex_state->guest_VSR28); VECZERO(vex_state->guest_VSR29); VECZERO(vex_state->guest_VSR30); VECZERO(vex_state->guest_VSR31); VECZERO(vex_state->guest_VSR32); VECZERO(vex_state->guest_VSR33); VECZERO(vex_state->guest_VSR34); VECZERO(vex_state->guest_VSR35); VECZERO(vex_state->guest_VSR36); VECZERO(vex_state->guest_VSR37); VECZERO(vex_state->guest_VSR38); VECZERO(vex_state->guest_VSR39); VECZERO(vex_state->guest_VSR40); VECZERO(vex_state->guest_VSR41); VECZERO(vex_state->guest_VSR42); VECZERO(vex_state->guest_VSR43); VECZERO(vex_state->guest_VSR44); VECZERO(vex_state->guest_VSR45); VECZERO(vex_state->guest_VSR46); VECZERO(vex_state->guest_VSR47); VECZERO(vex_state->guest_VSR48); VECZERO(vex_state->guest_VSR49); VECZERO(vex_state->guest_VSR50); VECZERO(vex_state->guest_VSR51); VECZERO(vex_state->guest_VSR52); VECZERO(vex_state->guest_VSR53); VECZERO(vex_state->guest_VSR54); VECZERO(vex_state->guest_VSR55); VECZERO(vex_state->guest_VSR56); VECZERO(vex_state->guest_VSR57); VECZERO(vex_state->guest_VSR58); VECZERO(vex_state->guest_VSR59); VECZERO(vex_state->guest_VSR60); VECZERO(vex_state->guest_VSR61); VECZERO(vex_state->guest_VSR62); VECZERO(vex_state->guest_VSR63); # undef VECZERO vex_state->guest_CIA = 0; vex_state->guest_LR = 0; vex_state->guest_CTR = 0; vex_state->guest_XER_SO = 0; vex_state->guest_XER_OV = 0; vex_state->guest_XER_CA = 0; vex_state->guest_XER_BC = 0; vex_state->guest_CR0_321 = 0; vex_state->guest_CR0_0 = 0; vex_state->guest_CR1_321 = 0; vex_state->guest_CR1_0 = 0; vex_state->guest_CR2_321 = 0; vex_state->guest_CR2_0 = 0; vex_state->guest_CR3_321 = 0; vex_state->guest_CR3_0 = 0; vex_state->guest_CR4_321 = 0; vex_state->guest_CR4_0 = 0; vex_state->guest_CR5_321 = 0; vex_state->guest_CR5_0 = 0; vex_state->guest_CR6_321 = 0; vex_state->guest_CR6_0 = 0; vex_state->guest_CR7_321 = 0; vex_state->guest_CR7_0 = 0; vex_state->guest_FPROUND = PPCrm_NEAREST; vex_state->guest_DFPROUND = PPCrm_NEAREST; vex_state->guest_C_FPCC = 0; vex_state->pad2 = 0; vex_state->guest_VRSAVE = 0; # if defined(VGP_ppc64be_linux) /* By default, the HW for BE sets the VSCR[NJ] bit to 1. VSR is a 128-bit register, NJ bit is bit 111 (IBM numbering). However, VSCR is modeled as a 64-bit register. */ vex_state->guest_VSCR = 0x1 << (127 - 111); # else /* LE API requires NJ be set to 0. */ vex_state->guest_VSCR = 0x0; #endif vex_state->guest_EMNOTE = EmNote_NONE; vex_state->padding = 0; vex_state->guest_CMSTART = 0; vex_state->guest_CMLEN = 0; vex_state->guest_NRADDR = 0; vex_state->guest_NRADDR_GPR2 = 0; vex_state->guest_REDIR_SP = -1; for (i = 0; i < VEX_GUEST_PPC64_REDIR_STACK_SIZE; i++) vex_state->guest_REDIR_STACK[i] = 0; vex_state->guest_IP_AT_SYSCALL = 0; vex_state->guest_SPRG3_RO = 0; vex_state->guest_TFHAR = 0; vex_state->guest_TFIAR = 0; vex_state->guest_TEXASR = 0; vex_state->guest_PPR = 0x4ULL << 50; // medium priority vex_state->guest_PSPB = 0x100; // an arbitrary non-zero value to start with vex_state->guest_DSCR = 0; } /*-----------------------------------------------------------*/ /*--- Describing the ppc guest state, for the benefit ---*/ /*--- of iropt and instrumenters. ---*/ /*-----------------------------------------------------------*/ /* Figure out if any part of the guest state contained in minoff .. maxoff requires precise memory exceptions. If in doubt return True (but this is generates significantly slower code). By default we enforce precise exns for guest R1 (stack pointer), CIA (current insn address) and LR (link register). These are the minimum needed to extract correct stack backtraces from ppc code. [[NB: not sure if keeping LR up to date is actually necessary.]] Only R1 is needed in mode VexRegUpdSpAtMemAccess. */ Bool guest_ppc32_state_requires_precise_mem_exns ( Int minoff, Int maxoff, VexRegisterUpdates pxControl ) { Int lr_min = offsetof(VexGuestPPC32State, guest_LR); Int lr_max = lr_min + 4 - 1; Int r1_min = offsetof(VexGuestPPC32State, guest_GPR1); Int r1_max = r1_min + 4 - 1; Int cia_min = offsetof(VexGuestPPC32State, guest_CIA); Int cia_max = cia_min + 4 - 1; if (maxoff < r1_min || minoff > r1_max) { /* no overlap with R1 */ if (pxControl == VexRegUpdSpAtMemAccess) return False; // We only need to check stack pointer. } else { return True; } if (maxoff < lr_min || minoff > lr_max) { /* no overlap with LR */ } else { return True; } if (maxoff < cia_min || minoff > cia_max) { /* no overlap with CIA */ } else { return True; } return False; } Bool guest_ppc64_state_requires_precise_mem_exns ( Int minoff, Int maxoff, VexRegisterUpdates pxControl ) { /* Given that R2 is a Big Deal in the ELF ppc64 ABI, it seems prudent to be conservative with it, even though thus far there is no evidence to suggest that it actually needs to be kept up to date wrt possible exceptions. */ Int lr_min = offsetof(VexGuestPPC64State, guest_LR); Int lr_max = lr_min + 8 - 1; Int r1_min = offsetof(VexGuestPPC64State, guest_GPR1); Int r1_max = r1_min + 8 - 1; Int r2_min = offsetof(VexGuestPPC64State, guest_GPR2); Int r2_max = r2_min + 8 - 1; Int cia_min = offsetof(VexGuestPPC64State, guest_CIA); Int cia_max = cia_min + 8 - 1; if (maxoff < r1_min || minoff > r1_max) { /* no overlap with R1 */ if (pxControl == VexRegUpdSpAtMemAccess) return False; // We only need to check stack pointer. } else { return True; } if (maxoff < lr_min || minoff > lr_max) { /* no overlap with LR */ } else { return True; } if (maxoff < r2_min || minoff > r2_max) { /* no overlap with R2 */ } else { return True; } if (maxoff < cia_min || minoff > cia_max) { /* no overlap with CIA */ } else { return True; } return False; } #define ALWAYSDEFD32(field) \ { offsetof(VexGuestPPC32State, field), \ (sizeof ((VexGuestPPC32State*)0)->field) } VexGuestLayout ppc32Guest_layout = { /* Total size of the guest state, in bytes. */ .total_sizeB = sizeof(VexGuestPPC32State), /* Describe the stack pointer. */ .offset_SP = offsetof(VexGuestPPC32State,guest_GPR1), .sizeof_SP = 4, /* Describe the frame pointer. */ .offset_FP = offsetof(VexGuestPPC32State,guest_GPR1), .sizeof_FP = 4, /* Describe the instruction pointer. */ .offset_IP = offsetof(VexGuestPPC32State,guest_CIA), .sizeof_IP = 4, /* Describe any sections to be regarded by Memcheck as 'always-defined'. */ .n_alwaysDefd = 12, .alwaysDefd = { /* 0 */ ALWAYSDEFD32(guest_CIA), /* 1 */ ALWAYSDEFD32(guest_EMNOTE), /* 2 */ ALWAYSDEFD32(guest_CMSTART), /* 3 */ ALWAYSDEFD32(guest_CMLEN), /* 4 */ ALWAYSDEFD32(guest_VSCR), /* 5 */ ALWAYSDEFD32(guest_FPROUND), /* 6 */ ALWAYSDEFD32(guest_NRADDR), /* 7 */ ALWAYSDEFD32(guest_NRADDR_GPR2), /* 8 */ ALWAYSDEFD32(guest_REDIR_SP), /* 9 */ ALWAYSDEFD32(guest_REDIR_STACK), /* 10 */ ALWAYSDEFD32(guest_IP_AT_SYSCALL), /* 11 */ ALWAYSDEFD32(guest_C_FPCC) } }; #define ALWAYSDEFD64(field) \ { offsetof(VexGuestPPC64State, field), \ (sizeof ((VexGuestPPC64State*)0)->field) } VexGuestLayout ppc64Guest_layout = { /* Total size of the guest state, in bytes. */ .total_sizeB = sizeof(VexGuestPPC64State), /* Describe the stack pointer. */ .offset_SP = offsetof(VexGuestPPC64State,guest_GPR1), .sizeof_SP = 8, /* Describe the frame pointer. */ .offset_FP = offsetof(VexGuestPPC64State,guest_GPR1), .sizeof_FP = 8, /* Describe the instruction pointer. */ .offset_IP = offsetof(VexGuestPPC64State,guest_CIA), .sizeof_IP = 8, /* Describe any sections to be regarded by Memcheck as 'always-defined'. */ .n_alwaysDefd = 12, .alwaysDefd = { /* 0 */ ALWAYSDEFD64(guest_CIA), /* 1 */ ALWAYSDEFD64(guest_EMNOTE), /* 2 */ ALWAYSDEFD64(guest_CMSTART), /* 3 */ ALWAYSDEFD64(guest_CMLEN), /* 4 */ ALWAYSDEFD64(guest_VSCR), /* 5 */ ALWAYSDEFD64(guest_FPROUND), /* 6 */ ALWAYSDEFD64(guest_NRADDR), /* 7 */ ALWAYSDEFD64(guest_NRADDR_GPR2), /* 8 */ ALWAYSDEFD64(guest_REDIR_SP), /* 9 */ ALWAYSDEFD64(guest_REDIR_STACK), /* 10 */ ALWAYSDEFD64(guest_IP_AT_SYSCALL), /* 11 */ ALWAYSDEFD64(guest_C_FPCC) } }; /*---------------------------------------------------------------*/ /*--- end guest_ppc_helpers.c ---*/ /*---------------------------------------------------------------*/