/* * Copyright (c) 2014 Google Inc. All rights reserved * Contributed by Stephane Eranian * * Permission is hereby granted, free of charge, to any person obtaining a copy * of this software and associated documentation files (the "Software"), to deal * in the Software without restriction, including without limitation the rights * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies * of the Software, and to permit persons to whom the Software is furnished to do so, * subject to the following conditions: * * The above copyright notice and this permission notice shall be included in all * copies or substantial portions of the Software. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, * INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A * PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT * HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF * CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE * OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. * * This file is part of libpfm, a performance monitoring support library for * applications on Linux. * * This file has been automatically generated. * * PMU: ivbep_unc_ha (Intel IvyBridge-EP HA uncore PMU) */ static const intel_x86_umask_t ivbep_unc_h_conflict_cycles[]={ { .uname = "CONFLICT", .udesc = "Number of cycles that we are handling conflicts", .ucode = 0x200, }, { .uname = "LAST", .udesc = "Count every last conflictor in conflict chain. Can be used to compute average conflict chain length", .ucode = 0x400, }, { .uname = "CMP_FWDS", .udesc = "Count the number of cmp_fwd. This gives the number of late conflicts", .ucode = 0x1000, }, { .uname = "ACKCNFLTS", .udesc = "Count the number Acknflts", .ucode = 0x800, }, }; static const intel_x86_umask_t ivbep_unc_h_directory_lookup[]={ { .uname = "NO_SNP", .udesc = "Snoop not needed", .ucode = 0x200, }, { .uname = "SNOOP", .udesc = "SNooop needed", .ucode = 0x100, }, }; static const intel_x86_umask_t ivbep_unc_h_bypass_imc[]={ { .uname = "TAKEN", .udesc = "Bypass taken", .ucode = 0x200, }, { .uname = "NOT_TAKEN", .udesc = "Bypass not taken", .ucode = 0x100, }, }; static const intel_x86_umask_t ivbep_unc_h_directory_update[]={ { .uname = "ANY", .udesc = "Counts any directory update", .ucode = 0x300, .uflags = INTEL_X86_NCOMBO | INTEL_X86_DFL, }, { .uname = "CLEAR", .udesc = "Directory clears", .ucode = 0x200, .uflags = INTEL_X86_NCOMBO, }, { .uname = "SET", .udesc = "Directory set", .ucode = 0x100, .uflags = INTEL_X86_NCOMBO, }, }; static const intel_x86_umask_t ivbep_unc_h_igr_no_credit_cycles[]={ { .uname = "AD_QPI0", .udesc = "AD to QPI link 0", .ucode = 0x100, .uflags = INTEL_X86_NCOMBO, }, { .uname = "AD_QPI1", .udesc = "AD to QPI link 1", .ucode = 0x200, .uflags = INTEL_X86_NCOMBO, }, { .uname = "BL_QPI0", .udesc = "BL to QPI link 0", .ucode = 0x400, .uflags = INTEL_X86_NCOMBO, }, { .uname = "BL_QPI1", .udesc = "BL to QPI link 1", .ucode = 0x800, .uflags = INTEL_X86_NCOMBO, }, }; static const intel_x86_umask_t ivbep_unc_h_imc_writes[]={ { .uname = "ALL", .udesc = "Counts all writes", .ucode = 0xf00, .uflags = INTEL_X86_NCOMBO | INTEL_X86_DFL, }, { .uname = "FULL", .udesc = "Counts full line non ISOCH", .ucode = 0x100, .uflags = INTEL_X86_NCOMBO, }, { .uname = "FULL_ISOCH", .udesc = "Counts ISOCH full line", .ucode = 0x400, .uflags = INTEL_X86_NCOMBO, }, { .uname = "PARTIAL", .udesc = "Counts partial non-ISOCH", .ucode = 0x200, .uflags = INTEL_X86_NCOMBO, }, { .uname = "PARTIAL_ISOCH", .udesc = "Counts ISOCH partial", .ucode = 0x800, .uflags = INTEL_X86_NCOMBO, }, }; static const intel_x86_umask_t ivbep_unc_h_imc_reads[]={ { .uname = "NORMAL", .udesc = "Normal priority", .ucode = 0x100, .uflags = INTEL_X86_DFL, }, }; static const intel_x86_umask_t ivbep_unc_h_requests[]={ { .uname = "READS", .udesc = "Counts incoming read requests. Good proxy for LLC read misses, incl. RFOs", .ucode = 0x300, .uflags = INTEL_X86_NCOMBO, }, { .uname = "READS_LOCAL", .udesc = "Counts incoming read requests coming from local socket. Good proxy for LLC read misses, incl. RFOs from the local socket", .ucode = 0x100, .uflags = INTEL_X86_NCOMBO, }, { .uname = "READS_REMOTE", .udesc = "Counts incoming read requests coming from remote socket. Good proxy for LLC read misses, incl. RFOs from the remote socket", .ucode = 0x200, .uflags = INTEL_X86_NCOMBO, }, { .uname = "WRITES", .udesc = "Counts incoming writes", .ucode = 0xc00, .uflags = INTEL_X86_NCOMBO, }, { .uname = "WRITES_LOCAL", .udesc = "Counts incoming writes from local socket", .ucode = 0x400, .uflags = INTEL_X86_NCOMBO, }, { .uname = "WRITES_REMOTE", .udesc = "Counts incoming writes from remote socket", .ucode = 0x800, .uflags = INTEL_X86_NCOMBO, }, { .uname = "INVITOE_LOCAL", .udesc = "Counts InvItoE coming from local socket", .ucode = 0x1000, .uflags = INTEL_X86_NCOMBO, }, { .uname = "INVITOE_REMOTE", .udesc = "Counts InvItoE coming from remote socket", .ucode = 0x2000, .uflags = INTEL_X86_NCOMBO, } }; static const intel_x86_umask_t ivbep_unc_h_rpq_cycles_no_reg_credits[]={ { .uname = "CHN0", .udesc = "Channel 0", .ucode = 0x100, .uflags = INTEL_X86_NCOMBO, }, { .uname = "CHN1", .udesc = "Channel 1", .ucode = 0x200, .uflags = INTEL_X86_NCOMBO, }, { .uname = "CHN2", .udesc = "channel 2", .ucode = 0x400, .uflags = INTEL_X86_NCOMBO, }, { .uname = "CHN3", .udesc = "Chanell 3", .ucode = 0x800, .uflags = INTEL_X86_NCOMBO, }, }; static const intel_x86_umask_t ivbep_unc_h_tad_requests_g0[]={ { .uname = "REGION0", .udesc = "Counts for TAD Region 0", .ucode = 0x100, .uflags = INTEL_X86_NCOMBO, }, { .uname = "REGION1", .udesc = "Counts for TAD Region 1", .ucode = 0x200, .uflags = INTEL_X86_NCOMBO, }, { .uname = "REGION2", .udesc = "Counts for TAD Region 2", .ucode = 0x400, .uflags = INTEL_X86_NCOMBO, }, { .uname = "REGION3", .udesc = "Counts for TAD Region 3", .ucode = 0x800, .uflags = INTEL_X86_NCOMBO, }, { .uname = "REGION4", .udesc = "Counts for TAD Region 4", .ucode = 0x1000, .uflags = INTEL_X86_NCOMBO, }, { .uname = "REGION5", .udesc = "Counts for TAD Region 5", .ucode = 0x2000, .uflags = INTEL_X86_NCOMBO, }, { .uname = "REGION6", .udesc = "Counts for TAD Region 6", .ucode = 0x4000, .uflags = INTEL_X86_NCOMBO, }, { .uname = "REGION7", .udesc = "Counts for TAD Region 7", .ucode = 0x8000, .uflags = INTEL_X86_NCOMBO, }, }; static const intel_x86_umask_t ivbep_unc_h_tad_requests_g1[]={ { .uname = "REGION8", .udesc = "Counts for TAD Region 8", .ucode = 0x100, .uflags = INTEL_X86_NCOMBO, }, { .uname = "REGION9", .udesc = "Counts for TAD Region 9", .ucode = 0x200, .uflags = INTEL_X86_NCOMBO, }, { .uname = "REGION10", .udesc = "Counts for TAD Region 10", .ucode = 0x400, .uflags = INTEL_X86_NCOMBO, }, { .uname = "REGION11", .udesc = "Counts for TAD Region 11", .ucode = 0x800, .uflags = INTEL_X86_NCOMBO, }, }; static const intel_x86_umask_t ivbep_unc_h_snoop_resp[]={ { .uname = "RSPI", .udesc = "Filters for snoop responses of RspI. RspI is returned when the remote cache does not have the data or when the remote cache silently evicts data (e.g. RFO hit non-modified line)", .ucode = 0x100, }, { .uname = "RSPS", .udesc = "Filters for snoop responses of RspS. RspS is returned when the remote cache has the data but is not forwarding it. It is a way to let the requesting socket know that it cannot allocate the data in E-state", .ucode = 0x200, }, { .uname = "RSPIFWD", .udesc = "Filters for snoop responses of RspIFwd. RspIFwd is returned when the remote cache agent forwards data and the requesting agent is able to acquire the data in E or M state. This is commonly returned with RFO transacations. It can be either HitM or HitFE", .ucode = 0x400, }, { .uname = "RSPSFWD", .udesc = "Filters for snoop responses of RspSFwd. RspSFwd is returned when the remote cache agent forwards data but holds on to its current copy. This is common for data and code reads that hit in a remote socket in E or F state", .ucode = 0x800, }, { .uname = "RSP_WB", .udesc = "Filters for snoop responses of RspIWB or RspSWB. This is returned when a non-RFO requests hits in M-state. Data and code reads can return either RspIWB or RspSWB depending on how the system has been configured. InvItoE transactions will also return RspIWB because they must acquire ownership", .ucode = 0x1000, }, { .uname = "RSP_FWD_WB", .udesc = "Filters for snoop responses of RspxFwdxWB. This snoop response is only used in 4s systems. It is used when a snoop HITM in a remote caching agent and it directly forwards data to a requester and simultaneously returns data to the home to be written back to memory", .ucode = 0x2000, }, { .uname = "RSPCNFLCT", .udesc = "Filters for snoop responses of RspConflict. This is returned when a snoop finds an existing outstanding transaction in a remote caching agent when it CMAs that caching agent. This triggers the conflict resolution hardware. This covers both RspConflct and RspCnflctWBI", .ucode = 0x4000, }, }; static const intel_x86_umask_t ivbep_unc_h_txr_ad_cycles_full[]={ { .uname = "ALL", .udesc = "Counts cycles full from both schedulers", .ucode = 0x300, .uflags = INTEL_X86_NCOMBO | INTEL_X86_DFL, }, { .uname = "SCHED0", .udesc = "Counts cycles full from scheduler bank 0", .ucode = 0x100, .uflags = INTEL_X86_NCOMBO, }, { .uname = "SCHED1", .udesc = "Counts cycles full from scheduler bank 1", .ucode = 0x200, .uflags = INTEL_X86_NCOMBO, }, }; static const intel_x86_umask_t ivbep_unc_h_txr_bl_occupancy[]={ { .uname = "SCHED0", .udesc = "Counts cycles full from scheduler bank 0", .ucode = 0x100, .uflags = INTEL_X86_NCOMBO, }, { .uname = "SCHED1", .udesc = "Counts cycles full from scheduler bank 1", .ucode = 0x200, .uflags = INTEL_X86_NCOMBO, }, }; static const intel_x86_umask_t ivbep_unc_h_txr_ak_cycles_full[]={ { .uname = "ALL", .udesc = "Counts cycles from both schedulers", .ucode = 0x300, .uflags = INTEL_X86_NCOMBO | INTEL_X86_DFL, }, { .uname = "SCHED0", .udesc = "Counts cycles from scheduler bank 0", .ucode = 0x100, .uflags = INTEL_X86_NCOMBO, }, { .uname = "SCHED1", .udesc = "Counts cycles from scheduler bank 1", .ucode = 0x200, .uflags = INTEL_X86_NCOMBO, }, }; static const intel_x86_umask_t ivbep_unc_h_txr_bl[]={ { .uname = "DRS_CACHE", .udesc = "Counts data being sent to the cache", .ucode = 0x100, .uflags = INTEL_X86_NCOMBO, }, { .uname = "DRS_CORE", .udesc = "Counts data being sent directly to the requesting core", .ucode = 0x200, .uflags = INTEL_X86_NCOMBO, }, { .uname = "DRS_QPI", .udesc = "Counts data being sent to a remote socket over QPI", .ucode = 0x400, .uflags = INTEL_X86_NCOMBO, }, }; #if 0 static const intel_x86_umask_t ivbep_unc_h_addr_opc_match[]={ { .uname = "FILT", .udesc = "Number of addr and opcode matches (opc via opc= or address via addr= modifiers)", .ucode = 0x300, .uflags = INTEL_X86_NCOMBO | INTEL_X86_DFL | INTEL_X86_ADDR, }, }; #endif static const intel_x86_umask_t ivbep_unc_h_bt_occupancy[]={ { .uname = "LOCAL", .udesc = "Local", .ucode = 0x100, }, { .uname = "REMOTE", .udesc = "Remote", .ucode = 0x200, }, { .uname = "READS_REMOTE", .udesc = "Reads remote", .ucode = 0x800, }, { .uname = "WRITES_LOCAL", .udesc = "Writes local", .ucode = 0x1000, }, { .uname = "WRITES_REMOTE", .udesc = "Writes remote", .ucode = 0x2000, }, }; static const intel_x86_umask_t ivbep_unc_h_osb[]={ { .uname = "REMOTE", .udesc = "Remote", .ucode = 0x800, }, { .uname = "READS_LOCAL", .udesc = "Local reads", .ucode = 0x200, }, { .uname = "INVITOE_LOCAL", .udesc = "Local InvItoE", .ucode = 0x400, } }; static const intel_x86_umask_t ivbep_unc_h_osb_edr[]={ { .uname = "ALL", .udesc = "All data returns", .ucode = 0x100, .uflags = INTEL_X86_DFL | INTEL_X86_NCOMBO, }, { .uname = "READS_LOCAL_I", .udesc = "Reads to local I", .ucode = 0x200, }, { .uname = "READS_REMOTE_I", .udesc = "Reads to remote I", .ucode = 0x400, }, { .uname = "READS_LOCAL_S", .udesc = "Reads to local S", .ucode = 0x800, }, { .uname = "READS_REMOTE_S", .udesc = "Reads to remote S", .ucode = 0x1000, } }; static const intel_x86_umask_t ivbep_unc_h_ring_ad_used[]={ { .uname = "CCW_VR0_EVEN", .udesc = "Counter-clockwise and even ring polarity on virtual ring 0", .ucode = 0x400, }, { .uname = "CCW_VR0_ODD", .udesc = "Counter-clockwise and odd ring polarity on virtual ring 0", .ucode = 0x800, }, { .uname = "CW_VR0_EVEN", .udesc = "Clockwise and even ring polarity on virtual ring 0", .ucode = 0x100, }, { .uname = "CW_VR0_ODD", .udesc = "Clockwise and odd ring polarity on virtual ring 0", .ucode = 0x200, }, { .uname = "CCW_VR1_EVEN", .udesc = "Counter-clockwise and even ring polarity on virtual ring 1", .ucode = 0x400, }, { .uname = "CCW_VR1_ODD", .udesc = "Counter-clockwise and odd ring polarity on virtual ring 1", .ucode = 0x800, }, { .uname = "CW_VR1_EVEN", .udesc = "Clockwise and even ring polarity on virtual ring 1", .ucode = 0x100, }, { .uname = "CW_VR1_ODD", .udesc = "Clockwise and odd ring polarity on virtual ring 1", .ucode = 0x200, }, { .uname = "CW", .udesc = "Clockwise with any polarity on either virtual rings", .ucode = 0x3300, }, { .uname = "CCW", .udesc = "Counter-clockwise with any polarity on either virtual rings", .ucode = 0xcc00, }, }; static const intel_x86_umask_t ivbep_unc_h_snp_resp_recv_local[]={ { .uname = "RSPI", .udesc = "Filters for snoop responses of RspI. RspI is returned when the remote cache does not have the data or when the remote cache silently evicts data (e.g. RFO hit non-modified line)", .ucode = 0x100, }, { .uname = "RSPS", .udesc = "Filters for snoop responses of RspS. RspS is returned when the remote cache has the data but is not forwarding it. It is a way to let the requesting socket know that it cannot allocate the data in E-state", .ucode = 0x200, }, { .uname = "RSPIFWD", .udesc = "Filters for snoop responses of RspIFwd. RspIFwd is returned when the remote cache agent forwards data and the requesting agent is able to acquire the data in E or M state. This is commonly returned with RFO transacations. It can be either HitM or HitFE", .ucode = 0x400, }, { .uname = "RSPSFWD", .udesc = "Filters for snoop responses of RspSFwd. RspSFwd is returned when the remote cache agent forwards data but holds on to its current copy. This is common for data and code reads that hit in a remote socket in E or F state", .ucode = 0x800, }, { .uname = "RSP_WB", .udesc = "Filters for snoop responses of RspIWB or RspSWB. This is returned when a non-RFO requests hits in M-state. Data and code reads can return either RspIWB or RspSWB depending on how the system has been configured. InvItoE transactions will also return RspIWB because they must acquire ownership", .ucode = 0x1000, }, { .uname = "RSP_FWD_WB", .udesc = "Filters for snoop responses of RspxFwdxWB. This snoop response is only used in 4s systems. It is used when a snoop HITM in a remote caching agent and it directly forwards data to a requester and simultaneously returns data to the home to be written back to memory", .ucode = 0x2000, }, { .uname = "RSPCNFLCT", .udesc = "Filters for snoop responses of RspConflict. This is returned when a snoop finds an existing outstanding transaction in a remote caching agent when it CMAs that caching agent. This triggers the conflict resolution hardware. This covers both RspConflct and RspCnflctWBI", .ucode = 0x4000, }, { .uname = "OTHER", .udesc = "Filters all other snoop responses", .ucode = 0x8000, }, }; static const intel_x86_umask_t ivbep_unc_h_txr_ak[]={ { .uname = "NDR", .udesc = "Number of outbound NDR (non-data response) transactions send on the AK ring. AK NDR is used for messages to the local socket", .ucode = 0x100, }, { .uname = "CRD_CBO", .udesc = "Number of outbound CDR transactions send on the AK ring to CBO", .ucode = 0x200, }, { .uname = "CRD_QPI", .udesc = "Number of outbound CDR transactions send on the AK ring to QPI", .ucode = 0x400, }, }; static const intel_x86_umask_t ivbep_unc_h_iodc_conflicts[]={ { .uname = "ANY", .udesc = "Any conflict", .ucode = 0x100, .uflags = INTEL_X86_DFL | INTEL_X86_NCOMBO, }, { .uname = "LAST", .udesc = "Last conflict", .ucode = 0x400, } }; static const intel_x86_entry_t intel_ivbep_unc_h_pe[]={ { .name = "UNC_H_CLOCKTICKS", .desc = "HA Uncore clockticks", .modmsk = IVBEP_UNC_HA_ATTRS, .cntmsk = 0xf, .code = 0x00, }, { .name = "UNC_H_CONFLICT_CYCLES", .desc = "Conflict Checks", .code = 0xb, .cntmsk = 0xf, .ngrp = 1, .modmsk = IVBEP_UNC_HA_ATTRS, .numasks = LIBPFM_ARRAY_SIZE(ivbep_unc_h_conflict_cycles), .umasks = ivbep_unc_h_conflict_cycles, }, { .name = "UNC_H_DIRECT2CORE_COUNT", .desc = "Direct2Core Messages Sent", .code = 0x11, .cntmsk = 0xf, .modmsk = IVBEP_UNC_HA_ATTRS, }, { .name = "UNC_H_DIRECT2CORE_CYCLES_DISABLED", .desc = "Cycles when Direct2Core was Disabled", .code = 0x12, .cntmsk = 0xf, .modmsk = IVBEP_UNC_HA_ATTRS, }, { .name = "UNC_H_DIRECT2CORE_TXN_OVERRIDE", .desc = "Number of Reads that had Direct2Core Overridden", .code = 0x13, .cntmsk = 0xf, .modmsk = IVBEP_UNC_HA_ATTRS, }, { .name = "UNC_H_DIRECTORY_LOOKUP", .desc = "Directory Lookups", .code = 0xc, .cntmsk = 0xf, .ngrp = 1, .modmsk = IVBEP_UNC_HA_ATTRS, .numasks = LIBPFM_ARRAY_SIZE(ivbep_unc_h_directory_lookup), .umasks = ivbep_unc_h_directory_lookup }, { .name = "UNC_H_DIRECTORY_UPDATE", .desc = "Directory Updates", .code = 0xd, .cntmsk = 0xf, .ngrp = 1, .modmsk = IVBEP_UNC_HA_ATTRS, .numasks = LIBPFM_ARRAY_SIZE(ivbep_unc_h_directory_update), .umasks = ivbep_unc_h_directory_update }, { .name = "UNC_H_IGR_NO_CREDIT_CYCLES", .desc = "Cycles without QPI Ingress Credits", .code = 0x22, .cntmsk = 0xf, .ngrp = 1, .modmsk = IVBEP_UNC_HA_ATTRS, .numasks = LIBPFM_ARRAY_SIZE(ivbep_unc_h_igr_no_credit_cycles), .umasks = ivbep_unc_h_igr_no_credit_cycles }, { .name = "UNC_H_IMC_RETRY", .desc = "Retry Events", .code = 0x1e, .cntmsk = 0xf, .modmsk = IVBEP_UNC_HA_ATTRS, }, { .name = "UNC_H_IMC_WRITES", .desc = "HA to IMC Full Line Writes Issued", .code = 0x1a, .cntmsk = 0xf, .ngrp = 1, .modmsk = IVBEP_UNC_HA_ATTRS, .numasks = LIBPFM_ARRAY_SIZE(ivbep_unc_h_imc_writes), .umasks = ivbep_unc_h_imc_writes }, { .name = "UNC_H_IMC_READS", .desc = "HA to IMC normal priority reads issued", .code = 0x17, .cntmsk = 0xf, .ngrp = 1, .modmsk = IVBEP_UNC_HA_ATTRS, .numasks = LIBPFM_ARRAY_SIZE(ivbep_unc_h_imc_reads), .umasks = ivbep_unc_h_imc_reads }, { .name = "UNC_H_REQUESTS", .desc = "Read and Write Requests", .code = 0x1, .cntmsk = 0xf, .ngrp = 1, .modmsk = IVBEP_UNC_HA_ATTRS, .numasks = LIBPFM_ARRAY_SIZE(ivbep_unc_h_requests), .umasks = ivbep_unc_h_requests }, { .name = "UNC_H_RPQ_CYCLES_NO_REG_CREDITS", .desc = "IMC RPQ Credits Empty", .code = 0x15, .cntmsk = 0xf, .ngrp = 1, .modmsk = IVBEP_UNC_HA_ATTRS, .numasks = LIBPFM_ARRAY_SIZE(ivbep_unc_h_rpq_cycles_no_reg_credits), .umasks = ivbep_unc_h_rpq_cycles_no_reg_credits }, { .name = "UNC_H_TAD_REQUESTS_G0", .desc = "HA Requests to a TAD Region", .code = 0x1b, .cntmsk = 0xf, .ngrp = 1, .modmsk = IVBEP_UNC_HA_ATTRS, .numasks = LIBPFM_ARRAY_SIZE(ivbep_unc_h_tad_requests_g0), .umasks = ivbep_unc_h_tad_requests_g0 }, { .name = "UNC_H_TAD_REQUESTS_G1", .desc = "HA Requests to a TAD Region", .code = 0x1c, .cntmsk = 0xf, .ngrp = 1, .modmsk = IVBEP_UNC_HA_ATTRS, .numasks = LIBPFM_ARRAY_SIZE(ivbep_unc_h_tad_requests_g1), .umasks = ivbep_unc_h_tad_requests_g1 }, { .name = "UNC_H_TXR_AD_CYCLES_FULL", .desc = "AD Egress Full", .code = 0x2a, .cntmsk = 0xf, .ngrp = 1, .modmsk = IVBEP_UNC_HA_ATTRS, .numasks = LIBPFM_ARRAY_SIZE(ivbep_unc_h_txr_ad_cycles_full), .umasks = ivbep_unc_h_txr_ad_cycles_full }, { .name = "UNC_H_TXR_AK_CYCLES_FULL", .desc = "AK Egress Full", .code = 0x32, .cntmsk = 0xf, .ngrp = 1, .modmsk = IVBEP_UNC_HA_ATTRS, .numasks = LIBPFM_ARRAY_SIZE(ivbep_unc_h_txr_ak_cycles_full), .umasks = ivbep_unc_h_txr_ak_cycles_full }, { .name = "UNC_H_TXR_AK", .desc = "Outbound Ring Transactions on AK", .code = 0xe, .cntmsk = 0xf, .ngrp = 1, .modmsk = IVBEP_UNC_HA_ATTRS, .numasks = LIBPFM_ARRAY_SIZE(ivbep_unc_h_txr_ak), .umasks = ivbep_unc_h_txr_ak }, { .name = "UNC_H_TXR_BL", .desc = "Outbound DRS Ring Transactions to Cache", .code = 0x10, .cntmsk = 0xf, .ngrp = 1, .modmsk = IVBEP_UNC_HA_ATTRS, .numasks = LIBPFM_ARRAY_SIZE(ivbep_unc_h_txr_bl), .umasks = ivbep_unc_h_txr_bl }, { .name = "UNC_H_TXR_BL_CYCLES_FULL", .desc = "BL Egress Full", .code = 0x36, .cntmsk = 0xf, .ngrp = 1, .modmsk = IVBEP_UNC_HA_ATTRS, .numasks = LIBPFM_ARRAY_SIZE(ivbep_unc_h_txr_ak_cycles_full), .umasks = ivbep_unc_h_txr_ak_cycles_full, /* identical to snbep_unc_h_txr_ak_cycles_full */ }, { .name = "UNC_H_WPQ_CYCLES_NO_REG_CREDITS", .desc = "HA IMC CHN0 WPQ Credits Empty", .code = 0x18, .cntmsk = 0xf, .ngrp = 1, .modmsk = IVBEP_UNC_HA_ATTRS, .numasks = LIBPFM_ARRAY_SIZE(ivbep_unc_h_rpq_cycles_no_reg_credits), .umasks = ivbep_unc_h_rpq_cycles_no_reg_credits, /* shared */ }, { .name = "UNC_H_BT_BYPASS", .desc = "Backup Tracker bypass", .code = 0x52, .cntmsk = 0xf, .modmsk = IVBEP_UNC_HA_ATTRS, }, { .name = "UNC_H_BYPASS_IMC", .desc = "HA to IMC bypass", .code = 0x14, .cntmsk = 0xf, .ngrp = 1, .modmsk = IVBEP_UNC_HA_ATTRS, .numasks = LIBPFM_ARRAY_SIZE(ivbep_unc_h_bypass_imc), .umasks = ivbep_unc_h_bypass_imc, }, { .name = "UNC_H_BT_CYCLES_NE", .desc = "Backup Tracker cycles not empty", .code = 0x42, .cntmsk = 0xf, .modmsk = IVBEP_UNC_HA_ATTRS, }, { .name = "UNC_H_BT_OCCUPANCY", .desc = "Backup Tracker inserts", .code = 0x43, .cntmsk = 0xf, .ngrp = 1, .modmsk = IVBEP_UNC_HA_ATTRS, .numasks = LIBPFM_ARRAY_SIZE(ivbep_unc_h_bt_occupancy), .umasks = ivbep_unc_h_bt_occupancy, }, { .name = "UNC_H_IGR_AD_QPI2", .desc = "AD QPI Link 2 credit accumulator", .code = 0x59, .cntmsk = 0xf, .modmsk = IVBEP_UNC_HA_ATTRS, }, { .name = "UNC_H_IGR_BL_QPI2", .desc = "BL QPI Link 2 credit accumulator", .code = 0x5a, .cntmsk = 0xf, .modmsk = IVBEP_UNC_HA_ATTRS, }, { .name = "UNC_H_IODC_INSERTS", .desc = "IODC inserts", .code = 0x56, .cntmsk = 0xf, .modmsk = IVBEP_UNC_HA_ATTRS, }, { .name = "UNC_H_IODC_CONFLICTS", .desc = "IODC conflicts", .code = 0x57, .cntmsk = 0xf, .ngrp = 1, .modmsk = IVBEP_UNC_HA_ATTRS, .numasks = LIBPFM_ARRAY_SIZE(ivbep_unc_h_iodc_conflicts), .umasks = ivbep_unc_h_iodc_conflicts, }, { .name = "UNC_H_IODC_OLEN_WBMTOI", .desc = "IODC zero length writes", .code = 0x58, .cntmsk = 0xf, .modmsk = IVBEP_UNC_HA_ATTRS, }, { .name = "UNC_H_OSB", .desc = "OSB snoop broadcast", .code = 0x53, .cntmsk = 0xf, .ngrp = 1, .modmsk = IVBEP_UNC_HA_ATTRS, .numasks = LIBPFM_ARRAY_SIZE(ivbep_unc_h_osb), .umasks = ivbep_unc_h_osb, }, { .name = "UNC_H_OSB_EDR", .desc = "OSB early data return", .code = 0x54, .cntmsk = 0xf, .ngrp = 1, .modmsk = IVBEP_UNC_HA_ATTRS, .numasks = LIBPFM_ARRAY_SIZE(ivbep_unc_h_osb_edr), .umasks = ivbep_unc_h_osb_edr, }, { .name = "UNC_H_RING_AD_USED", .desc = "AD ring in use", .code = 0x3e, .cntmsk = 0xf, .ngrp = 1, .modmsk = IVBEP_UNC_HA_ATTRS, .numasks = LIBPFM_ARRAY_SIZE(ivbep_unc_h_ring_ad_used), .umasks = ivbep_unc_h_ring_ad_used, }, { .name = "UNC_H_RING_AK_USED", .desc = "AK ring in use", .code = 0x3f, .cntmsk = 0xf, .ngrp = 1, .modmsk = IVBEP_UNC_HA_ATTRS, .numasks = LIBPFM_ARRAY_SIZE(ivbep_unc_h_ring_ad_used), /* shared */ .umasks = ivbep_unc_h_ring_ad_used, }, { .name = "UNC_H_RING_BL_USED", .desc = "BL ring in use", .code = 0x40, .cntmsk = 0xf, .ngrp = 1, .modmsk = IVBEP_UNC_HA_ATTRS, .numasks = LIBPFM_ARRAY_SIZE(ivbep_unc_h_ring_ad_used), /* shared */ .umasks = ivbep_unc_h_ring_ad_used, }, { .name = "UNC_H_DIRECTORY_LAT_OPT", .desc = "Directory latency optimization data return path taken", .code = 0x41, .cntmsk = 0xf, .modmsk = IVBEP_UNC_HA_ATTRS, }, { .name = "UNC_H_SNP_RESP_RECV_LOCAL", .desc = "Snoop responses received local", .code = 0x60, .cntmsk = 0xf, .ngrp = 1, .modmsk = IVBEP_UNC_HA_ATTRS, .numasks = LIBPFM_ARRAY_SIZE(ivbep_unc_h_snp_resp_recv_local), .umasks = ivbep_unc_h_snp_resp_recv_local, }, { .name = "UNC_H_TXR_BL_OCCUPANCY", .desc = "BL Egress occupancy", .code = 0x34, .cntmsk = 0xf, .ngrp = 1, .modmsk = IVBEP_UNC_HA_ATTRS, .numasks = LIBPFM_ARRAY_SIZE(ivbep_unc_h_txr_bl_occupancy), .umasks = ivbep_unc_h_txr_bl_occupancy, }, { .name = "UNC_H_SNOOP_RESP", .desc = "Snoop responses received", .code = 0x21, .cntmsk = 0xf, .ngrp = 1, .modmsk = IVBEP_UNC_HA_ATTRS, .numasks = LIBPFM_ARRAY_SIZE(ivbep_unc_h_snoop_resp), .umasks = ivbep_unc_h_snoop_resp }, };