xref: /illumos-gate/usr/src/uts/intel/pcbe/opteron_pcbe.c (revision 1a065e93eee983124652c3eb0cfdcb4776cd89ab)
1 /*
2  * CDDL HEADER START
3  *
4  * The contents of this file are subject to the terms of the
5  * Common Development and Distribution License (the "License").
6  * You may not use this file except in compliance with the License.
7  *
8  * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
9  * or http://www.opensolaris.org/os/licensing.
10  * See the License for the specific language governing permissions
11  * and limitations under the License.
12  *
13  * When distributing Covered Code, include this CDDL HEADER in each
14  * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
15  * If applicable, add the following below this CDDL HEADER, with the
16  * fields enclosed by brackets "[]" replaced with your own identifying
17  * information: Portions Copyright [yyyy] [name of copyright owner]
18  *
19  * CDDL HEADER END
20  */
21 /*
22  * Copyright 2009 Sun Microsystems, Inc.  All rights reserved.
23  * Use is subject to license terms.
24  */
25 
26 /*
27  * This file contains preset event names from the Performance Application
28  * Programming Interface v3.5 which included the following notice:
29  *
30  *                             Copyright (c) 2005,6
31  *                           Innovative Computing Labs
32  *                         Computer Science Department,
33  *                            University of Tennessee,
34  *                                 Knoxville, TN.
35  *                              All Rights Reserved.
36  *
37  *
38  * Redistribution and use in source and binary forms, with or without
39  * modification, are permitted provided that the following conditions are met:
40  *
41  *    * Redistributions of source code must retain the above copyright notice,
42  *      this list of conditions and the following disclaimer.
43  *    * Redistributions in binary form must reproduce the above copyright
44  *	notice, this list of conditions and the following disclaimer in the
45  *	documentation and/or other materials provided with the distribution.
46  *    * Neither the name of the University of Tennessee nor the names of its
47  *      contributors may be used to endorse or promote products derived from
48  *	this software without specific prior written permission.
49  *
50  * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
51  * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
52  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
53  * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
54  * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
55  * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
56  * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
57  * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
58  * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
59  * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
60  * POSSIBILITY OF SUCH DAMAGE.
61  *
62  *
63  * This open source software license conforms to the BSD License template.
64  */
65 
66 /*
67  * Portions Copyright 2009 Advanced Micro Devices, Inc.
68  * Copyright 2019 Joyent, Inc.
69  * Copyright 2021 Oxide Computer Company
70  */
71 
72 /*
73  * Performance Counter Back-End for AMD Opteron and AMD Athlon 64 processors.
74  */
75 
76 #include <sys/cpuvar.h>
77 #include <sys/param.h>
78 #include <sys/systm.h>
79 #include <sys/cpc_pcbe.h>
80 #include <sys/kmem.h>
81 #include <sys/sdt.h>
82 #include <sys/modctl.h>
83 #include <sys/errno.h>
84 #include <sys/debug.h>
85 #include <sys/archsystm.h>
86 #include <sys/x86_archext.h>
87 #include <sys/privregs.h>
88 #include <sys/ddi.h>
89 #include <sys/sunddi.h>
90 
91 #include "opteron_pcbe_table.h"
92 #include <opteron_pcbe_cpcgen.h>
93 
94 static int opt_pcbe_init(void);
95 static uint_t opt_pcbe_ncounters(void);
96 static const char *opt_pcbe_impl_name(void);
97 static const char *opt_pcbe_cpuref(void);
98 static char *opt_pcbe_list_events(uint_t picnum);
99 static char *opt_pcbe_list_attrs(void);
100 static uint64_t opt_pcbe_event_coverage(char *event);
101 static uint64_t opt_pcbe_overflow_bitmap(void);
102 static int opt_pcbe_configure(uint_t picnum, char *event, uint64_t preset,
103     uint32_t flags, uint_t nattrs, kcpc_attr_t *attrs, void **data,
104     void *token);
105 static void opt_pcbe_program(void *token);
106 static void opt_pcbe_allstop(void);
107 static void opt_pcbe_sample(void *token);
108 static void opt_pcbe_free(void *config);
109 
110 static pcbe_ops_t opt_pcbe_ops = {
111 	PCBE_VER_1,
112 	CPC_CAP_OVERFLOW_INTERRUPT,
113 	opt_pcbe_ncounters,
114 	opt_pcbe_impl_name,
115 	opt_pcbe_cpuref,
116 	opt_pcbe_list_events,
117 	opt_pcbe_list_attrs,
118 	opt_pcbe_event_coverage,
119 	opt_pcbe_overflow_bitmap,
120 	opt_pcbe_configure,
121 	opt_pcbe_program,
122 	opt_pcbe_allstop,
123 	opt_pcbe_sample,
124 	opt_pcbe_free
125 };
126 
127 /*
128  * Base MSR addresses for the PerfEvtSel registers and the counters themselves.
129  * Add counter number to base address to get corresponding MSR address.
130  */
131 #define	PES_BASE_ADDR	0xC0010000
132 #define	PIC_BASE_ADDR	0xC0010004
133 
134 /*
135  * Base MSR addresses for the PerfEvtSel registers and counters. The counter and
136  * event select registers are interleaved, so one needs to multiply the counter
137  * number by two to determine what they should be set to.
138  */
139 #define	PES_EXT_BASE_ADDR	0xC0010200
140 #define	PIC_EXT_BASE_ADDR	0xC0010201
141 
142 /*
143  * The number of counters present depends on which CPU features are present.
144  */
145 #define	OPT_PCBE_DEF_NCOUNTERS	4
146 #define	OPT_PCBE_EXT_NCOUNTERS	6
147 
148 /*
149  * Define offsets and masks for the fields in the Performance
150  * Event-Select (PES) registers.
151  */
152 #define	OPT_PES_HOST_SHIFT	41
153 #define	OPT_PES_GUEST_SHIFT	40
154 #define	OPT_PES_EVSELHI_SHIFT	32
155 #define	OPT_PES_CMASK_SHIFT	24
156 #define	OPT_PES_CMASK_MASK	0xFF
157 #define	OPT_PES_INV_SHIFT	23
158 #define	OPT_PES_ENABLE_SHIFT	22
159 #define	OPT_PES_INT_SHIFT	20
160 #define	OPT_PES_PC_SHIFT	19
161 #define	OPT_PES_EDGE_SHIFT	18
162 #define	OPT_PES_OS_SHIFT	17
163 #define	OPT_PES_USR_SHIFT	16
164 #define	OPT_PES_UMASK_SHIFT	8
165 #define	OPT_PES_UMASK_MASK	0xFF
166 
167 #define	OPT_PES_INV		(1ULL << OPT_PES_INV_SHIFT)
168 #define	OPT_PES_ENABLE		(1ULL << OPT_PES_ENABLE_SHIFT)
169 #define	OPT_PES_INT		(1ULL << OPT_PES_INT_SHIFT)
170 #define	OPT_PES_PC		(1ULL << OPT_PES_PC_SHIFT)
171 #define	OPT_PES_EDGE		(1ULL << OPT_PES_EDGE_SHIFT)
172 #define	OPT_PES_OS		(1ULL << OPT_PES_OS_SHIFT)
173 #define	OPT_PES_USR		(1ULL << OPT_PES_USR_SHIFT)
174 #define	OPT_PES_HOST		(1ULL << OPT_PES_HOST_SHIFT)
175 #define	OPT_PES_GUEST		(1ULL << OPT_PES_GUEST_SHIFT)
176 
177 typedef struct _opt_pcbe_config {
178 	uint8_t		opt_picno;	/* Counter number: 0, 1, 2, or 3 */
179 	uint64_t	opt_evsel;	/* Event Selection register */
180 	uint64_t	opt_rawpic;	/* Raw counter value */
181 } opt_pcbe_config_t;
182 
183 opt_pcbe_config_t nullcfgs[OPT_PCBE_EXT_NCOUNTERS] = {
184 	{ 0, 0, 0 },
185 	{ 1, 0, 0 },
186 	{ 2, 0, 0 },
187 	{ 3, 0, 0 },
188 	{ 4, 0, 0 },
189 	{ 5, 0, 0 },
190 };
191 
192 typedef uint64_t (*opt_pcbe_addr_f)(uint_t);
193 
194 typedef struct opt_pcbe_data {
195 	uint_t		opd_ncounters;
196 	uint_t		opd_cmask;
197 	opt_pcbe_addr_f	opd_pesf;
198 	opt_pcbe_addr_f	opd_picf;
199 } opt_pcbe_data_t;
200 
201 opt_pcbe_data_t opd;
202 
203 #define	MASK48		0xFFFFFFFFFFFF
204 
205 #define	EV_END {NULL, 0}
206 #define	GEN_EV_END {NULL, NULL, 0 }
207 
208 /*
209  * The following Macros are used to define tables of events that are used by
210  * various families and some generic classes of events.
211  *
212  * When programming a performance counter there are two different values that we
213  * need to set:
214  *
215  *   o Event - Determines the general class of event that is being used.
216  *   o Unit  - A further breakdown that gives more specific value.
217  *
218  * Prior to the introduction of family 17h support, all family specific events
219  * were programmed based on their event. The generic events, which tried to
220  * provide PAPI mappings to events specified an additional unit mask.
221  *
222  * Starting with Family 17h, CPU performance counters default to using both the
223  * unit mask and the event select. Generic events are always aliases to a
224  * specific event/unit pair, hence why the units for them are always zero. In
225  * addition, the naming of events in family 17h has been changed to reflect
226  * AMD's guide. While this is a departure from what people are used to, it is
227  * believed that matching the more detailed literature that folks are told to
228  * reference is more valuable.
229  */
230 
231 #define	AMD_cmn_events						\
232 	{ "FP_dispatched_fpu_ops",			0x0 },	\
233 	{ "FP_cycles_no_fpu_ops_retired",		0x1 },	\
234 	{ "FP_dispatched_fpu_ops_ff",			0x2 },	\
235 	{ "LS_seg_reg_load",				0x20 },	\
236 	{ "LS_uarch_resync_self_modify",		0x21 },	\
237 	{ "LS_uarch_resync_snoop",			0x22 },	\
238 	{ "LS_buffer_2_full",				0x23 },	\
239 	{ "LS_locked_operation",			0x24 },	\
240 	{ "LS_retired_cflush",				0x26 },	\
241 	{ "LS_retired_cpuid",				0x27 },	\
242 	{ "DC_access",					0x40 },	\
243 	{ "DC_miss",					0x41 },	\
244 	{ "DC_refill_from_L2",				0x42 },	\
245 	{ "DC_refill_from_system",			0x43 },	\
246 	{ "DC_copyback",				0x44 },	\
247 	{ "DC_dtlb_L1_miss_L2_hit",			0x45 },	\
248 	{ "DC_dtlb_L1_miss_L2_miss",			0x46 },	\
249 	{ "DC_misaligned_data_ref",			0x47 },	\
250 	{ "DC_uarch_late_cancel_access",		0x48 },	\
251 	{ "DC_uarch_early_cancel_access",		0x49 },	\
252 	{ "DC_1bit_ecc_error_found",			0x4A },	\
253 	{ "DC_dispatched_prefetch_instr",		0x4B },	\
254 	{ "DC_dcache_accesses_by_locks",		0x4C },	\
255 	{ "BU_memory_requests",				0x65 },	\
256 	{ "BU_data_prefetch",				0x67 },	\
257 	{ "BU_system_read_responses",			0x6C },	\
258 	{ "BU_cpu_clk_unhalted",			0x76 },	\
259 	{ "BU_internal_L2_req",				0x7D },	\
260 	{ "BU_fill_req_missed_L2",			0x7E },	\
261 	{ "BU_fill_into_L2",				0x7F },	\
262 	{ "IC_fetch",					0x80 },	\
263 	{ "IC_miss",					0x81 },	\
264 	{ "IC_refill_from_L2",				0x82 },	\
265 	{ "IC_refill_from_system",			0x83 },	\
266 	{ "IC_itlb_L1_miss_L2_hit",			0x84 },	\
267 	{ "IC_itlb_L1_miss_L2_miss",			0x85 },	\
268 	{ "IC_uarch_resync_snoop",			0x86 },	\
269 	{ "IC_instr_fetch_stall",			0x87 },	\
270 	{ "IC_return_stack_hit",			0x88 },	\
271 	{ "IC_return_stack_overflow",			0x89 },	\
272 	{ "FR_retired_x86_instr_w_excp_intr",		0xC0 },	\
273 	{ "FR_retired_uops",				0xC1 },	\
274 	{ "FR_retired_branches_w_excp_intr",		0xC2 },	\
275 	{ "FR_retired_branches_mispred",		0xC3 },	\
276 	{ "FR_retired_taken_branches",			0xC4 },	\
277 	{ "FR_retired_taken_branches_mispred",		0xC5 },	\
278 	{ "FR_retired_far_ctl_transfer",		0xC6 },	\
279 	{ "FR_retired_resyncs",				0xC7 },	\
280 	{ "FR_retired_near_rets",			0xC8 },	\
281 	{ "FR_retired_near_rets_mispred",		0xC9 },	\
282 	{ "FR_retired_taken_branches_mispred_addr_miscomp",	0xCA },\
283 	{ "FR_retired_fastpath_double_op_instr",	0xCC },	\
284 	{ "FR_intr_masked_cycles",			0xCD },	\
285 	{ "FR_intr_masked_while_pending_cycles",	0xCE },	\
286 	{ "FR_taken_hardware_intrs",			0xCF },	\
287 	{ "FR_nothing_to_dispatch",			0xD0 },	\
288 	{ "FR_dispatch_stalls",				0xD1 },	\
289 	{ "FR_dispatch_stall_branch_abort_to_retire",	0xD2 },	\
290 	{ "FR_dispatch_stall_serialization",		0xD3 },	\
291 	{ "FR_dispatch_stall_segment_load",		0xD4 },	\
292 	{ "FR_dispatch_stall_reorder_buffer_full",	0xD5 },	\
293 	{ "FR_dispatch_stall_resv_stations_full",	0xD6 },	\
294 	{ "FR_dispatch_stall_fpu_full",			0xD7 },	\
295 	{ "FR_dispatch_stall_ls_full",			0xD8 },	\
296 	{ "FR_dispatch_stall_waiting_all_quiet",	0xD9 },	\
297 	{ "FR_dispatch_stall_far_ctl_trsfr_resync_branch_pend",	0xDA },\
298 	{ "FR_fpu_exception",				0xDB },	\
299 	{ "FR_num_brkpts_dr0",				0xDC },	\
300 	{ "FR_num_brkpts_dr1",				0xDD },	\
301 	{ "FR_num_brkpts_dr2",				0xDE },	\
302 	{ "FR_num_brkpts_dr3",				0xDF },	\
303 	{ "NB_mem_ctrlr_page_access",			0xE0 },	\
304 	{ "NB_mem_ctrlr_turnaround",			0xE3 },	\
305 	{ "NB_mem_ctrlr_bypass_counter_saturation",	0xE4 },	\
306 	{ "NB_cpu_io_to_mem_io",			0xE9 },	\
307 	{ "NB_cache_block_commands",			0xEA },	\
308 	{ "NB_sized_commands",				0xEB },	\
309 	{ "NB_ht_bus0_bandwidth",			0xF6 }
310 
311 #define	AMD_FAMILY_f_events					\
312 	{ "BU_quadwords_written_to_system",		0x6D },	\
313 	{ "FR_retired_fpu_instr",			0xCB },	\
314 	{ "NB_mem_ctrlr_page_table_overflow",		0xE1 },	\
315 	{ "NB_sized_blocks",				0xE5 },	\
316 	{ "NB_ECC_errors",				0xE8 },	\
317 	{ "NB_probe_result",				0xEC },	\
318 	{ "NB_gart_events",				0xEE },	\
319 	{ "NB_ht_bus1_bandwidth",			0xF7 },	\
320 	{ "NB_ht_bus2_bandwidth",			0xF8 }
321 
322 #define	AMD_FAMILY_10h_events					\
323 	{ "FP_retired_sse_ops",				0x3 },	\
324 	{ "FP_retired_move_ops",			0x4 },	\
325 	{ "FP_retired_serialize_ops",			0x5 },	\
326 	{ "FP_serialize_ops_cycles",			0x6 },	\
327 	{ "LS_cancelled_store_to_load_fwd_ops",		0x2A },	\
328 	{ "LS_smi_received",				0x2B },	\
329 	{ "DC_dtlb_L1_hit",				0x4D },	\
330 	{ "LS_ineffective_prefetch",			0x52 },	\
331 	{ "LS_global_tlb_flush",			0x54 },	\
332 	{ "BU_octwords_written_to_system",		0x6D },	\
333 	{ "Page_size_mismatches",			0x165 },	\
334 	{ "IC_eviction",				0x8B },	\
335 	{ "IC_cache_lines_invalidate",			0x8C },	\
336 	{ "IC_itlb_reload",				0x99 },	\
337 	{ "IC_itlb_reload_aborted",			0x9A },	\
338 	{ "FR_retired_mmx_sse_fp_instr",		0xCB },	\
339 	{ "Retired_x87_fp_ops",				0x1C0 },	\
340 	{ "IBS_ops_tagged",				0x1CF },	\
341 	{ "LFENCE_inst_retired",			0x1D3 },	\
342 	{ "SFENCE_inst_retired",			0x1D4 },	\
343 	{ "MFENCE_inst_retired",			0x1D5 },	\
344 	{ "NB_mem_ctrlr_page_table_overflow",		0xE1 },	\
345 	{ "NB_mem_ctrlr_dram_cmd_slots_missed",		0xE2 },	\
346 	{ "NB_thermal_status",				0xE8 },	\
347 	{ "NB_probe_results_upstream_req",		0xEC },	\
348 	{ "NB_gart_events",				0xEE },	\
349 	{ "NB_mem_ctrlr_req",				0x1F0 },	\
350 	{ "CB_cpu_to_dram_req_to_target",		0x1E0 },	\
351 	{ "CB_io_to_dram_req_to_target",		0x1E1 },	\
352 	{ "CB_cpu_read_cmd_latency_to_target_0_to_3",	0x1E2 },	\
353 	{ "CB_cpu_read_cmd_req_to_target_0_to_3",	0x1E3 },	\
354 	{ "CB_cpu_read_cmd_latency_to_target_4_to_7",	0x1E4 },	\
355 	{ "CB_cpu_read_cmd_req_to_target_4_to_7",	0x1E5 },	\
356 	{ "CB_cpu_cmd_latency_to_target_0_to_7",	0x1E6 },	\
357 	{ "CB_cpu_req_to_target_0_to_7",		0x1E7 },	\
358 	{ "NB_ht_bus1_bandwidth",			0xF7 },	\
359 	{ "NB_ht_bus2_bandwidth",			0xF8 },	\
360 	{ "NB_ht_bus3_bandwidth",			0x1F9 },	\
361 	{ "L3_read_req",				0x4E0 },	\
362 	{ "L3_miss",					0x4E1 },	\
363 	{ "L3_l2_eviction_l3_fill",			0x4E2 },	\
364 	{ "L3_eviction",				0x4E3 }
365 
366 #define	AMD_FAMILY_11h_events					\
367 	{ "BU_quadwords_written_to_system",		0x6D },	\
368 	{ "FR_retired_mmx_fp_instr",			0xCB },	\
369 	{ "NB_mem_ctrlr_page_table_events",		0xE1 },	\
370 	{ "NB_thermal_status",				0xE8 },	\
371 	{ "NB_probe_results_upstream_req",		0xEC },	\
372 	{ "NB_dev_events",				0xEE },	\
373 	{ "NB_mem_ctrlr_req",				0x1F0 }
374 
375 #define	AMD_cmn_generic_events						\
376 	{ "PAPI_br_ins",	"FR_retired_branches_w_excp_intr", 0x0 },\
377 	{ "PAPI_br_msp",	"FR_retired_branches_mispred",	0x0 },	\
378 	{ "PAPI_br_tkn",	"FR_retired_taken_branches",	0x0 },	\
379 	{ "PAPI_fp_ops",	"FP_dispatched_fpu_ops",	0x3 },	\
380 	{ "PAPI_fad_ins",	"FP_dispatched_fpu_ops",	0x1 },	\
381 	{ "PAPI_fml_ins",	"FP_dispatched_fpu_ops",	0x2 },	\
382 	{ "PAPI_fpu_idl",	"FP_cycles_no_fpu_ops_retired",	0x0 },	\
383 	{ "PAPI_tot_cyc",	"BU_cpu_clk_unhalted",		0x0 },	\
384 	{ "PAPI_tot_ins",	"FR_retired_x86_instr_w_excp_intr", 0x0 }, \
385 	{ "PAPI_l1_dca",	"DC_access",			0x0 },	\
386 	{ "PAPI_l1_dcm",	"DC_miss",			0x0 },	\
387 	{ "PAPI_l1_ldm",	"DC_refill_from_L2",		0xe },	\
388 	{ "PAPI_l1_stm",	"DC_refill_from_L2",		0x10 },	\
389 	{ "PAPI_l1_ica",	"IC_fetch",			0x0 },	\
390 	{ "PAPI_l1_icm",	"IC_miss",			0x0 },	\
391 	{ "PAPI_l1_icr",	"IC_fetch",			0x0 },	\
392 	{ "PAPI_l2_dch",	"DC_refill_from_L2",		0x1e },	\
393 	{ "PAPI_l2_dcm",	"DC_refill_from_system",	0x1e },	\
394 	{ "PAPI_l2_dcr",	"DC_refill_from_L2",		0xe },	\
395 	{ "PAPI_l2_dcw",	"DC_refill_from_L2",		0x10 },	\
396 	{ "PAPI_l2_ich",	"IC_refill_from_L2",		0x0 },	\
397 	{ "PAPI_l2_icm",	"IC_refill_from_system",	0x0 },	\
398 	{ "PAPI_l2_ldm",	"DC_refill_from_system",	0xe },	\
399 	{ "PAPI_l2_stm",	"DC_refill_from_system",	0x10 },	\
400 	{ "PAPI_res_stl",	"FR_dispatch_stalls",		0x0 },	\
401 	{ "PAPI_stl_icy",	"FR_nothing_to_dispatch",	0x0 },	\
402 	{ "PAPI_hw_int",	"FR_taken_hardware_intrs",	0x0 }
403 
404 #define	OPT_cmn_generic_events						\
405 	{ "PAPI_tlb_dm",	"DC_dtlb_L1_miss_L2_miss",	0x0 },	\
406 	{ "PAPI_tlb_im",	"IC_itlb_L1_miss_L2_miss",	0x0 },	\
407 	{ "PAPI_fp_ins",	"FR_retired_fpu_instr",		0xd },	\
408 	{ "PAPI_vec_ins",	"FR_retired_fpu_instr",		0x4 }
409 
410 #define	AMD_FAMILY_10h_generic_events					\
411 	{ "PAPI_tlb_dm",	"DC_dtlb_L1_miss_L2_miss",	0x7 },	\
412 	{ "PAPI_tlb_im",	"IC_itlb_L1_miss_L2_miss",	0x3 },	\
413 	{ "PAPI_l3_dcr",	"L3_read_req",			0xf1 }, \
414 	{ "PAPI_l3_icr",	"L3_read_req",			0xf2 }, \
415 	{ "PAPI_l3_tcr",	"L3_read_req",			0xf7 }, \
416 	{ "PAPI_l3_stm",	"L3_miss",			0xf4 }, \
417 	{ "PAPI_l3_ldm",	"L3_miss",			0xf3 }, \
418 	{ "PAPI_l3_tcm",	"L3_miss",			0xf7 }
419 
420 static const amd_event_t family_f_events[] = {
421 	AMD_cmn_events,
422 	AMD_FAMILY_f_events,
423 	EV_END
424 };
425 
426 static const amd_event_t family_10h_events[] = {
427 	AMD_cmn_events,
428 	AMD_FAMILY_10h_events,
429 	EV_END
430 };
431 
432 static const amd_event_t family_11h_events[] = {
433 	AMD_cmn_events,
434 	AMD_FAMILY_11h_events,
435 	EV_END
436 };
437 
438 static const amd_generic_event_t opt_generic_events[] = {
439 	AMD_cmn_generic_events,
440 	OPT_cmn_generic_events,
441 	GEN_EV_END
442 };
443 
444 static const amd_generic_event_t family_10h_generic_events[] = {
445 	AMD_cmn_generic_events,
446 	AMD_FAMILY_10h_generic_events,
447 	GEN_EV_END
448 };
449 
450 /*
451  * For Family 17h and Family 19h, the cpcgen utility generates all of our events
452  * including ones that need specific unit codes, therefore we leave all unit
453  * codes out of these. Zen 1, Zen 2, and Zen 3 have different event sets that
454  * they support.
455  */
456 static const amd_generic_event_t family_17h_zen1_papi_events[] = {
457 	{ "PAPI_br_cn",		"ExRetCond" },
458 	{ "PAPI_br_ins",	"ExRetBrn" },
459 	{ "PAPI_fpu_idl",	"FpSchedEmpty" },
460 	{ "PAPI_tot_cyc",	"LsNotHaltedCyc" },
461 	{ "PAPI_tot_ins",	"ExRetInstr" },
462 	{ "PAPI_tlb_dm",	"LsL1DTlbMiss" },
463 	{ "PAPI_tlb_im",	"BpL1TlbMissL2Miss" },
464 	{ "PAPI_tot_cyc",	"LsNotHaltedCyc" },
465 	GEN_EV_END
466 };
467 
468 static const amd_generic_event_t family_17h_zen2_papi_events[] = {
469 	{ "PAPI_br_cn",		"ExRetCond" },
470 	{ "PAPI_br_ins",	"ExRetBrn" },
471 	{ "PAPI_tot_cyc",	"LsNotHaltedCyc" },
472 	{ "PAPI_tot_ins",	"ExRetInstr" },
473 	{ "PAPI_tlb_dm",	"LsL1DTlbMiss" },
474 	{ "PAPI_tlb_im",	"BpL1TlbMissL2Miss" },
475 	{ "PAPI_tot_cyc",	"LsNotHaltedCyc" },
476 	GEN_EV_END
477 };
478 
479 static const amd_generic_event_t family_19h_zen3_papi_events[] = {
480 	{ "PAPI_br_cn",		"ExRetCond" },
481 	{ "PAPI_br_ins",	"ExRetBrn" },
482 	{ "PAPI_tot_cyc",	"LsNotHaltedCyc" },
483 	{ "PAPI_tot_ins",	"ExRetInstr" },
484 	{ "PAPI_tlb_dm",	"LsL1DTlbMiss" },
485 	{ "PAPI_tlb_im",	"BpL1TlbMissL2TlbMiss" },
486 	{ "PAPI_tot_cyc",	"LsNotHaltedCyc" },
487 	GEN_EV_END
488 };
489 
490 
491 
492 static char	*evlist;
493 static size_t	evlist_sz;
494 static const amd_event_t *amd_events = NULL;
495 static uint_t amd_family, amd_model;
496 static const amd_generic_event_t *amd_generic_events = NULL;
497 
498 static char amd_fam_f_rev_ae_bkdg[] = "See \"BIOS and Kernel Developer's "
499 "Guide for AMD Athlon 64 and AMD Opteron Processors\" (AMD publication 26094)";
500 static char amd_fam_f_NPT_bkdg[] = "See \"BIOS and Kernel Developer's Guide "
501 "for AMD NPT Family 0Fh Processors\" (AMD publication 32559)";
502 static char amd_fam_10h_bkdg[] = "See \"BIOS and Kernel Developer's Guide "
503 "(BKDG) For AMD Family 10h Processors\" (AMD publication 31116)";
504 static char amd_fam_11h_bkdg[] = "See \"BIOS and Kernel Developer's Guide "
505 "(BKDG) For AMD Family 11h Processors\" (AMD publication 41256)";
506 static char amd_fam_17h_zen1_reg[] = "See \"Open-Source Register Reference For "
507 "AMD Family 17h Processors Models 00h-2Fh\" (AMD publication 56255) and "
508 "amd_f17h_zen1_events(3CPC)";
509 static char amd_fam_17h_zen2_reg[] = "See \"Preliminary Processor Programming "
510 "Reference (PPR) for AMD Family 17h Model 31h, Revision B0 Processors\" "
511 "(AMD publication 55803), \"Processor Programming Reference (PPR) for AMD "
512 "Family 17h Model 71h, Revision B0 Processors\" (AMD publication 56176), and "
513 "amd_f17h_zen2_events(3CPC)";
514 static char amd_fam_19h_zen3_reg[] = "See \"Preliminary Processor Programming "
515 "Reference (PPR) for AMD Family 19h Model 01h, Revision B1 Processors Volume "
516 "1 of 2 (AMD publication 55898) and amd_f17h_zen3_events(3CPC)";
517 
518 static char amd_pcbe_impl_name[64];
519 static char *amd_pcbe_cpuref;
520 
521 
522 #define	BITS(v, u, l)   \
523 	(((v) >> (l)) & ((1 << (1 + (u) - (l))) - 1))
524 
525 static uint64_t
526 opt_pcbe_pes_addr(uint_t counter)
527 {
528 	ASSERT3U(counter, <, opd.opd_ncounters);
529 	return (PES_BASE_ADDR + counter);
530 }
531 
532 static uint64_t
533 opt_pcbe_pes_ext_addr(uint_t counter)
534 {
535 	ASSERT3U(counter, <, opd.opd_ncounters);
536 	return (PES_EXT_BASE_ADDR + 2 * counter);
537 }
538 
539 static uint64_t
540 opt_pcbe_pic_addr(uint_t counter)
541 {
542 	ASSERT3U(counter, <, opd.opd_ncounters);
543 	return (PIC_BASE_ADDR + 2 * counter);
544 }
545 
546 static uint64_t
547 opt_pcbe_pic_ext_addr(uint_t counter)
548 {
549 	ASSERT3U(counter, <, opd.opd_ncounters);
550 	return (PIC_EXT_BASE_ADDR + 2 * counter);
551 }
552 
553 static int
554 opt_pcbe_init(void)
555 {
556 	const amd_event_t		*evp;
557 	const amd_generic_event_t	*gevp;
558 
559 	amd_family = cpuid_getfamily(CPU);
560 	amd_model = cpuid_getmodel(CPU);
561 
562 	/*
563 	 * Make sure this really _is_ an Opteron or Athlon 64 system. The kernel
564 	 * loads this module based on its name in the module directory, but it
565 	 * could have been renamed.
566 	 */
567 	if ((cpuid_getvendor(CPU) != X86_VENDOR_AMD || amd_family < 0xf) &&
568 	    cpuid_getvendor(CPU) != X86_VENDOR_HYGON)
569 		return (-1);
570 
571 	if (amd_family == 0xf) {
572 		/* Some tools expect this string for family 0fh */
573 		(void) snprintf(amd_pcbe_impl_name, sizeof (amd_pcbe_impl_name),
574 		    "AMD Opteron & Athlon64");
575 	} else {
576 		(void) snprintf(amd_pcbe_impl_name, sizeof (amd_pcbe_impl_name),
577 		    "%s Family %02xh",
578 		    cpuid_getvendor(CPU) == X86_VENDOR_HYGON ? "Hygon" : "AMD",
579 		    amd_family);
580 	}
581 
582 	/*
583 	 * Determine whether or not the extended counter set is supported on
584 	 * this processor.
585 	 */
586 	if (is_x86_feature(x86_featureset, X86FSET_AMD_PCEC)) {
587 		opd.opd_ncounters = OPT_PCBE_EXT_NCOUNTERS;
588 		opd.opd_pesf = opt_pcbe_pes_ext_addr;
589 		opd.opd_picf = opt_pcbe_pic_ext_addr;
590 	} else {
591 		opd.opd_ncounters = OPT_PCBE_DEF_NCOUNTERS;
592 		opd.opd_pesf = opt_pcbe_pes_addr;
593 		opd.opd_picf = opt_pcbe_pic_addr;
594 	}
595 	opd.opd_cmask = (1 << opd.opd_ncounters) - 1;
596 
597 	/*
598 	 * Figure out processor revision here and assign appropriate
599 	 * event configuration.
600 	 */
601 
602 	if (amd_family == 0xf) {
603 		uint32_t rev;
604 
605 		rev = cpuid_getchiprev(CPU);
606 
607 		if (X86_CHIPREV_ATLEAST(rev, X86_CHIPREV_AMD_F_REV_F))
608 			amd_pcbe_cpuref = amd_fam_f_NPT_bkdg;
609 		else
610 			amd_pcbe_cpuref = amd_fam_f_rev_ae_bkdg;
611 		amd_events = family_f_events;
612 		amd_generic_events = opt_generic_events;
613 	} else if (amd_family == 0x10) {
614 		amd_pcbe_cpuref = amd_fam_10h_bkdg;
615 		amd_events = family_10h_events;
616 		amd_generic_events = family_10h_generic_events;
617 	} else if (amd_family == 0x11) {
618 		amd_pcbe_cpuref = amd_fam_11h_bkdg;
619 		amd_events = family_11h_events;
620 		amd_generic_events = opt_generic_events;
621 	} else if ((amd_family == 0x17 && amd_model <= 0x2f) ||
622 	    amd_family == 0x18) {
623 		amd_pcbe_cpuref = amd_fam_17h_zen1_reg;
624 		amd_events = opteron_pcbe_f17h_zen1_events;
625 		amd_generic_events = family_17h_zen1_papi_events;
626 	} else if (amd_family == 0x17 && amd_model >= 0x30 &&
627 	    amd_model <= 0x7f) {
628 		amd_pcbe_cpuref = amd_fam_17h_zen2_reg;
629 		amd_events = opteron_pcbe_f17h_zen2_events;
630 		amd_generic_events = family_17h_zen2_papi_events;
631 	} else if (amd_family == 0x19 && amd_model <= 0xf) {
632 		amd_pcbe_cpuref = amd_fam_19h_zen3_reg;
633 		amd_events = opteron_pcbe_f19h_zen3_events;
634 		amd_generic_events = family_19h_zen3_papi_events;
635 	} else {
636 		/*
637 		 * Different families have different meanings on events and even
638 		 * worse (like family 15h), different constraints around
639 		 * programming these values.
640 		 */
641 		return (-1);
642 	}
643 
644 	/*
645 	 * Construct event list.
646 	 *
647 	 * First pass:  Calculate size needed. We'll need an additional byte
648 	 *		for the NULL pointer during the last strcat.
649 	 *
650 	 * Second pass: Copy strings.
651 	 */
652 	for (evp = amd_events; evp->name != NULL; evp++)
653 		evlist_sz += strlen(evp->name) + 1;
654 
655 	for (gevp = amd_generic_events; gevp->name != NULL; gevp++)
656 		evlist_sz += strlen(gevp->name) + 1;
657 
658 	evlist = kmem_alloc(evlist_sz + 1, KM_SLEEP);
659 	evlist[0] = '\0';
660 
661 	for (evp = amd_events; evp->name != NULL; evp++) {
662 		(void) strcat(evlist, evp->name);
663 		(void) strcat(evlist, ",");
664 	}
665 
666 	for (gevp = amd_generic_events; gevp->name != NULL; gevp++) {
667 		(void) strcat(evlist, gevp->name);
668 		(void) strcat(evlist, ",");
669 	}
670 
671 	/*
672 	 * Remove trailing comma.
673 	 */
674 	evlist[evlist_sz - 1] = '\0';
675 
676 	return (0);
677 }
678 
679 static uint_t
680 opt_pcbe_ncounters(void)
681 {
682 	return (opd.opd_ncounters);
683 }
684 
685 static const char *
686 opt_pcbe_impl_name(void)
687 {
688 	return (amd_pcbe_impl_name);
689 }
690 
691 static const char *
692 opt_pcbe_cpuref(void)
693 {
694 
695 	return (amd_pcbe_cpuref);
696 }
697 
698 /*ARGSUSED*/
699 static char *
700 opt_pcbe_list_events(uint_t picnum)
701 {
702 	return (evlist);
703 }
704 
705 static char *
706 opt_pcbe_list_attrs(void)
707 {
708 	return ("edge,pc,inv,cmask,umask");
709 }
710 
711 static const amd_generic_event_t *
712 find_generic_event(char *name)
713 {
714 	const amd_generic_event_t	*gevp;
715 
716 	for (gevp = amd_generic_events; gevp->name != NULL; gevp++)
717 		if (strcmp(name, gevp->name) == 0)
718 			return (gevp);
719 
720 	return (NULL);
721 }
722 
723 static const amd_event_t *
724 find_event(char *name)
725 {
726 	const amd_event_t	*evp;
727 
728 	for (evp = amd_events; evp->name != NULL; evp++)
729 		if (strcmp(name, evp->name) == 0)
730 			return (evp);
731 
732 	return (NULL);
733 }
734 
735 /*ARGSUSED*/
736 static uint64_t
737 opt_pcbe_event_coverage(char *event)
738 {
739 	/*
740 	 * Check whether counter event is supported
741 	 */
742 	if (find_event(event) == NULL && find_generic_event(event) == NULL)
743 		return (0);
744 
745 	/*
746 	 * Fortunately, all counters can count all events.
747 	 */
748 	return (opd.opd_cmask);
749 }
750 
751 static uint64_t
752 opt_pcbe_overflow_bitmap(void)
753 {
754 	/*
755 	 * Unfortunately, this chip cannot detect which counter overflowed, so
756 	 * we must act as if they all did.
757 	 */
758 	return (opd.opd_cmask);
759 }
760 
761 /*ARGSUSED*/
762 static int
763 opt_pcbe_configure(uint_t picnum, char *event, uint64_t preset, uint32_t flags,
764     uint_t nattrs, kcpc_attr_t *attrs, void **data, void *token)
765 {
766 	opt_pcbe_config_t		*cfg;
767 	const amd_event_t		*evp;
768 	amd_event_t			ev_raw = { "raw", 0};
769 	const amd_generic_event_t	*gevp;
770 	int				i;
771 	uint64_t			evsel = 0, evsel_tmp = 0;
772 
773 	/*
774 	 * If we've been handed an existing configuration, we need only preset
775 	 * the counter value.
776 	 */
777 	if (*data != NULL) {
778 		cfg = *data;
779 		cfg->opt_rawpic = preset & MASK48;
780 		return (0);
781 	}
782 
783 	if (picnum >= opd.opd_ncounters)
784 		return (CPC_INVALID_PICNUM);
785 
786 	if ((evp = find_event(event)) == NULL) {
787 		if ((gevp = find_generic_event(event)) != NULL) {
788 			evp = find_event(gevp->event);
789 			ASSERT(evp != NULL);
790 
791 			if (nattrs > 0)
792 				return (CPC_ATTRIBUTE_OUT_OF_RANGE);
793 
794 			evsel |= gevp->umask << OPT_PES_UMASK_SHIFT;
795 		} else {
796 			long tmp;
797 
798 			/*
799 			 * If ddi_strtol() likes this event, use it as a raw
800 			 * event code.
801 			 */
802 			if (ddi_strtol(event, NULL, 0, &tmp) != 0)
803 				return (CPC_INVALID_EVENT);
804 
805 			ev_raw.emask = tmp;
806 			evp = &ev_raw;
807 		}
808 	}
809 
810 	/*
811 	 * Configuration of EventSelect register. While on some families
812 	 * certain bits might not be supported (e.g. Guest/Host on family
813 	 * 11h), setting these bits is harmless
814 	 */
815 
816 	/* Set GuestOnly bit to 0 and HostOnly bit to 1 */
817 	evsel &= ~OPT_PES_HOST;
818 	evsel &= ~OPT_PES_GUEST;
819 
820 	/* Set bits [35:32] for extended part of Event Select field */
821 	evsel_tmp = evp->emask & 0x0f00;
822 	evsel |= evsel_tmp << OPT_PES_EVSELHI_SHIFT;
823 
824 	evsel |= evp->emask & 0x00ff;
825 	evsel |= evp->unit << OPT_PES_UMASK_SHIFT;
826 
827 	if (flags & CPC_COUNT_USER)
828 		evsel |= OPT_PES_USR;
829 	if (flags & CPC_COUNT_SYSTEM)
830 		evsel |= OPT_PES_OS;
831 	if (flags & CPC_OVF_NOTIFY_EMT)
832 		evsel |= OPT_PES_INT;
833 
834 	for (i = 0; i < nattrs; i++) {
835 		if (strcmp(attrs[i].ka_name, "edge") == 0) {
836 			if (attrs[i].ka_val != 0)
837 				evsel |= OPT_PES_EDGE;
838 		} else if (strcmp(attrs[i].ka_name, "pc") == 0) {
839 			if (attrs[i].ka_val != 0)
840 				evsel |= OPT_PES_PC;
841 		} else if (strcmp(attrs[i].ka_name, "inv") == 0) {
842 			if (attrs[i].ka_val != 0)
843 				evsel |= OPT_PES_INV;
844 		} else if (strcmp(attrs[i].ka_name, "cmask") == 0) {
845 			if ((attrs[i].ka_val | OPT_PES_CMASK_MASK) !=
846 			    OPT_PES_CMASK_MASK)
847 				return (CPC_ATTRIBUTE_OUT_OF_RANGE);
848 			evsel |= attrs[i].ka_val << OPT_PES_CMASK_SHIFT;
849 		} else if (strcmp(attrs[i].ka_name, "umask") == 0) {
850 			if ((attrs[i].ka_val | OPT_PES_UMASK_MASK) !=
851 			    OPT_PES_UMASK_MASK)
852 				return (CPC_ATTRIBUTE_OUT_OF_RANGE);
853 			evsel |= attrs[i].ka_val << OPT_PES_UMASK_SHIFT;
854 		} else
855 			return (CPC_INVALID_ATTRIBUTE);
856 	}
857 
858 	cfg = kmem_alloc(sizeof (*cfg), KM_SLEEP);
859 
860 	cfg->opt_picno = picnum;
861 	cfg->opt_evsel = evsel;
862 	cfg->opt_rawpic = preset & MASK48;
863 
864 	*data = cfg;
865 	return (0);
866 }
867 
868 static void
869 opt_pcbe_program(void *token)
870 {
871 	opt_pcbe_config_t	*cfgs[OPT_PCBE_EXT_NCOUNTERS] = { &nullcfgs[0],
872 						&nullcfgs[1], &nullcfgs[2],
873 						&nullcfgs[3], &nullcfgs[4],
874 						&nullcfgs[5] };
875 	opt_pcbe_config_t	*pcfg = NULL;
876 	int			i;
877 	ulong_t			curcr4 = getcr4();
878 
879 	/*
880 	 * Allow nonprivileged code to read the performance counters if desired.
881 	 */
882 	if (kcpc_allow_nonpriv(token))
883 		setcr4(curcr4 | CR4_PCE);
884 	else
885 		setcr4(curcr4 & ~CR4_PCE);
886 
887 	/*
888 	 * Query kernel for all configs which will be co-programmed.
889 	 */
890 	do {
891 		pcfg = (opt_pcbe_config_t *)kcpc_next_config(token, pcfg, NULL);
892 
893 		if (pcfg != NULL) {
894 			ASSERT(pcfg->opt_picno < opd.opd_ncounters);
895 			cfgs[pcfg->opt_picno] = pcfg;
896 		}
897 	} while (pcfg != NULL);
898 
899 	/*
900 	 * Program in two loops. The first configures and presets the counter,
901 	 * and the second loop enables the counters. This ensures that the
902 	 * counters are all enabled as closely together in time as possible.
903 	 */
904 
905 	for (i = 0; i < opd.opd_ncounters; i++) {
906 		wrmsr(opd.opd_pesf(i), cfgs[i]->opt_evsel);
907 		wrmsr(opd.opd_picf(i), cfgs[i]->opt_rawpic);
908 	}
909 
910 	for (i = 0; i < opd.opd_ncounters; i++) {
911 		wrmsr(opd.opd_pesf(i), cfgs[i]->opt_evsel |
912 		    (uint64_t)(uintptr_t)OPT_PES_ENABLE);
913 	}
914 }
915 
916 static void
917 opt_pcbe_allstop(void)
918 {
919 	int		i;
920 
921 	for (i = 0; i < opd.opd_ncounters; i++)
922 		wrmsr(opd.opd_pesf(i), 0ULL);
923 
924 	/*
925 	 * Disable non-privileged access to the counter registers.
926 	 */
927 	setcr4(getcr4() & ~CR4_PCE);
928 }
929 
930 static void
931 opt_pcbe_sample(void *token)
932 {
933 	opt_pcbe_config_t	*cfgs[OPT_PCBE_EXT_NCOUNTERS] = { NULL, NULL,
934 						NULL, NULL, NULL, NULL };
935 	opt_pcbe_config_t	*pcfg = NULL;
936 	int			i;
937 	uint64_t		curpic[OPT_PCBE_EXT_NCOUNTERS];
938 	uint64_t		*addrs[OPT_PCBE_EXT_NCOUNTERS];
939 	uint64_t		*tmp;
940 	int64_t			diff;
941 
942 	for (i = 0; i < opd.opd_ncounters; i++)
943 		curpic[i] = rdmsr(opd.opd_picf(i));
944 
945 	/*
946 	 * Query kernel for all configs which are co-programmed.
947 	 */
948 	do {
949 		pcfg = (opt_pcbe_config_t *)kcpc_next_config(token, pcfg, &tmp);
950 
951 		if (pcfg != NULL) {
952 			ASSERT3U(pcfg->opt_picno, <, opd.opd_ncounters);
953 			cfgs[pcfg->opt_picno] = pcfg;
954 			addrs[pcfg->opt_picno] = tmp;
955 		}
956 	} while (pcfg != NULL);
957 
958 	for (i = 0; i < opd.opd_ncounters; i++) {
959 		if (cfgs[i] == NULL)
960 			continue;
961 
962 		diff = (curpic[i] - cfgs[i]->opt_rawpic) & MASK48;
963 		*addrs[i] += diff;
964 		DTRACE_PROBE4(opt__pcbe__sample, int, i, uint64_t, *addrs[i],
965 		    uint64_t, curpic[i], uint64_t, cfgs[i]->opt_rawpic);
966 		cfgs[i]->opt_rawpic = *addrs[i] & MASK48;
967 	}
968 }
969 
970 static void
971 opt_pcbe_free(void *config)
972 {
973 	kmem_free(config, sizeof (opt_pcbe_config_t));
974 }
975 
976 
977 static struct modlpcbe modlpcbe = {
978 	&mod_pcbeops,
979 	"AMD Performance Counters",
980 	&opt_pcbe_ops
981 };
982 
983 static struct modlinkage modl = {
984 	MODREV_1,
985 	&modlpcbe,
986 };
987 
988 int
989 _init(void)
990 {
991 	int ret;
992 
993 	if (opt_pcbe_init() != 0)
994 		return (ENOTSUP);
995 
996 	if ((ret = mod_install(&modl)) != 0)
997 		kmem_free(evlist, evlist_sz + 1);
998 
999 	return (ret);
1000 }
1001 
1002 int
1003 _fini(void)
1004 {
1005 	int ret;
1006 
1007 	if ((ret = mod_remove(&modl)) == 0)
1008 		kmem_free(evlist, evlist_sz + 1);
1009 	return (ret);
1010 }
1011 
1012 int
1013 _info(struct modinfo *mi)
1014 {
1015 	return (mod_info(&modl, mi));
1016 }
1017