xref: /linux/arch/powerpc/include/asm/plpar_wrappers.h (revision 8dd765a5d769c521d73931850d1c8708fbc490cb)
1 /* SPDX-License-Identifier: GPL-2.0 */
2 #ifndef _ASM_POWERPC_PLPAR_WRAPPERS_H
3 #define _ASM_POWERPC_PLPAR_WRAPPERS_H
4 
5 #ifdef CONFIG_PPC_PSERIES
6 
7 #include <linux/string.h>
8 #include <linux/irqflags.h>
9 
10 #include <asm/hvcall.h>
11 #include <asm/paca.h>
12 #include <asm/lppaca.h>
13 #include <asm/page.h>
14 
15 static inline long poll_pending(void)
16 {
17 	return plpar_hcall_norets(H_POLL_PENDING);
18 }
19 
20 static inline u8 get_cede_latency_hint(void)
21 {
22 	return get_lppaca()->cede_latency_hint;
23 }
24 
25 static inline void set_cede_latency_hint(u8 latency_hint)
26 {
27 	get_lppaca()->cede_latency_hint = latency_hint;
28 }
29 
30 static inline long cede_processor(void)
31 {
32 	/*
33 	 * We cannot call tracepoints inside RCU idle regions which
34 	 * means we must not trace H_CEDE.
35 	 */
36 	return plpar_hcall_norets_notrace(H_CEDE);
37 }
38 
39 static inline long extended_cede_processor(unsigned long latency_hint)
40 {
41 	long rc;
42 	u8 old_latency_hint = get_cede_latency_hint();
43 
44 	set_cede_latency_hint(latency_hint);
45 
46 	rc = cede_processor();
47 
48 	/* Ensure that H_CEDE returns with IRQs on */
49 	if (WARN_ON(IS_ENABLED(CONFIG_PPC_IRQ_SOFT_MASK_DEBUG) && !(mfmsr() & MSR_EE)))
50 		__hard_irq_enable();
51 
52 	set_cede_latency_hint(old_latency_hint);
53 
54 	return rc;
55 }
56 
57 static inline long vpa_call(unsigned long flags, unsigned long cpu,
58 		unsigned long vpa)
59 {
60 	flags = flags << H_VPA_FUNC_SHIFT;
61 
62 	return plpar_hcall_norets(H_REGISTER_VPA, flags, cpu, vpa);
63 }
64 
65 static inline long unregister_vpa(unsigned long cpu)
66 {
67 	return vpa_call(H_VPA_DEREG_VPA, cpu, 0);
68 }
69 
70 static inline long register_vpa(unsigned long cpu, unsigned long vpa)
71 {
72 	return vpa_call(H_VPA_REG_VPA, cpu, vpa);
73 }
74 
75 static inline long unregister_slb_shadow(unsigned long cpu)
76 {
77 	return vpa_call(H_VPA_DEREG_SLB, cpu, 0);
78 }
79 
80 static inline long register_slb_shadow(unsigned long cpu, unsigned long vpa)
81 {
82 	return vpa_call(H_VPA_REG_SLB, cpu, vpa);
83 }
84 
85 static inline long unregister_dtl(unsigned long cpu)
86 {
87 	return vpa_call(H_VPA_DEREG_DTL, cpu, 0);
88 }
89 
90 static inline long register_dtl(unsigned long cpu, unsigned long vpa)
91 {
92 	return vpa_call(H_VPA_REG_DTL, cpu, vpa);
93 }
94 
95 extern void vpa_init(int cpu);
96 
97 static inline long plpar_pte_enter(unsigned long flags,
98 		unsigned long hpte_group, unsigned long hpte_v,
99 		unsigned long hpte_r, unsigned long *slot)
100 {
101 	long rc;
102 	unsigned long retbuf[PLPAR_HCALL_BUFSIZE];
103 
104 	rc = plpar_hcall(H_ENTER, retbuf, flags, hpte_group, hpte_v, hpte_r);
105 
106 	*slot = retbuf[0];
107 
108 	return rc;
109 }
110 
111 static inline long plpar_pte_remove(unsigned long flags, unsigned long ptex,
112 		unsigned long avpn, unsigned long *old_pteh_ret,
113 		unsigned long *old_ptel_ret)
114 {
115 	long rc;
116 	unsigned long retbuf[PLPAR_HCALL_BUFSIZE];
117 
118 	rc = plpar_hcall(H_REMOVE, retbuf, flags, ptex, avpn);
119 
120 	*old_pteh_ret = retbuf[0];
121 	*old_ptel_ret = retbuf[1];
122 
123 	return rc;
124 }
125 
126 /* plpar_pte_remove_raw can be called in real mode. It calls plpar_hcall_raw */
127 static inline long plpar_pte_remove_raw(unsigned long flags, unsigned long ptex,
128 		unsigned long avpn, unsigned long *old_pteh_ret,
129 		unsigned long *old_ptel_ret)
130 {
131 	long rc;
132 	unsigned long retbuf[PLPAR_HCALL_BUFSIZE];
133 
134 	rc = plpar_hcall_raw(H_REMOVE, retbuf, flags, ptex, avpn);
135 
136 	*old_pteh_ret = retbuf[0];
137 	*old_ptel_ret = retbuf[1];
138 
139 	return rc;
140 }
141 
142 static inline long plpar_pte_read(unsigned long flags, unsigned long ptex,
143 		unsigned long *old_pteh_ret, unsigned long *old_ptel_ret)
144 {
145 	long rc;
146 	unsigned long retbuf[PLPAR_HCALL_BUFSIZE];
147 
148 	rc = plpar_hcall(H_READ, retbuf, flags, ptex);
149 
150 	*old_pteh_ret = retbuf[0];
151 	*old_ptel_ret = retbuf[1];
152 
153 	return rc;
154 }
155 
156 /* plpar_pte_read_raw can be called in real mode. It calls plpar_hcall_raw */
157 static inline long plpar_pte_read_raw(unsigned long flags, unsigned long ptex,
158 		unsigned long *old_pteh_ret, unsigned long *old_ptel_ret)
159 {
160 	long rc;
161 	unsigned long retbuf[PLPAR_HCALL_BUFSIZE];
162 
163 	rc = plpar_hcall_raw(H_READ, retbuf, flags, ptex);
164 
165 	*old_pteh_ret = retbuf[0];
166 	*old_ptel_ret = retbuf[1];
167 
168 	return rc;
169 }
170 
171 /*
172  * ptes must be 8*sizeof(unsigned long)
173  */
174 static inline long plpar_pte_read_4(unsigned long flags, unsigned long ptex,
175 				    unsigned long *ptes)
176 
177 {
178 	long rc;
179 	unsigned long retbuf[PLPAR_HCALL9_BUFSIZE];
180 
181 	rc = plpar_hcall9(H_READ, retbuf, flags | H_READ_4, ptex);
182 
183 	memcpy(ptes, retbuf, 8*sizeof(unsigned long));
184 
185 	return rc;
186 }
187 
188 /*
189  * plpar_pte_read_4_raw can be called in real mode.
190  * ptes must be 8*sizeof(unsigned long)
191  */
192 static inline long plpar_pte_read_4_raw(unsigned long flags, unsigned long ptex,
193 					unsigned long *ptes)
194 
195 {
196 	long rc;
197 	unsigned long retbuf[PLPAR_HCALL9_BUFSIZE];
198 
199 	rc = plpar_hcall9_raw(H_READ, retbuf, flags | H_READ_4, ptex);
200 
201 	memcpy(ptes, retbuf, 8*sizeof(unsigned long));
202 
203 	return rc;
204 }
205 
206 static inline long plpar_pte_protect(unsigned long flags, unsigned long ptex,
207 		unsigned long avpn)
208 {
209 	return plpar_hcall_norets(H_PROTECT, flags, ptex, avpn);
210 }
211 
212 static inline long plpar_resize_hpt_prepare(unsigned long flags,
213 					    unsigned long shift)
214 {
215 	return plpar_hcall_norets(H_RESIZE_HPT_PREPARE, flags, shift);
216 }
217 
218 static inline long plpar_resize_hpt_commit(unsigned long flags,
219 					   unsigned long shift)
220 {
221 	return plpar_hcall_norets(H_RESIZE_HPT_COMMIT, flags, shift);
222 }
223 
224 static inline long plpar_tce_get(unsigned long liobn, unsigned long ioba,
225 		unsigned long *tce_ret)
226 {
227 	long rc;
228 	unsigned long retbuf[PLPAR_HCALL_BUFSIZE];
229 
230 	rc = plpar_hcall(H_GET_TCE, retbuf, liobn, ioba);
231 
232 	*tce_ret = retbuf[0];
233 
234 	return rc;
235 }
236 
237 static inline long plpar_tce_put(unsigned long liobn, unsigned long ioba,
238 		unsigned long tceval)
239 {
240 	return plpar_hcall_norets(H_PUT_TCE, liobn, ioba, tceval);
241 }
242 
243 static inline long plpar_tce_put_indirect(unsigned long liobn,
244 		unsigned long ioba, unsigned long page, unsigned long count)
245 {
246 	return plpar_hcall_norets(H_PUT_TCE_INDIRECT, liobn, ioba, page, count);
247 }
248 
249 static inline long plpar_tce_stuff(unsigned long liobn, unsigned long ioba,
250 		unsigned long tceval, unsigned long count)
251 {
252 	return plpar_hcall_norets(H_STUFF_TCE, liobn, ioba, tceval, count);
253 }
254 
255 /* Set various resource mode parameters */
256 static inline long plpar_set_mode(unsigned long mflags, unsigned long resource,
257 		unsigned long value1, unsigned long value2)
258 {
259 	return plpar_hcall_norets(H_SET_MODE, mflags, resource, value1, value2);
260 }
261 
262 /*
263  * Enable relocation on exceptions on this partition
264  *
265  * Note: this call has a partition wide scope and can take a while to complete.
266  * If it returns H_LONG_BUSY_* it should be retried periodically until it
267  * returns H_SUCCESS.
268  */
269 static inline long enable_reloc_on_exceptions(void)
270 {
271 	/* mflags = 3: Exceptions at 0xC000000000004000 */
272 	return plpar_set_mode(3, H_SET_MODE_RESOURCE_ADDR_TRANS_MODE, 0, 0);
273 }
274 
275 /*
276  * Disable relocation on exceptions on this partition
277  *
278  * Note: this call has a partition wide scope and can take a while to complete.
279  * If it returns H_LONG_BUSY_* it should be retried periodically until it
280  * returns H_SUCCESS.
281  */
282 static inline long disable_reloc_on_exceptions(void) {
283 	return plpar_set_mode(0, H_SET_MODE_RESOURCE_ADDR_TRANS_MODE, 0, 0);
284 }
285 
286 /*
287  * Take exceptions in big endian mode on this partition
288  *
289  * Note: this call has a partition wide scope and can take a while to complete.
290  * If it returns H_LONG_BUSY_* it should be retried periodically until it
291  * returns H_SUCCESS.
292  */
293 static inline long enable_big_endian_exceptions(void)
294 {
295 	/* mflags = 0: big endian exceptions */
296 	return plpar_set_mode(0, H_SET_MODE_RESOURCE_LE, 0, 0);
297 }
298 
299 /*
300  * Take exceptions in little endian mode on this partition
301  *
302  * Note: this call has a partition wide scope and can take a while to complete.
303  * If it returns H_LONG_BUSY_* it should be retried periodically until it
304  * returns H_SUCCESS.
305  */
306 static inline long enable_little_endian_exceptions(void)
307 {
308 	/* mflags = 1: little endian exceptions */
309 	return plpar_set_mode(1, H_SET_MODE_RESOURCE_LE, 0, 0);
310 }
311 
312 static inline long plpar_set_ciabr(unsigned long ciabr)
313 {
314 	return plpar_set_mode(0, H_SET_MODE_RESOURCE_SET_CIABR, ciabr, 0);
315 }
316 
317 static inline long plpar_set_watchpoint0(unsigned long dawr0, unsigned long dawrx0)
318 {
319 	return plpar_set_mode(0, H_SET_MODE_RESOURCE_SET_DAWR0, dawr0, dawrx0);
320 }
321 
322 static inline long plpar_set_watchpoint1(unsigned long dawr1, unsigned long dawrx1)
323 {
324 	return plpar_set_mode(0, H_SET_MODE_RESOURCE_SET_DAWR1, dawr1, dawrx1);
325 }
326 
327 static inline long plpar_signal_sys_reset(long cpu)
328 {
329 	return plpar_hcall_norets(H_SIGNAL_SYS_RESET, cpu);
330 }
331 
332 static inline long plpar_get_cpu_characteristics(struct h_cpu_char_result *p)
333 {
334 	unsigned long retbuf[PLPAR_HCALL_BUFSIZE];
335 	long rc;
336 
337 	rc = plpar_hcall(H_GET_CPU_CHARACTERISTICS, retbuf);
338 	if (rc == H_SUCCESS) {
339 		p->character = retbuf[0];
340 		p->behaviour = retbuf[1];
341 	}
342 
343 	return rc;
344 }
345 
346 /*
347  * Wrapper to H_RPT_INVALIDATE hcall that handles return values appropriately
348  *
349  * - Returns H_SUCCESS on success
350  * - For H_BUSY return value, we retry the hcall.
351  * - For any other hcall failures, attempt a full flush once before
352  *   resorting to BUG().
353  *
354  * Note: This hcall is expected to fail only very rarely. The correct
355  * error recovery of killing the process/guest will be eventually
356  * needed.
357  */
358 static inline long pseries_rpt_invalidate(u32 pid, u64 target, u64 type,
359 					  u64 page_sizes, u64 start, u64 end)
360 {
361 	long rc;
362 	unsigned long all;
363 
364 	while (true) {
365 		rc = plpar_hcall_norets(H_RPT_INVALIDATE, pid, target, type,
366 					page_sizes, start, end);
367 		if (rc == H_BUSY) {
368 			cpu_relax();
369 			continue;
370 		} else if (rc == H_SUCCESS)
371 			return rc;
372 
373 		/* Flush request failed, try with a full flush once */
374 		if (type & H_RPTI_TYPE_NESTED)
375 			all = H_RPTI_TYPE_NESTED | H_RPTI_TYPE_NESTED_ALL;
376 		else
377 			all = H_RPTI_TYPE_ALL;
378 retry:
379 		rc = plpar_hcall_norets(H_RPT_INVALIDATE, pid, target,
380 					all, page_sizes, 0, -1UL);
381 		if (rc == H_BUSY) {
382 			cpu_relax();
383 			goto retry;
384 		} else if (rc == H_SUCCESS)
385 			return rc;
386 
387 		BUG();
388 	}
389 }
390 
391 #else /* !CONFIG_PPC_PSERIES */
392 
393 static inline long plpar_set_ciabr(unsigned long ciabr)
394 {
395 	return 0;
396 }
397 
398 static inline long plpar_pte_read_4(unsigned long flags, unsigned long ptex,
399 				    unsigned long *ptes)
400 {
401 	return 0;
402 }
403 
404 static inline long pseries_rpt_invalidate(u32 pid, u64 target, u64 type,
405 					  u64 page_sizes, u64 start, u64 end)
406 {
407 	return 0;
408 }
409 
410 #endif /* CONFIG_PPC_PSERIES */
411 
412 #endif /* _ASM_POWERPC_PLPAR_WRAPPERS_H */
413