xref: /illumos-gate/usr/src/uts/common/xen/io/evtchn_dev.c (revision 2b208da6458bdbaad9f45832216da636f9184765)
1 /*
2  * CDDL HEADER START
3  *
4  * The contents of this file are subject to the terms of the
5  * Common Development and Distribution License (the "License").
6  * You may not use this file except in compliance with the License.
7  *
8  * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
9  * or http://www.opensolaris.org/os/licensing.
10  * See the License for the specific language governing permissions
11  * and limitations under the License.
12  *
13  * When distributing Covered Code, include this CDDL HEADER in each
14  * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
15  * If applicable, add the following below this CDDL HEADER, with the
16  * fields enclosed by brackets "[]" replaced with your own identifying
17  * information: Portions Copyright [yyyy] [name of copyright owner]
18  *
19  * CDDL HEADER END
20  */
21 
22 /*
23  * Copyright 2009 Sun Microsystems, Inc.  All rights reserved.
24  * Use is subject to license terms.
25  * Copyright 2017 Joyent, Inc.
26  */
27 
28 
29 /*
30  * evtchn.c
31  *
32  * Driver for receiving and demuxing event-channel signals.
33  *
34  * Copyright (c) 2004-2005, K A Fraser
35  * Multi-process extensions Copyright (c) 2004, Steven Smith
36  *
37  * This file may be distributed separately from the Linux kernel, or
38  * incorporated into other software packages, subject to the following license:
39  *
40  * Permission is hereby granted, free of charge, to any person obtaining a copy
41  * of this source file (the "Software"), to deal in the Software without
42  * restriction, including without limitation the rights to use, copy, modify,
43  * merge, publish, distribute, sublicense, and/or sell copies of the Software,
44  * and to permit persons to whom the Software is furnished to do so, subject to
45  * the following conditions:
46  *
47  * The above copyright notice and this permission notice shall be included in
48  * all copies or substantial portions of the Software.
49  *
50  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
51  * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
52  * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
53  * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
54  * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
55  * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
56  * IN THE SOFTWARE.
57  */
58 
59 #include <sys/types.h>
60 #include <sys/hypervisor.h>
61 #include <sys/machsystm.h>
62 #include <sys/mutex.h>
63 #include <sys/evtchn_impl.h>
64 #include <sys/ddi_impldefs.h>
65 #include <sys/avintr.h>
66 #include <sys/cpuvar.h>
67 #include <sys/smp_impldefs.h>
68 #include <sys/archsystm.h>
69 #include <sys/sysmacros.h>
70 #include <sys/fcntl.h>
71 #include <sys/open.h>
72 #include <sys/stat.h>
73 #include <sys/psm.h>
74 #include <sys/cpu.h>
75 #include <sys/cmn_err.h>
76 #include <sys/xen_errno.h>
77 #include <sys/policy.h>
78 #include <xen/sys/evtchn.h>
79 
80 /* Some handy macros */
81 #define	EVTCHNDRV_MINOR2INST(minor)	((int)(minor))
82 #define	EVTCHNDRV_DEFAULT_NCLONES 	256
83 #define	EVTCHNDRV_INST2SOFTS(inst)	\
84 	(ddi_get_soft_state(evtchndrv_statep, (inst)))
85 
86 /* Soft state data structure for evtchn driver */
87 struct evtsoftdata {
88 	dev_info_t *dip;
89 	/* Notification ring, accessed via /dev/xen/evtchn. */
90 #define	EVTCHN_RING_SIZE	(PAGESIZE / sizeof (evtchn_port_t))
91 #define	EVTCHN_RING_MASK(_i)	((_i) & (EVTCHN_RING_SIZE - 1))
92 	evtchn_port_t *ring;
93 	unsigned int ring_cons, ring_prod, ring_overflow;
94 
95 	kcondvar_t evtchn_wait; /* Processes wait on this when ring is empty. */
96 	kmutex_t evtchn_lock;
97 	struct pollhead evtchn_pollhead;
98 
99 	pid_t pid;		/* last pid to bind to this event channel. */
100 	processorid_t cpu;	/* cpu thread/evtchn is bound to */
101 };
102 
103 static void *evtchndrv_statep;
104 int evtchndrv_nclones = EVTCHNDRV_DEFAULT_NCLONES;
105 static int *evtchndrv_clone_tab;
106 static dev_info_t *evtchndrv_dip;
107 static kmutex_t evtchndrv_clone_tab_mutex;
108 
109 static int evtchndrv_detach(dev_info_t *, ddi_detach_cmd_t);
110 
111 /* Who's bound to each port? */
112 static struct evtsoftdata *port_user[NR_EVENT_CHANNELS];
113 static kmutex_t port_user_lock;
114 
115 void
116 evtchn_device_upcall()
117 {
118 	struct evtsoftdata *ep;
119 	int port;
120 
121 	/*
122 	 * This is quite gross, we had to leave the evtchn that led to this
123 	 * invocation in a per-cpu mailbox, retrieve it now.
124 	 * We do this because the interface doesn't offer us a way to pass
125 	 * a dynamic argument up through the generic interrupt service layer.
126 	 * The mailbox is safe since we either run with interrupts disabled or
127 	 * non-preemptable till we reach here.
128 	 */
129 	port = CPU->cpu_m.mcpu_ec_mbox;
130 	ASSERT(port != 0);
131 	CPU->cpu_m.mcpu_ec_mbox = 0;
132 	ec_clear_evtchn(port);
133 	mutex_enter(&port_user_lock);
134 
135 	if ((ep = port_user[port]) != NULL) {
136 		mutex_enter(&ep->evtchn_lock);
137 		if ((ep->ring_prod - ep->ring_cons) < EVTCHN_RING_SIZE) {
138 			ep->ring[EVTCHN_RING_MASK(ep->ring_prod)] = port;
139 			/*
140 			 * Wake up reader when ring goes non-empty
141 			 */
142 			if (ep->ring_cons == ep->ring_prod++) {
143 				cv_signal(&ep->evtchn_wait);
144 				mutex_exit(&ep->evtchn_lock);
145 				pollwakeup(&ep->evtchn_pollhead,
146 				    POLLIN | POLLRDNORM);
147 				goto done;
148 			}
149 		} else {
150 			ep->ring_overflow = 1;
151 		}
152 		mutex_exit(&ep->evtchn_lock);
153 	}
154 
155 done:
156 	mutex_exit(&port_user_lock);
157 }
158 
159 /* ARGSUSED */
160 static int
161 evtchndrv_read(dev_t dev, struct uio *uio, cred_t *cr)
162 {
163 	int rc = 0;
164 	ssize_t count;
165 	unsigned int c, p, bytes1 = 0, bytes2 = 0;
166 	struct evtsoftdata *ep;
167 	minor_t minor = getminor(dev);
168 
169 	if (secpolicy_xvm_control(cr))
170 		return (EPERM);
171 
172 	ep = EVTCHNDRV_INST2SOFTS(EVTCHNDRV_MINOR2INST(minor));
173 
174 	/* Whole number of ports. */
175 	count = uio->uio_resid;
176 	count &= ~(sizeof (evtchn_port_t) - 1);
177 
178 	if (count == 0)
179 		return (0);
180 
181 	if (count > PAGESIZE)
182 		count = PAGESIZE;
183 
184 	mutex_enter(&ep->evtchn_lock);
185 	for (;;) {
186 		if (ep->ring_overflow) {
187 			rc = EFBIG;
188 			goto done;
189 		}
190 
191 		if ((c = ep->ring_cons) != (p = ep->ring_prod))
192 			break;
193 
194 		if (uio->uio_fmode & O_NONBLOCK) {
195 			rc = EAGAIN;
196 			goto done;
197 		}
198 
199 		if (cv_wait_sig(&ep->evtchn_wait, &ep->evtchn_lock) == 0) {
200 			rc = EINTR;
201 			goto done;
202 		}
203 	}
204 
205 	/* Byte lengths of two chunks. Chunk split (if any) is at ring wrap. */
206 	if (((c ^ p) & EVTCHN_RING_SIZE) != 0) {
207 		bytes1 = (EVTCHN_RING_SIZE - EVTCHN_RING_MASK(c)) *
208 		    sizeof (evtchn_port_t);
209 		bytes2 = EVTCHN_RING_MASK(p) * sizeof (evtchn_port_t);
210 	} else {
211 		bytes1 = (p - c) * sizeof (evtchn_port_t);
212 		bytes2 = 0;
213 	}
214 
215 	/* Truncate chunks according to caller's maximum byte count. */
216 	if (bytes1 > count) {
217 		bytes1 = count;
218 		bytes2 = 0;
219 	} else if ((bytes1 + bytes2) > count) {
220 		bytes2 = count - bytes1;
221 	}
222 
223 	if (uiomove(&ep->ring[EVTCHN_RING_MASK(c)], bytes1, UIO_READ, uio) ||
224 	    ((bytes2 != 0) && uiomove(&ep->ring[0], bytes2, UIO_READ, uio))) {
225 		rc = EFAULT;
226 		goto done;
227 	}
228 
229 	ep->ring_cons += (bytes1 + bytes2) / sizeof (evtchn_port_t);
230 done:
231 	mutex_exit(&ep->evtchn_lock);
232 	return (rc);
233 }
234 
235 /* ARGSUSED */
236 static int
237 evtchndrv_write(dev_t dev, struct uio *uio, cred_t *cr)
238 {
239 	int  rc, i;
240 	ssize_t count;
241 	evtchn_port_t *kbuf;
242 	struct evtsoftdata *ep;
243 	ulong_t flags;
244 	minor_t minor = getminor(dev);
245 	evtchn_port_t sbuf[32];
246 
247 	if (secpolicy_xvm_control(cr))
248 		return (EPERM);
249 
250 	ep = EVTCHNDRV_INST2SOFTS(EVTCHNDRV_MINOR2INST(minor));
251 
252 
253 	/* Whole number of ports. */
254 	count = uio->uio_resid;
255 	count &= ~(sizeof (evtchn_port_t) - 1);
256 
257 	if (count == 0)
258 		return (0);
259 
260 	if (count > PAGESIZE)
261 		count = PAGESIZE;
262 
263 	if (count <= sizeof (sbuf))
264 		kbuf = sbuf;
265 	else
266 		kbuf = kmem_alloc(PAGESIZE, KM_SLEEP);
267 	if ((rc = uiomove(kbuf, count, UIO_WRITE, uio)) != 0)
268 		goto out;
269 
270 	mutex_enter(&port_user_lock);
271 	for (i = 0; i < (count / sizeof (evtchn_port_t)); i++)
272 		if ((kbuf[i] < NR_EVENT_CHANNELS) &&
273 		    (port_user[kbuf[i]] == ep)) {
274 			flags = intr_clear();
275 			ec_unmask_evtchn(kbuf[i]);
276 			intr_restore(flags);
277 		}
278 	mutex_exit(&port_user_lock);
279 
280 out:
281 	if (kbuf != sbuf)
282 		kmem_free(kbuf, PAGESIZE);
283 	return (rc);
284 }
285 
286 static void
287 evtchn_bind_to_user(struct evtsoftdata *u, int port)
288 {
289 	ulong_t flags;
290 
291 	/*
292 	 * save away the PID of the last process to bind to this event channel.
293 	 * Useful for debugging.
294 	 */
295 	u->pid = ddi_get_pid();
296 
297 	mutex_enter(&port_user_lock);
298 	ASSERT(port_user[port] == NULL);
299 	port_user[port] = u;
300 	ec_irq_add_evtchn(ec_dev_irq, port);
301 	flags = intr_clear();
302 	ec_unmask_evtchn(port);
303 	intr_restore(flags);
304 	mutex_exit(&port_user_lock);
305 }
306 
307 static void
308 evtchndrv_close_evtchn(int port)
309 {
310 	struct evtsoftdata *ep;
311 
312 	ASSERT(MUTEX_HELD(&port_user_lock));
313 	ep = port_user[port];
314 	ASSERT(ep != NULL);
315 	(void) ec_mask_evtchn(port);
316 	/*
317 	 * It is possible the event is in transit to us.
318 	 * If it is already in the ring buffer, then a client may
319 	 * get a spurious event notification on the next read of
320 	 * of the evtchn device.  Clients will need to be able to
321 	 * handle getting a spurious event notification.
322 	 */
323 	port_user[port] = NULL;
324 	/*
325 	 * The event is masked and should stay so, clean it up.
326 	 */
327 	ec_irq_rm_evtchn(ec_dev_irq, port);
328 }
329 
330 /* ARGSUSED */
331 static int
332 evtchndrv_ioctl(dev_t dev, int cmd, intptr_t data, int flag, cred_t *cr,
333     int *rvalp)
334 {
335 	int err = 0;
336 	struct evtsoftdata *ep;
337 	minor_t minor = getminor(dev);
338 
339 	if (secpolicy_xvm_control(cr))
340 		return (EPERM);
341 
342 	ep = EVTCHNDRV_INST2SOFTS(EVTCHNDRV_MINOR2INST(minor));
343 
344 	*rvalp = 0;
345 
346 	switch (cmd) {
347 	case IOCTL_EVTCHN_BIND_VIRQ: {
348 		struct ioctl_evtchn_bind_virq bind;
349 
350 		if (copyin((void *)data, &bind, sizeof (bind))) {
351 			err = EFAULT;
352 			break;
353 		}
354 
355 		if ((err = xen_bind_virq(bind.virq, 0, rvalp)) != 0)
356 			break;
357 
358 		evtchn_bind_to_user(ep, *rvalp);
359 		break;
360 	}
361 
362 	case IOCTL_EVTCHN_BIND_INTERDOMAIN: {
363 		struct ioctl_evtchn_bind_interdomain bind;
364 
365 		if (copyin((void *)data, &bind, sizeof (bind))) {
366 			err = EFAULT;
367 			break;
368 		}
369 
370 		if ((err = xen_bind_interdomain(bind.remote_domain,
371 		    bind.remote_port, rvalp)) != 0)
372 			break;
373 
374 		ec_bind_vcpu(*rvalp, 0);
375 		evtchn_bind_to_user(ep, *rvalp);
376 		break;
377 	}
378 
379 	case IOCTL_EVTCHN_BIND_UNBOUND_PORT: {
380 		struct ioctl_evtchn_bind_unbound_port bind;
381 
382 		if (copyin((void *)data, &bind, sizeof (bind))) {
383 			err = EFAULT;
384 			break;
385 		}
386 
387 		if ((err = xen_alloc_unbound_evtchn(bind.remote_domain,
388 		    rvalp)) != 0)
389 			break;
390 
391 		evtchn_bind_to_user(ep, *rvalp);
392 		break;
393 	}
394 
395 	case IOCTL_EVTCHN_UNBIND: {
396 		struct ioctl_evtchn_unbind unbind;
397 
398 		if (copyin((void *)data, &unbind, sizeof (unbind))) {
399 			err = EFAULT;
400 			break;
401 		}
402 
403 		if (unbind.port >= NR_EVENT_CHANNELS) {
404 			err = EFAULT;
405 			break;
406 		}
407 
408 		mutex_enter(&port_user_lock);
409 
410 		if (port_user[unbind.port] != ep) {
411 			mutex_exit(&port_user_lock);
412 			err = ENOTCONN;
413 			break;
414 		}
415 
416 		evtchndrv_close_evtchn(unbind.port);
417 		mutex_exit(&port_user_lock);
418 		break;
419 	}
420 
421 	case IOCTL_EVTCHN_NOTIFY: {
422 		struct ioctl_evtchn_notify notify;
423 
424 		if (copyin((void *)data, &notify, sizeof (notify))) {
425 			err = EFAULT;
426 			break;
427 		}
428 
429 		if (notify.port >= NR_EVENT_CHANNELS) {
430 			err = EINVAL;
431 		} else if (port_user[notify.port] != ep) {
432 			err = ENOTCONN;
433 		} else {
434 			ec_notify_via_evtchn(notify.port);
435 		}
436 		break;
437 	}
438 
439 	default:
440 		err = ENOSYS;
441 	}
442 
443 	return (err);
444 }
445 
446 static int
447 evtchndrv_poll(dev_t dev, short ev, int anyyet, short *revp, pollhead_t **phpp)
448 {
449 	struct evtsoftdata *ep;
450 	minor_t minor = getminor(dev);
451 	short mask = 0;
452 
453 	ep = EVTCHNDRV_INST2SOFTS(EVTCHNDRV_MINOR2INST(minor));
454 
455 	if (ev & POLLOUT)
456 		mask |= POLLOUT;
457 	if (ep->ring_overflow)
458 		mask |= POLLERR;
459 	if (ev & (POLLIN | POLLRDNORM)) {
460 		mutex_enter(&ep->evtchn_lock);
461 		if (ep->ring_cons != ep->ring_prod) {
462 			mask |= (POLLIN | POLLRDNORM) & ev;
463 		}
464 		mutex_exit(&ep->evtchn_lock);
465 	}
466 	if ((mask == 0 && !anyyet) || (ev & POLLET)) {
467 		*phpp = &ep->evtchn_pollhead;
468 	}
469 	*revp = mask;
470 	return (0);
471 }
472 
473 
474 /* ARGSUSED */
475 static int
476 evtchndrv_open(dev_t *devp, int flag, int otyp, cred_t *credp)
477 {
478 	struct evtsoftdata *ep;
479 	minor_t minor = getminor(*devp);
480 
481 	if (otyp == OTYP_BLK)
482 		return (ENXIO);
483 
484 	/*
485 	 * only allow open on minor = 0 - the clone device
486 	 */
487 	if (minor != 0)
488 		return (ENXIO);
489 
490 	/*
491 	 * find a free slot and grab it
492 	 */
493 	mutex_enter(&evtchndrv_clone_tab_mutex);
494 	for (minor = 1; minor < evtchndrv_nclones; minor++) {
495 		if (evtchndrv_clone_tab[minor] == 0) {
496 			evtchndrv_clone_tab[minor] = 1;
497 			break;
498 		}
499 	}
500 	mutex_exit(&evtchndrv_clone_tab_mutex);
501 	if (minor == evtchndrv_nclones)
502 		return (EAGAIN);
503 
504 	/* Allocate softstate structure */
505 	if (ddi_soft_state_zalloc(evtchndrv_statep,
506 	    EVTCHNDRV_MINOR2INST(minor)) != DDI_SUCCESS) {
507 		mutex_enter(&evtchndrv_clone_tab_mutex);
508 		evtchndrv_clone_tab[minor] = 0;
509 		mutex_exit(&evtchndrv_clone_tab_mutex);
510 		return (EAGAIN);
511 	}
512 	ep = EVTCHNDRV_INST2SOFTS(EVTCHNDRV_MINOR2INST(minor));
513 
514 	/* ... and init it */
515 	ep->dip = evtchndrv_dip;
516 
517 	cv_init(&ep->evtchn_wait, NULL, CV_DEFAULT, NULL);
518 	mutex_init(&ep->evtchn_lock, NULL, MUTEX_DEFAULT, NULL);
519 
520 	ep->ring = kmem_alloc(PAGESIZE, KM_SLEEP);
521 
522 	/* clone driver */
523 	*devp = makedevice(getmajor(*devp), minor);
524 
525 	return (0);
526 }
527 
528 /* ARGSUSED */
529 static int
530 evtchndrv_close(dev_t dev, int flag, int otyp, struct cred *credp)
531 {
532 	struct evtsoftdata *ep;
533 	minor_t minor = getminor(dev);
534 	int i;
535 
536 	ep = EVTCHNDRV_INST2SOFTS(EVTCHNDRV_MINOR2INST(minor));
537 	if (ep == NULL)
538 		return (ENXIO);
539 
540 	mutex_enter(&port_user_lock);
541 
542 
543 	for (i = 0; i < NR_EVENT_CHANNELS; i++) {
544 		if (port_user[i] != ep)
545 			continue;
546 
547 		evtchndrv_close_evtchn(i);
548 	}
549 
550 	mutex_exit(&port_user_lock);
551 
552 	kmem_free(ep->ring, PAGESIZE);
553 	ddi_soft_state_free(evtchndrv_statep, EVTCHNDRV_MINOR2INST(minor));
554 
555 	/*
556 	 * free clone tab slot
557 	 */
558 	mutex_enter(&evtchndrv_clone_tab_mutex);
559 	evtchndrv_clone_tab[minor] = 0;
560 	mutex_exit(&evtchndrv_clone_tab_mutex);
561 
562 	return (0);
563 }
564 
565 /* ARGSUSED */
566 static int
567 evtchndrv_info(dev_info_t *dip, ddi_info_cmd_t cmd, void *arg, void **result)
568 {
569 	dev_t	dev = (dev_t)arg;
570 	minor_t	minor = getminor(dev);
571 	int	retval;
572 
573 	switch (cmd) {
574 	case DDI_INFO_DEVT2DEVINFO:
575 		if (minor != 0 || evtchndrv_dip == NULL) {
576 			*result = (void *)NULL;
577 			retval = DDI_FAILURE;
578 		} else {
579 			*result = (void *)evtchndrv_dip;
580 			retval = DDI_SUCCESS;
581 		}
582 		break;
583 	case DDI_INFO_DEVT2INSTANCE:
584 		*result = (void *)0;
585 		retval = DDI_SUCCESS;
586 		break;
587 	default:
588 		retval = DDI_FAILURE;
589 	}
590 	return (retval);
591 }
592 
593 
594 static int
595 evtchndrv_attach(dev_info_t *dip, ddi_attach_cmd_t cmd)
596 {
597 	int	error;
598 	int	unit = ddi_get_instance(dip);
599 
600 
601 	switch (cmd) {
602 	case DDI_ATTACH:
603 		break;
604 	case DDI_RESUME:
605 		return (DDI_SUCCESS);
606 	default:
607 		cmn_err(CE_WARN, "evtchn_attach: unknown cmd 0x%x\n", cmd);
608 		return (DDI_FAILURE);
609 	}
610 
611 	/* DDI_ATTACH */
612 
613 	/*
614 	 * only one instance - but we clone using the open routine
615 	 */
616 	if (ddi_get_instance(dip) > 0)
617 		return (DDI_FAILURE);
618 
619 	mutex_init(&evtchndrv_clone_tab_mutex, NULL, MUTEX_DRIVER,
620 	    NULL);
621 
622 	error = ddi_create_minor_node(dip, "evtchn", S_IFCHR, unit,
623 	    DDI_PSEUDO, NULL);
624 	if (error != DDI_SUCCESS)
625 		goto fail;
626 
627 	/*
628 	 * save dip for getinfo
629 	 */
630 	evtchndrv_dip = dip;
631 	ddi_report_dev(dip);
632 
633 	mutex_init(&port_user_lock, NULL, MUTEX_DRIVER, NULL);
634 	(void) memset(port_user, 0, sizeof (port_user));
635 
636 	ec_dev_irq = ec_dev_alloc_irq();
637 	(void) add_avintr(NULL, IPL_EVTCHN, (avfunc)evtchn_device_upcall,
638 	    "evtchn_driver", ec_dev_irq, NULL, NULL, NULL, dip);
639 
640 	return (DDI_SUCCESS);
641 
642 fail:
643 	(void) evtchndrv_detach(dip, DDI_DETACH);
644 	return (error);
645 }
646 
647 /*ARGSUSED*/
648 static int
649 evtchndrv_detach(dev_info_t *dip, ddi_detach_cmd_t cmd)
650 {
651 	/*
652 	 * Don't allow detach for now.
653 	 */
654 	return (DDI_FAILURE);
655 }
656 
657 /* Solaris driver framework */
658 
659 static 	struct cb_ops evtchndrv_cb_ops = {
660 	evtchndrv_open,		/* cb_open */
661 	evtchndrv_close,	/* cb_close */
662 	nodev,			/* cb_strategy */
663 	nodev,			/* cb_print */
664 	nodev,			/* cb_dump */
665 	evtchndrv_read,		/* cb_read */
666 	evtchndrv_write,	/* cb_write */
667 	evtchndrv_ioctl,	/* cb_ioctl */
668 	nodev,			/* cb_devmap */
669 	nodev,			/* cb_mmap */
670 	nodev,			/* cb_segmap */
671 	evtchndrv_poll,		/* cb_chpoll */
672 	ddi_prop_op,		/* cb_prop_op */
673 	0,			/* cb_stream */
674 	D_NEW | D_MP | D_64BIT	/* cb_flag */
675 };
676 
677 static struct dev_ops evtchndrv_dev_ops = {
678 	DEVO_REV,		/* devo_rev */
679 	0,			/* devo_refcnt */
680 	evtchndrv_info,		/* devo_getinfo */
681 	nulldev,		/* devo_identify */
682 	nulldev,		/* devo_probe */
683 	evtchndrv_attach,	/* devo_attach */
684 	evtchndrv_detach,	/* devo_detach */
685 	nodev,			/* devo_reset */
686 	&evtchndrv_cb_ops,	/* devo_cb_ops */
687 	NULL,			/* devo_bus_ops */
688 	NULL,			/* power */
689 	ddi_quiesce_not_needed,		/* devo_quiesce */
690 };
691 
692 static struct modldrv modldrv = {
693 	&mod_driverops,		/* Type of module.  This one is a driver */
694 	"Evtchn driver",	/* Name of the module. */
695 	&evtchndrv_dev_ops	/* driver ops */
696 };
697 
698 static struct modlinkage modlinkage = {
699 	MODREV_1,
700 	&modldrv,
701 	NULL
702 };
703 
704 int
705 _init(void)
706 {
707 	int err;
708 
709 	err = ddi_soft_state_init(&evtchndrv_statep,
710 	    sizeof (struct evtsoftdata), 1);
711 	if (err)
712 		return (err);
713 
714 	err = mod_install(&modlinkage);
715 	if (err)
716 		ddi_soft_state_fini(&evtchndrv_statep);
717 	else
718 		evtchndrv_clone_tab = kmem_zalloc(
719 		    sizeof (int) * evtchndrv_nclones, KM_SLEEP);
720 	return (err);
721 }
722 
723 int
724 _fini(void)
725 {
726 	int e;
727 
728 	e = mod_remove(&modlinkage);
729 	if (e)
730 		return (e);
731 
732 	ddi_soft_state_fini(&evtchndrv_statep);
733 
734 	return (0);
735 }
736 
737 int
738 _info(struct modinfo *modinfop)
739 {
740 	return (mod_info(&modlinkage, modinfop));
741 }
742