diff options
-rw-r--r-- | sys/dev/pv/if_vio.c | 4 | ||||
-rw-r--r-- | sys/dev/pv/vioblk.c | 6 | ||||
-rw-r--r-- | sys/dev/pv/viocon.c | 4 | ||||
-rw-r--r-- | sys/dev/pv/viogpu.c | 19 | ||||
-rw-r--r-- | sys/dev/pv/virtio.c | 50 |
5 files changed, 59 insertions, 24 deletions
diff --git a/sys/dev/pv/if_vio.c b/sys/dev/pv/if_vio.c index eee59d81232..e6168ca3823 100644 --- a/sys/dev/pv/if_vio.c +++ b/sys/dev/pv/if_vio.c @@ -1,4 +1,4 @@ -/* $OpenBSD: if_vio.c,v 1.44 2024/07/26 07:55:23 sf Exp $ */ +/* $OpenBSD: if_vio.c,v 1.45 2024/08/01 11:13:19 sf Exp $ */ /* * Copyright (c) 2012 Stefan Fritsch, Alexander Fiveg. @@ -1303,7 +1303,7 @@ vio_txtick(void *arg) { struct virtqueue *vq = arg; int s = splnet(); - vio_tx_intr(vq); + virtio_check_vq(vq->vq_owner, vq); splx(s); } diff --git a/sys/dev/pv/vioblk.c b/sys/dev/pv/vioblk.c index 3be22b9309e..108309f4afb 100644 --- a/sys/dev/pv/vioblk.c +++ b/sys/dev/pv/vioblk.c @@ -1,4 +1,4 @@ -/* $OpenBSD: vioblk.c,v 1.40 2024/07/26 07:55:23 sf Exp $ */ +/* $OpenBSD: vioblk.c,v 1.41 2024/08/01 11:13:19 sf Exp $ */ /* * Copyright (c) 2012 Stefan Fritsch. @@ -375,7 +375,7 @@ vioblk_reset(struct vioblk_softc *sc) virtio_reset(sc->sc_virtio); /* finish requests that have been completed */ - vioblk_vq_done(&sc->sc_vq[0]); + virtio_check_vq(sc->sc_virtio, &sc->sc_vq[0]); /* abort all remaining requests */ for (i = 0; i < sc->sc_nreqs; i++) { @@ -535,7 +535,7 @@ vioblk_scsi_cmd(struct scsi_xfer *xs) if (!ISSET(xs->flags, SCSI_POLL)) { /* check if some xfers are done: */ if (sc->sc_queued > 1) - vioblk_vq_done(vq); + virtio_check_vq(sc->sc_virtio, vq); splx(s); return; } diff --git a/sys/dev/pv/viocon.c b/sys/dev/pv/viocon.c index e81bc952004..40992228573 100644 --- a/sys/dev/pv/viocon.c +++ b/sys/dev/pv/viocon.c @@ -1,4 +1,4 @@ -/* $OpenBSD: viocon.c,v 1.12 2024/06/26 01:40:49 jsg Exp $ */ +/* $OpenBSD: viocon.c,v 1.13 2024/08/01 11:13:19 sf Exp $ */ /* * Copyright (c) 2013-2015 Stefan Fritsch <sf@sfritsch.de> @@ -463,7 +463,7 @@ vioconhwiflow(struct tty *tp, int stop) virtio_stop_vq_intr(vp->vp_sc->sc_virtio, vp->vp_rx); } else { virtio_start_vq_intr(vp->vp_sc->sc_virtio, vp->vp_rx); - softintr_schedule(vp->vp_si); + virtio_check_vq(vp->vp_sc->sc_virtio, vp->vp_rx); } splx(s); return 1; diff --git a/sys/dev/pv/viogpu.c b/sys/dev/pv/viogpu.c index 81844e393fc..aa9f1596e70 100644 --- a/sys/dev/pv/viogpu.c +++ b/sys/dev/pv/viogpu.c @@ -1,4 +1,4 @@ -/* $OpenBSD: viogpu.c,v 1.6 2024/05/24 10:05:55 jsg Exp $ */ +/* $OpenBSD: viogpu.c,v 1.7 2024/08/01 11:13:19 sf Exp $ */ /* * Copyright (c) 2021-2023 joshua stein <jcs@openbsd.org> @@ -42,7 +42,7 @@ struct viogpu_softc; int viogpu_match(struct device *, void *, void *); void viogpu_attach(struct device *, struct device *, void *); int viogpu_send_cmd(struct viogpu_softc *, void *, size_t, void *, size_t); -int viogpu_vq_wait(struct virtqueue *vq); +int viogpu_vq_done(struct virtqueue *vq); void viogpu_rx_soft(void *arg); int viogpu_get_display_info(struct viogpu_softc *); @@ -178,7 +178,7 @@ viogpu_attach(struct device *parent, struct device *self, void *aux) printf(": alloc_vq failed\n"); return; } - sc->sc_vqs[VQCTRL].vq_done = viogpu_vq_wait; + sc->sc_vqs[VQCTRL].vq_done = viogpu_vq_done; if (virtio_alloc_vq(vsc, &sc->sc_vqs[VQCURS], VQCURS, NBPG, 1, "cursor")) { @@ -211,6 +211,8 @@ viogpu_attach(struct device *parent, struct device *self, void *aux) goto unmap; } + virtio_set_status(vsc, VIRTIO_CONFIG_DEVICE_STATUS_DRIVER_OK); + if (viogpu_get_display_info(sc) != 0) goto unmap; @@ -230,8 +232,6 @@ viogpu_attach(struct device *parent, struct device *self, void *aux) sc->sc_fb_dma_kva, sc->sc_fb_dma_size, NULL, BUS_DMA_NOWAIT) != 0) goto fb_unmap; - virtio_set_status(vsc, VIRTIO_CONFIG_DEVICE_STATUS_DRIVER_OK); - if (viogpu_create_2d(sc, 1, sc->sc_fb_width, sc->sc_fb_height) != 0) goto fb_unmap; @@ -325,14 +325,14 @@ viogpu_repaint(void *arg) } int -viogpu_vq_wait(struct virtqueue *vq) +viogpu_vq_done(struct virtqueue *vq) { struct virtio_softc *vsc = vq->vq_owner; struct viogpu_softc *sc = (struct viogpu_softc *)vsc->sc_child; int slot, len; - while (virtio_dequeue(vsc, vq, &slot, &len) != 0) - ; + if (virtio_dequeue(vsc, vq, &slot, &len) != 0) + return 0; bus_dmamap_sync(vsc->sc_dmat, sc->sc_dma_map, 0, sc->sc_dma_size, BUS_DMASYNC_POSTREAD); @@ -402,7 +402,8 @@ viogpu_send_cmd(struct viogpu_softc *sc, void *cmd, size_t cmd_size, void *ret, virtio_enqueue_p(vq, slot, sc->sc_dma_map, cmd_size, ret_size, 0); virtio_enqueue_commit(vsc, vq, slot, 1); - viogpu_vq_wait(vq); + while (virtio_check_vq(vsc, vq) == 0) + ; bus_dmamap_sync(vsc->sc_dmat, sc->sc_dma_map, 0, cmd_size, BUS_DMASYNC_POSTWRITE); diff --git a/sys/dev/pv/virtio.c b/sys/dev/pv/virtio.c index 7386d0ca12c..4c834b67d82 100644 --- a/sys/dev/pv/virtio.c +++ b/sys/dev/pv/virtio.c @@ -1,4 +1,4 @@ -/* $OpenBSD: virtio.c,v 1.28 2024/07/26 07:55:23 sf Exp $ */ +/* $OpenBSD: virtio.c,v 1.29 2024/08/01 11:13:19 sf Exp $ */ /* $NetBSD: virtio.c,v 1.3 2011/11/02 23:05:52 njoly Exp $ */ /* @@ -185,6 +185,9 @@ virtio_reinit_end(struct virtio_softc *sc) /* * dmamap sync operations for a virtqueue. + * + * XXX These should be more fine grained. Syncing the whole ring if we + * XXX only need a few bytes is inefficient if we use bounce buffers. */ static inline void vq_sync_descs(struct virtio_softc *sc, struct virtqueue *vq, int ops) @@ -203,6 +206,15 @@ vq_sync_aring(struct virtio_softc *sc, struct virtqueue *vq, int ops) } static inline void +vq_sync_aring_used_event(struct virtio_softc *sc, struct virtqueue *vq, int ops) +{ + bus_dmamap_sync(sc->sc_dmat, vq->vq_dmamap, vq->vq_availoffset + + offsetof(struct vring_avail, ring) + vq->vq_num * sizeof(uint16_t), + sizeof(uint16_t), ops); +} + + +static inline void vq_sync_uring(struct virtio_softc *sc, struct virtqueue *vq, int ops) { bus_dmamap_sync(sc->sc_dmat, vq->vq_dmamap, vq->vq_usedoffset, @@ -211,6 +223,16 @@ vq_sync_uring(struct virtio_softc *sc, struct virtqueue *vq, int ops) } static inline void +vq_sync_uring_avail_event(struct virtio_softc *sc, struct virtqueue *vq, int ops) +{ + bus_dmamap_sync(sc->sc_dmat, vq->vq_dmamap, + vq->vq_usedoffset + offsetof(struct vring_used, ring) + + vq->vq_num * sizeof(struct vring_used_elem), sizeof(uint16_t), + ops); +} + + +static inline void vq_sync_indirect(struct virtio_softc *sc, struct virtqueue *vq, int slot, int ops) { @@ -672,11 +694,13 @@ virtio_enqueue_p(struct virtqueue *vq, int slot, bus_dmamap_t dmamap, static void publish_avail_idx(struct virtio_softc *sc, struct virtqueue *vq) { + /* first make sure the avail ring entries are visible to the device */ vq_sync_aring(sc, vq, BUS_DMASYNC_PREWRITE); virtio_membar_producer(); vq->vq_avail->idx = vq->vq_avail_idx; - vq_sync_aring(sc, vq, BUS_DMASYNC_POSTWRITE); + /* make the avail idx visible to the device */ + vq_sync_aring(sc, vq, BUS_DMASYNC_PREWRITE); vq->vq_queued = 1; } @@ -706,6 +730,7 @@ notify: publish_avail_idx(sc, vq); virtio_membar_sync(); + vq_sync_uring_avail_event(sc, vq, BUS_DMASYNC_POSTREAD); t = VQ_AVAIL_EVENT(vq) + 1; if ((uint16_t)(n - t) < (uint16_t)(n - o)) sc->sc_ops->kick(sc, vq->vq_index); @@ -713,6 +738,7 @@ notify: publish_avail_idx(sc, vq); virtio_membar_sync(); + vq_sync_uring(sc, vq, BUS_DMASYNC_POSTREAD); if (!(vq->vq_used->flags & VRING_USED_F_NO_NOTIFY)) sc->sc_ops->kick(sc, vq->vq_index); } @@ -781,8 +807,10 @@ virtio_enqueue_trim(struct virtqueue *vq, int slot, int nsegs) * Dequeue a request. */ /* - * dequeue: dequeue a request from uring; dmamap_sync for uring is - * already done in the interrupt handler. + * dequeue: dequeue a request from uring; bus_dmamap_sync for uring must + * already have been done, usually by virtio_check_vq() + * in the interrupt handler. This means that polling virtio_dequeue() + * repeatedly until it returns 0 does not work. */ int virtio_dequeue(struct virtio_softc *sc, struct virtqueue *vq, @@ -797,6 +825,7 @@ virtio_dequeue(struct virtio_softc *sc, struct virtqueue *vq, usedidx &= vq->vq_mask; virtio_membar_consumer(); + vq_sync_uring(sc, vq, BUS_DMASYNC_POSTREAD); slot = vq->vq_used->ring[usedidx].id; qe = &vq->vq_entries[slot]; @@ -852,7 +881,7 @@ virtio_postpone_intr(struct virtqueue *vq, uint16_t nslots) VQ_USED_EVENT(vq) = idx; virtio_membar_sync(); - vq_sync_aring(vq->vq_owner, vq, BUS_DMASYNC_PREWRITE); + vq_sync_aring_used_event(vq->vq_owner, vq, BUS_DMASYNC_PREWRITE); vq->vq_queued++; if (nslots < virtio_nused(vq)) @@ -905,6 +934,7 @@ virtio_stop_vq_intr(struct virtio_softc *sc, struct virtqueue *vq) * interrupt. */ VQ_USED_EVENT(vq) = vq->vq_used_idx + 0x8000; + vq_sync_aring_used_event(sc, vq, BUS_DMASYNC_PREWRITE); } else { vq->vq_avail->flags |= VRING_AVAIL_F_NO_INTERRUPT; } @@ -920,16 +950,19 @@ virtio_start_vq_intr(struct virtio_softc *sc, struct virtqueue *vq) * interrupts is done through setting the latest * consumed index in the used_event field */ - if (virtio_has_feature(sc, VIRTIO_F_RING_EVENT_IDX)) + if (virtio_has_feature(sc, VIRTIO_F_RING_EVENT_IDX)) { VQ_USED_EVENT(vq) = vq->vq_used_idx; - else + vq_sync_aring_used_event(sc, vq, BUS_DMASYNC_PREWRITE); + } else { vq->vq_avail->flags &= ~VRING_AVAIL_F_NO_INTERRUPT; + vq_sync_aring(sc, vq, BUS_DMASYNC_PREWRITE); + } virtio_membar_sync(); - vq_sync_aring(sc, vq, BUS_DMASYNC_PREWRITE); vq->vq_queued++; + vq_sync_uring(sc, vq, BUS_DMASYNC_POSTREAD); if (vq->vq_used_idx != vq->vq_used->idx) return 1; @@ -945,6 +978,7 @@ virtio_nused(struct virtqueue *vq) { uint16_t n; + vq_sync_uring(vq->vq_owner, vq, BUS_DMASYNC_POSTREAD); n = (uint16_t)(vq->vq_used->idx - vq->vq_used_idx); VIRTIO_ASSERT(n <= vq->vq_num); |