[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]
Re: [Qemu-devel] [PATCH v6 2/8] virtio: allow byte swapping for vring an
From: |
Thomas Huth |
Subject: |
Re: [Qemu-devel] [PATCH v6 2/8] virtio: allow byte swapping for vring and config access |
Date: |
Fri, 28 Mar 2014 17:07:31 +0100 |
On Fri, 28 Mar 2014 11:57:25 +0100
Greg Kurz <address@hidden> wrote:
> From: Rusty Russell <address@hidden>
>
> This is based on a simpler patch by Anthony Liguouri, which only handled
> the vring accesses. We also need some drivers to access these helpers,
> eg. for data which contains headers.
>
> Signed-off-by: Rusty Russell <address@hidden>
> [ ldq_phys() API change,
> use per-device needs_byteswap flag,
> Greg Kurz <address@hidden> ]
> Signed-off-by: Greg Kurz <address@hidden>
> ---
> hw/virtio/virtio.c | 93
> ++++++++++++++++++++++++++++++----------------------
> 1 file changed, 53 insertions(+), 40 deletions(-)
>
> diff --git a/hw/virtio/virtio.c b/hw/virtio/virtio.c
> index 24b565f..1877b46 100644
> --- a/hw/virtio/virtio.c
> +++ b/hw/virtio/virtio.c
> @@ -102,53 +102,57 @@ static void virtqueue_init(VirtQueue *vq)
> vq->vring.align);
> }
>
> -static inline uint64_t vring_desc_addr(hwaddr desc_pa, int i)
> +static inline uint64_t vring_desc_addr(hwaddr desc_pa, int i,
> + struct VirtIODevice *vdev)
> {
> hwaddr pa;
> pa = desc_pa + sizeof(VRingDesc) * i + offsetof(VRingDesc, addr);
> - return ldq_phys(&address_space_memory, pa);
> + return virtio_ldq_phys(&address_space_memory, pa, vdev);
> }
>
> -static inline uint32_t vring_desc_len(hwaddr desc_pa, int i)
> +static inline uint32_t vring_desc_len(hwaddr desc_pa, int i,
> + struct VirtIODevice *vdev)
> {
> hwaddr pa;
> pa = desc_pa + sizeof(VRingDesc) * i + offsetof(VRingDesc, len);
> - return ldl_phys(&address_space_memory, pa);
> + return virtio_ldl_phys(&address_space_memory, pa, vdev);
> }
>
> -static inline uint16_t vring_desc_flags(hwaddr desc_pa, int i)
> +static inline uint16_t vring_desc_flags(hwaddr desc_pa, int i,
> + struct VirtIODevice *vdev)
> {
> hwaddr pa;
> pa = desc_pa + sizeof(VRingDesc) * i + offsetof(VRingDesc, flags);
> - return lduw_phys(&address_space_memory, pa);
> + return virtio_lduw_phys(&address_space_memory, pa, vdev);
> }
>
> -static inline uint16_t vring_desc_next(hwaddr desc_pa, int i)
> +static inline uint16_t vring_desc_next(hwaddr desc_pa, int i,
> + struct VirtIODevice *vdev)
> {
> hwaddr pa;
> pa = desc_pa + sizeof(VRingDesc) * i + offsetof(VRingDesc, next);
> - return lduw_phys(&address_space_memory, pa);
> + return virtio_lduw_phys(&address_space_memory, pa, vdev);
> }
>
> static inline uint16_t vring_avail_flags(VirtQueue *vq)
> {
> hwaddr pa;
> pa = vq->vring.avail + offsetof(VRingAvail, flags);
> - return lduw_phys(&address_space_memory, pa);
> + return virtio_lduw_phys(&address_space_memory, pa, vq->vdev);
> }
>
> static inline uint16_t vring_avail_idx(VirtQueue *vq)
> {
> hwaddr pa;
> pa = vq->vring.avail + offsetof(VRingAvail, idx);
> - return lduw_phys(&address_space_memory, pa);
> + return virtio_lduw_phys(&address_space_memory, pa, vq->vdev);
> }
>
> static inline uint16_t vring_avail_ring(VirtQueue *vq, int i)
> {
> hwaddr pa;
> pa = vq->vring.avail + offsetof(VRingAvail, ring[i]);
> - return lduw_phys(&address_space_memory, pa);
> + return virtio_lduw_phys(&address_space_memory, pa, vq->vdev);
> }
>
> static inline uint16_t vring_used_event(VirtQueue *vq)
> @@ -160,44 +164,46 @@ static inline void vring_used_ring_id(VirtQueue *vq,
> int i, uint32_t val)
> {
> hwaddr pa;
> pa = vq->vring.used + offsetof(VRingUsed, ring[i].id);
> - stl_phys(&address_space_memory, pa, val);
> + virtio_stl_phys(&address_space_memory, pa, val, vq->vdev);
> }
>
> static inline void vring_used_ring_len(VirtQueue *vq, int i, uint32_t val)
> {
> hwaddr pa;
> pa = vq->vring.used + offsetof(VRingUsed, ring[i].len);
> - stl_phys(&address_space_memory, pa, val);
> + virtio_stl_phys(&address_space_memory, pa, val, vq->vdev);
> }
>
> static uint16_t vring_used_idx(VirtQueue *vq)
> {
> hwaddr pa;
> pa = vq->vring.used + offsetof(VRingUsed, idx);
> - return lduw_phys(&address_space_memory, pa);
> + return virtio_lduw_phys(&address_space_memory, pa, vq->vdev);
> }
>
> static inline void vring_used_idx_set(VirtQueue *vq, uint16_t val)
> {
> hwaddr pa;
> pa = vq->vring.used + offsetof(VRingUsed, idx);
> - stw_phys(&address_space_memory, pa, val);
> + virtio_stw_phys(&address_space_memory, pa, val, vq->vdev);
> }
>
> static inline void vring_used_flags_set_bit(VirtQueue *vq, int mask)
> {
> hwaddr pa;
> pa = vq->vring.used + offsetof(VRingUsed, flags);
> - stw_phys(&address_space_memory,
> - pa, lduw_phys(&address_space_memory, pa) | mask);
> + virtio_stw_phys(&address_space_memory, pa,
> + virtio_lduw_phys(&address_space_memory, pa,
> + vq->vdev) | mask, vq->vdev);
> }
>
> static inline void vring_used_flags_unset_bit(VirtQueue *vq, int mask)
> {
> hwaddr pa;
> pa = vq->vring.used + offsetof(VRingUsed, flags);
> - stw_phys(&address_space_memory,
> - pa, lduw_phys(&address_space_memory, pa) & ~mask);
> + virtio_stw_phys(&address_space_memory, pa,
> + virtio_lduw_phys(&address_space_memory, pa,
> + vq->vdev) & ~mask, vq->vdev);
> }
>
> static inline void vring_avail_event(VirtQueue *vq, uint16_t val)
> @@ -207,7 +213,7 @@ static inline void vring_avail_event(VirtQueue *vq,
> uint16_t val)
> return;
> }
> pa = vq->vring.used + offsetof(VRingUsed, ring[vq->vring.num]);
> - stw_phys(&address_space_memory, pa, val);
> + virtio_stw_phys(&address_space_memory, pa, val, vq->vdev);
> }
>
> void virtio_queue_set_notification(VirtQueue *vq, int enable)
> @@ -325,16 +331,18 @@ static unsigned int virtqueue_get_head(VirtQueue *vq,
> unsigned int idx)
> }
>
> static unsigned virtqueue_next_desc(hwaddr desc_pa,
> - unsigned int i, unsigned int max)
> + unsigned int i, unsigned int max,
> + struct VirtIODevice *vdev)
> {
> unsigned int next;
>
> /* If this descriptor says it doesn't chain, we're done. */
> - if (!(vring_desc_flags(desc_pa, i) & VRING_DESC_F_NEXT))
> + if (!(vring_desc_flags(desc_pa, i, vdev) & VRING_DESC_F_NEXT)) {
> return max;
> + }
>
> /* Check they're not leading us off end of descriptors. */
> - next = vring_desc_next(desc_pa, i);
> + next = vring_desc_next(desc_pa, i, vdev);
> /* Make sure compiler knows to grab that: we don't want it changing! */
> smp_wmb();
>
> @@ -366,8 +374,9 @@ void virtqueue_get_avail_bytes(VirtQueue *vq, unsigned
> int *in_bytes,
> i = virtqueue_get_head(vq, idx++);
> desc_pa = vq->vring.desc;
>
> - if (vring_desc_flags(desc_pa, i) & VRING_DESC_F_INDIRECT) {
> - if (vring_desc_len(desc_pa, i) % sizeof(VRingDesc)) {
> + if (vring_desc_flags(desc_pa, i, vq->vdev)
> + & VRING_DESC_F_INDIRECT) {
I think the above if-statement would still fit into one line with the
80-columns limit.
> + if (vring_desc_len(desc_pa, i, vq->vdev) % sizeof(VRingDesc)) {
> error_report("Invalid size for indirect buffer table");
> exit(1);
> }
> @@ -380,8 +389,8 @@ void virtqueue_get_avail_bytes(VirtQueue *vq, unsigned
> int *in_bytes,
>
> /* loop over the indirect descriptor table */
> indirect = 1;
> - max = vring_desc_len(desc_pa, i) / sizeof(VRingDesc);
> - desc_pa = vring_desc_addr(desc_pa, i);
> + max = vring_desc_len(desc_pa, i, vq->vdev) / sizeof(VRingDesc);
> + desc_pa = vring_desc_addr(desc_pa, i, vq->vdev);
> num_bufs = i = 0;
> }
>
> @@ -392,15 +401,17 @@ void virtqueue_get_avail_bytes(VirtQueue *vq, unsigned
> int *in_bytes,
> exit(1);
> }
>
> - if (vring_desc_flags(desc_pa, i) & VRING_DESC_F_WRITE) {
> - in_total += vring_desc_len(desc_pa, i);
> + if (vring_desc_flags(desc_pa, i, vq->vdev)
> + & VRING_DESC_F_WRITE) {
Again, no need for two lines?
> + in_total += vring_desc_len(desc_pa, i, vq->vdev);
> } else {
> - out_total += vring_desc_len(desc_pa, i);
> + out_total += vring_desc_len(desc_pa, i, vq->vdev);
> }
> if (in_total >= max_in_bytes && out_total >= max_out_bytes) {
> goto done;
> }
> - } while ((i = virtqueue_next_desc(desc_pa, i, max)) != max);
> + } while ((i = virtqueue_next_desc(desc_pa, i, max, vq->vdev))
> + != max);
Doesn't that while statement also still fit into 80 columns?
> if (!indirect)
> total_bufs = num_bufs;
> @@ -459,15 +470,15 @@ int virtqueue_pop(VirtQueue *vq, VirtQueueElement *elem)
> vring_avail_event(vq, vring_avail_idx(vq));
> }
>
> - if (vring_desc_flags(desc_pa, i) & VRING_DESC_F_INDIRECT) {
> - if (vring_desc_len(desc_pa, i) % sizeof(VRingDesc)) {
> + if (vring_desc_flags(desc_pa, i, vq->vdev) & VRING_DESC_F_INDIRECT) {
> + if (vring_desc_len(desc_pa, i, vq->vdev) % sizeof(VRingDesc)) {
> error_report("Invalid size for indirect buffer table");
> exit(1);
> }
>
> /* loop over the indirect descriptor table */
> - max = vring_desc_len(desc_pa, i) / sizeof(VRingDesc);
> - desc_pa = vring_desc_addr(desc_pa, i);
> + max = vring_desc_len(desc_pa, i, vq->vdev) / sizeof(VRingDesc);
> + desc_pa = vring_desc_addr(desc_pa, i, vq->vdev);
> i = 0;
> }
>
> @@ -475,30 +486,32 @@ int virtqueue_pop(VirtQueue *vq, VirtQueueElement *elem)
> do {
> struct iovec *sg;
>
> - if (vring_desc_flags(desc_pa, i) & VRING_DESC_F_WRITE) {
> + if (vring_desc_flags(desc_pa, i, vq->vdev) & VRING_DESC_F_WRITE) {
> if (elem->in_num >= ARRAY_SIZE(elem->in_sg)) {
> error_report("Too many write descriptors in indirect table");
> exit(1);
> }
> - elem->in_addr[elem->in_num] = vring_desc_addr(desc_pa, i);
> + elem->in_addr[elem->in_num] = vring_desc_addr(desc_pa, i,
> + vq->vdev);
That one seems to be pretty close, but could also still fit into 80
columns?
> sg = &elem->in_sg[elem->in_num++];
> } else {
> if (elem->out_num >= ARRAY_SIZE(elem->out_sg)) {
> error_report("Too many read descriptors in indirect table");
> exit(1);
> }
> - elem->out_addr[elem->out_num] = vring_desc_addr(desc_pa, i);
> + elem->out_addr[elem->out_num] = vring_desc_addr(desc_pa, i,
> + vq->vdev);
> sg = &elem->out_sg[elem->out_num++];
> }
>
> - sg->iov_len = vring_desc_len(desc_pa, i);
> + sg->iov_len = vring_desc_len(desc_pa, i, vq->vdev);
>
> /* If we've got too many, that implies a descriptor loop. */
> if ((elem->in_num + elem->out_num) > max) {
> error_report("Looped descriptor");
> exit(1);
> }
> - } while ((i = virtqueue_next_desc(desc_pa, i, max)) != max);
> + } while ((i = virtqueue_next_desc(desc_pa, i, max, vq->vdev)) != max);
>
> /* Now map what we have collected */
> virtqueue_map_sg(elem->in_sg, elem->in_addr, elem->in_num, 1);
Apart from the cosmetic nits, patch looks fine to me.
Reviewed-by: Thomas Huth <address@hidden>
- [Qemu-devel] [PATCH v6 0/8] virtio endian-ambivalent target fixes, Greg Kurz, 2014/03/28
- [Qemu-devel] [PATCH v6 1/8] virtio_legacy_get_byteswap: endian-ambivalent targets using legacy virtio, Greg Kurz, 2014/03/28
- [Qemu-devel] [PATCH v6 2/8] virtio: allow byte swapping for vring and config access, Greg Kurz, 2014/03/28
- [Qemu-devel] [PATCH v6 3/8] virtio-net: use virtio wrappers to access headers, Greg Kurz, 2014/03/28
- [Qemu-devel] [PATCH v6 4/8] virtio-balloon: use virtio wrappers to access page frame numbers, Greg Kurz, 2014/03/28
- [Qemu-devel] [PATCH v6 5/8] virtio-blk: use virtio wrappers to access headers, Greg Kurz, 2014/03/28
- [Qemu-devel] [PATCH v6 6/8] virtio-scsi: use virtio wrappers to access headers, Greg Kurz, 2014/03/28