On Fri, Mar 10, 2017 at 06:25:51PM +0200, Michael S. Tsirkin wrote: > On Tue, Mar 07, 2017 at 10:47:05AM +0100, Guillaume Missonnier wrote: > > Move vhost.h to include/linux to allow vhost driver out-of-tree compilation. > > Currently, this cannot be done properly because the vhost header file is in > > driver/vhost. > > > > To distribute a new vhost driver before it is included in the kernel tree, > > we need to package it using kmod, dkms, ..., and to compile it out-of-tree > > using headers provided by the distribution's kernel development package. > > > > Signed-off-by: Guillaume Missonnier <guillaume.missonnier@xxxxxxxx> > > FYI, I won't merge infrastructure patches before the first user was even > posted. Let's see that new driver first. > Ok, I will eventually resend this patch when our driver is ready. Thanks > > --- > > MAINTAINERS | 1 + > > drivers/vhost/net.c | 2 - > > drivers/vhost/scsi.c | 2 - > > drivers/vhost/test.c | 1 - > > drivers/vhost/vhost.c | 2 - > > drivers/vhost/vhost.h | 292 -------------------------------------------- > > drivers/vhost/vsock.c | 1 - > > include/linux/vhost.h | 293 +++++++++++++++++++++++++++++++++++++++++++++ > > include/uapi/linux/vhost.h | 4 +- > > 9 files changed, 296 insertions(+), 302 deletions(-) > > delete mode 100644 drivers/vhost/vhost.h > > create mode 100644 include/linux/vhost.h > > > > diff --git a/MAINTAINERS b/MAINTAINERS > > index 527d137..0788737 100644 > > --- a/MAINTAINERS > > +++ b/MAINTAINERS > > @@ -13113,6 +13113,7 @@ L: netdev@xxxxxxxxxxxxxxx > > T: git git://git.kernel.org/pub/scm/linux/kernel/git/mst/vhost.git > > S: Maintained > > F: drivers/vhost/ > > +F: include/linux/vhost.h > > F: include/uapi/linux/vhost.h > > > > VIRTIO INPUT DRIVER > > diff --git a/drivers/vhost/net.c b/drivers/vhost/net.c > > index 5dc3465..5de59ad 100644 > > --- a/drivers/vhost/net.c > > +++ b/drivers/vhost/net.c > > @@ -28,8 +28,6 @@ > > > > #include <net/sock.h> > > > > -#include "vhost.h" > > - > > static int experimental_zcopytx = 1; > > module_param(experimental_zcopytx, int, 0444); > > MODULE_PARM_DESC(experimental_zcopytx, "Enable Zero Copy TX;" > > diff --git a/drivers/vhost/scsi.c b/drivers/vhost/scsi.c > > index fd6c8b6..5b41456 100644 > > --- a/drivers/vhost/scsi.c > > +++ b/drivers/vhost/scsi.c > > @@ -48,8 +48,6 @@ > > #include <linux/bitmap.h> > > #include <linux/percpu_ida.h> > > > > -#include "vhost.h" > > - > > #define VHOST_SCSI_VERSION "v0.1" > > #define VHOST_SCSI_NAMELEN 256 > > #define VHOST_SCSI_MAX_CDB_SIZE 32 > > diff --git a/drivers/vhost/test.c b/drivers/vhost/test.c > > index 3cc98c0..d515cbd 100644 > > --- a/drivers/vhost/test.c > > +++ b/drivers/vhost/test.c > > @@ -17,7 +17,6 @@ > > #include <linux/slab.h> > > > > #include "test.h" > > -#include "vhost.h" > > > > /* Max number of bytes transferred before requeueing the job. > > * Using this limit prevents one virtqueue from starving others. */ > > diff --git a/drivers/vhost/vhost.c b/drivers/vhost/vhost.c > > index 998bed5..d5c5f3c 100644 > > --- a/drivers/vhost/vhost.c > > +++ b/drivers/vhost/vhost.c > > @@ -29,8 +29,6 @@ > > #include <linux/sort.h> > > #include <linux/interval_tree_generic.h> > > > > -#include "vhost.h" > > - > > static ushort max_mem_regions = 64; > > module_param(max_mem_regions, ushort, 0444); > > MODULE_PARM_DESC(max_mem_regions, > > diff --git a/drivers/vhost/vhost.h b/drivers/vhost/vhost.h > > deleted file mode 100644 > > index f55671d..0000000 > > --- a/drivers/vhost/vhost.h > > +++ /dev/null > > @@ -1,292 +0,0 @@ > > -#ifndef _VHOST_H > > -#define _VHOST_H > > - > > -#include <linux/eventfd.h> > > -#include <linux/vhost.h> > > -#include <linux/mm.h> > > -#include <linux/mutex.h> > > -#include <linux/poll.h> > > -#include <linux/file.h> > > -#include <linux/uio.h> > > -#include <linux/virtio_config.h> > > -#include <linux/virtio_ring.h> > > -#include <linux/atomic.h> > > - > > -struct vhost_work; > > -typedef void (*vhost_work_fn_t)(struct vhost_work *work); > > - > > -#define VHOST_WORK_QUEUED 1 > > -struct vhost_work { > > - struct llist_node node; > > - vhost_work_fn_t fn; > > - wait_queue_head_t done; > > - int flushing; > > - unsigned queue_seq; > > - unsigned done_seq; > > - unsigned long flags; > > -}; > > - > > -/* Poll a file (eventfd or socket) */ > > -/* Note: there's nothing vhost specific about this structure. */ > > -struct vhost_poll { > > - poll_table table; > > - wait_queue_head_t *wqh; > > - wait_queue_t wait; > > - struct vhost_work work; > > - unsigned long mask; > > - struct vhost_dev *dev; > > -}; > > - > > -void vhost_work_init(struct vhost_work *work, vhost_work_fn_t fn); > > -void vhost_work_queue(struct vhost_dev *dev, struct vhost_work *work); > > -bool vhost_has_work(struct vhost_dev *dev); > > - > > -void vhost_poll_init(struct vhost_poll *poll, vhost_work_fn_t fn, > > - unsigned long mask, struct vhost_dev *dev); > > -int vhost_poll_start(struct vhost_poll *poll, struct file *file); > > -void vhost_poll_stop(struct vhost_poll *poll); > > -void vhost_poll_flush(struct vhost_poll *poll); > > -void vhost_poll_queue(struct vhost_poll *poll); > > -void vhost_work_flush(struct vhost_dev *dev, struct vhost_work *work); > > -long vhost_vring_ioctl(struct vhost_dev *d, int ioctl, void __user *argp); > > - > > -struct vhost_log { > > - u64 addr; > > - u64 len; > > -}; > > - > > -#define START(node) ((node)->start) > > -#define LAST(node) ((node)->last) > > - > > -struct vhost_umem_node { > > - struct rb_node rb; > > - struct list_head link; > > - __u64 start; > > - __u64 last; > > - __u64 size; > > - __u64 userspace_addr; > > - __u32 perm; > > - __u32 flags_padding; > > - __u64 __subtree_last; > > -}; > > - > > -struct vhost_umem { > > - struct rb_root umem_tree; > > - struct list_head umem_list; > > - int numem; > > -}; > > - > > -enum vhost_uaddr_type { > > - VHOST_ADDR_DESC = 0, > > - VHOST_ADDR_AVAIL = 1, > > - VHOST_ADDR_USED = 2, > > - VHOST_NUM_ADDRS = 3, > > -}; > > - > > -/* The virtqueue structure describes a queue attached to a device. */ > > -struct vhost_virtqueue { > > - struct vhost_dev *dev; > > - > > - /* The actual ring of buffers. */ > > - struct mutex mutex; > > - unsigned int num; > > - struct vring_desc __user *desc; > > - struct vring_avail __user *avail; > > - struct vring_used __user *used; > > - const struct vhost_umem_node *meta_iotlb[VHOST_NUM_ADDRS]; > > - struct file *kick; > > - struct file *call; > > - struct file *error; > > - struct eventfd_ctx *call_ctx; > > - struct eventfd_ctx *error_ctx; > > - struct eventfd_ctx *log_ctx; > > - > > - struct vhost_poll poll; > > - > > - /* The routine to call when the Guest pings us, or timeout. */ > > - vhost_work_fn_t handle_kick; > > - > > - /* Last available index we saw. */ > > - u16 last_avail_idx; > > - > > - /* Caches available index value from user. */ > > - u16 avail_idx; > > - > > - /* Last index we used. */ > > - u16 last_used_idx; > > - > > - /* Last used evet we've seen */ > > - u16 last_used_event; > > - > > - /* Used flags */ > > - u16 used_flags; > > - > > - /* Last used index value we have signalled on */ > > - u16 signalled_used; > > - > > - /* Last used index value we have signalled on */ > > - bool signalled_used_valid; > > - > > - /* Log writes to used structure. */ > > - bool log_used; > > - u64 log_addr; > > - > > - struct iovec iov[UIO_MAXIOV]; > > - struct iovec iotlb_iov[64]; > > - struct iovec *indirect; > > - struct vring_used_elem *heads; > > - /* Protected by virtqueue mutex. */ > > - struct vhost_umem *umem; > > - struct vhost_umem *iotlb; > > - void *private_data; > > - u64 acked_features; > > - /* Log write descriptors */ > > - void __user *log_base; > > - struct vhost_log *log; > > - > > - /* Ring endianness. Defaults to legacy native endianness. > > - * Set to true when starting a modern virtio device. */ > > - bool is_le; > > -#ifdef CONFIG_VHOST_CROSS_ENDIAN_LEGACY > > - /* Ring endianness requested by userspace for cross-endian support. */ > > - bool user_be; > > -#endif > > - u32 busyloop_timeout; > > -}; > > - > > -struct vhost_msg_node { > > - struct vhost_msg msg; > > - struct vhost_virtqueue *vq; > > - struct list_head node; > > -}; > > - > > -struct vhost_dev { > > - struct mm_struct *mm; > > - struct mutex mutex; > > - struct vhost_virtqueue **vqs; > > - int nvqs; > > - struct file *log_file; > > - struct eventfd_ctx *log_ctx; > > - struct llist_head work_list; > > - struct task_struct *worker; > > - struct vhost_umem *umem; > > - struct vhost_umem *iotlb; > > - spinlock_t iotlb_lock; > > - struct list_head read_list; > > - struct list_head pending_list; > > - wait_queue_head_t wait; > > -}; > > - > > -void vhost_dev_init(struct vhost_dev *, struct vhost_virtqueue **vqs, int nvqs); > > -long vhost_dev_set_owner(struct vhost_dev *dev); > > -bool vhost_dev_has_owner(struct vhost_dev *dev); > > -long vhost_dev_check_owner(struct vhost_dev *); > > -struct vhost_umem *vhost_dev_reset_owner_prepare(void); > > -void vhost_dev_reset_owner(struct vhost_dev *, struct vhost_umem *); > > -void vhost_dev_cleanup(struct vhost_dev *, bool locked); > > -void vhost_dev_stop(struct vhost_dev *); > > -long vhost_dev_ioctl(struct vhost_dev *, unsigned int ioctl, void __user *argp); > > -long vhost_vring_ioctl(struct vhost_dev *d, int ioctl, void __user *argp); > > -int vhost_vq_access_ok(struct vhost_virtqueue *vq); > > -int vhost_log_access_ok(struct vhost_dev *); > > - > > -int vhost_get_vq_desc(struct vhost_virtqueue *, > > - struct iovec iov[], unsigned int iov_count, > > - unsigned int *out_num, unsigned int *in_num, > > - struct vhost_log *log, unsigned int *log_num); > > -void vhost_discard_vq_desc(struct vhost_virtqueue *, int n); > > - > > -int vhost_vq_init_access(struct vhost_virtqueue *); > > -int vhost_add_used(struct vhost_virtqueue *, unsigned int head, int len); > > -int vhost_add_used_n(struct vhost_virtqueue *, struct vring_used_elem *heads, > > - unsigned count); > > -void vhost_add_used_and_signal(struct vhost_dev *, struct vhost_virtqueue *, > > - unsigned int id, int len); > > -void vhost_add_used_and_signal_n(struct vhost_dev *, struct vhost_virtqueue *, > > - struct vring_used_elem *heads, unsigned count); > > -void vhost_signal(struct vhost_dev *, struct vhost_virtqueue *); > > -void vhost_disable_notify(struct vhost_dev *, struct vhost_virtqueue *); > > -bool vhost_vq_avail_empty(struct vhost_dev *, struct vhost_virtqueue *); > > -bool vhost_enable_notify(struct vhost_dev *, struct vhost_virtqueue *); > > - > > -int vhost_log_write(struct vhost_virtqueue *vq, struct vhost_log *log, > > - unsigned int log_num, u64 len); > > -int vq_iotlb_prefetch(struct vhost_virtqueue *vq); > > - > > -struct vhost_msg_node *vhost_new_msg(struct vhost_virtqueue *vq, int type); > > -void vhost_enqueue_msg(struct vhost_dev *dev, > > - struct list_head *head, > > - struct vhost_msg_node *node); > > -struct vhost_msg_node *vhost_dequeue_msg(struct vhost_dev *dev, > > - struct list_head *head); > > -unsigned int vhost_chr_poll(struct file *file, struct vhost_dev *dev, > > - poll_table *wait); > > -ssize_t vhost_chr_read_iter(struct vhost_dev *dev, struct iov_iter *to, > > - int noblock); > > -ssize_t vhost_chr_write_iter(struct vhost_dev *dev, > > - struct iov_iter *from); > > -int vhost_init_device_iotlb(struct vhost_dev *d, bool enabled); > > - > > -#define vq_err(vq, fmt, ...) do { \ > > - pr_debug(pr_fmt(fmt), ##__VA_ARGS__); \ > > - if ((vq)->error_ctx) \ > > - eventfd_signal((vq)->error_ctx, 1);\ > > - } while (0) > > - > > -enum { > > - VHOST_FEATURES = (1ULL << VIRTIO_F_NOTIFY_ON_EMPTY) | > > - (1ULL << VIRTIO_RING_F_INDIRECT_DESC) | > > - (1ULL << VIRTIO_RING_F_EVENT_IDX) | > > - (1ULL << VHOST_F_LOG_ALL) | > > - (1ULL << VIRTIO_F_ANY_LAYOUT) | > > - (1ULL << VIRTIO_F_VERSION_1) > > -}; > > - > > -static inline bool vhost_has_feature(struct vhost_virtqueue *vq, int bit) > > -{ > > - return vq->acked_features & (1ULL << bit); > > -} > > - > > -#ifdef CONFIG_VHOST_CROSS_ENDIAN_LEGACY > > -static inline bool vhost_is_little_endian(struct vhost_virtqueue *vq) > > -{ > > - return vq->is_le; > > -} > > -#else > > -static inline bool vhost_is_little_endian(struct vhost_virtqueue *vq) > > -{ > > - return virtio_legacy_is_little_endian() || vq->is_le; > > -} > > -#endif > > - > > -/* Memory accessors */ > > -static inline u16 vhost16_to_cpu(struct vhost_virtqueue *vq, __virtio16 val) > > -{ > > - return __virtio16_to_cpu(vhost_is_little_endian(vq), val); > > -} > > - > > -static inline __virtio16 cpu_to_vhost16(struct vhost_virtqueue *vq, u16 val) > > -{ > > - return __cpu_to_virtio16(vhost_is_little_endian(vq), val); > > -} > > - > > -static inline u32 vhost32_to_cpu(struct vhost_virtqueue *vq, __virtio32 val) > > -{ > > - return __virtio32_to_cpu(vhost_is_little_endian(vq), val); > > -} > > - > > -static inline __virtio32 cpu_to_vhost32(struct vhost_virtqueue *vq, u32 val) > > -{ > > - return __cpu_to_virtio32(vhost_is_little_endian(vq), val); > > -} > > - > > -static inline u64 vhost64_to_cpu(struct vhost_virtqueue *vq, __virtio64 val) > > -{ > > - return __virtio64_to_cpu(vhost_is_little_endian(vq), val); > > -} > > - > > -static inline __virtio64 cpu_to_vhost64(struct vhost_virtqueue *vq, u64 val) > > -{ > > - return __cpu_to_virtio64(vhost_is_little_endian(vq), val); > > -} > > -#endif > > diff --git a/drivers/vhost/vsock.c b/drivers/vhost/vsock.c > > index ce5e63d..f4c6fd8 100644 > > --- a/drivers/vhost/vsock.c > > +++ b/drivers/vhost/vsock.c > > @@ -17,7 +17,6 @@ > > #include <linux/vhost.h> > > > > #include <net/af_vsock.h> > > -#include "vhost.h" > > > > #define VHOST_VSOCK_DEFAULT_HOST_CID 2 > > > > diff --git a/include/linux/vhost.h b/include/linux/vhost.h > > new file mode 100644 > > index 0000000..7168c40 > > --- /dev/null > > +++ b/include/linux/vhost.h > > @@ -0,0 +1,293 @@ > > +#ifndef _LINUX_VHOST_H > > +#define _LINUX_VHOST_H > > + > > +#include <uapi/linux/vhost.h> > > + > > +#include <linux/eventfd.h> > > +#include <linux/mm.h> > > +#include <linux/mutex.h> > > +#include <linux/poll.h> > > +#include <linux/file.h> > > +#include <linux/uio.h> > > +#include <linux/virtio_config.h> > > +#include <linux/virtio_ring.h> > > +#include <linux/atomic.h> > > + > > +struct vhost_work; > > +typedef void (*vhost_work_fn_t)(struct vhost_work *work); > > + > > +#define VHOST_WORK_QUEUED 1 > > +struct vhost_work { > > + struct llist_node node; > > + vhost_work_fn_t fn; > > + wait_queue_head_t done; > > + int flushing; > > + unsigned queue_seq; > > + unsigned done_seq; > > + unsigned long flags; > > +}; > > + > > +/* Poll a file (eventfd or socket) */ > > +/* Note: there's nothing vhost specific about this structure. */ > > +struct vhost_poll { > > + poll_table table; > > + wait_queue_head_t *wqh; > > + wait_queue_t wait; > > + struct vhost_work work; > > + unsigned long mask; > > + struct vhost_dev *dev; > > +}; > > + > > +void vhost_work_init(struct vhost_work *work, vhost_work_fn_t fn); > > +void vhost_work_queue(struct vhost_dev *dev, struct vhost_work *work); > > +bool vhost_has_work(struct vhost_dev *dev); > > + > > +void vhost_poll_init(struct vhost_poll *poll, vhost_work_fn_t fn, > > + unsigned long mask, struct vhost_dev *dev); > > +int vhost_poll_start(struct vhost_poll *poll, struct file *file); > > +void vhost_poll_stop(struct vhost_poll *poll); > > +void vhost_poll_flush(struct vhost_poll *poll); > > +void vhost_poll_queue(struct vhost_poll *poll); > > +void vhost_work_flush(struct vhost_dev *dev, struct vhost_work *work); > > +long vhost_vring_ioctl(struct vhost_dev *d, int ioctl, void __user *argp); > > + > > +struct vhost_log { > > + u64 addr; > > + u64 len; > > +}; > > + > > +#define START(node) ((node)->start) > > +#define LAST(node) ((node)->last) > > + > > +struct vhost_umem_node { > > + struct rb_node rb; > > + struct list_head link; > > + __u64 start; > > + __u64 last; > > + __u64 size; > > + __u64 userspace_addr; > > + __u32 perm; > > + __u32 flags_padding; > > + __u64 __subtree_last; > > +}; > > + > > +struct vhost_umem { > > + struct rb_root umem_tree; > > + struct list_head umem_list; > > + int numem; > > +}; > > + > > +enum vhost_uaddr_type { > > + VHOST_ADDR_DESC = 0, > > + VHOST_ADDR_AVAIL = 1, > > + VHOST_ADDR_USED = 2, > > + VHOST_NUM_ADDRS = 3, > > +}; > > + > > +/* The virtqueue structure describes a queue attached to a device. */ > > +struct vhost_virtqueue { > > + struct vhost_dev *dev; > > + > > + /* The actual ring of buffers. */ > > + struct mutex mutex; > > + unsigned int num; > > + struct vring_desc __user *desc; > > + struct vring_avail __user *avail; > > + struct vring_used __user *used; > > + const struct vhost_umem_node *meta_iotlb[VHOST_NUM_ADDRS]; > > + struct file *kick; > > + struct file *call; > > + struct file *error; > > + struct eventfd_ctx *call_ctx; > > + struct eventfd_ctx *error_ctx; > > + struct eventfd_ctx *log_ctx; > > + > > + struct vhost_poll poll; > > + > > + /* The routine to call when the Guest pings us, or timeout. */ > > + vhost_work_fn_t handle_kick; > > + > > + /* Last available index we saw. */ > > + u16 last_avail_idx; > > + > > + /* Caches available index value from user. */ > > + u16 avail_idx; > > + > > + /* Last index we used. */ > > + u16 last_used_idx; > > + > > + /* Last used evet we've seen */ > > + u16 last_used_event; > > + > > + /* Used flags */ > > + u16 used_flags; > > + > > + /* Last used index value we have signalled on */ > > + u16 signalled_used; > > + > > + /* Last used index value we have signalled on */ > > + bool signalled_used_valid; > > + > > + /* Log writes to used structure. */ > > + bool log_used; > > + u64 log_addr; > > + > > + struct iovec iov[UIO_MAXIOV]; > > + struct iovec iotlb_iov[64]; > > + struct iovec *indirect; > > + struct vring_used_elem *heads; > > + /* Protected by virtqueue mutex. */ > > + struct vhost_umem *umem; > > + struct vhost_umem *iotlb; > > + void *private_data; > > + u64 acked_features; > > + /* Log write descriptors */ > > + void __user *log_base; > > + struct vhost_log *log; > > + > > + /* Ring endianness. Defaults to legacy native endianness. > > + * Set to true when starting a modern virtio device. */ > > + bool is_le; > > +#ifdef CONFIG_VHOST_CROSS_ENDIAN_LEGACY > > + /* Ring endianness requested by userspace for cross-endian support. */ > > + bool user_be; > > +#endif > > + u32 busyloop_timeout; > > +}; > > + > > +struct vhost_msg_node { > > + struct vhost_msg msg; > > + struct vhost_virtqueue *vq; > > + struct list_head node; > > +}; > > + > > +struct vhost_dev { > > + struct mm_struct *mm; > > + struct mutex mutex; > > + struct vhost_virtqueue **vqs; > > + int nvqs; > > + struct file *log_file; > > + struct eventfd_ctx *log_ctx; > > + struct llist_head work_list; > > + struct task_struct *worker; > > + struct vhost_umem *umem; > > + struct vhost_umem *iotlb; > > + spinlock_t iotlb_lock; > > + struct list_head read_list; > > + struct list_head pending_list; > > + wait_queue_head_t wait; > > +}; > > + > > +void vhost_dev_init(struct vhost_dev *, struct vhost_virtqueue **vqs, int nvqs); > > +long vhost_dev_set_owner(struct vhost_dev *dev); > > +bool vhost_dev_has_owner(struct vhost_dev *dev); > > +long vhost_dev_check_owner(struct vhost_dev *); > > +struct vhost_umem *vhost_dev_reset_owner_prepare(void); > > +void vhost_dev_reset_owner(struct vhost_dev *, struct vhost_umem *); > > +void vhost_dev_cleanup(struct vhost_dev *, bool locked); > > +void vhost_dev_stop(struct vhost_dev *); > > +long vhost_dev_ioctl(struct vhost_dev *, unsigned int ioctl, void __user *argp); > > +long vhost_vring_ioctl(struct vhost_dev *d, int ioctl, void __user *argp); > > +int vhost_vq_access_ok(struct vhost_virtqueue *vq); > > +int vhost_log_access_ok(struct vhost_dev *); > > + > > +int vhost_get_vq_desc(struct vhost_virtqueue *, > > + struct iovec iov[], unsigned int iov_count, > > + unsigned int *out_num, unsigned int *in_num, > > + struct vhost_log *log, unsigned int *log_num); > > +void vhost_discard_vq_desc(struct vhost_virtqueue *, int n); > > + > > +int vhost_vq_init_access(struct vhost_virtqueue *); > > +int vhost_add_used(struct vhost_virtqueue *, unsigned int head, int len); > > +int vhost_add_used_n(struct vhost_virtqueue *, struct vring_used_elem *heads, > > + unsigned count); > > +void vhost_add_used_and_signal(struct vhost_dev *, struct vhost_virtqueue *, > > + unsigned int id, int len); > > +void vhost_add_used_and_signal_n(struct vhost_dev *, struct vhost_virtqueue *, > > + struct vring_used_elem *heads, unsigned count); > > +void vhost_signal(struct vhost_dev *, struct vhost_virtqueue *); > > +void vhost_disable_notify(struct vhost_dev *, struct vhost_virtqueue *); > > +bool vhost_vq_avail_empty(struct vhost_dev *, struct vhost_virtqueue *); > > +bool vhost_enable_notify(struct vhost_dev *, struct vhost_virtqueue *); > > + > > +int vhost_log_write(struct vhost_virtqueue *vq, struct vhost_log *log, > > + unsigned int log_num, u64 len); > > +int vq_iotlb_prefetch(struct vhost_virtqueue *vq); > > + > > +struct vhost_msg_node *vhost_new_msg(struct vhost_virtqueue *vq, int type); > > +void vhost_enqueue_msg(struct vhost_dev *dev, > > + struct list_head *head, > > + struct vhost_msg_node *node); > > +struct vhost_msg_node *vhost_dequeue_msg(struct vhost_dev *dev, > > + struct list_head *head); > > +unsigned int vhost_chr_poll(struct file *file, struct vhost_dev *dev, > > + poll_table *wait); > > +ssize_t vhost_chr_read_iter(struct vhost_dev *dev, struct iov_iter *to, > > + int noblock); > > +ssize_t vhost_chr_write_iter(struct vhost_dev *dev, > > + struct iov_iter *from); > > +int vhost_init_device_iotlb(struct vhost_dev *d, bool enabled); > > + > > +#define vq_err(vq, fmt, ...) do { \ > > + pr_debug(pr_fmt(fmt), ##__VA_ARGS__); \ > > + if ((vq)->error_ctx) \ > > + eventfd_signal((vq)->error_ctx, 1);\ > > + } while (0) > > + > > +enum { > > + VHOST_FEATURES = (1ULL << VIRTIO_F_NOTIFY_ON_EMPTY) | > > + (1ULL << VIRTIO_RING_F_INDIRECT_DESC) | > > + (1ULL << VIRTIO_RING_F_EVENT_IDX) | > > + (1ULL << VHOST_F_LOG_ALL) | > > + (1ULL << VIRTIO_F_ANY_LAYOUT) | > > + (1ULL << VIRTIO_F_VERSION_1) > > +}; > > + > > +static inline bool vhost_has_feature(struct vhost_virtqueue *vq, int bit) > > +{ > > + return vq->acked_features & (1ULL << bit); > > +} > > + > > +#ifdef CONFIG_VHOST_CROSS_ENDIAN_LEGACY > > +static inline bool vhost_is_little_endian(struct vhost_virtqueue *vq) > > +{ > > + return vq->is_le; > > +} > > +#else > > +static inline bool vhost_is_little_endian(struct vhost_virtqueue *vq) > > +{ > > + return virtio_legacy_is_little_endian() || vq->is_le; > > +} > > +#endif > > + > > +/* Memory accessors */ > > +static inline u16 vhost16_to_cpu(struct vhost_virtqueue *vq, __virtio16 val) > > +{ > > + return __virtio16_to_cpu(vhost_is_little_endian(vq), val); > > +} > > + > > +static inline __virtio16 cpu_to_vhost16(struct vhost_virtqueue *vq, u16 val) > > +{ > > + return __cpu_to_virtio16(vhost_is_little_endian(vq), val); > > +} > > + > > +static inline u32 vhost32_to_cpu(struct vhost_virtqueue *vq, __virtio32 val) > > +{ > > + return __virtio32_to_cpu(vhost_is_little_endian(vq), val); > > +} > > + > > +static inline __virtio32 cpu_to_vhost32(struct vhost_virtqueue *vq, u32 val) > > +{ > > + return __cpu_to_virtio32(vhost_is_little_endian(vq), val); > > +} > > + > > +static inline u64 vhost64_to_cpu(struct vhost_virtqueue *vq, __virtio64 val) > > +{ > > + return __virtio64_to_cpu(vhost_is_little_endian(vq), val); > > +} > > + > > +static inline __virtio64 cpu_to_vhost64(struct vhost_virtqueue *vq, u64 val) > > +{ > > + return __cpu_to_virtio64(vhost_is_little_endian(vq), val); > > +} > > +#endif > > diff --git a/include/uapi/linux/vhost.h b/include/uapi/linux/vhost.h > > index 60180c0..678a716 100644 > > --- a/include/uapi/linux/vhost.h > > +++ b/include/uapi/linux/vhost.h > > @@ -1,5 +1,5 @@ > > -#ifndef _LINUX_VHOST_H > > -#define _LINUX_VHOST_H > > +#ifndef _UAPI_LINUX_VHOST_H > > +#define _UAPI_LINUX_VHOST_H > > /* Userspace interface for in-kernel virtio accelerators. */ > > > > /* vhost is used to reduce the number of system calls involved in virtio. > > -- > > 1.8.3.1 _______________________________________________ Virtualization mailing list Virtualization@xxxxxxxxxxxxxxxxxxxxxxxxxx https://lists.linuxfoundation.org/mailman/listinfo/virtualization