Re: [PATCH net-next 2/2] virtio-net: get rid of unnecessary container of rq stats

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Tue, Jul 31, 2018 at 05:43:39PM +0800, Jason Wang wrote:
> We don't maintain tx counters in rx stats any more. There's no need
> for an extra container of rq stats.
> 
> Cc: Toshiaki Makita <makita.toshiaki@xxxxxxxxxxxxx>
> Signed-off-by: Jason Wang <jasowang@xxxxxxxxxx>

Acked-by: Michael S. Tsirkin <mst@xxxxxxxxxx>

> ---
>  drivers/net/virtio_net.c | 80 ++++++++++++++++++++++--------------------------
>  1 file changed, 36 insertions(+), 44 deletions(-)
> 
> diff --git a/drivers/net/virtio_net.c b/drivers/net/virtio_net.c
> index 72d3f68..14f661c 100644
> --- a/drivers/net/virtio_net.c
> +++ b/drivers/net/virtio_net.c
> @@ -87,7 +87,8 @@ struct virtnet_sq_stats {
>  	u64 kicks;
>  };
>  
> -struct virtnet_rq_stat_items {
> +struct virtnet_rq_stats {
> +	struct u64_stats_sync syncp;
>  	u64 packets;
>  	u64 bytes;
>  	u64 drops;
> @@ -98,17 +99,8 @@ struct virtnet_rq_stat_items {
>  	u64 kicks;
>  };
>  
> -struct virtnet_rq_stats {
> -	struct u64_stats_sync syncp;
> -	struct virtnet_rq_stat_items items;
> -};
> -
> -struct virtnet_rx_stats {
> -	struct virtnet_rq_stat_items rx;
> -};
> -
>  #define VIRTNET_SQ_STAT(m)	offsetof(struct virtnet_sq_stats, m)
> -#define VIRTNET_RQ_STAT(m)	offsetof(struct virtnet_rq_stat_items, m)
> +#define VIRTNET_RQ_STAT(m)	offsetof(struct virtnet_rq_stats, m)
>  
>  static const struct virtnet_stat_desc virtnet_sq_stats_desc[] = {
>  	{ "packets",		VIRTNET_SQ_STAT(packets) },
> @@ -617,7 +609,7 @@ static struct sk_buff *receive_small(struct net_device *dev,
>  				     void *buf, void *ctx,
>  				     unsigned int len,
>  				     unsigned int *xdp_xmit,
> -				     struct virtnet_rx_stats *stats)
> +				     struct virtnet_rq_stats *stats)
>  {
>  	struct sk_buff *skb;
>  	struct bpf_prog *xdp_prog;
> @@ -632,7 +624,7 @@ static struct sk_buff *receive_small(struct net_device *dev,
>  	int err;
>  
>  	len -= vi->hdr_len;
> -	stats->rx.bytes += len;
> +	stats->bytes += len;
>  
>  	rcu_read_lock();
>  	xdp_prog = rcu_dereference(rq->xdp_prog);
> @@ -674,7 +666,7 @@ static struct sk_buff *receive_small(struct net_device *dev,
>  		xdp.rxq = &rq->xdp_rxq;
>  		orig_data = xdp.data;
>  		act = bpf_prog_run_xdp(xdp_prog, &xdp);
> -		stats->rx.xdp_packets++;
> +		stats->xdp_packets++;
>  
>  		switch (act) {
>  		case XDP_PASS:
> @@ -683,7 +675,7 @@ static struct sk_buff *receive_small(struct net_device *dev,
>  			len = xdp.data_end - xdp.data;
>  			break;
>  		case XDP_TX:
> -			stats->rx.xdp_tx++;
> +			stats->xdp_tx++;
>  			xdpf = convert_to_xdp_frame(&xdp);
>  			if (unlikely(!xdpf))
>  				goto err_xdp;
> @@ -696,7 +688,7 @@ static struct sk_buff *receive_small(struct net_device *dev,
>  			rcu_read_unlock();
>  			goto xdp_xmit;
>  		case XDP_REDIRECT:
> -			stats->rx.xdp_redirects++;
> +			stats->xdp_redirects++;
>  			err = xdp_do_redirect(dev, &xdp, xdp_prog);
>  			if (err)
>  				goto err_xdp;
> @@ -730,8 +722,8 @@ static struct sk_buff *receive_small(struct net_device *dev,
>  
>  err_xdp:
>  	rcu_read_unlock();
> -	stats->rx.xdp_drops++;
> -	stats->rx.drops++;
> +	stats->xdp_drops++;
> +	stats->drops++;
>  	put_page(page);
>  xdp_xmit:
>  	return NULL;
> @@ -742,19 +734,19 @@ static struct sk_buff *receive_big(struct net_device *dev,
>  				   struct receive_queue *rq,
>  				   void *buf,
>  				   unsigned int len,
> -				   struct virtnet_rx_stats *stats)
> +				   struct virtnet_rq_stats *stats)
>  {
>  	struct page *page = buf;
>  	struct sk_buff *skb = page_to_skb(vi, rq, page, 0, len, PAGE_SIZE);
>  
> -	stats->rx.bytes += len - vi->hdr_len;
> +	stats->bytes += len - vi->hdr_len;
>  	if (unlikely(!skb))
>  		goto err;
>  
>  	return skb;
>  
>  err:
> -	stats->rx.drops++;
> +	stats->drops++;
>  	give_pages(rq, page);
>  	return NULL;
>  }
> @@ -766,7 +758,7 @@ static struct sk_buff *receive_mergeable(struct net_device *dev,
>  					 void *ctx,
>  					 unsigned int len,
>  					 unsigned int *xdp_xmit,
> -					 struct virtnet_rx_stats *stats)
> +					 struct virtnet_rq_stats *stats)
>  {
>  	struct virtio_net_hdr_mrg_rxbuf *hdr = buf;
>  	u16 num_buf = virtio16_to_cpu(vi->vdev, hdr->num_buffers);
> @@ -779,7 +771,7 @@ static struct sk_buff *receive_mergeable(struct net_device *dev,
>  	int err;
>  
>  	head_skb = NULL;
> -	stats->rx.bytes += len - vi->hdr_len;
> +	stats->bytes += len - vi->hdr_len;
>  
>  	rcu_read_lock();
>  	xdp_prog = rcu_dereference(rq->xdp_prog);
> @@ -828,7 +820,7 @@ static struct sk_buff *receive_mergeable(struct net_device *dev,
>  		xdp.rxq = &rq->xdp_rxq;
>  
>  		act = bpf_prog_run_xdp(xdp_prog, &xdp);
> -		stats->rx.xdp_packets++;
> +		stats->xdp_packets++;
>  
>  		switch (act) {
>  		case XDP_PASS:
> @@ -853,7 +845,7 @@ static struct sk_buff *receive_mergeable(struct net_device *dev,
>  			}
>  			break;
>  		case XDP_TX:
> -			stats->rx.xdp_tx++;
> +			stats->xdp_tx++;
>  			xdpf = convert_to_xdp_frame(&xdp);
>  			if (unlikely(!xdpf))
>  				goto err_xdp;
> @@ -870,7 +862,7 @@ static struct sk_buff *receive_mergeable(struct net_device *dev,
>  			rcu_read_unlock();
>  			goto xdp_xmit;
>  		case XDP_REDIRECT:
> -			stats->rx.xdp_redirects++;
> +			stats->xdp_redirects++;
>  			err = xdp_do_redirect(dev, &xdp, xdp_prog);
>  			if (err) {
>  				if (unlikely(xdp_page != page))
> @@ -920,7 +912,7 @@ static struct sk_buff *receive_mergeable(struct net_device *dev,
>  			goto err_buf;
>  		}
>  
> -		stats->rx.bytes += len;
> +		stats->bytes += len;
>  		page = virt_to_head_page(buf);
>  
>  		truesize = mergeable_ctx_to_truesize(ctx);
> @@ -966,7 +958,7 @@ static struct sk_buff *receive_mergeable(struct net_device *dev,
>  
>  err_xdp:
>  	rcu_read_unlock();
> -	stats->rx.xdp_drops++;
> +	stats->xdp_drops++;
>  err_skb:
>  	put_page(page);
>  	while (num_buf-- > 1) {
> @@ -977,12 +969,12 @@ static struct sk_buff *receive_mergeable(struct net_device *dev,
>  			dev->stats.rx_length_errors++;
>  			break;
>  		}
> -		stats->rx.bytes += len;
> +		stats->bytes += len;
>  		page = virt_to_head_page(buf);
>  		put_page(page);
>  	}
>  err_buf:
> -	stats->rx.drops++;
> +	stats->drops++;
>  	dev_kfree_skb(head_skb);
>  xdp_xmit:
>  	return NULL;
> @@ -991,7 +983,7 @@ static struct sk_buff *receive_mergeable(struct net_device *dev,
>  static void receive_buf(struct virtnet_info *vi, struct receive_queue *rq,
>  			void *buf, unsigned int len, void **ctx,
>  			unsigned int *xdp_xmit,
> -			struct virtnet_rx_stats *stats)
> +			struct virtnet_rq_stats *stats)
>  {
>  	struct net_device *dev = vi->dev;
>  	struct sk_buff *skb;
> @@ -1212,7 +1204,7 @@ static bool try_fill_recv(struct virtnet_info *vi, struct receive_queue *rq,
>  	} while (rq->vq->num_free);
>  	if (virtqueue_kick_prepare(rq->vq) && virtqueue_notify(rq->vq)) {
>  		u64_stats_update_begin(&rq->stats.syncp);
> -		rq->stats.items.kicks++;
> +		rq->stats.kicks++;
>  		u64_stats_update_end(&rq->stats.syncp);
>  	}
>  
> @@ -1290,7 +1282,7 @@ static int virtnet_receive(struct receive_queue *rq, int budget,
>  			   unsigned int *xdp_xmit)
>  {
>  	struct virtnet_info *vi = rq->vq->vdev->priv;
> -	struct virtnet_rx_stats stats = {};
> +	struct virtnet_rq_stats stats = {};
>  	unsigned int len;
>  	void *buf;
>  	int i;
> @@ -1298,16 +1290,16 @@ static int virtnet_receive(struct receive_queue *rq, int budget,
>  	if (!vi->big_packets || vi->mergeable_rx_bufs) {
>  		void *ctx;
>  
> -		while (stats.rx.packets < budget &&
> +		while (stats.packets < budget &&
>  		       (buf = virtqueue_get_buf_ctx(rq->vq, &len, &ctx))) {
>  			receive_buf(vi, rq, buf, len, ctx, xdp_xmit, &stats);
> -			stats.rx.packets++;
> +			stats.packets++;
>  		}
>  	} else {
> -		while (stats.rx.packets < budget &&
> +		while (stats.packets < budget &&
>  		       (buf = virtqueue_get_buf(rq->vq, &len)) != NULL) {
>  			receive_buf(vi, rq, buf, len, NULL, xdp_xmit, &stats);
> -			stats.rx.packets++;
> +			stats.packets++;
>  		}
>  	}
>  
> @@ -1321,12 +1313,12 @@ static int virtnet_receive(struct receive_queue *rq, int budget,
>  		size_t offset = virtnet_rq_stats_desc[i].offset;
>  		u64 *item;
>  
> -		item = (u64 *)((u8 *)&rq->stats.items + offset);
> -		*item += *(u64 *)((u8 *)&stats.rx + offset);
> +		item = (u64 *)((u8 *)&rq->stats + offset);
> +		*item += *(u64 *)((u8 *)&stats + offset);
>  	}
>  	u64_stats_update_end(&rq->stats.syncp);
>  
> -	return stats.rx.packets;
> +	return stats.packets;
>  }
>  
>  static void free_old_xmit_skbs(struct send_queue *sq)
> @@ -1686,9 +1678,9 @@ static void virtnet_stats(struct net_device *dev,
>  
>  		do {
>  			start = u64_stats_fetch_begin_irq(&rq->stats.syncp);
> -			rpackets = rq->stats.items.packets;
> -			rbytes   = rq->stats.items.bytes;
> -			rdrops   = rq->stats.items.drops;
> +			rpackets = rq->stats.packets;
> +			rbytes   = rq->stats.bytes;
> +			rdrops   = rq->stats.drops;
>  		} while (u64_stats_fetch_retry_irq(&rq->stats.syncp, start));
>  
>  		tot->rx_packets += rpackets;
> @@ -2078,7 +2070,7 @@ static void virtnet_get_ethtool_stats(struct net_device *dev,
>  	for (i = 0; i < vi->curr_queue_pairs; i++) {
>  		struct receive_queue *rq = &vi->rq[i];
>  
> -		stats_base = (u8 *)&rq->stats.items;
> +		stats_base = (u8 *)&rq->stats;
>  		do {
>  			start = u64_stats_fetch_begin_irq(&rq->stats.syncp);
>  			for (j = 0; j < VIRTNET_RQ_STATS_LEN; j++) {
> -- 
> 2.7.4
_______________________________________________
Virtualization mailing list
Virtualization@xxxxxxxxxxxxxxxxxxxxxxxxxx
https://lists.linuxfoundation.org/mailman/listinfo/virtualization



[Index of Archives]     [KVM Development]     [Libvirt Development]     [Libvirt Users]     [CentOS Virtualization]     [Netdev]     [Ethernet Bridging]     [Linux Wireless]     [Kernel Newbies]     [Security]     [Linux for Hams]     [Netfilter]     [Bugtraq]     [Yosemite Forum]     [MIPS Linux]     [ARM Linux]     [Linux RAID]     [Linux Admin]     [Samba]

  Powered by Linux