Thanks, I have queued this change. On Sat, Feb 19, 2011 at 06:05:08PM +0800, Changli Gao wrote: > Signed-off-by: Changli Gao <xiaosuo@xxxxxxxxx> > --- > include/net/ip_vs.h | 2 - > net/netfilter/ipvs/ip_vs_conn.c | 52 ++++++++++++++++++++++------------------ > 2 files changed, 30 insertions(+), 24 deletions(-) > diff --git a/include/net/ip_vs.h b/include/net/ip_vs.h > index e80ffb7..2078a47 100644 > --- a/include/net/ip_vs.h > +++ b/include/net/ip_vs.h > @@ -494,7 +494,7 @@ struct ip_vs_conn_param { > * IP_VS structure allocated for each dynamically scheduled connection > */ > struct ip_vs_conn { > - struct list_head c_list; /* hashed list heads */ > + struct hlist_node c_list; /* hashed list heads */ > #ifdef CONFIG_NET_NS > struct net *net; /* Name space */ > #endif > diff --git a/net/netfilter/ipvs/ip_vs_conn.c b/net/netfilter/ipvs/ip_vs_conn.c > index 83233fe..9c2a517 100644 > --- a/net/netfilter/ipvs/ip_vs_conn.c > +++ b/net/netfilter/ipvs/ip_vs_conn.c > @@ -59,7 +59,7 @@ static int ip_vs_conn_tab_mask __read_mostly; > /* > * Connection hash table: for input and output packets lookups of IPVS > */ > -static struct list_head *ip_vs_conn_tab __read_mostly; > +static struct hlist_head *ip_vs_conn_tab __read_mostly; > > /* SLAB cache for IPVS connections */ > static struct kmem_cache *ip_vs_conn_cachep __read_mostly; > @@ -201,7 +201,7 @@ static inline int ip_vs_conn_hash(struct ip_vs_conn *cp) > spin_lock(&cp->lock); > > if (!(cp->flags & IP_VS_CONN_F_HASHED)) { > - list_add(&cp->c_list, &ip_vs_conn_tab[hash]); > + hlist_add_head(&cp->c_list, &ip_vs_conn_tab[hash]); > cp->flags |= IP_VS_CONN_F_HASHED; > atomic_inc(&cp->refcnt); > ret = 1; > @@ -234,7 +234,7 @@ static inline int ip_vs_conn_unhash(struct ip_vs_conn *cp) > spin_lock(&cp->lock); > > if (cp->flags & IP_VS_CONN_F_HASHED) { > - list_del(&cp->c_list); > + hlist_del(&cp->c_list); > cp->flags &= ~IP_VS_CONN_F_HASHED; > atomic_dec(&cp->refcnt); > ret = 1; > @@ -259,12 +259,13 @@ __ip_vs_conn_in_get(const struct ip_vs_conn_param *p) > { > unsigned hash; > struct ip_vs_conn *cp; > + struct hlist_node *n; > > hash = ip_vs_conn_hashkey_param(p, false); > > ct_read_lock(hash); > > - list_for_each_entry(cp, &ip_vs_conn_tab[hash], c_list) { > + hlist_for_each_entry(cp, n, &ip_vs_conn_tab[hash], c_list) { > if (cp->af == p->af && > p->cport == cp->cport && p->vport == cp->vport && > ip_vs_addr_equal(p->af, p->caddr, &cp->caddr) && > @@ -345,12 +346,13 @@ struct ip_vs_conn *ip_vs_ct_in_get(const struct ip_vs_conn_param *p) > { > unsigned hash; > struct ip_vs_conn *cp; > + struct hlist_node *n; > > hash = ip_vs_conn_hashkey_param(p, false); > > ct_read_lock(hash); > > - list_for_each_entry(cp, &ip_vs_conn_tab[hash], c_list) { > + hlist_for_each_entry(cp, n, &ip_vs_conn_tab[hash], c_list) { > if (!ip_vs_conn_net_eq(cp, p->net)) > continue; > if (p->pe_data && p->pe->ct_match) { > @@ -394,6 +396,7 @@ struct ip_vs_conn *ip_vs_conn_out_get(const struct ip_vs_conn_param *p) > { > unsigned hash; > struct ip_vs_conn *cp, *ret=NULL; > + struct hlist_node *n; > > /* > * Check for "full" addressed entries > @@ -402,7 +405,7 @@ struct ip_vs_conn *ip_vs_conn_out_get(const struct ip_vs_conn_param *p) > > ct_read_lock(hash); > > - list_for_each_entry(cp, &ip_vs_conn_tab[hash], c_list) { > + hlist_for_each_entry(cp, n, &ip_vs_conn_tab[hash], c_list) { > if (cp->af == p->af && > p->vport == cp->cport && p->cport == cp->dport && > ip_vs_addr_equal(p->af, p->vaddr, &cp->caddr) && > @@ -818,7 +821,7 @@ ip_vs_conn_new(const struct ip_vs_conn_param *p, > return NULL; > } > > - INIT_LIST_HEAD(&cp->c_list); > + INIT_HLIST_NODE(&cp->c_list); > setup_timer(&cp->timer, ip_vs_conn_expire, (unsigned long)cp); > ip_vs_conn_net_set(cp, p->net); > cp->af = p->af; > @@ -894,8 +897,8 @@ ip_vs_conn_new(const struct ip_vs_conn_param *p, > */ > #ifdef CONFIG_PROC_FS > struct ip_vs_iter_state { > - struct seq_net_private p; > - struct list_head *l; > + struct seq_net_private p; > + struct hlist_head *l; > }; > > static void *ip_vs_conn_array(struct seq_file *seq, loff_t pos) > @@ -903,13 +906,14 @@ static void *ip_vs_conn_array(struct seq_file *seq, loff_t pos) > int idx; > struct ip_vs_conn *cp; > struct ip_vs_iter_state *iter = seq->private; > + struct hlist_node *n; > > for (idx = 0; idx < ip_vs_conn_tab_size; idx++) { > ct_read_lock_bh(idx); > - list_for_each_entry(cp, &ip_vs_conn_tab[idx], c_list) { > + hlist_for_each_entry(cp, n, &ip_vs_conn_tab[idx], c_list) { > if (pos-- == 0) { > iter->l = &ip_vs_conn_tab[idx]; > - return cp; > + return cp; > } > } > ct_read_unlock_bh(idx); > @@ -930,7 +934,8 @@ static void *ip_vs_conn_seq_next(struct seq_file *seq, void *v, loff_t *pos) > { > struct ip_vs_conn *cp = v; > struct ip_vs_iter_state *iter = seq->private; > - struct list_head *e, *l = iter->l; > + struct hlist_node *e; > + struct hlist_head *l = iter->l; > int idx; > > ++*pos; > @@ -938,15 +943,15 @@ static void *ip_vs_conn_seq_next(struct seq_file *seq, void *v, loff_t *pos) > return ip_vs_conn_array(seq, 0); > > /* more on same hash chain? */ > - if ((e = cp->c_list.next) != l) > - return list_entry(e, struct ip_vs_conn, c_list); > + if ((e = cp->c_list.next)) > + return hlist_entry(e, struct ip_vs_conn, c_list); > > idx = l - ip_vs_conn_tab; > ct_read_unlock_bh(idx); > > while (++idx < ip_vs_conn_tab_size) { > ct_read_lock_bh(idx); > - list_for_each_entry(cp, &ip_vs_conn_tab[idx], c_list) { > + hlist_for_each_entry(cp, e, &ip_vs_conn_tab[idx], c_list) { > iter->l = &ip_vs_conn_tab[idx]; > return cp; > } > @@ -959,7 +964,7 @@ static void *ip_vs_conn_seq_next(struct seq_file *seq, void *v, loff_t *pos) > static void ip_vs_conn_seq_stop(struct seq_file *seq, void *v) > { > struct ip_vs_iter_state *iter = seq->private; > - struct list_head *l = iter->l; > + struct hlist_head *l = iter->l; > > if (l) > ct_read_unlock_bh(l - ip_vs_conn_tab); > @@ -1148,13 +1153,14 @@ void ip_vs_random_dropentry(struct net *net) > */ > for (idx = 0; idx < (ip_vs_conn_tab_size>>5); idx++) { > unsigned hash = net_random() & ip_vs_conn_tab_mask; > + struct hlist_node *n; > > /* > * Lock is actually needed in this loop. > */ > ct_write_lock_bh(hash); > > - list_for_each_entry(cp, &ip_vs_conn_tab[hash], c_list) { > + hlist_for_each_entry(cp, n, &ip_vs_conn_tab[hash], c_list) { > if (cp->flags & IP_VS_CONN_F_TEMPLATE) > /* connection template */ > continue; > @@ -1202,12 +1208,14 @@ static void ip_vs_conn_flush(struct net *net) > > flush_again: > for (idx = 0; idx < ip_vs_conn_tab_size; idx++) { > + struct hlist_node *n; > + > /* > * Lock is actually needed in this loop. > */ > ct_write_lock_bh(idx); > > - list_for_each_entry(cp, &ip_vs_conn_tab[idx], c_list) { > + hlist_for_each_entry(cp, n, &ip_vs_conn_tab[idx], c_list) { > if (!ip_vs_conn_net_eq(cp, net)) > continue; > IP_VS_DBG(4, "del connection\n"); > @@ -1265,8 +1273,7 @@ int __init ip_vs_conn_init(void) > /* > * Allocate the connection hash table and initialize its list heads > */ > - ip_vs_conn_tab = vmalloc(ip_vs_conn_tab_size * > - sizeof(struct list_head)); > + ip_vs_conn_tab = vmalloc(ip_vs_conn_tab_size * sizeof(*ip_vs_conn_tab)); > if (!ip_vs_conn_tab) > return -ENOMEM; > > @@ -1286,9 +1293,8 @@ int __init ip_vs_conn_init(void) > IP_VS_DBG(0, "Each connection entry needs %Zd bytes at least\n", > sizeof(struct ip_vs_conn)); > > - for (idx = 0; idx < ip_vs_conn_tab_size; idx++) { > - INIT_LIST_HEAD(&ip_vs_conn_tab[idx]); > - } > + for (idx = 0; idx < ip_vs_conn_tab_size; idx++) > + INIT_HLIST_HEAD(&ip_vs_conn_tab[idx]); > > for (idx = 0; idx < CT_LOCKARRAY_SIZE; idx++) { > rwlock_init(&__ip_vs_conntbl_lock_array[idx].l); > -- > To unsubscribe from this list: send the line "unsubscribe lvs-devel" in > the body of a message to majordomo@xxxxxxxxxxxxxxx > More majordomo info at http://vger.kernel.org/majordomo-info.html > -- To unsubscribe from this list: send the line "unsubscribe lvs-devel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html