The indr_block infrastructure will depend on the tcf_block_cb, move this code on top to avoid forward declarations. Signed-off-by: Pablo Neira Ayuso <pablo@xxxxxxxxxxxxx> --- net/sched/cls_api.c | 484 ++++++++++++++++++++++++++-------------------------- 1 file changed, 242 insertions(+), 242 deletions(-) diff --git a/net/sched/cls_api.c b/net/sched/cls_api.c index 9115f053883f..71ec0a34100d 100644 --- a/net/sched/cls_api.c +++ b/net/sched/cls_api.c @@ -568,6 +568,248 @@ static struct tcf_block *tc_dev_ingress_block(struct net_device *dev) return cops->tcf_block(qdisc, TC_H_MIN_INGRESS, NULL); } +static struct tcf_chain * +__tcf_get_next_chain(struct tcf_block *block, struct tcf_chain *chain) +{ + mutex_lock(&block->lock); + if (chain) + chain = list_is_last(&chain->list, &block->chain_list) ? + NULL : list_next_entry(chain, list); + else + chain = list_first_entry_or_null(&block->chain_list, + struct tcf_chain, list); + + /* skip all action-only chains */ + while (chain && tcf_chain_held_by_acts_only(chain)) + chain = list_is_last(&chain->list, &block->chain_list) ? + NULL : list_next_entry(chain, list); + + if (chain) + tcf_chain_hold(chain); + mutex_unlock(&block->lock); + + return chain; +} + +/* Function to be used by all clients that want to iterate over all chains on + * block. It properly obtains block->lock and takes reference to chain before + * returning it. Users of this function must be tolerant to concurrent chain + * insertion/deletion or ensure that no concurrent chain modification is + * possible. Note that all netlink dump callbacks cannot guarantee to provide + * consistent dump because rtnl lock is released each time skb is filled with + * data and sent to user-space. + */ + +struct tcf_chain * +tcf_get_next_chain(struct tcf_block *block, struct tcf_chain *chain) +{ + struct tcf_chain *chain_next = __tcf_get_next_chain(block, chain); + + if (chain) + tcf_chain_put(chain); + + return chain_next; +} +EXPORT_SYMBOL(tcf_get_next_chain); + +static struct tcf_proto * +__tcf_get_next_proto(struct tcf_chain *chain, struct tcf_proto *tp) +{ + u32 prio = 0; + + ASSERT_RTNL(); + mutex_lock(&chain->filter_chain_lock); + + if (!tp) { + tp = tcf_chain_dereference(chain->filter_chain, chain); + } else if (tcf_proto_is_deleting(tp)) { + /* 'deleting' flag is set and chain->filter_chain_lock was + * unlocked, which means next pointer could be invalid. Restart + * search. + */ + prio = tp->prio + 1; + tp = tcf_chain_dereference(chain->filter_chain, chain); + + for (; tp; tp = tcf_chain_dereference(tp->next, chain)) + if (!tp->deleting && tp->prio >= prio) + break; + } else { + tp = tcf_chain_dereference(tp->next, chain); + } + + if (tp) + tcf_proto_get(tp); + + mutex_unlock(&chain->filter_chain_lock); + + return tp; +} + +/* Function to be used by all clients that want to iterate over all tp's on + * chain. Users of this function must be tolerant to concurrent tp + * insertion/deletion or ensure that no concurrent chain modification is + * possible. Note that all netlink dump callbacks cannot guarantee to provide + * consistent dump because rtnl lock is released each time skb is filled with + * data and sent to user-space. + */ + +struct tcf_proto * +tcf_get_next_proto(struct tcf_chain *chain, struct tcf_proto *tp, + bool rtnl_held) +{ + struct tcf_proto *tp_next = __tcf_get_next_proto(chain, tp); + + if (tp) + tcf_proto_put(tp, rtnl_held, NULL); + + return tp_next; +} +EXPORT_SYMBOL(tcf_get_next_proto); + +static int +tcf_block_playback_offloads(struct tcf_block *block, tc_setup_cb_t *cb, + void *cb_priv, bool add, bool offload_in_use, + struct netlink_ext_ack *extack) +{ + struct tcf_chain *chain, *chain_prev; + struct tcf_proto *tp, *tp_prev; + int err; + + for (chain = __tcf_get_next_chain(block, NULL); + chain; + chain_prev = chain, + chain = __tcf_get_next_chain(block, chain), + tcf_chain_put(chain_prev)) { + for (tp = __tcf_get_next_proto(chain, NULL); tp; + tp_prev = tp, + tp = __tcf_get_next_proto(chain, tp), + tcf_proto_put(tp_prev, true, NULL)) { + if (tp->ops->reoffload) { + err = tp->ops->reoffload(tp, add, cb, cb_priv, + extack); + if (err && add) + goto err_playback_remove; + } else if (add && offload_in_use) { + err = -EOPNOTSUPP; + NL_SET_ERR_MSG(extack, "Filter HW offload failed - classifier without re-offloading support"); + goto err_playback_remove; + } + } + } + + return 0; + +err_playback_remove: + tcf_proto_put(tp, true, NULL); + tcf_chain_put(chain); + tcf_block_playback_offloads(block, cb, cb_priv, false, offload_in_use, + extack); + return err; +} + +struct tcf_block_cb { + struct list_head list; + tc_setup_cb_t *cb; + void *cb_ident; + void *cb_priv; + unsigned int refcnt; +}; + +static bool tcf_block_offload_in_use(struct tcf_block *block) +{ + return block->offloadcnt; +} + +void *tcf_block_cb_priv(struct tcf_block_cb *block_cb) +{ + return block_cb->cb_priv; +} +EXPORT_SYMBOL(tcf_block_cb_priv); + +struct tcf_block_cb *tcf_block_cb_lookup(struct tcf_block *block, + tc_setup_cb_t *cb, void *cb_ident) +{ struct tcf_block_cb *block_cb; + + list_for_each_entry(block_cb, &block->cb_list, list) + if (block_cb->cb == cb && block_cb->cb_ident == cb_ident) + return block_cb; + return NULL; +} +EXPORT_SYMBOL(tcf_block_cb_lookup); + +void tcf_block_cb_incref(struct tcf_block_cb *block_cb) +{ + block_cb->refcnt++; +} +EXPORT_SYMBOL(tcf_block_cb_incref); + +unsigned int tcf_block_cb_decref(struct tcf_block_cb *block_cb) +{ + return --block_cb->refcnt; +} +EXPORT_SYMBOL(tcf_block_cb_decref); + +struct tcf_block_cb *__tcf_block_cb_register(struct tcf_block *block, + tc_setup_cb_t *cb, void *cb_ident, + void *cb_priv, + struct netlink_ext_ack *extack) +{ + struct tcf_block_cb *block_cb; + int err; + + /* Replay any already present rules */ + err = tcf_block_playback_offloads(block, cb, cb_priv, true, + tcf_block_offload_in_use(block), + extack); + if (err) + return ERR_PTR(err); + + block_cb = kzalloc(sizeof(*block_cb), GFP_KERNEL); + if (!block_cb) + return ERR_PTR(-ENOMEM); + block_cb->cb = cb; + block_cb->cb_ident = cb_ident; + block_cb->cb_priv = cb_priv; + list_add(&block_cb->list, &block->cb_list); + return block_cb; +} +EXPORT_SYMBOL(__tcf_block_cb_register); + +int tcf_block_cb_register(struct tcf_block *block, + tc_setup_cb_t *cb, void *cb_ident, + void *cb_priv, struct netlink_ext_ack *extack) +{ + struct tcf_block_cb *block_cb; + + block_cb = __tcf_block_cb_register(block, cb, cb_ident, cb_priv, + extack); + return PTR_ERR_OR_ZERO(block_cb); +} +EXPORT_SYMBOL(tcf_block_cb_register); + +void __tcf_block_cb_unregister(struct tcf_block *block, + struct tcf_block_cb *block_cb) +{ + tcf_block_playback_offloads(block, block_cb->cb, block_cb->cb_priv, + false, tcf_block_offload_in_use(block), + NULL); + list_del(&block_cb->list); + kfree(block_cb); +} +EXPORT_SYMBOL(__tcf_block_cb_unregister); + +void tcf_block_cb_unregister(struct tcf_block *block, + tc_setup_cb_t *cb, void *cb_ident) +{ + struct tcf_block_cb *block_cb; + + block_cb = tcf_block_cb_lookup(block, cb, cb_ident); + if (!block_cb) + return; + __tcf_block_cb_unregister(block, block_cb); +} +EXPORT_SYMBOL(tcf_block_cb_unregister); + static struct rhashtable indr_setup_block_ht; struct tc_indr_block_dev { @@ -785,11 +1027,6 @@ static void tc_indr_block_call(struct tcf_block *block, struct net_device *dev, &bo); } -static bool tcf_block_offload_in_use(struct tcf_block *block) -{ - return block->offloadcnt; -} - static int tcf_block_offload_cmd(struct tcf_block *block, struct net_device *dev, struct tcf_block_ext_info *ei, @@ -1006,104 +1243,6 @@ static struct tcf_block *tcf_block_refcnt_get(struct net *net, u32 block_index) return block; } -static struct tcf_chain * -__tcf_get_next_chain(struct tcf_block *block, struct tcf_chain *chain) -{ - mutex_lock(&block->lock); - if (chain) - chain = list_is_last(&chain->list, &block->chain_list) ? - NULL : list_next_entry(chain, list); - else - chain = list_first_entry_or_null(&block->chain_list, - struct tcf_chain, list); - - /* skip all action-only chains */ - while (chain && tcf_chain_held_by_acts_only(chain)) - chain = list_is_last(&chain->list, &block->chain_list) ? - NULL : list_next_entry(chain, list); - - if (chain) - tcf_chain_hold(chain); - mutex_unlock(&block->lock); - - return chain; -} - -/* Function to be used by all clients that want to iterate over all chains on - * block. It properly obtains block->lock and takes reference to chain before - * returning it. Users of this function must be tolerant to concurrent chain - * insertion/deletion or ensure that no concurrent chain modification is - * possible. Note that all netlink dump callbacks cannot guarantee to provide - * consistent dump because rtnl lock is released each time skb is filled with - * data and sent to user-space. - */ - -struct tcf_chain * -tcf_get_next_chain(struct tcf_block *block, struct tcf_chain *chain) -{ - struct tcf_chain *chain_next = __tcf_get_next_chain(block, chain); - - if (chain) - tcf_chain_put(chain); - - return chain_next; -} -EXPORT_SYMBOL(tcf_get_next_chain); - -static struct tcf_proto * -__tcf_get_next_proto(struct tcf_chain *chain, struct tcf_proto *tp) -{ - u32 prio = 0; - - ASSERT_RTNL(); - mutex_lock(&chain->filter_chain_lock); - - if (!tp) { - tp = tcf_chain_dereference(chain->filter_chain, chain); - } else if (tcf_proto_is_deleting(tp)) { - /* 'deleting' flag is set and chain->filter_chain_lock was - * unlocked, which means next pointer could be invalid. Restart - * search. - */ - prio = tp->prio + 1; - tp = tcf_chain_dereference(chain->filter_chain, chain); - - for (; tp; tp = tcf_chain_dereference(tp->next, chain)) - if (!tp->deleting && tp->prio >= prio) - break; - } else { - tp = tcf_chain_dereference(tp->next, chain); - } - - if (tp) - tcf_proto_get(tp); - - mutex_unlock(&chain->filter_chain_lock); - - return tp; -} - -/* Function to be used by all clients that want to iterate over all tp's on - * chain. Users of this function must be tolerant to concurrent tp - * insertion/deletion or ensure that no concurrent chain modification is - * possible. Note that all netlink dump callbacks cannot guarantee to provide - * consistent dump because rtnl lock is released each time skb is filled with - * data and sent to user-space. - */ - -struct tcf_proto * -tcf_get_next_proto(struct tcf_chain *chain, struct tcf_proto *tp, - bool rtnl_held) -{ - struct tcf_proto *tp_next = __tcf_get_next_proto(chain, tp); - - if (tp) - tcf_proto_put(tp, rtnl_held, NULL); - - return tp_next; -} -EXPORT_SYMBOL(tcf_get_next_proto); - static void tcf_block_flush_all_chains(struct tcf_block *block, bool rtnl_held) { struct tcf_chain *chain; @@ -1497,145 +1636,6 @@ void tcf_block_put(struct tcf_block *block) EXPORT_SYMBOL(tcf_block_put); -struct tcf_block_cb { - struct list_head list; - tc_setup_cb_t *cb; - void *cb_ident; - void *cb_priv; - unsigned int refcnt; -}; - -void *tcf_block_cb_priv(struct tcf_block_cb *block_cb) -{ - return block_cb->cb_priv; -} -EXPORT_SYMBOL(tcf_block_cb_priv); - -struct tcf_block_cb *tcf_block_cb_lookup(struct tcf_block *block, - tc_setup_cb_t *cb, void *cb_ident) -{ struct tcf_block_cb *block_cb; - - list_for_each_entry(block_cb, &block->cb_list, list) - if (block_cb->cb == cb && block_cb->cb_ident == cb_ident) - return block_cb; - return NULL; -} -EXPORT_SYMBOL(tcf_block_cb_lookup); - -void tcf_block_cb_incref(struct tcf_block_cb *block_cb) -{ - block_cb->refcnt++; -} -EXPORT_SYMBOL(tcf_block_cb_incref); - -unsigned int tcf_block_cb_decref(struct tcf_block_cb *block_cb) -{ - return --block_cb->refcnt; -} -EXPORT_SYMBOL(tcf_block_cb_decref); - -static int -tcf_block_playback_offloads(struct tcf_block *block, tc_setup_cb_t *cb, - void *cb_priv, bool add, bool offload_in_use, - struct netlink_ext_ack *extack) -{ - struct tcf_chain *chain, *chain_prev; - struct tcf_proto *tp, *tp_prev; - int err; - - for (chain = __tcf_get_next_chain(block, NULL); - chain; - chain_prev = chain, - chain = __tcf_get_next_chain(block, chain), - tcf_chain_put(chain_prev)) { - for (tp = __tcf_get_next_proto(chain, NULL); tp; - tp_prev = tp, - tp = __tcf_get_next_proto(chain, tp), - tcf_proto_put(tp_prev, true, NULL)) { - if (tp->ops->reoffload) { - err = tp->ops->reoffload(tp, add, cb, cb_priv, - extack); - if (err && add) - goto err_playback_remove; - } else if (add && offload_in_use) { - err = -EOPNOTSUPP; - NL_SET_ERR_MSG(extack, "Filter HW offload failed - classifier without re-offloading support"); - goto err_playback_remove; - } - } - } - - return 0; - -err_playback_remove: - tcf_proto_put(tp, true, NULL); - tcf_chain_put(chain); - tcf_block_playback_offloads(block, cb, cb_priv, false, offload_in_use, - extack); - return err; -} - -struct tcf_block_cb *__tcf_block_cb_register(struct tcf_block *block, - tc_setup_cb_t *cb, void *cb_ident, - void *cb_priv, - struct netlink_ext_ack *extack) -{ - struct tcf_block_cb *block_cb; - int err; - - /* Replay any already present rules */ - err = tcf_block_playback_offloads(block, cb, cb_priv, true, - tcf_block_offload_in_use(block), - extack); - if (err) - return ERR_PTR(err); - - block_cb = kzalloc(sizeof(*block_cb), GFP_KERNEL); - if (!block_cb) - return ERR_PTR(-ENOMEM); - block_cb->cb = cb; - block_cb->cb_ident = cb_ident; - block_cb->cb_priv = cb_priv; - list_add(&block_cb->list, &block->cb_list); - return block_cb; -} -EXPORT_SYMBOL(__tcf_block_cb_register); - -int tcf_block_cb_register(struct tcf_block *block, - tc_setup_cb_t *cb, void *cb_ident, - void *cb_priv, struct netlink_ext_ack *extack) -{ - struct tcf_block_cb *block_cb; - - block_cb = __tcf_block_cb_register(block, cb, cb_ident, cb_priv, - extack); - return PTR_ERR_OR_ZERO(block_cb); -} -EXPORT_SYMBOL(tcf_block_cb_register); - -void __tcf_block_cb_unregister(struct tcf_block *block, - struct tcf_block_cb *block_cb) -{ - tcf_block_playback_offloads(block, block_cb->cb, block_cb->cb_priv, - false, tcf_block_offload_in_use(block), - NULL); - list_del(&block_cb->list); - kfree(block_cb); -} -EXPORT_SYMBOL(__tcf_block_cb_unregister); - -void tcf_block_cb_unregister(struct tcf_block *block, - tc_setup_cb_t *cb, void *cb_ident) -{ - struct tcf_block_cb *block_cb; - - block_cb = tcf_block_cb_lookup(block, cb, cb_ident); - if (!block_cb) - return; - __tcf_block_cb_unregister(block, block_cb); -} -EXPORT_SYMBOL(tcf_block_cb_unregister); - /* Main classifier routine: scans classifier chain attached * to this qdisc, (optionally) tests for protocol and asks * specific classifiers. -- 2.11.0