On Sat, Jan 25, 2025 at 9:09 AM Martin KaFai Lau <martin.lau@xxxxxxxxx> wrote: > > On 1/20/25 5:29 PM, Jason Xing wrote: > > Introducing the lock to avoid affecting the applications which > > are not using timestamping bpf feature. > > > > Signed-off-by: Jason Xing <kerneljasonxing@xxxxxxxxx> > > --- > > net/core/skbuff.c | 6 ++++-- > > net/ipv4/tcp.c | 3 ++- > > net/ipv4/tcp_input.c | 3 ++- > > net/ipv4/tcp_output.c | 3 ++- > > 4 files changed, 10 insertions(+), 5 deletions(-) > > > > diff --git a/net/core/skbuff.c b/net/core/skbuff.c > > index 33340e0b094f..db5b4b653351 100644 > > --- a/net/core/skbuff.c > > +++ b/net/core/skbuff.c > > @@ -5605,11 +5605,13 @@ void __skb_tstamp_tx(struct sk_buff *orig_skb, > > return; > > > > /* bpf extension feature entry */ > > - if (skb_shinfo(orig_skb)->tx_flags & SKBTX_BPF) > > + if (cgroup_bpf_enabled(CGROUP_SOCK_OPS) && > > I wonder if it is really needed. The caller has just tested the tx_flags. > > > + skb_shinfo(orig_skb)->tx_flags & SKBTX_BPF) > > skb_tstamp_tx_bpf(orig_skb, sk, tstype, sw, hwtstamps); > > > > /* application feature entry */ > > - if (!skb_enable_app_tstamp(orig_skb, tstype, sw)) > > + if (cgroup_bpf_enabled(CGROUP_SOCK_OPS) && > > Same here and this one looks wrong also. The userspace may get something > unexpected in the err queue. The bpf prog may have already detached here after > setting the SKBTX_BPF earlier. Oh, thanks for spotting this case. > > > + !skb_enable_app_tstamp(orig_skb, tstype, sw)) > > return; > > > > tsflags = READ_ONCE(sk->sk_tsflags); > > diff --git a/net/ipv4/tcp.c b/net/ipv4/tcp.c > > index 49e489c346ea..d88160af00c4 100644 > > --- a/net/ipv4/tcp.c > > +++ b/net/ipv4/tcp.c > > @@ -493,7 +493,8 @@ static void tcp_tx_timestamp(struct sock *sk, struct sockcm_cookie *sockc) > > shinfo->tskey = TCP_SKB_CB(skb)->seq + skb->len - 1; > > } > > > > - if (SK_BPF_CB_FLAG_TEST(sk, SK_BPF_CB_TX_TIMESTAMPING) && skb) { > > + if (cgroup_bpf_enabled(CGROUP_SOCK_OPS) && > > This looks ok considering SK_BPF_CB_FLAG_TEST may get to another cacheline. > > > + SK_BPF_CB_FLAG_TEST(sk, SK_BPF_CB_TX_TIMESTAMPING) && skb) { > > struct skb_shared_info *shinfo = skb_shinfo(skb); > > struct tcp_skb_cb *tcb = TCP_SKB_CB(skb); > > > > diff --git a/net/ipv4/tcp_input.c b/net/ipv4/tcp_input.c > > index c8945f5be31b..e30607ba41e5 100644 > > --- a/net/ipv4/tcp_input.c > > +++ b/net/ipv4/tcp_input.c > > @@ -3324,7 +3324,8 @@ static void tcp_ack_tstamp(struct sock *sk, struct sk_buff *skb, > > > > /* Avoid cache line misses to get skb_shinfo() and shinfo->tx_flags */ > > if (likely(!TCP_SKB_CB(skb)->txstamp_ack && > > - !TCP_SKB_CB(skb)->txstamp_ack_bpf)) > > + !(cgroup_bpf_enabled(CGROUP_SOCK_OPS) && > > Same here. txtstamp_ack has just been tested.... txstamp_ack_bpf is the next bit. > > > + TCP_SKB_CB(skb)->txstamp_ack_bpf))) > > return; > > > > shinfo = skb_shinfo(skb); > > diff --git a/net/ipv4/tcp_output.c b/net/ipv4/tcp_output.c > > index fc84ca669b76..483f19c2083e 100644 > > --- a/net/ipv4/tcp_output.c > > +++ b/net/ipv4/tcp_output.c > > @@ -1556,7 +1556,8 @@ static void tcp_adjust_pcount(struct sock *sk, const struct sk_buff *skb, int de > > static bool tcp_has_tx_tstamp(const struct sk_buff *skb) > > { > > return TCP_SKB_CB(skb)->txstamp_ack || > > - TCP_SKB_CB(skb)->txstamp_ack_bpf || > > + (cgroup_bpf_enabled(CGROUP_SOCK_OPS) && > > Same here. I thought the cgroup_bpf_enabled which nearly doesn't consume any resources can protect the timestamping use everywhere. I have no strong preference. I will remove them as you suggested. Thanks, Jason > > > + TCP_SKB_CB(skb)->txstamp_ack_bpf) || > > (skb_shinfo(skb)->tx_flags & SKBTX_ANY_TSTAMP); > > } > > >