summaryrefslogtreecommitdiff
path: root/drivers/net/ethernet/intel/ice/ice_flow.h
diff options
context:
space:
mode:
authorJakub Kicinski <kuba@kernel.org>2025-11-07 19:02:42 -0800
committerJakub Kicinski <kuba@kernel.org>2025-11-07 19:02:43 -0800
commit86b721bb0b4c44499a29253b9fe26fa9ec5d2d82 (patch)
tree3262c2483a8503e1ee04820311157fb6db8726d7 /drivers/net/ethernet/intel/ice/ice_flow.h
parentfd9557c3606bb683c01a6c7627e915b539b9a8df (diff)
parentb61785852ed0a0e7dc16b606157e4a0228cd76cf (diff)
Merge branch 'net-use-skb_attempt_defer_free-in-napi_consume_skb'
Eric Dumazet says: ==================== net: use skb_attempt_defer_free() in napi_consume_skb() There is a lack of NUMA awareness and more generally lack of slab caches affinity on TX completion path. Modern drivers are using napi_consume_skb(), hoping to cache sk_buff in per-cpu caches so that they can be recycled in RX path. Only use this if the skb was allocated on the same cpu, otherwise use skb_attempt_defer_free() so that the skb is freed on the original cpu. This removes contention on SLUB spinlocks and data structures, and this makes sure that recycled sk_buff have correct NUMA locality. After this series, I get ~50% improvement for an UDP tx workload on an AMD EPYC 9B45 (IDPF 200Gbit NIC with 32 TX queues). I will later refactor skb_attempt_defer_free() to no longer have to care of skb_shared() and skb_release_head_state(). ==================== Link: https://patch.msgid.link/20251106202935.1776179-1-edumazet@google.com Signed-off-by: Jakub Kicinski <kuba@kernel.org>
Diffstat (limited to 'drivers/net/ethernet/intel/ice/ice_flow.h')
0 files changed, 0 insertions, 0 deletions