mirror of
https://mirrors.bfsu.edu.cn/git/linux.git
synced 2024-12-27 04:54:41 +08:00
d27a662290
zero_copy_allocator has been removed back when Bjorn Topel introduced xsk_buff_pool. Remove references to it that were dangling in the tree. Signed-off-by: Maciej Fijalkowski <maciej.fijalkowski@intel.com> Signed-off-by: Daniel Borkmann <daniel@iogearbox.net> Acked-by: Magnus Karlsson <magnus.karlsson@intel.com> Link: https://lore.kernel.org/bpf/20211210171511.11574-1-maciej.fijalkowski@intel.com
39 lines
1.3 KiB
C
39 lines
1.3 KiB
C
/* SPDX-License-Identifier: GPL-2.0 */
|
|
/* Copyright(c) 2018 Intel Corporation. */
|
|
|
|
#ifndef _I40E_XSK_H_
|
|
#define _I40E_XSK_H_
|
|
|
|
/* This value should match the pragma in the loop_unrolled_for
|
|
* macro. Why 4? It is strictly empirical. It seems to be a good
|
|
* compromise between the advantage of having simultaneous outstanding
|
|
* reads to the DMA array that can hide each others latency and the
|
|
* disadvantage of having a larger code path.
|
|
*/
|
|
#define PKTS_PER_BATCH 4
|
|
|
|
#ifdef __clang__
|
|
#define loop_unrolled_for _Pragma("clang loop unroll_count(4)") for
|
|
#elif __GNUC__ >= 8
|
|
#define loop_unrolled_for _Pragma("GCC unroll 4") for
|
|
#else
|
|
#define loop_unrolled_for for
|
|
#endif
|
|
|
|
struct i40e_vsi;
|
|
struct xsk_buff_pool;
|
|
|
|
int i40e_queue_pair_disable(struct i40e_vsi *vsi, int queue_pair);
|
|
int i40e_queue_pair_enable(struct i40e_vsi *vsi, int queue_pair);
|
|
int i40e_xsk_pool_setup(struct i40e_vsi *vsi, struct xsk_buff_pool *pool,
|
|
u16 qid);
|
|
bool i40e_alloc_rx_buffers_zc(struct i40e_ring *rx_ring, u16 cleaned_count);
|
|
int i40e_clean_rx_irq_zc(struct i40e_ring *rx_ring, int budget);
|
|
|
|
bool i40e_clean_xdp_tx_irq(struct i40e_vsi *vsi, struct i40e_ring *tx_ring);
|
|
int i40e_xsk_wakeup(struct net_device *dev, u32 queue_id, u32 flags);
|
|
int i40e_alloc_rx_bi_zc(struct i40e_ring *rx_ring);
|
|
void i40e_clear_rx_bi_zc(struct i40e_ring *rx_ring);
|
|
|
|
#endif /* _I40E_XSK_H_ */
|