agent-enviroments/builder/libs/seastar/dpdk/drivers/net/nfp/nfp_rxtx.h
2024-09-10 17:06:08 +03:00

424 lines
13 KiB
C

/* SPDX-License-Identifier: BSD-3-Clause
* Copyright (c) 2014-2021 Netronome Systems, Inc.
* All rights reserved.
*/
#ifndef _NFP_RXTX_H_
#define _NFP_RXTX_H_
#include <rte_io.h>
#define NFP_DESC_META_LEN(d) ((d)->rxd.meta_len_dd & PCIE_DESC_RX_META_LEN_MASK)
#define NFP_HASH_OFFSET ((uint8_t *)mbuf->buf_addr + mbuf->data_off - 4)
#define NFP_HASH_TYPE_OFFSET ((uint8_t *)mbuf->buf_addr + mbuf->data_off - 8)
#define RTE_MBUF_DMA_ADDR_DEFAULT(mb) \
((uint64_t)((mb)->buf_iova + RTE_PKTMBUF_HEADROOM))
/* Maximum number of supported VLANs in parsed form packet metadata. */
#define NFP_META_MAX_VLANS 2
/* Maximum number of NFP packet metadata fields. */
#define NFP_META_MAX_FIELDS 8
/*
* struct nfp_net_meta_raw - Raw memory representation of packet metadata
*
* Describe the raw metadata format, useful when preparing metadata for a
* transmission mbuf.
*
* @header: NFD3 or NFDk field type header (see format in nfp.rst)
* @data: Array of each fields data member
* @length: Keep track of number of valid fields in @header and data. Not part
* of the raw metadata.
*/
struct nfp_net_meta_raw {
uint32_t header;
uint32_t data[NFP_META_MAX_FIELDS];
uint8_t length;
};
/*
* struct nfp_meta_parsed - Record metadata parsed from packet
*
* Parsed NFP packet metadata are recorded in this struct. The content is
* read-only after it have been recorded during parsing by nfp_net_parse_meta().
*
* @port_id: Port id value
* @hash: RSS hash value
* @hash_type: RSS hash type
* @vlan_layer: The layers of VLAN info which are passed from nic.
* Only this number of entries of the @vlan array are valid.
*
* @vlan: Holds information parses from NFP_NET_META_VLAN. The inner most vlan
* starts at position 0 and only @vlan_layer entries contain valid
* information.
*
* Currently only 2 layers of vlan are supported,
* vlan[0] - vlan strip info
* vlan[1] - qinq strip info
*
* @vlan.offload: Flag indicates whether VLAN is offloaded
* @vlan.tpid: Vlan TPID
* @vlan.tci: Vlan TCI including PCP + Priority + VID
*/
struct nfp_meta_parsed {
uint32_t port_id;
uint32_t hash;
uint8_t hash_type;
uint8_t vlan_layer;
struct {
uint8_t offload;
uint8_t tpid;
uint16_t tci;
} vlan[NFP_META_MAX_VLANS];
};
/*
* The maximum number of descriptors is limited by design as
* DPDK uses uint16_t variables for these values
*/
#define NFP_NET_MAX_TX_DESC (32 * 1024)
#define NFP_NET_MIN_TX_DESC 256
#define NFP3800_NET_MIN_TX_DESC 512
#define NFP_NET_MAX_RX_DESC (32 * 1024)
#define NFP_NET_MIN_RX_DESC 256
#define NFP3800_NET_MIN_RX_DESC 512
/* Descriptor alignment */
#define NFP_ALIGN_RING_DESC 128
#define DIV_ROUND_UP(n, d) (((n) + (d) - 1) / (d))
struct nfp_net_dp_buf {
struct rte_mbuf *mbuf;
};
struct nfp_net_txq {
/** Backpointer to nfp_net structure */
struct nfp_net_hw *hw;
/** Point to the base of the queue structure on the NFP. */
uint8_t *qcp_q;
/**
* Host side read and write pointer, they are free running and
* have little relation to the QCP pointers.
*/
uint32_t wr_p;
uint32_t rd_p;
/** The size of the queue in number of descriptors. */
uint32_t tx_count;
uint32_t tx_free_thresh;
/**
* For each descriptor keep a reference to the mbuf and
* DMA address used until completion is signalled.
*/
struct nfp_net_dp_buf *txbufs;
/**
* Information about the host side queue location.
* It is the virtual address for the queue.
*/
union {
struct nfp_net_nfd3_tx_desc *txds;
struct nfp_net_nfdk_tx_desc *ktxds;
};
/** The index of the QCP queue relative to the TX queue BAR. */
uint32_t tx_qcidx;
/** The queue index from Linux's perspective. */
uint16_t qidx;
uint16_t port_id;
/** Used by NFDk only */
uint16_t data_pending;
/**
* At this point 58 bytes have been used for all the fields in the
* TX critical path. We have room for 6 bytes and still all placed
* in a cache line.
*/
uint64_t dma;
} __rte_aligned(64);
/* RX and freelist descriptor format */
#define PCIE_DESC_RX_DD (1 << 7)
#define PCIE_DESC_RX_META_LEN_MASK (0x7f)
/* Flags in the RX descriptor */
#define PCIE_DESC_RX_RSS (1 << 15)
#define PCIE_DESC_RX_I_IP4_CSUM (1 << 14)
#define PCIE_DESC_RX_I_IP4_CSUM_OK (1 << 13)
#define PCIE_DESC_RX_I_TCP_CSUM (1 << 12)
#define PCIE_DESC_RX_I_TCP_CSUM_OK (1 << 11)
#define PCIE_DESC_RX_I_UDP_CSUM (1 << 10)
#define PCIE_DESC_RX_I_UDP_CSUM_OK (1 << 9)
#define PCIE_DESC_RX_SPARE (1 << 8)
#define PCIE_DESC_RX_EOP (1 << 7)
#define PCIE_DESC_RX_IP4_CSUM (1 << 6)
#define PCIE_DESC_RX_IP4_CSUM_OK (1 << 5)
#define PCIE_DESC_RX_TCP_CSUM (1 << 4)
#define PCIE_DESC_RX_TCP_CSUM_OK (1 << 3)
#define PCIE_DESC_RX_UDP_CSUM (1 << 2)
#define PCIE_DESC_RX_UDP_CSUM_OK (1 << 1)
#define PCIE_DESC_RX_VLAN (1 << 0)
#define PCIE_DESC_RX_L4_CSUM_OK (PCIE_DESC_RX_TCP_CSUM_OK | \
PCIE_DESC_RX_UDP_CSUM_OK)
/*
* The bit format and map of nfp packet type for rxd.offload_info in Rx descriptor.
*
* Bit format about nfp packet type refers to the following:
* ---------------------------------
* 1 0
* 5 4 3 2 1 0 9 8 7 6 5 4 3 2 1 0
* +-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+
* | |ol3|tunnel | l3 | l4 |
* +-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+
*
* Bit map about nfp packet type refers to the following:
*
* L4: bit 0~2, used for layer 4 or inner layer 4.
* 000: NFP_NET_PTYPE_L4_NONE
* 001: NFP_NET_PTYPE_L4_TCP
* 010: NFP_NET_PTYPE_L4_UDP
* 011: NFP_NET_PTYPE_L4_FRAG
* 100: NFP_NET_PTYPE_L4_NONFRAG
* 101: NFP_NET_PTYPE_L4_ICMP
* 110: NFP_NET_PTYPE_L4_SCTP
* 111: reserved
*
* L3: bit 3~5, used for layer 3 or inner layer 3.
* 000: NFP_NET_PTYPE_L3_NONE
* 001: NFP_NET_PTYPE_L3_IPV6
* 010: NFP_NET_PTYPE_L3_IPV4
* 011: NFP_NET_PTYPE_L3_IPV4_EXT
* 100: NFP_NET_PTYPE_L3_IPV6_EXT
* 101: NFP_NET_PTYPE_L3_IPV4_EXT_UNKNOWN
* 110: NFP_NET_PTYPE_L3_IPV6_EXT_UNKNOWN
* 111: reserved
*
* Tunnel: bit 6~9, used for tunnel.
* 0000: NFP_NET_PTYPE_TUNNEL_NONE
* 0001: NFP_NET_PTYPE_TUNNEL_VXLAN
* 0100: NFP_NET_PTYPE_TUNNEL_NVGRE
* 0101: NFP_NET_PTYPE_TUNNEL_GENEVE
* 0010, 0011, 0110~1111: reserved
*
* Outer L3: bit 10~11, used for outer layer 3.
* 00: NFP_NET_PTYPE_OUTER_L3_NONE
* 01: NFP_NET_PTYPE_OUTER_L3_IPV6
* 10: NFP_NET_PTYPE_OUTER_L3_IPV4
* 11: reserved
*
* Reserved: bit 10~15, used for extension.
*/
/* Mask and offset about nfp packet type based on the bit map above. */
#define NFP_NET_PTYPE_L4_MASK 0x0007
#define NFP_NET_PTYPE_L3_MASK 0x0038
#define NFP_NET_PTYPE_TUNNEL_MASK 0x03c0
#define NFP_NET_PTYPE_OUTER_L3_MASK 0x0c00
#define NFP_NET_PTYPE_L4_OFFSET 0
#define NFP_NET_PTYPE_L3_OFFSET 3
#define NFP_NET_PTYPE_TUNNEL_OFFSET 6
#define NFP_NET_PTYPE_OUTER_L3_OFFSET 10
/* Case about nfp packet type based on the bit map above. */
#define NFP_NET_PTYPE_L4_NONE 0
#define NFP_NET_PTYPE_L4_TCP 1
#define NFP_NET_PTYPE_L4_UDP 2
#define NFP_NET_PTYPE_L4_FRAG 3
#define NFP_NET_PTYPE_L4_NONFRAG 4
#define NFP_NET_PTYPE_L4_ICMP 5
#define NFP_NET_PTYPE_L4_SCTP 6
#define NFP_NET_PTYPE_L3_NONE 0
#define NFP_NET_PTYPE_L3_IPV6 1
#define NFP_NET_PTYPE_L3_IPV4 2
#define NFP_NET_PTYPE_L3_IPV4_EXT 3
#define NFP_NET_PTYPE_L3_IPV6_EXT 4
#define NFP_NET_PTYPE_L3_IPV4_EXT_UNKNOWN 5
#define NFP_NET_PTYPE_L3_IPV6_EXT_UNKNOWN 6
#define NFP_NET_PTYPE_TUNNEL_NONE 0
#define NFP_NET_PTYPE_TUNNEL_VXLAN 1
#define NFP_NET_PTYPE_TUNNEL_NVGRE 4
#define NFP_NET_PTYPE_TUNNEL_GENEVE 5
#define NFP_NET_PTYPE_OUTER_L3_NONE 0
#define NFP_NET_PTYPE_OUTER_L3_IPV6 1
#define NFP_NET_PTYPE_OUTER_L3_IPV4 2
#define NFP_PTYPE2RTE(tunnel, type) ((tunnel) ? RTE_PTYPE_INNER_##type : RTE_PTYPE_##type)
/* Record NFP packet type parsed from rxd.offload_info. */
struct nfp_ptype_parsed {
uint8_t l4_ptype; /**< Packet type of layer 4, or inner layer 4. */
uint8_t l3_ptype; /**< Packet type of layer 3, or inner layer 3. */
uint8_t tunnel_ptype; /**< Packet type of tunnel. */
uint8_t outer_l3_ptype; /**< Packet type of outer layer 3. */
};
struct nfp_net_rx_desc {
union {
/** Freelist descriptor. */
struct {
uint16_t dma_addr_hi; /**< High bits of buffer address. */
uint8_t spare; /**< Reserved, must be zero. */
uint8_t dd; /**< Whether descriptor available. */
uint32_t dma_addr_lo; /**< Low bits of buffer address. */
} __rte_packed fld;
/** RX descriptor. */
struct {
uint16_t data_len; /**< Length of frame + metadata. */
uint8_t reserved; /**< Reserved, must be zero. */
uint8_t meta_len_dd; /**< Length of metadata + done flag. */
uint16_t flags; /**< RX flags. */
uint16_t offload_info; /**< Offloading info. */
} __rte_packed rxd;
/** Reserved. */
uint32_t vals[2];
};
};
struct nfp_net_rxq {
/** Backpointer to nfp_net structure */
struct nfp_net_hw *hw;
/**
* Point to the base addresses of the freelist queue
* controller peripheral queue structures on the NFP.
*/
uint8_t *qcp_fl;
/**
* Host side read pointer, free running and have little relation
* to the QCP pointers. It is where the driver start reading
* descriptors for newly arrive packets from.
*/
uint32_t rd_p;
/**
* The index of the QCP queue relative to the RX queue BAR
* used for the freelist.
*/
uint32_t fl_qcidx;
/**
* For each buffer placed on the freelist, record the
* associated mbuf.
*/
struct nfp_net_dp_buf *rxbufs;
/**
* Information about the host side queue location.
* It is the virtual address for the queue.
*/
struct nfp_net_rx_desc *rxds;
/**
* The mempool is created by the user specifying a mbuf size.
* We save here the reference of the mempool needed in the RX
* path and the mbuf size for checking received packets can be
* safely copied to the mbuf using the NFP_NET_RX_OFFSET.
*/
struct rte_mempool *mem_pool;
uint16_t mbuf_size;
/**
* Next two fields are used for giving more free descriptors
* to the NFP.
*/
uint16_t rx_free_thresh;
uint16_t nb_rx_hold;
/** The size of the queue in number of descriptors */
uint16_t rx_count;
/** Referencing dev->data->port_id */
uint16_t port_id;
/** The queue index from Linux's perspective */
uint16_t qidx;
/**
* At this point 60 bytes have been used for all the fields in the
* RX critical path. We have room for 4 bytes and still all placed
* in a cache line.
*/
/** DMA address of the queue */
uint64_t dma;
} __rte_aligned(64);
static inline void
nfp_net_mbuf_alloc_failed(struct nfp_net_rxq *rxq)
{
rte_eth_devices[rxq->port_id].data->rx_mbuf_alloc_failed++;
}
/* set mbuf checksum flags based on RX descriptor flags */
static inline void
nfp_net_rx_cksum(struct nfp_net_rxq *rxq, struct nfp_net_rx_desc *rxd,
struct rte_mbuf *mb)
{
struct nfp_net_hw *hw = rxq->hw;
if (!(hw->ctrl & NFP_NET_CFG_CTRL_RXCSUM))
return;
/* If IPv4 and IP checksum error, fail */
if (unlikely((rxd->rxd.flags & PCIE_DESC_RX_IP4_CSUM) &&
!(rxd->rxd.flags & PCIE_DESC_RX_IP4_CSUM_OK)))
mb->ol_flags |= RTE_MBUF_F_RX_IP_CKSUM_BAD;
else
mb->ol_flags |= RTE_MBUF_F_RX_IP_CKSUM_GOOD;
/* If neither UDP nor TCP return */
if (!(rxd->rxd.flags & PCIE_DESC_RX_TCP_CSUM) &&
!(rxd->rxd.flags & PCIE_DESC_RX_UDP_CSUM))
return;
if (likely(rxd->rxd.flags & PCIE_DESC_RX_L4_CSUM_OK))
mb->ol_flags |= RTE_MBUF_F_RX_L4_CKSUM_GOOD;
else
mb->ol_flags |= RTE_MBUF_F_RX_L4_CKSUM_BAD;
}
int nfp_net_rx_freelist_setup(struct rte_eth_dev *dev);
uint32_t nfp_net_rx_queue_count(void *rx_queue);
uint16_t nfp_net_recv_pkts(void *rx_queue, struct rte_mbuf **rx_pkts,
uint16_t nb_pkts);
void nfp_net_rx_queue_release(struct rte_eth_dev *dev, uint16_t queue_idx);
void nfp_net_reset_rx_queue(struct nfp_net_rxq *rxq);
int nfp_net_rx_queue_setup(struct rte_eth_dev *dev, uint16_t queue_idx,
uint16_t nb_desc, unsigned int socket_id,
const struct rte_eth_rxconf *rx_conf,
struct rte_mempool *mp);
void nfp_net_tx_queue_release(struct rte_eth_dev *dev, uint16_t queue_idx);
void nfp_net_reset_tx_queue(struct nfp_net_txq *txq);
int nfp_net_tx_queue_setup(struct rte_eth_dev *dev,
uint16_t queue_idx,
uint16_t nb_desc,
unsigned int socket_id,
const struct rte_eth_txconf *tx_conf);
int nfp_net_tx_free_bufs(struct nfp_net_txq *txq);
void nfp_net_set_meta_vlan(struct nfp_net_meta_raw *meta_data,
struct rte_mbuf *pkt,
uint8_t layer);
#endif /* _NFP_RXTX_H_ */