mirror of
https://git.kernel.org/pub/scm/linux/kernel/git/stable/linux.git
synced 2024-10-02 07:04:24 +00:00
net: mana: Refactor RX buffer allocation code to prepare for various MTU
Move out common buffer allocation code from mana_process_rx_cqe() and mana_alloc_rx_wqe() to helper functions. Refactor related variables so they can be changed in one place, and buffer sizes are in sync. Signed-off-by: Haiyang Zhang <haiyangz@microsoft.com> Reviewed-by: Jesse Brandeburg <jesse.brandeburg@intel.com> Signed-off-by: David S. Miller <davem@davemloft.net>
This commit is contained in:
parent
ce518bc3e9
commit
a2917b2349
2 changed files with 91 additions and 69 deletions
|
@ -1282,14 +1282,64 @@ static void mana_rx_skb(void *buf_va, struct mana_rxcomp_oob *cqe,
|
||||||
u64_stats_update_end(&rx_stats->syncp);
|
u64_stats_update_end(&rx_stats->syncp);
|
||||||
|
|
||||||
drop:
|
drop:
|
||||||
WARN_ON_ONCE(rxq->xdp_save_page);
|
WARN_ON_ONCE(rxq->xdp_save_va);
|
||||||
rxq->xdp_save_page = virt_to_page(buf_va);
|
/* Save for reuse */
|
||||||
|
rxq->xdp_save_va = buf_va;
|
||||||
|
|
||||||
++ndev->stats.rx_dropped;
|
++ndev->stats.rx_dropped;
|
||||||
|
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static void *mana_get_rxfrag(struct mana_rxq *rxq, struct device *dev,
|
||||||
|
dma_addr_t *da, bool is_napi)
|
||||||
|
{
|
||||||
|
struct page *page;
|
||||||
|
void *va;
|
||||||
|
|
||||||
|
/* Reuse XDP dropped page if available */
|
||||||
|
if (rxq->xdp_save_va) {
|
||||||
|
va = rxq->xdp_save_va;
|
||||||
|
rxq->xdp_save_va = NULL;
|
||||||
|
} else {
|
||||||
|
page = dev_alloc_page();
|
||||||
|
if (!page)
|
||||||
|
return NULL;
|
||||||
|
|
||||||
|
va = page_to_virt(page);
|
||||||
|
}
|
||||||
|
|
||||||
|
*da = dma_map_single(dev, va + XDP_PACKET_HEADROOM, rxq->datasize,
|
||||||
|
DMA_FROM_DEVICE);
|
||||||
|
|
||||||
|
if (dma_mapping_error(dev, *da)) {
|
||||||
|
put_page(virt_to_head_page(va));
|
||||||
|
return NULL;
|
||||||
|
}
|
||||||
|
|
||||||
|
return va;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* Allocate frag for rx buffer, and save the old buf */
|
||||||
|
static void mana_refill_rxoob(struct device *dev, struct mana_rxq *rxq,
|
||||||
|
struct mana_recv_buf_oob *rxoob, void **old_buf)
|
||||||
|
{
|
||||||
|
dma_addr_t da;
|
||||||
|
void *va;
|
||||||
|
|
||||||
|
va = mana_get_rxfrag(rxq, dev, &da, true);
|
||||||
|
|
||||||
|
if (!va)
|
||||||
|
return;
|
||||||
|
|
||||||
|
dma_unmap_single(dev, rxoob->sgl[0].address, rxq->datasize,
|
||||||
|
DMA_FROM_DEVICE);
|
||||||
|
*old_buf = rxoob->buf_va;
|
||||||
|
|
||||||
|
rxoob->buf_va = va;
|
||||||
|
rxoob->sgl[0].address = da;
|
||||||
|
}
|
||||||
|
|
||||||
static void mana_process_rx_cqe(struct mana_rxq *rxq, struct mana_cq *cq,
|
static void mana_process_rx_cqe(struct mana_rxq *rxq, struct mana_cq *cq,
|
||||||
struct gdma_comp *cqe)
|
struct gdma_comp *cqe)
|
||||||
{
|
{
|
||||||
|
@ -1299,10 +1349,8 @@ static void mana_process_rx_cqe(struct mana_rxq *rxq, struct mana_cq *cq,
|
||||||
struct mana_recv_buf_oob *rxbuf_oob;
|
struct mana_recv_buf_oob *rxbuf_oob;
|
||||||
struct mana_port_context *apc;
|
struct mana_port_context *apc;
|
||||||
struct device *dev = gc->dev;
|
struct device *dev = gc->dev;
|
||||||
void *new_buf, *old_buf;
|
void *old_buf = NULL;
|
||||||
struct page *new_page;
|
|
||||||
u32 curr, pktlen;
|
u32 curr, pktlen;
|
||||||
dma_addr_t da;
|
|
||||||
|
|
||||||
apc = netdev_priv(ndev);
|
apc = netdev_priv(ndev);
|
||||||
|
|
||||||
|
@ -1345,40 +1393,11 @@ static void mana_process_rx_cqe(struct mana_rxq *rxq, struct mana_cq *cq,
|
||||||
rxbuf_oob = &rxq->rx_oobs[curr];
|
rxbuf_oob = &rxq->rx_oobs[curr];
|
||||||
WARN_ON_ONCE(rxbuf_oob->wqe_inf.wqe_size_in_bu != 1);
|
WARN_ON_ONCE(rxbuf_oob->wqe_inf.wqe_size_in_bu != 1);
|
||||||
|
|
||||||
/* Reuse XDP dropped page if available */
|
mana_refill_rxoob(dev, rxq, rxbuf_oob, &old_buf);
|
||||||
if (rxq->xdp_save_page) {
|
|
||||||
new_page = rxq->xdp_save_page;
|
|
||||||
rxq->xdp_save_page = NULL;
|
|
||||||
} else {
|
|
||||||
new_page = alloc_page(GFP_ATOMIC);
|
|
||||||
}
|
|
||||||
|
|
||||||
if (new_page) {
|
|
||||||
da = dma_map_page(dev, new_page, XDP_PACKET_HEADROOM, rxq->datasize,
|
|
||||||
DMA_FROM_DEVICE);
|
|
||||||
|
|
||||||
if (dma_mapping_error(dev, da)) {
|
|
||||||
__free_page(new_page);
|
|
||||||
new_page = NULL;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
new_buf = new_page ? page_to_virt(new_page) : NULL;
|
|
||||||
|
|
||||||
if (new_buf) {
|
|
||||||
dma_unmap_page(dev, rxbuf_oob->buf_dma_addr, rxq->datasize,
|
|
||||||
DMA_FROM_DEVICE);
|
|
||||||
|
|
||||||
old_buf = rxbuf_oob->buf_va;
|
|
||||||
|
|
||||||
/* refresh the rxbuf_oob with the new page */
|
|
||||||
rxbuf_oob->buf_va = new_buf;
|
|
||||||
rxbuf_oob->buf_dma_addr = da;
|
|
||||||
rxbuf_oob->sgl[0].address = rxbuf_oob->buf_dma_addr;
|
|
||||||
} else {
|
|
||||||
old_buf = NULL; /* drop the packet if no memory */
|
|
||||||
}
|
|
||||||
|
|
||||||
|
/* Unsuccessful refill will have old_buf == NULL.
|
||||||
|
* In this case, mana_rx_skb() will drop the packet.
|
||||||
|
*/
|
||||||
mana_rx_skb(old_buf, oob, rxq);
|
mana_rx_skb(old_buf, oob, rxq);
|
||||||
|
|
||||||
drop:
|
drop:
|
||||||
|
@ -1659,8 +1678,8 @@ static void mana_destroy_rxq(struct mana_port_context *apc,
|
||||||
|
|
||||||
mana_deinit_cq(apc, &rxq->rx_cq);
|
mana_deinit_cq(apc, &rxq->rx_cq);
|
||||||
|
|
||||||
if (rxq->xdp_save_page)
|
if (rxq->xdp_save_va)
|
||||||
__free_page(rxq->xdp_save_page);
|
put_page(virt_to_head_page(rxq->xdp_save_va));
|
||||||
|
|
||||||
for (i = 0; i < rxq->num_rx_buf; i++) {
|
for (i = 0; i < rxq->num_rx_buf; i++) {
|
||||||
rx_oob = &rxq->rx_oobs[i];
|
rx_oob = &rxq->rx_oobs[i];
|
||||||
|
@ -1668,10 +1687,10 @@ static void mana_destroy_rxq(struct mana_port_context *apc,
|
||||||
if (!rx_oob->buf_va)
|
if (!rx_oob->buf_va)
|
||||||
continue;
|
continue;
|
||||||
|
|
||||||
dma_unmap_page(dev, rx_oob->buf_dma_addr, rxq->datasize,
|
dma_unmap_single(dev, rx_oob->sgl[0].address,
|
||||||
DMA_FROM_DEVICE);
|
rx_oob->sgl[0].size, DMA_FROM_DEVICE);
|
||||||
|
|
||||||
free_page((unsigned long)rx_oob->buf_va);
|
put_page(virt_to_head_page(rx_oob->buf_va));
|
||||||
rx_oob->buf_va = NULL;
|
rx_oob->buf_va = NULL;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -1681,6 +1700,26 @@ static void mana_destroy_rxq(struct mana_port_context *apc,
|
||||||
kfree(rxq);
|
kfree(rxq);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static int mana_fill_rx_oob(struct mana_recv_buf_oob *rx_oob, u32 mem_key,
|
||||||
|
struct mana_rxq *rxq, struct device *dev)
|
||||||
|
{
|
||||||
|
dma_addr_t da;
|
||||||
|
void *va;
|
||||||
|
|
||||||
|
va = mana_get_rxfrag(rxq, dev, &da, false);
|
||||||
|
|
||||||
|
if (!va)
|
||||||
|
return -ENOMEM;
|
||||||
|
|
||||||
|
rx_oob->buf_va = va;
|
||||||
|
|
||||||
|
rx_oob->sgl[0].address = da;
|
||||||
|
rx_oob->sgl[0].size = rxq->datasize;
|
||||||
|
rx_oob->sgl[0].mem_key = mem_key;
|
||||||
|
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
#define MANA_WQE_HEADER_SIZE 16
|
#define MANA_WQE_HEADER_SIZE 16
|
||||||
#define MANA_WQE_SGE_SIZE 16
|
#define MANA_WQE_SGE_SIZE 16
|
||||||
|
|
||||||
|
@ -1690,9 +1729,8 @@ static int mana_alloc_rx_wqe(struct mana_port_context *apc,
|
||||||
struct gdma_context *gc = apc->ac->gdma_dev->gdma_context;
|
struct gdma_context *gc = apc->ac->gdma_dev->gdma_context;
|
||||||
struct mana_recv_buf_oob *rx_oob;
|
struct mana_recv_buf_oob *rx_oob;
|
||||||
struct device *dev = gc->dev;
|
struct device *dev = gc->dev;
|
||||||
struct page *page;
|
|
||||||
dma_addr_t da;
|
|
||||||
u32 buf_idx;
|
u32 buf_idx;
|
||||||
|
int ret;
|
||||||
|
|
||||||
WARN_ON(rxq->datasize == 0 || rxq->datasize > PAGE_SIZE);
|
WARN_ON(rxq->datasize == 0 || rxq->datasize > PAGE_SIZE);
|
||||||
|
|
||||||
|
@ -1703,25 +1741,12 @@ static int mana_alloc_rx_wqe(struct mana_port_context *apc,
|
||||||
rx_oob = &rxq->rx_oobs[buf_idx];
|
rx_oob = &rxq->rx_oobs[buf_idx];
|
||||||
memset(rx_oob, 0, sizeof(*rx_oob));
|
memset(rx_oob, 0, sizeof(*rx_oob));
|
||||||
|
|
||||||
page = alloc_page(GFP_KERNEL);
|
|
||||||
if (!page)
|
|
||||||
return -ENOMEM;
|
|
||||||
|
|
||||||
da = dma_map_page(dev, page, XDP_PACKET_HEADROOM, rxq->datasize,
|
|
||||||
DMA_FROM_DEVICE);
|
|
||||||
|
|
||||||
if (dma_mapping_error(dev, da)) {
|
|
||||||
__free_page(page);
|
|
||||||
return -ENOMEM;
|
|
||||||
}
|
|
||||||
|
|
||||||
rx_oob->buf_va = page_to_virt(page);
|
|
||||||
rx_oob->buf_dma_addr = da;
|
|
||||||
|
|
||||||
rx_oob->num_sge = 1;
|
rx_oob->num_sge = 1;
|
||||||
rx_oob->sgl[0].address = rx_oob->buf_dma_addr;
|
|
||||||
rx_oob->sgl[0].size = rxq->datasize;
|
ret = mana_fill_rx_oob(rx_oob, apc->ac->gdma_dev->gpa_mkey, rxq,
|
||||||
rx_oob->sgl[0].mem_key = apc->ac->gdma_dev->gpa_mkey;
|
dev);
|
||||||
|
if (ret)
|
||||||
|
return ret;
|
||||||
|
|
||||||
rx_oob->wqe_req.sgl = rx_oob->sgl;
|
rx_oob->wqe_req.sgl = rx_oob->sgl;
|
||||||
rx_oob->wqe_req.num_sge = rx_oob->num_sge;
|
rx_oob->wqe_req.num_sge = rx_oob->num_sge;
|
||||||
|
@ -1780,9 +1805,10 @@ static struct mana_rxq *mana_create_rxq(struct mana_port_context *apc,
|
||||||
rxq->ndev = ndev;
|
rxq->ndev = ndev;
|
||||||
rxq->num_rx_buf = RX_BUFFERS_PER_QUEUE;
|
rxq->num_rx_buf = RX_BUFFERS_PER_QUEUE;
|
||||||
rxq->rxq_idx = rxq_idx;
|
rxq->rxq_idx = rxq_idx;
|
||||||
rxq->datasize = ALIGN(MAX_FRAME_SIZE, 64);
|
|
||||||
rxq->rxobj = INVALID_MANA_HANDLE;
|
rxq->rxobj = INVALID_MANA_HANDLE;
|
||||||
|
|
||||||
|
rxq->datasize = ALIGN(ETH_FRAME_LEN, 64);
|
||||||
|
|
||||||
err = mana_alloc_rx_wqe(apc, rxq, &rq_size, &cq_size);
|
err = mana_alloc_rx_wqe(apc, rxq, &rq_size, &cq_size);
|
||||||
if (err)
|
if (err)
|
||||||
goto out;
|
goto out;
|
||||||
|
|
|
@ -36,9 +36,6 @@ enum TRI_STATE {
|
||||||
|
|
||||||
#define COMP_ENTRY_SIZE 64
|
#define COMP_ENTRY_SIZE 64
|
||||||
|
|
||||||
#define ADAPTER_MTU_SIZE 1500
|
|
||||||
#define MAX_FRAME_SIZE (ADAPTER_MTU_SIZE + 14)
|
|
||||||
|
|
||||||
#define RX_BUFFERS_PER_QUEUE 512
|
#define RX_BUFFERS_PER_QUEUE 512
|
||||||
|
|
||||||
#define MAX_SEND_BUFFERS_PER_QUEUE 256
|
#define MAX_SEND_BUFFERS_PER_QUEUE 256
|
||||||
|
@ -282,7 +279,6 @@ struct mana_recv_buf_oob {
|
||||||
struct gdma_wqe_request wqe_req;
|
struct gdma_wqe_request wqe_req;
|
||||||
|
|
||||||
void *buf_va;
|
void *buf_va;
|
||||||
dma_addr_t buf_dma_addr;
|
|
||||||
|
|
||||||
/* SGL of the buffer going to be sent has part of the work request. */
|
/* SGL of the buffer going to be sent has part of the work request. */
|
||||||
u32 num_sge;
|
u32 num_sge;
|
||||||
|
@ -322,7 +318,7 @@ struct mana_rxq {
|
||||||
|
|
||||||
struct bpf_prog __rcu *bpf_prog;
|
struct bpf_prog __rcu *bpf_prog;
|
||||||
struct xdp_rxq_info xdp_rxq;
|
struct xdp_rxq_info xdp_rxq;
|
||||||
struct page *xdp_save_page;
|
void *xdp_save_va; /* for reusing */
|
||||||
bool xdp_flush;
|
bool xdp_flush;
|
||||||
int xdp_rc; /* XDP redirect return code */
|
int xdp_rc; /* XDP redirect return code */
|
||||||
|
|
||||||
|
|
Loading…
Reference in a new issue