Lines Matching refs:tx_buf
34 struct ice_tx_buf *tx_buf, *first; in ice_prgm_fdir_fltr() local
66 first = &tx_ring->tx_buf[i]; in ice_prgm_fdir_fltr()
73 tx_buf = &tx_ring->tx_buf[i]; in ice_prgm_fdir_fltr()
78 memset(tx_buf, 0, sizeof(*tx_buf)); in ice_prgm_fdir_fltr()
79 dma_unmap_len_set(tx_buf, len, ICE_FDIR_MAX_RAW_PKT_SIZE); in ice_prgm_fdir_fltr()
80 dma_unmap_addr_set(tx_buf, dma, dma); in ice_prgm_fdir_fltr()
86 tx_buf->tx_flags = ICE_TX_FLAGS_DUMMY_PKT; in ice_prgm_fdir_fltr()
87 tx_buf->raw_buf = raw_packet; in ice_prgm_fdir_fltr()
111 ice_unmap_and_free_tx_buf(struct ice_tx_ring *ring, struct ice_tx_buf *tx_buf) in ice_unmap_and_free_tx_buf() argument
113 if (tx_buf->skb) { in ice_unmap_and_free_tx_buf()
114 if (tx_buf->tx_flags & ICE_TX_FLAGS_DUMMY_PKT) in ice_unmap_and_free_tx_buf()
115 devm_kfree(ring->dev, tx_buf->raw_buf); in ice_unmap_and_free_tx_buf()
117 page_frag_free(tx_buf->raw_buf); in ice_unmap_and_free_tx_buf()
119 dev_kfree_skb_any(tx_buf->skb); in ice_unmap_and_free_tx_buf()
120 if (dma_unmap_len(tx_buf, len)) in ice_unmap_and_free_tx_buf()
122 dma_unmap_addr(tx_buf, dma), in ice_unmap_and_free_tx_buf()
123 dma_unmap_len(tx_buf, len), in ice_unmap_and_free_tx_buf()
125 } else if (dma_unmap_len(tx_buf, len)) { in ice_unmap_and_free_tx_buf()
127 dma_unmap_addr(tx_buf, dma), in ice_unmap_and_free_tx_buf()
128 dma_unmap_len(tx_buf, len), in ice_unmap_and_free_tx_buf()
132 tx_buf->next_to_watch = NULL; in ice_unmap_and_free_tx_buf()
133 tx_buf->skb = NULL; in ice_unmap_and_free_tx_buf()
134 dma_unmap_len_set(tx_buf, len, 0); in ice_unmap_and_free_tx_buf()
158 if (!tx_ring->tx_buf) in ice_clean_tx_ring()
163 ice_unmap_and_free_tx_buf(tx_ring, &tx_ring->tx_buf[i]); in ice_clean_tx_ring()
166 memset(tx_ring->tx_buf, 0, sizeof(*tx_ring->tx_buf) * tx_ring->count); in ice_clean_tx_ring()
194 devm_kfree(tx_ring->dev, tx_ring->tx_buf); in ice_free_tx_ring()
195 tx_ring->tx_buf = NULL; in ice_free_tx_ring()
220 struct ice_tx_buf *tx_buf; in ice_clean_tx_irq() local
222 tx_buf = &tx_ring->tx_buf[i]; in ice_clean_tx_irq()
229 struct ice_tx_desc *eop_desc = tx_buf->next_to_watch; in ice_clean_tx_irq()
237 ice_trace(clean_tx_irq, tx_ring, tx_desc, tx_buf); in ice_clean_tx_irq()
244 tx_buf->next_to_watch = NULL; in ice_clean_tx_irq()
247 total_bytes += tx_buf->bytecount; in ice_clean_tx_irq()
248 total_pkts += tx_buf->gso_segs; in ice_clean_tx_irq()
251 napi_consume_skb(tx_buf->skb, napi_budget); in ice_clean_tx_irq()
255 dma_unmap_addr(tx_buf, dma), in ice_clean_tx_irq()
256 dma_unmap_len(tx_buf, len), in ice_clean_tx_irq()
260 tx_buf->skb = NULL; in ice_clean_tx_irq()
261 dma_unmap_len_set(tx_buf, len, 0); in ice_clean_tx_irq()
265 ice_trace(clean_tx_irq_unmap, tx_ring, tx_desc, tx_buf); in ice_clean_tx_irq()
266 tx_buf++; in ice_clean_tx_irq()
271 tx_buf = tx_ring->tx_buf; in ice_clean_tx_irq()
276 if (dma_unmap_len(tx_buf, len)) { in ice_clean_tx_irq()
278 dma_unmap_addr(tx_buf, dma), in ice_clean_tx_irq()
279 dma_unmap_len(tx_buf, len), in ice_clean_tx_irq()
281 dma_unmap_len_set(tx_buf, len, 0); in ice_clean_tx_irq()
284 ice_trace(clean_tx_irq_unmap_eop, tx_ring, tx_desc, tx_buf); in ice_clean_tx_irq()
287 tx_buf++; in ice_clean_tx_irq()
292 tx_buf = tx_ring->tx_buf; in ice_clean_tx_irq()
344 WARN_ON(tx_ring->tx_buf); in ice_setup_tx_ring()
345 tx_ring->tx_buf = in ice_setup_tx_ring()
346 devm_kcalloc(dev, sizeof(*tx_ring->tx_buf), tx_ring->count, in ice_setup_tx_ring()
348 if (!tx_ring->tx_buf) in ice_setup_tx_ring()
368 devm_kfree(dev, tx_ring->tx_buf); in ice_setup_tx_ring()
369 tx_ring->tx_buf = NULL; in ice_setup_tx_ring()
1574 struct ice_tx_buf *tx_buf; in ice_tx_map() local
1597 tx_buf = first; in ice_tx_map()
1606 dma_unmap_len_set(tx_buf, len, size); in ice_tx_map()
1607 dma_unmap_addr_set(tx_buf, dma, dma); in ice_tx_map()
1656 tx_buf = &tx_ring->tx_buf[i]; in ice_tx_map()
1698 tx_buf = &tx_ring->tx_buf[i]; in ice_tx_map()
1699 ice_unmap_and_free_tx_buf(tx_ring, tx_buf); in ice_tx_map()
1700 if (tx_buf == first) in ice_tx_map()
2278 first = &tx_ring->tx_buf[tx_ring->next_to_use]; in ice_xmit_frame_ring()
2403 struct ice_tx_buf *tx_buf; in ice_clean_ctrl_tx_irq() local
2405 tx_buf = &tx_ring->tx_buf[i]; in ice_clean_ctrl_tx_irq()
2410 struct ice_tx_desc *eop_desc = tx_buf->next_to_watch; in ice_clean_ctrl_tx_irq()
2425 tx_buf->next_to_watch = NULL; in ice_clean_ctrl_tx_irq()
2430 tx_buf++; in ice_clean_ctrl_tx_irq()
2435 tx_buf = tx_ring->tx_buf; in ice_clean_ctrl_tx_irq()
2440 if (dma_unmap_len(tx_buf, len)) in ice_clean_ctrl_tx_irq()
2442 dma_unmap_addr(tx_buf, dma), in ice_clean_ctrl_tx_irq()
2443 dma_unmap_len(tx_buf, len), in ice_clean_ctrl_tx_irq()
2445 if (tx_buf->tx_flags & ICE_TX_FLAGS_DUMMY_PKT) in ice_clean_ctrl_tx_irq()
2446 devm_kfree(tx_ring->dev, tx_buf->raw_buf); in ice_clean_ctrl_tx_irq()
2449 tx_buf->raw_buf = NULL; in ice_clean_ctrl_tx_irq()
2450 tx_buf->tx_flags = 0; in ice_clean_ctrl_tx_irq()
2451 tx_buf->next_to_watch = NULL; in ice_clean_ctrl_tx_irq()
2452 dma_unmap_len_set(tx_buf, len, 0); in ice_clean_ctrl_tx_irq()
2457 tx_buf++; in ice_clean_ctrl_tx_irq()
2462 tx_buf = tx_ring->tx_buf; in ice_clean_ctrl_tx_irq()