Searched refs:tcp_seq (Results 1 – 9 of 9) sorted by relevance
| /linux/net/tls/ |
| A D | trace.h | 25 __field( u32, tcp_seq ) 33 __entry->tcp_seq = tcp_seq; 54 __field( u32, tcp_seq ) 63 __entry->tcp_seq = tcp_seq; 86 __field( u32, tcp_seq ) 93 __entry->tcp_seq = tcp_seq; 155 __field( u32, tcp_seq ) 161 __entry->tcp_seq = tcp_seq; 175 TP_ARGS(sk, tcp_seq, rec_no), 180 __field( u32, tcp_seq ) [all …]
|
| A D | tls_device_fallback.c | 237 u32 tcp_seq = ntohl(tcp_hdr(skb)->seq); in fill_sg_in() local 244 record = tls_get_record(ctx, tcp_seq, rcd_sn); in fill_sg_in() 250 *sync_size = tcp_seq - tls_record_start_seq(record); in fill_sg_in()
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/en_accel/ |
| A D | tls_rxtx.c | 102 u32 tcp_seq, struct sync_info *info) in mlx5e_tls_get_sync_data() argument 110 record = tls_get_record(&context->base, tcp_seq, &info->rcd_sn); in mlx5e_tls_get_sync_data() 115 sync_size = tcp_seq - tls_record_start_seq(record); in mlx5e_tls_get_sync_data() 144 struct sk_buff *nskb, u32 tcp_seq, in mlx5e_tls_complete_sync_skb() argument 164 tcp_seq -= data_len; in mlx5e_tls_complete_sync_skb() 165 th->seq = htonl(tcp_seq); in mlx5e_tls_complete_sync_skb() 177 pet->content.send.first_seq = htons(tcp_seq); in mlx5e_tls_complete_sync_skb() 191 u32 tcp_seq = ntohl(tcp_hdr(skb)->seq); in mlx5e_tls_handle_ooo() local 200 if (mlx5e_tls_get_sync_data(context, tcp_seq, &info)) { in mlx5e_tls_handle_ooo() 236 context->expected_seq = tcp_seq + skb->len - headln; in mlx5e_tls_handle_ooo() [all …]
|
| A D | ktls_tx.c | 233 u32 tcp_seq, int datalen, struct tx_sync_info *info) in tx_sync_info_get() argument 243 record = tls_get_record(tx_ctx, tcp_seq, &info->rcd_sn); in tx_sync_info_get() 258 ends_before = before(tcp_seq + datalen - 1, tls_record_start_seq(record)); in tx_sync_info_get() 268 info->sync_len = tcp_seq - tls_record_start_seq(record); in tx_sync_info_get()
|
| /linux/drivers/net/ethernet/chelsio/inline_crypto/ch_ktls/ |
| A D | chcr_ktls.c | 880 TCB_TX_MAX_V(tcp_seq), 0); in chcr_ktls_xmit_tcb_cpls() 1228 tx_data->rsvd = htonl(tcp_seq); in chcr_ktls_xmit_wr_complete() 1404 tx_data->rsvd = htonl(tcp_seq); in chcr_ktls_xmit_wr_short() 1538 tx_data->rsvd = htonl(tcp_seq); in chcr_ktls_tx_plaintxt() 1876 tcp_seq -= prior_data_len; in chcr_short_record_handler() 1940 tcp_seq = ntohl(th->seq); in chcr_ktls_xmit() 2001 u32 tx_max = tcp_seq; in chcr_ktls_xmit() 2050 tcp_seq = record->end_seq; in chcr_ktls_xmit() 2058 tcp_seq, mss, in chcr_ktls_xmit() 2068 tcp_seq += tls_end_offset; in chcr_ktls_xmit() [all …]
|
| /linux/drivers/net/ethernet/netronome/nfp/crypto/ |
| A D | fw.h | 14 __be32 tcp_seq; member 54 __be32 tcp_seq; member 90 __be32 tcp_seq; member
|
| A D | tls.c | 335 back->tcp_seq = cpu_to_be32(start_offload_tcp_sn); in nfp_net_tls_add() 445 req->tcp_seq = cpu_to_be32(seq); in nfp_net_tls_resync() 481 __be32 tcp_seq; in nfp_net_tls_rx_resync_req() local 530 memcpy(&tcp_seq, &req->tcp_seq, sizeof(tcp_seq)); in nfp_net_tls_rx_resync_req() 531 tls_offload_rx_resync_request(sk, tcp_seq); in nfp_net_tls_rx_resync_req()
|
| /linux/net/core/ |
| A D | tso.c | 38 put_unaligned_be32(tso->tcp_seq, &tcph->seq); in tso_build_hdr() 56 tso->tcp_seq += size; /* not worth avoiding this operation for UDP */ in tso_build_data() 79 tso->tcp_seq = (tlen != sizeof(struct udphdr)) ? ntohl(tcp_hdr(skb)->seq) : 0; in tso_start()
|
| /linux/include/net/ |
| A D | tso.h | 16 u32 tcp_seq; member
|
Completed in 22 milliseconds