/linux/net/rds/ |
A D | ib_send.c | 159 if (send->s_op) { in rds_ib_send_unmap_op() 166 if (send->s_op) { in rds_ib_send_unmap_op() 201 send->s_wr.sg_list = send->s_sge; in rds_ib_send_init_ring() 220 if (send->s_op && send->s_wr.opcode != 0xdead) in rds_ib_send_clear_ring() 620 first = send; in rds_ib_xmit() 683 &send->s_wr, send->s_wr.num_sge, send->s_wr.next); in rds_ib_xmit() 701 prev = send; in rds_ib_xmit() 825 send->s_sge[0].addr, send->s_sge[0].length); in rds_ib_xmit_atomic() 833 send, &send->s_atomic_wr, ret, failed_wr); in rds_ib_xmit_atomic() 911 first = send; in rds_ib_xmit_rdma() [all …]
|
/linux/drivers/isdn/hardware/mISDN/ |
A D | netjet.c | 76 struct tiger_dma send; member 260 card->send.idx = (card->send.dmacur - card->send.dmastart) >> 2; in mode_tiger() 267 card->send.idx, in mode_tiger() 326 card->send.dmaend = card->send.dmastart + in inittiger() 328 card->send.dmairq = card->send.dmastart + in inittiger() 335 card->send.dmastart, card->send.dmairq, in inittiger() 336 card->send.dmaend, card->send.start, card->send.size); in inittiger() 468 card->send.idx = (card->send.dmacur - card->send.dmastart) >> 2; in resync() 475 if (card->send.idx < ((card->send.size / 2) - 1)) in resync() 711 card->send.idx = (card->send.dmacur - card->send.dmastart) >> 2; in nj_irq() [all …]
|
/linux/net/atm/ |
A D | raw.c | 59 return vcc->dev->ops->send(vcc, skb); in atm_send_aal0() 67 vcc->send = atm_send_aal0; in atm_init_aal0() 77 vcc->send = vcc->dev->ops->send_bh; in atm_init_aal34() 79 vcc->send = vcc->dev->ops->send; in atm_init_aal34() 89 vcc->send = vcc->dev->ops->send_bh; in atm_init_aal5() 91 vcc->send = vcc->dev->ops->send; in atm_init_aal5()
|
/linux/drivers/usb/usbip/ |
A D | usbip_common.c | 412 if (send) { in correct_endian_basic() 428 int send) in correct_endian_cmd_submit() argument 430 if (send) { in correct_endian_cmd_submit() 448 int send) in correct_endian_ret_submit() argument 450 if (send) { in correct_endian_ret_submit() 466 int send) in correct_endian_cmd_unlink() argument 468 if (send) in correct_endian_cmd_unlink() 477 if (send) in correct_endian_ret_unlink() 487 if (send) in usbip_header_correct_endian() 492 if (!send) in usbip_header_correct_endian() [all …]
|
/linux/drivers/gpu/drm/amd/display/dc/dcn30/ |
A D | dcn30_dio_stream_encoder.c | 74 uint32_t cont, send, line; in enc3_update_hdmi_info_packet() local 87 send = 1; in enc3_update_hdmi_info_packet() 92 send = 0; in enc3_update_hdmi_info_packet() 103 HDMI_GENERIC0_SEND, send); in enc3_update_hdmi_info_packet() 110 HDMI_GENERIC1_SEND, send); in enc3_update_hdmi_info_packet() 117 HDMI_GENERIC2_SEND, send); in enc3_update_hdmi_info_packet() 124 HDMI_GENERIC3_SEND, send); in enc3_update_hdmi_info_packet() 131 HDMI_GENERIC4_SEND, send); in enc3_update_hdmi_info_packet() 138 HDMI_GENERIC5_SEND, send); in enc3_update_hdmi_info_packet() 145 HDMI_GENERIC6_SEND, send); in enc3_update_hdmi_info_packet() [all …]
|
/linux/tools/testing/selftests/net/ |
A D | tls.c | 423 EXPECT_GE(send(self->fd, buf, send_len, 0), 0); in TEST_F() 672 EXPECT_GE(send(self->fd, mem_send, send_len, 0), 0); in TEST_F() 1020 EXPECT_EQ(send(self->fd, send_mem, 10, 0), 10); in TEST_F() 1021 EXPECT_EQ(send(self->fd, send_mem, 5, 0), 5); in TEST_F() 1144 int res = send(self->fd, buf, in TEST_F() 1198 int res = send(self->fd, buf, in TEST_F() 1306 res = send(self->fd, buf, in test_mutliproc() 1505 EXPECT_EQ(send(self->fd2, buf, n, 0), n); in TEST_F() 1532 EXPECT_EQ(send(self->fd2, cip[i], n, 0), n); in TEST_F() 1563 EXPECT_EQ(send(self->fd2, cip, n, 0), n); in TEST_F() [all …]
|
/linux/Documentation/devicetree/bindings/mailbox/ |
A D | apple,mailbox.yaml | 17 One of the two FIFOs is used to send data to a co-processor while the other 44 - description: send fifo is empty interrupt 45 - description: send fifo is not empty interrupt 51 - const: send-empty 52 - const: send-not-empty 74 interrupt-names = "send-empty", "send-not-empty",
|
/linux/drivers/infiniband/hw/hfi1/ |
A D | pio_copy.c | 34 void __iomem *send = dest + PIO_BLOCK_SIZE; in pio_copy() local 44 if (dend < send) { in pio_copy() 66 while (dest < send) { in pio_copy() 275 if (dend < send) { in seg_pio_copy_start() 297 while (dest < send) { in seg_pio_copy_start() 371 void __iomem *send; /* SOP end */ in mid_copy_mix() local 378 send = pbuf->start + PIO_BLOCK_SIZE; in mid_copy_mix() 379 xend = min(send, dend); in mid_copy_mix() 497 void __iomem *send; /* SOP end */ in mid_copy_straight() local 504 send = pbuf->start + PIO_BLOCK_SIZE; in mid_copy_straight() [all …]
|
/linux/drivers/mmc/host/ |
A D | meson-mx-sdio.c | 161 u32 mult, send = 0, ext = 0; in meson_mx_mmc_start_cmd() local 180 send |= MESON_MX_SDIO_SEND_RESP_CRC7_FROM_8; in meson_mx_mmc_start_cmd() 187 send |= MESON_MX_SDIO_SEND_RESP_WITHOUT_CRC7; in meson_mx_mmc_start_cmd() 190 send |= MESON_MX_SDIO_SEND_CHECK_DAT0_BUSY; in meson_mx_mmc_start_cmd() 206 send |= MESON_MX_SDIO_SEND_DATA; in meson_mx_mmc_start_cmd() 208 send |= MESON_MX_SDIO_SEND_RESP_HAS_DATA; in meson_mx_mmc_start_cmd() 236 writel(send, host->base + MESON_MX_SDIO_SEND); in meson_mx_mmc_start_cmd() 389 u32 irqs, u32 send) in meson_mx_mmc_process_cmd_irq() argument 410 (send & MESON_MX_SDIO_SEND_RESP_WITHOUT_CRC7))) in meson_mx_mmc_process_cmd_irq() 420 u32 irqs, send; in meson_mx_mmc_irq() local [all …]
|
/linux/drivers/gpu/drm/amd/display/dc/dcn20/ |
A D | dcn20_stream_encoder.c | 56 uint32_t cont, send, line; in enc2_update_hdmi_info_packet() local 68 send = 1; in enc2_update_hdmi_info_packet() 73 send = 0; in enc2_update_hdmi_info_packet() 84 HDMI_GENERIC0_SEND, send); in enc2_update_hdmi_info_packet() 91 HDMI_GENERIC1_SEND, send); in enc2_update_hdmi_info_packet() 98 HDMI_GENERIC2_SEND, send); in enc2_update_hdmi_info_packet() 105 HDMI_GENERIC3_SEND, send); in enc2_update_hdmi_info_packet() 112 HDMI_GENERIC4_SEND, send); in enc2_update_hdmi_info_packet() 119 HDMI_GENERIC5_SEND, send); in enc2_update_hdmi_info_packet() 126 HDMI_GENERIC6_SEND, send); in enc2_update_hdmi_info_packet() [all …]
|
/linux/drivers/mailbox/ |
A D | arm_mhuv2.c | 176 struct mhu2_send_frame_reg __iomem *send; member 306 &mhu->send->ch_wn[priv->ch_wn_idx].stat_set); in mhuv2_doorbell_send_data() 392 writel_relaxed(0x1, &mhu->send->ch_wn[i].int_clr); in mhuv2_data_transfer_tx_startup() 393 writel_relaxed(0x1, &mhu->send->ch_wn[i].int_en); in mhuv2_data_transfer_tx_startup() 404 writel_relaxed(0x0, &mhu->send->ch_wn[i].int_en); in mhuv2_data_transfer_tx_shutdown() 414 return !readl_relaxed(&mhu->send->ch_wn[i].stat); in mhuv2_data_transfer_last_tx_done() 972 mhu->send = reg; in mhuv2_tx_init() 999 writel_relaxed(0x0, &mhu->send->ch_wn[i].int_en); in mhuv2_tx_init() 1011 writel_relaxed(0x1, &mhu->send->access_request); in mhuv2_tx_init() 1012 while (!readl_relaxed(&mhu->send->access_ready)) in mhuv2_tx_init() [all …]
|
/linux/drivers/gpu/drm/nouveau/nvkm/subdev/pmu/ |
A D | gt215.c | 37 mutex_lock(&pmu->send.mutex); in gt215_pmu_send() 45 mutex_unlock(&pmu->send.mutex); in gt215_pmu_send() 65 pmu->send.base)); in gt215_pmu_send() 82 mutex_unlock(&pmu->send.mutex); in gt215_pmu_send() 228 pmu->send.base = nvkm_rd32(device, 0x10a4d0) & 0x0000ffff; in gt215_pmu_init() 229 pmu->send.size = nvkm_rd32(device, 0x10a4d0) >> 16; in gt215_pmu_init() 274 .send = gt215_pmu_send,
|
/linux/Documentation/input/devices/ |
A D | sentelic.rst | 529 1. send 0xf3 PS/2 command to FSP; 531 2. send 0x66 PS/2 command to FSP; 533 3. send 0x88 PS/2 command to FSP; 535 4. send 0xf3 PS/2 command to FSP; 553 7. send 0x66 PS/2 command to FSP; 574 1. send 0xf3 PS/2 command to FSP; 592 4. send 0x55 PS/2 command to FSP; 596 5. send 0xf3 PS/2 command to FSP; 614 8. send 0x33 PS/2 command to FSP; 644 1. send 0xf3 PS/2 command to FSP; [all …]
|
/linux/drivers/gpu/drm/i915/gt/uc/ |
A D | intel_guc_ct.c | 105 spin_lock_init(&ct->ctbs.send.lock); in intel_guc_ct_init_early() 324 GEM_BUG_ON(blob != ct->ctbs.send.desc); in intel_guc_ct_enable() 327 guc_ct_buffer_reset(&ct->ctbs.send); in intel_guc_ct_enable() 345 desc, cmds, ct->ctbs.send.size * 4); in intel_guc_ct_enable() 520 struct guc_ct_buffer_desc *send = ct->ctbs.send.desc; in ct_deadlocked() local 525 send->status, recv->status); in ct_deadlocked() 526 ct->ctbs.send.broken = true; in ct_deadlocked() 545 lockdep_assert_held(&ct->ctbs.send.lock); in g2h_reserve_space() 754 if (unlikely(ct->ctbs.send.broken)) in intel_guc_ct_send() 1182 ct->ctbs.send.desc->head); in intel_guc_ct_print_info() [all …]
|
/linux/Documentation/crypto/ |
A D | userspace-if.rst | 79 particular cipher instance. When invoking send/write or recv/read 80 system calls to send data to the kernel or obtain data from the 88 the input buffer used for the send/write system call and the output 120 Using the send() system call, the application provides the data that 124 - MSG_MORE: If this flag is set, the send system call acts like a 165 send/recv system call family. That cmsghdr data structure holds the 178 - MSG_MORE: If this flag is set, the send system call acts like a 180 subsequent invocation of the send system call. 241 - MSG_MORE: If this flag is set, the send system call acts like a 243 subsequent invocation of the send system call. [all …]
|
/linux/Documentation/infiniband/ |
A D | tag_matching.rst | 6 source send operations to destination receives. The following parameters must 14 The ordering rules require that when more than one pair of send and receive 15 message envelopes may match, the pair that includes the earliest posted-send 31 1. The Eager protocol- the complete message is sent when the send is 32 processed by the sender. A completion send is received in the send_cq 46 to the MPI receive routines in the posted receive list and posts send messages 47 using the MPI send routines. The head of the posted receive list may be 50 When send is initiated and arrives at the receive side, if there is no
|
/linux/Documentation/networking/ |
A D | tls.rst | 69 For example, we can send an encrypted hello world record as follows: 74 send(sock, msg, strlen(msg)); 76 send() data is directly encrypted from the userspace buffer provided 77 to the encrypted kernel send buffer if possible. 88 TLS records are created and sent after each send() call, unless 93 This buffer is allocated at the time send() is called, such that 94 either the entire send() call will return -ENOMEM (or block waiting 95 for memory), or the encryption will always succeed. If send() returns 134 /* send TLS control message using record_type */ 142 struct iovec msg_iov; /* Vector of data to send/receive into. */ [all …]
|
A D | rds.rst | 55 Sockets must be bound before you can send or receive data. 79 setsockopt(SOL_SOCKET): send and receive buffer size 80 RDS honors the send and receive buffer size socket options. 104 An attempt to send a message that exceeds SO_SNDSIZE will 135 POLLOUT is a little harder. Since you can essentially send 215 more send room) 336 The send path 343 - rds_message placed on send queue 344 - send worker awoken 356 - allocs work requests from send ring [all …]
|
/linux/drivers/char/tpm/st33zp24/ |
A D | st33zp24.c | 73 tpm_dev->ops->send(tpm_dev->phy_id, TPM_INT_STATUS, &interrupt, 1); in clear_interruption() 88 tpm_dev->ops->send(tpm_dev->phy_id, TPM_STS, &data, 1); in st33zp24_cancel() 141 ret = tpm_dev->ops->send(tpm_dev->phy_id, TPM_ACCESS, &data, 1); in request_locality() 169 tpm_dev->ops->send(tpm_dev->phy_id, TPM_ACCESS, &data, 1); in release_locality() 387 ret = tpm_dev->ops->send(tpm_dev->phy_id, TPM_DATA_FIFO, in st33zp24_send() 401 ret = tpm_dev->ops->send(tpm_dev->phy_id, TPM_DATA_FIFO, in st33zp24_send() 413 ret = tpm_dev->ops->send(tpm_dev->phy_id, TPM_STS, &data, 1); in st33zp24_send() 494 .send = st33zp24_send, 562 ret = tpm_dev->ops->send(tpm_dev->phy_id, TPM_INT_ENABLE, in st33zp24_probe() 568 ret = tpm_dev->ops->send(tpm_dev->phy_id, (TPM_INT_ENABLE + 3), in st33zp24_probe()
|
/linux/net/ipv6/ |
A D | ip6_icmp.c | 39 ip6_icmp_send_t *send; in __icmpv6_send() local 42 send = rcu_dereference(ip6_icmp_send); in __icmpv6_send() 43 if (send) in __icmpv6_send() 44 send(skb, type, code, info, NULL, parm); in __icmpv6_send()
|
/linux/tools/testing/selftests/net/af_unix/ |
A D | test_unix_oob.c | 138 send(cfd, buf, sizeof(buf), MSG_OOB); in producer() 145 send(cfd, buf, sizeof(buf), MSG_OOB); in producer() 149 send(cfd, buf, sizeof(buf), MSG_OOB); in producer() 156 send(cfd, buf, sizeof(buf), MSG_OOB); in producer() 160 send(cfd, buf, sizeof(buf), MSG_OOB); in producer() 163 send(cfd, buf, sizeof(buf), 0); in producer() 170 send(cfd, buf, 1, MSG_OOB); in producer()
|
/linux/drivers/cxl/core/ |
A D | mbox.c | 452 struct cxl_send_command send; in cxl_send_cmd() local 458 if (copy_from_user(&send, s, sizeof(send))) in cxl_send_cmd() 461 rc = cxl_validate_cmd_from_user(cxlmd->cxlm, &send, &c); in cxl_send_cmd() 469 rc = handle_mailbox_cmd_from_user(cxlm, &c, send.in.payload, in cxl_send_cmd() 470 send.out.payload, &send.out.size, in cxl_send_cmd() 471 &send.retval); in cxl_send_cmd() 475 if (copy_to_user(s, &send, sizeof(send))) in cxl_send_cmd()
|
/linux/Documentation/userspace-api/media/rc/ |
A D | lirc-func.rst | 15 lirc-get-send-mode 18 lirc-set-send-duty-cycle 23 lirc-set-send-carrier
|
/linux/drivers/net/wireless/intel/iwlwifi/dvm/ |
A D | rxon.c | 81 struct iwl_rxon_cmd *send) in iwlagn_disable_bss() argument 88 0, sizeof(*send), send); in iwlagn_disable_bss() 90 send->filter_flags = old_filter; in iwlagn_disable_bss() 101 struct iwl_rxon_cmd *send) in iwlagn_disable_pan() argument 105 u8 old_dev_type = send->dev_type; in iwlagn_disable_pan() 118 0, sizeof(*send), send); in iwlagn_disable_pan() 120 send->filter_flags = old_filter; in iwlagn_disable_pan() 121 send->dev_type = old_dev_type; in iwlagn_disable_pan() 138 struct iwl_rxon_cmd *send) in iwlagn_disconn_pan() argument 145 sizeof(*send), send); in iwlagn_disconn_pan() [all …]
|
/linux/Documentation/hid/ |
A D | intel-ish-hid.rst | 129 internal queues to sequence messages and send them in order to the firmware. 149 The transport layer uses them to send and receive messages. 178 independently. Every peer may send a message only if it has received a 181 Either side can send disconnect request bus message to end communication. Also 192 ISHTP client from either host or FW side wants to send something, it decides 193 whether to send over IPC or over DMA; for each transfer the decision is 203 send DMA; if FW doesn't support DMA then it won't respond with 213 need not wait for previous DMA_XFER to be ack'ed, and may send another message 225 TX and RX buffers respectively. On send request from client, the data to be 226 sent is copied to one of the send ring buffer and scheduled to be sent using [all …]
|