| /linux/drivers/infiniband/hw/mthca/ |
| A D | mthca_provider.c | 413 if (ib_copy_from_udata(&ucmd, udata, sizeof(ucmd))) in mthca_create_srq() 418 ucmd.db_page); in mthca_create_srq() 482 if (ib_copy_from_udata(&ucmd, udata, sizeof(ucmd))) in mthca_create_qp() 487 ucmd.sq_db_index, in mthca_create_qp() 488 ucmd.sq_db_page); in mthca_create_qp() 494 ucmd.rq_db_index, in mthca_create_qp() 603 if (ib_copy_from_udata(&ucmd, udata, sizeof(ucmd))) in mthca_create_cq() 729 if (ib_copy_from_udata(&ucmd, udata, sizeof ucmd)) { in mthca_resize_cq() 733 lkey = ucmd.lkey; in mthca_resize_cq() 855 ucmd.mr_attrs = 0; in mthca_reg_user_mr() [all …]
|
| /linux/drivers/infiniband/hw/mlx5/ |
| A D | qp.c | 370 if (ucmd) { in set_rq_size() 372 if (ucmd->rq_wqe_shift > BITS_PER_BYTE * sizeof(ucmd->rq_wqe_shift)) in set_rq_size() 568 if (ucmd->sq_wqe_count && !is_power_of_2(ucmd->sq_wqe_count)) { in set_user_buf_size() 1622 void *ucmd; member 1634 struct mlx5_ib_create_qp_rss *ucmd = params->ucmd; in create_rss_raw_qp_tir() local 1987 struct mlx5_ib_create_qp *ucmd = params->ucmd; in create_dci() local 2148 struct mlx5_ib_create_qp *ucmd = params->ucmd; in create_user_qp() local 2644 struct mlx5_ib_create_qp *ucmd = params->ucmd; in create_dct() local 2985 if (inlen > ucmd && !ib_is_udata_cleared(udata, ucmd, inlen - ucmd)) in process_udata_size() 3087 struct mlx5_ib_create_qp *ucmd = params->ucmd; in get_qp_uidx() local [all …]
|
| A D | cq.c | 732 if ((ucmd.cqe_size != 64 && ucmd.cqe_size != 128) || in create_cq_user() 733 ucmd.reserved0 || ucmd.reserved1) in create_cq_user() 736 *cqe_size = ucmd.cqe_size; in create_cq_user() 762 ucmd.buf_addr, entries * ucmd.cqe_size, in create_cq_user() 782 *index = ucmd.uar_page_index; in create_cq_user() 790 if (ucmd.cqe_comp_en == 1) { in create_cq_user() 1146 struct mlx5_ib_resize_cq ucmd; in resize_user() local 1150 err = ib_copy_from_udata(&ucmd, udata, sizeof(ucmd)); in resize_user() 1154 if (ucmd.reserved0 || ucmd.reserved1) in resize_user() 1158 if (ucmd.cqe_size && SIZE_MAX / ucmd.cqe_size <= entries - 1) in resize_user() [all …]
|
| A D | srq.c | 49 struct mlx5_ib_create_srq ucmd = {}; in create_srq_user() local 56 ucmdlen = min(udata->inlen, sizeof(ucmd)); in create_srq_user() 58 if (ib_copy_from_udata(&ucmd, udata, ucmdlen)) { in create_srq_user() 63 if (ucmd.reserved0 || ucmd.reserved1) in create_srq_user() 66 if (udata->inlen > sizeof(ucmd) && in create_srq_user() 67 !ib_is_udata_cleared(udata, sizeof(ucmd), in create_srq_user() 68 udata->inlen - sizeof(ucmd))) in create_srq_user() 72 err = get_srq_user_index(ucontext, &ucmd, udata->inlen, &uidx); in create_srq_user() 77 srq->wq_sig = !!(ucmd.flags & MLX5_SRQ_FLAG_SIGNATURE); in create_srq_user() 79 srq->umem = ib_umem_get(pd->device, ucmd.buf_addr, buf_size, 0); in create_srq_user() [all …]
|
| A D | mlx5_ib.h | 1510 struct mlx5_ib_create_qp *ucmd, in get_qp_user_index() argument 1516 if ((offsetofend(typeof(*ucmd), uidx) <= inlen) && !cqe_version && in get_qp_user_index() 1517 (ucmd->uidx == MLX5_IB_DEFAULT_UIDX)) in get_qp_user_index() 1520 if ((offsetofend(typeof(*ucmd), uidx) <= inlen) != !!cqe_version) in get_qp_user_index() 1523 return verify_assign_uidx(cqe_version, ucmd->uidx, user_index); in get_qp_user_index() 1527 struct mlx5_ib_create_srq *ucmd, in get_srq_user_index() argument 1533 if ((offsetofend(typeof(*ucmd), uidx) <= inlen) && !cqe_version && in get_srq_user_index() 1534 (ucmd->uidx == MLX5_IB_DEFAULT_UIDX)) in get_srq_user_index() 1537 if ((offsetofend(typeof(*ucmd), uidx) <= inlen) != !!cqe_version) in get_srq_user_index() 1540 return verify_assign_uidx(cqe_version, ucmd->uidx, user_index); in get_srq_user_index()
|
| A D | counters.h | 15 struct mlx5_ib_create_flow *ucmd);
|
| A D | counters.c | 744 struct mlx5_ib_create_flow *ucmd) in mlx5_ib_flow_counters_set_data() argument 752 if (ucmd && ucmd->ncounters_data != 0) { in mlx5_ib_flow_counters_set_data() 753 cntrs_data = ucmd->data; in mlx5_ib_flow_counters_set_data()
|
| A D | fs.c | 1094 struct mlx5_ib_create_flow *ucmd) in _create_flow_rule() argument 1166 err = mlx5_ib_flow_counters_set_data(flow_act.counters, ucmd); in _create_flow_rule() 1344 struct mlx5_ib_create_flow *ucmd = NULL, ucmd_hdr; in mlx5_ib_create_flow() local 1370 ucmd = kzalloc(required_ucmd_sz, GFP_KERNEL); in mlx5_ib_create_flow() 1371 if (!ucmd) in mlx5_ib_create_flow() 1374 err = ib_copy_from_udata(ucmd, udata, required_ucmd_sz); in mlx5_ib_create_flow() 1436 underlay_qpn, ucmd); in mlx5_ib_create_flow() 1458 kfree(ucmd); in mlx5_ib_create_flow() 1470 kfree(ucmd); in mlx5_ib_create_flow()
|
| /linux/drivers/infiniband/hw/hns/ |
| A D | hns_roce_cq.c | 299 struct hns_roce_ib_create_cq *ucmd) in get_cq_ucmd() argument 304 ret = ib_copy_from_udata(ucmd, udata, min(udata->inlen, sizeof(*ucmd))); in get_cq_ucmd() 314 struct hns_roce_ib_create_cq *ucmd) in set_cq_param() argument 330 struct hns_roce_ib_create_cq *ucmd) in set_cqe_size() argument 340 if (ucmd->cqe_size != HNS_ROCE_V2_CQE_SIZE && in set_cqe_size() 341 ucmd->cqe_size != HNS_ROCE_V3_CQE_SIZE) { in set_cqe_size() 343 "invalid cqe size %u.\n", ucmd->cqe_size); in set_cqe_size() 347 hr_cq->cqe_size = ucmd->cqe_size; in set_cqe_size() 362 struct hns_roce_ib_create_cq ucmd = {}; in hns_roce_create_cq() local 373 ret = get_cq_ucmd(hr_cq, udata, &ucmd); in hns_roce_create_cq() [all …]
|
| A D | hns_roce_qp.c | 559 struct hns_roce_ib_create_qp *ucmd) in check_sq_size_with_integrity() argument 565 if (ucmd->log_sq_stride > max_sq_stride || in check_sq_size_with_integrity() 582 struct hns_roce_ib_create_qp *ucmd) in set_user_sq_size() argument 601 hr_qp->sq.wqe_shift = ucmd->log_sq_stride; in set_user_sq_size() 800 struct hns_roce_ib_create_qp *ucmd) in user_qp_has_sdb() argument 829 struct hns_roce_ib_create_qp *ucmd, in alloc_user_qp_db() argument 903 struct hns_roce_ib_create_qp *ucmd, in alloc_qp_db() argument 983 struct hns_roce_ib_create_qp *ucmd) in set_qp_param() argument 1007 ret = ib_copy_from_udata(ucmd, udata, in set_qp_param() 1008 min(udata->inlen, sizeof(*ucmd))); in set_qp_param() [all …]
|
| A D | hns_roce_srq.c | 344 struct hns_roce_ib_create_srq ucmd = {}; in alloc_srq_buf() local 348 ret = ib_copy_from_udata(&ucmd, udata, in alloc_srq_buf() 349 min(udata->inlen, sizeof(ucmd))); in alloc_srq_buf() 358 ret = alloc_srq_idx(hr_dev, srq, udata, ucmd.que_addr); in alloc_srq_buf() 362 ret = alloc_srq_wqe_buf(hr_dev, srq, udata, ucmd.buf_addr); in alloc_srq_buf()
|
| /linux/drivers/infiniband/hw/mlx4/ |
| A D | qp.c | 417 ucmd->log_sq_stride > in set_user_sq_size() 539 ucmd->rx_hash_fields_mask); in set_qp_rss() 688 sizeof(ucmd.reserved1); in _mlx4_ib_create_qp_rss() 694 if (ib_copy_from_udata(&ucmd, udata, min(sizeof(ucmd), udata->inlen))) { in _mlx4_ib_create_qp_rss() 699 if (memchr_inv(ucmd.reserved, 0, sizeof(ucmd.reserved))) in _mlx4_ib_create_qp_rss() 702 if (ucmd.comp_mask || ucmd.reserved1) in _mlx4_ib_create_qp_rss() 1028 struct mlx4_ib_create_qp ucmd; in create_qp_common() local 4094 sizeof(ucmd.comp_mask); in mlx4_ib_create_wq() 4231 sizeof(ucmd.reserved); in mlx4_ib_modify_wq() 4240 if (ib_copy_from_udata(&ucmd, udata, min(sizeof(ucmd), udata->inlen))) in mlx4_ib_modify_wq() [all …]
|
| A D | srq.c | 112 struct mlx4_ib_create_srq ucmd; in mlx4_ib_create_srq() local 114 if (ib_copy_from_udata(&ucmd, udata, sizeof(ucmd))) in mlx4_ib_create_srq() 118 ib_umem_get(ib_srq->device, ucmd.buf_addr, buf_size, 0); in mlx4_ib_create_srq() 132 err = mlx4_ib_db_map_user(udata, ucmd.db_addr, &srq->db); in mlx4_ib_create_srq()
|
| A D | cq.c | 205 struct mlx4_ib_create_cq ucmd; in mlx4_ib_create_cq() local 207 if (ib_copy_from_udata(&ucmd, udata, sizeof ucmd)) { in mlx4_ib_create_cq() 212 buf_addr = (void *)(unsigned long)ucmd.buf_addr; in mlx4_ib_create_cq() 214 ucmd.buf_addr, entries); in mlx4_ib_create_cq() 218 err = mlx4_ib_db_map_user(udata, ucmd.db_addr, &cq->db); in mlx4_ib_create_cq() 317 struct mlx4_ib_resize_cq ucmd; in mlx4_alloc_resize_umem() local 323 if (ib_copy_from_udata(&ucmd, udata, sizeof ucmd)) in mlx4_alloc_resize_umem() 331 ucmd.buf_addr, entries); in mlx4_alloc_resize_umem()
|
| /linux/drivers/dma-buf/ |
| A D | dma-heap.c | 125 static long dma_heap_ioctl(struct file *file, unsigned int ucmd, in dma_heap_ioctl() argument 132 int nr = _IOC_NR(ucmd); in dma_heap_ioctl() 143 out_size = _IOC_SIZE(ucmd); in dma_heap_ioctl() 145 if ((ucmd & kcmd & IOC_IN) == 0) in dma_heap_ioctl() 147 if ((ucmd & kcmd & IOC_OUT) == 0) in dma_heap_ioctl()
|
| /linux/drivers/infiniband/hw/vmw_pvrdma/ |
| A D | pvrdma_srq.c | 109 struct pvrdma_create_srq ucmd; in pvrdma_create_srq() local 144 if (ib_copy_from_udata(&ucmd, udata, sizeof(ucmd))) { in pvrdma_create_srq() 149 srq->umem = ib_umem_get(ibsrq->device, ucmd.buf_addr, ucmd.buf_size, 0); in pvrdma_create_srq()
|
| A D | pvrdma_cq.c | 116 struct pvrdma_create_cq ucmd; in pvrdma_create_cq() local 136 if (ib_copy_from_udata(&ucmd, udata, sizeof(ucmd))) { in pvrdma_create_cq() 141 cq->umem = ib_umem_get(ibdev, ucmd.buf_addr, ucmd.buf_size, in pvrdma_create_cq()
|
| A D | pvrdma_qp.c | 201 struct pvrdma_create_qp ucmd; in pvrdma_create_qp() local 255 if (ib_copy_from_udata(&ucmd, udata, sizeof(ucmd))) { in pvrdma_create_qp() 272 ucmd.rbuf_addr, in pvrdma_create_qp() 273 ucmd.rbuf_size, 0); in pvrdma_create_qp() 284 qp->sumem = ib_umem_get(ibqp->device, ucmd.sbuf_addr, in pvrdma_create_qp() 285 ucmd.sbuf_size, 0); in pvrdma_create_qp()
|
| /linux/arch/mips/cavium-octeon/ |
| A D | octeon-platform.c | 34 u32 ucmd; in octeon2_usb_reset() local 41 ucmd = cvmx_read64_uint32(CVMX_UAHCX_EHCI_USBCMD); in octeon2_usb_reset() 42 ucmd &= ~CMD_RUN; in octeon2_usb_reset() 43 cvmx_write64_uint32(CVMX_UAHCX_EHCI_USBCMD, ucmd); in octeon2_usb_reset() 45 ucmd |= CMD_RESET; in octeon2_usb_reset() 46 cvmx_write64_uint32(CVMX_UAHCX_EHCI_USBCMD, ucmd); in octeon2_usb_reset() 47 ucmd = cvmx_read64_uint32(CVMX_UAHCX_OHCI_USBCMD); in octeon2_usb_reset() 48 ucmd |= CMD_RUN; in octeon2_usb_reset() 49 cvmx_write64_uint32(CVMX_UAHCX_OHCI_USBCMD, ucmd); in octeon2_usb_reset()
|
| /linux/drivers/nvme/host/ |
| A D | ioctl.c | 191 struct nvme_passthru_cmd __user *ucmd) in nvme_user_cmd() argument 201 if (copy_from_user(&cmd, ucmd, sizeof(cmd))) in nvme_user_cmd() 230 if (put_user(result, &ucmd->result)) in nvme_user_cmd() 238 struct nvme_passthru_cmd64 __user *ucmd) in nvme_user_cmd64() argument 247 if (copy_from_user(&cmd, ucmd, sizeof(cmd))) in nvme_user_cmd64() 276 if (put_user(cmd.result, &ucmd->result)) in nvme_user_cmd64()
|
| /linux/drivers/infiniband/sw/rxe/ |
| A D | rxe_srq.c | 126 struct rxe_modify_srq_cmd *ucmd, struct ib_udata *udata) in rxe_srq_from_attr() argument 137 mi = u64_to_user_ptr(ucmd->mmap_info_addr); in rxe_srq_from_attr()
|
| A D | rxe_verbs.c | 332 struct rxe_modify_srq_cmd ucmd = {}; in rxe_modify_srq() local 335 if (udata->inlen < sizeof(ucmd)) in rxe_modify_srq() 338 err = ib_copy_from_udata(&ucmd, udata, sizeof(ucmd)); in rxe_modify_srq() 347 err = rxe_srq_from_attr(rxe, srq, attr, mask, &ucmd, udata); in rxe_modify_srq()
|
| /linux/drivers/infiniband/hw/qib/ |
| A D | qib_file_ops.c | 2037 const struct qib_cmd __user *ucmd; in qib_write() local 2070 src = &ucmd->cmd.user_info; in qib_write() 2076 src = &ucmd->cmd.recv_ctrl; in qib_write() 2082 src = &ucmd->cmd.ctxt_info; in qib_write() 2089 src = &ucmd->cmd.tid_info; in qib_write() 2095 src = &ucmd->cmd.part_key; in qib_write() 2108 src = &ucmd->cmd.poll_type; in qib_write() 2114 src = &ucmd->cmd.armlaunch_ctrl; in qib_write() 2120 src = &ucmd->cmd.sdma_inflight; in qib_write() 2126 src = &ucmd->cmd.sdma_complete; in qib_write() [all …]
|
| /linux/drivers/block/rsxx/ |
| A D | cregs.c | 664 struct rsxx_reg_access __user *ucmd, in rsxx_reg_access() argument 670 st = copy_from_user(&cmd, ucmd, sizeof(cmd)); in rsxx_reg_access() 681 st = put_user(cmd.stat, &ucmd->stat); in rsxx_reg_access() 686 st = copy_to_user(ucmd->data, cmd.data, cmd.cnt); in rsxx_reg_access()
|
| /linux/drivers/scsi/sym53c8xx_2/ |
| A D | sym_glue.c | 129 struct sym_ucmd *ucmd = SYM_UCMD_PTR(cmd); in sym_xpt_done() local 132 if (ucmd->eh_done) in sym_xpt_done() 133 complete(ucmd->eh_done); in sym_xpt_done() 572 struct sym_ucmd *ucmd = SYM_UCMD_PTR(cmd); in sym_eh_handler() local 652 ucmd->eh_done = &eh_done; in sym_eh_handler() 655 ucmd->eh_done = NULL; in sym_eh_handler()
|