| /linux/drivers/net/ethernet/mellanox/mlx4/ |
| A D | pd.c | 161 uar->map = NULL; in mlx4_uar_alloc() 190 uar = kmalloc_node(sizeof(*uar), GFP_KERNEL, node); in mlx4_bf_alloc() 191 if (!uar) { in mlx4_bf_alloc() 192 uar = kmalloc(sizeof(*uar), GFP_KERNEL); in mlx4_bf_alloc() 193 if (!uar) { in mlx4_bf_alloc() 202 uar->map = ioremap(uar->pfn << PAGE_SHIFT, PAGE_SIZE); in mlx4_bf_alloc() 221 bf->uar = uar; in mlx4_bf_alloc() 231 bf->uar = NULL; in mlx4_bf_alloc() 238 kfree(uar); in mlx4_bf_alloc() 251 if (!bf->uar || !bf->uar->bf_map) in mlx4_bf_free() [all …]
|
| A D | cq.c | 343 struct mlx4_mtt *mtt, struct mlx4_uar *uar, u64 db_rec, in mlx4_cq_alloc() argument 383 mlx4_to_hw_uar_index(dev, uar->index)); in mlx4_cq_alloc() 412 cq->uar = uar; in mlx4_cq_alloc()
|
| /linux/drivers/infiniband/hw/mthca/ |
| A D | mthca_uar.c | 38 int mthca_uar_alloc(struct mthca_dev *dev, struct mthca_uar *uar) in mthca_uar_alloc() argument 40 uar->index = mthca_alloc(&dev->uar_table.alloc); in mthca_uar_alloc() 41 if (uar->index == -1) in mthca_uar_alloc() 44 uar->pfn = (pci_resource_start(dev->pdev, 2) >> PAGE_SHIFT) + uar->index; in mthca_uar_alloc() 49 void mthca_uar_free(struct mthca_dev *dev, struct mthca_uar *uar) in mthca_uar_free() argument 51 mthca_free(&dev->uar_table.alloc, uar->index); in mthca_uar_free()
|
| A D | mthca_provider.c | 318 mthca_uar_free(to_mdev(ibdev), &context->uar); in mthca_alloc_ucontext() 324 mthca_uar_free(to_mdev(ibdev), &context->uar); in mthca_alloc_ucontext() 349 to_mucontext(context)->uar.pfn, in mthca_mmap_uar() 485 err = mthca_map_user_db(dev, &context->uar, in mthca_create_qp() 492 err = mthca_map_user_db(dev, &context->uar, in mthca_create_qp() 497 mthca_unmap_user_db(dev, &context->uar, in mthca_create_qp() 570 &context->uar, in mthca_destroy_qp() 574 &context->uar, in mthca_destroy_qp() 648 mthca_unmap_user_db(to_mdev(ibdev), &context->uar, in mthca_create_cq() 790 &context->uar, in mthca_destroy_cq() [all …]
|
| A D | mthca_memfree.h | 165 int mthca_map_user_db(struct mthca_dev *dev, struct mthca_uar *uar, 167 void mthca_unmap_user_db(struct mthca_dev *dev, struct mthca_uar *uar, 170 void mthca_cleanup_user_db_tab(struct mthca_dev *dev, struct mthca_uar *uar,
|
| A D | mthca_memfree.c | 439 static u64 mthca_uarc_virt(struct mthca_dev *dev, struct mthca_uar *uar, int page) in mthca_uarc_virt() argument 442 uar->index * dev->uar_table.uarc_size + in mthca_uarc_virt() 446 int mthca_map_user_db(struct mthca_dev *dev, struct mthca_uar *uar, in mthca_map_user_db() argument 491 mthca_uarc_virt(dev, uar, i)); in mthca_map_user_db() 507 void mthca_unmap_user_db(struct mthca_dev *dev, struct mthca_uar *uar, in mthca_unmap_user_db() argument 549 void mthca_cleanup_user_db_tab(struct mthca_dev *dev, struct mthca_uar *uar, in mthca_cleanup_user_db_tab() argument 559 mthca_UNMAP_ICM(dev, mthca_uarc_virt(dev, uar, i), 1); in mthca_cleanup_user_db_tab()
|
| A D | mthca_srq.c | 54 __be32 uar; member 113 context->uar = cpu_to_be32(ucontext->uar.index); in mthca_tavor_init_srq_context() 115 context->uar = cpu_to_be32(dev->driver_uar.index); in mthca_tavor_init_srq_context() 141 context->logstride_usrpage |= cpu_to_be32(ucontext->uar.index); in mthca_arbel_init_srq_context()
|
| /linux/drivers/infiniband/hw/hns/ |
| A D | hns_roce_pd.c | 86 int hns_roce_uar_alloc(struct hns_roce_dev *hr_dev, struct hns_roce_uar *uar) in hns_roce_uar_alloc() argument 99 uar->logic_idx = (unsigned long)id; in hns_roce_uar_alloc() 101 if (uar->logic_idx > 0 && hr_dev->caps.phy_num_uars > 1) in hns_roce_uar_alloc() 102 uar->index = (uar->logic_idx - 1) % in hns_roce_uar_alloc() 105 uar->index = 0; in hns_roce_uar_alloc() 114 uar->pfn = ((res->start) >> PAGE_SHIFT) + uar->index; in hns_roce_uar_alloc() 116 uar->pfn = ((pci_resource_start(hr_dev->pci_dev, 2)) in hns_roce_uar_alloc()
|
| A D | hns_roce_main.c | 339 address = context->uar.pfn << PAGE_SHIFT; in hns_roce_alloc_uar_entry() 382 ret = hns_roce_uar_alloc(hr_dev, &context->uar); in hns_roce_alloc_ucontext() 409 ida_free(&hr_dev->uar_ida.ida, (int)context->uar.logic_idx); in hns_roce_alloc_ucontext() 422 ida_free(&hr_dev->uar_ida.ida, (int)context->uar.logic_idx); in hns_roce_dealloc_ucontext()
|
| /linux/drivers/infiniband/hw/vmw_pvrdma/ |
| A D | pvrdma_doorbell.c | 83 int pvrdma_uar_alloc(struct pvrdma_dev *dev, struct pvrdma_uar_map *uar) in pvrdma_uar_alloc() argument 108 uar->index = obj; in pvrdma_uar_alloc() 109 uar->pfn = (pci_resource_start(dev->pdev, PVRDMA_PCI_RESOURCE_UAR) >> in pvrdma_uar_alloc() 110 PAGE_SHIFT) + uar->index; in pvrdma_uar_alloc() 115 void pvrdma_uar_free(struct pvrdma_dev *dev, struct pvrdma_uar_map *uar) in pvrdma_uar_free() argument 121 obj = uar->index & (tbl->max - 1); in pvrdma_uar_free()
|
| A D | pvrdma_verbs.c | 329 ret = pvrdma_uar_alloc(vdev, &context->uar); in pvrdma_alloc_ucontext() 335 cmd->pfn = context->uar.pfn; in pvrdma_alloc_ucontext() 337 cmd->pfn64 = context->uar.pfn; in pvrdma_alloc_ucontext() 353 pvrdma_uar_free(vdev, &context->uar); in pvrdma_alloc_ucontext() 361 pvrdma_uar_free(vdev, &context->uar); in pvrdma_alloc_ucontext() 385 pvrdma_uar_free(to_vdev(ibcontext->device), &context->uar); in pvrdma_dealloc_ucontext() 413 if (io_remap_pfn_range(vma, start, context->uar.pfn, size, in pvrdma_mmap()
|
| A D | pvrdma.h | 90 struct pvrdma_uar_map *uar; member 123 struct pvrdma_uar_map uar; member 535 int pvrdma_uar_alloc(struct pvrdma_dev *dev, struct pvrdma_uar_map *uar); 536 void pvrdma_uar_free(struct pvrdma_dev *dev, struct pvrdma_uar_map *uar);
|
| A D | pvrdma_cq.c | 203 cq->uar = &context->uar; in pvrdma_create_cq()
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/steering/ |
| A D | dr_send.c | 48 struct mlx5_uars_page *uar; member 164 MLX5_SET(qpc, qpc, uar_page, attr->uar->index); in dr_create_rc_qp() 189 dr_qp->uar = attr->uar; in dr_create_rc_qp() 224 mlx5_write64(ctrl, dr_qp->uar->map + MLX5_BF_OFFSET); in dr_cmd_notify_hw() 756 struct mlx5_uars_page *uar, in dr_create_cq() argument 806 MLX5_SET(cqc, cqc, uar_page, uar->index); in dr_create_cq() 833 cq->mcq.uar = uar; in dr_create_cq() 928 dmn->send_ring->cq = dr_create_cq(dmn->mdev, dmn->uar, cq_size); in mlx5dr_send_ring_alloc() 937 init_attr.uar = dmn->uar; in mlx5dr_send_ring_alloc()
|
| A D | dr_domain.c | 75 dmn->uar = mlx5_get_uars_page(dmn->mdev); in dr_domain_init_resources() 76 if (IS_ERR(dmn->uar)) { in dr_domain_init_resources() 78 ret = PTR_ERR(dmn->uar); in dr_domain_init_resources() 109 mlx5_put_uars_page(dmn->mdev, dmn->uar); in dr_domain_init_resources() 121 mlx5_put_uars_page(dmn->mdev, dmn->uar); in dr_domain_uninit_resources()
|
| /linux/drivers/vdpa/mlx5/core/ |
| A D | resources.c | 259 res->uar = mlx5_get_uars_page(mdev); in mlx5_vdpa_alloc_resources() 260 if (IS_ERR(res->uar)) { in mlx5_vdpa_alloc_resources() 261 err = PTR_ERR(res->uar); in mlx5_vdpa_alloc_resources() 301 mlx5_put_uars_page(mdev, res->uar); in mlx5_vdpa_alloc_resources() 319 mlx5_put_uars_page(mvdev->mdev, res->uar); in mlx5_vdpa_free_resources()
|
| A D | mlx5_vdpa.h | 43 struct mlx5_uars_page *uar; member
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/fpga/ |
| A D | conn.c | 138 mlx5_write64(wqe, conn->fdev->conn_res.uar->map + MLX5_BF_OFFSET); in mlx5_fpga_conn_notify_hw() 362 conn->fdev->conn_res.uar->map, conn->cq.wq.cc); in mlx5_fpga_conn_arm_cq() 457 MLX5_SET(cqc, cqc, uar_page, fdev->conn_res.uar->index); in mlx5_fpga_conn_create_cq() 478 conn->cq.mcq.uar = fdev->conn_res.uar; in mlx5_fpga_conn_create_cq() 562 MLX5_SET(qpc, qpc, uar_page, fdev->conn_res.uar->index); in mlx5_fpga_conn_create_qp() 959 fdev->conn_res.uar = mlx5_get_uars_page(fdev->mdev); in mlx5_fpga_conn_device_init() 960 if (IS_ERR(fdev->conn_res.uar)) { in mlx5_fpga_conn_device_init() 961 err = PTR_ERR(fdev->conn_res.uar); in mlx5_fpga_conn_device_init() 966 fdev->conn_res.uar->index); in mlx5_fpga_conn_device_init() 988 mlx5_put_uars_page(fdev->mdev, fdev->conn_res.uar); in mlx5_fpga_conn_device_init() [all …]
|
| A D | core.h | 58 struct mlx5_uars_page *uar; member
|
| /linux/drivers/infiniband/hw/efa/ |
| A D | efa_admin_cmds_defs.h | 147 u16 uar; member 479 u16 uar; member 821 u16 uar; member 831 u16 uar; member
|
| A D | efa_com_cmd.c | 33 create_qp_cmd.uar = params->uarn; in efa_com_create_qp() 165 create_cmd.uar = params->uarn; in efa_com_create_cq() 716 result->uarn = resp.uar; in efa_com_alloc_uar() 730 cmd.uar = params->uarn; in efa_com_dealloc_uar() 740 cmd.uar, err); in efa_com_dealloc_uar()
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/ |
| A D | cq.c | 137 cq->uar = dev->priv.uar; in mlx5_core_create_cq()
|
| A D | uar.c | 50 *uarn = MLX5_GET(alloc_uar_out, out, uar); in mlx5_cmd_alloc_uar() 59 MLX5_SET(dealloc_uar_in, in, uar, uarn); in mlx5_cmd_free_uar()
|
| /linux/drivers/infiniband/hw/mlx5/ |
| A D | cmd.c | 222 *uarn = MLX5_GET(alloc_uar_out, out, uar); in mlx5_cmd_uar_alloc() 231 MLX5_SET(dealloc_uar_in, in, uar, uarn); in mlx5_cmd_uar_dealloc()
|
| /linux/drivers/infiniband/hw/mlx4/ |
| A D | cq.c | 182 struct mlx4_uar *uar; in mlx4_ib_create_cq() local 222 uar = &context->uar; in mlx4_ib_create_cq() 240 uar = &dev->priv_uar; in mlx4_ib_create_cq() 247 err = mlx4_cq_alloc(dev->dev, entries, &cq->buf.mtt, uar, cq->db.dma, in mlx4_ib_create_cq()
|