/linux/drivers/gpu/drm/i915/ |
A D | i915_query.c | 24 if (copy_from_user(query_hdr, u64_to_user_ptr(query_item->data_ptr), in copy_query_item() 71 if (copy_to_user(u64_to_user_ptr(query_item->data_ptr), in query_topology_info() 79 if (copy_to_user(u64_to_user_ptr(query_item->data_ptr + in query_topology_info() 84 if (copy_to_user(u64_to_user_ptr(query_item->data_ptr + in query_topology_info() 98 u64_to_user_ptr(query_item->data_ptr); in query_engine_info() 166 u32 __user *p = u64_to_user_ptr(user_regs_ptr); in copy_perf_config_registers_or_number() 196 u64_to_user_ptr(query_item->data_ptr); in query_perf_config_data() 198 u64_to_user_ptr(query_item->data_ptr + in query_perf_config_data() 341 u64_to_user_ptr(query_item->data_ptr); in query_perf_config_list() 428 u64_to_user_ptr(query_item->data_ptr); in query_memregion_info() [all …]
|
A D | i915_gem.c | 237 user_data = u64_to_user_ptr(args->data_ptr); in i915_gem_shmem_pread() 385 user_data = u64_to_user_ptr(args->data_ptr); in i915_gem_gtt_pread() 451 if (!access_ok(u64_to_user_ptr(args->data_ptr), in i915_gem_pread_ioctl() 558 user_data = u64_to_user_ptr(args->data_ptr); in i915_gem_gtt_pwrite_fast() 669 user_data = u64_to_user_ptr(args->data_ptr); in i915_gem_shmem_pwrite() 725 if (!access_ok(u64_to_user_ptr(args->data_ptr), args->size)) in i915_gem_pwrite_ioctl()
|
A D | i915_user_extensions.c | 57 ext = u64_to_user_ptr(next); in i915_user_extensions()
|
/linux/net/bpf/ |
A D | bpf_dummy_struct_ops.c | 36 ctx_in = u64_to_user_ptr(kattr->test.ctx_in); in dummy_ops_init_args() 41 u_state = u64_to_user_ptr(args->args[0]); in dummy_ops_init_args() 56 u_state = u64_to_user_ptr(args->args[0]); in dummy_ops_copy_args()
|
A D | test_run.c | 135 void __user *data_out = u64_to_user_ptr(kattr->test.data_out); in bpf_test_finish() 255 void __user *data_in = u64_to_user_ptr(kattr->test.data_in); in bpf_test_init() 341 void __user *ctx_in = u64_to_user_ptr(kattr->test.ctx_in); in bpf_prog_test_run_raw_tp() 397 void __user *data_in = u64_to_user_ptr(kattr->test.ctx_in); in bpf_ctx_init() 398 void __user *data_out = u64_to_user_ptr(kattr->test.ctx_out); in bpf_ctx_init() 430 void __user *data_out = u64_to_user_ptr(kattr->test.ctx_out); in bpf_ctx_finish() 1033 void __user *ctx_in = u64_to_user_ptr(kattr->test.ctx_in); in bpf_prog_test_run_syscall()
|
/linux/drivers/gpu/drm/ |
A D | drm_syncobj.c | 1162 u64_to_user_ptr(timeline_wait->points), in drm_syncobj_array_wait() 1249 u64_to_user_ptr(args->handles), in drm_syncobj_wait_ioctl() 1283 u64_to_user_ptr(args->handles), in drm_syncobj_timeline_wait_ioctl() 1317 u64_to_user_ptr(args->handles), in drm_syncobj_reset_ioctl() 1350 u64_to_user_ptr(args->handles), in drm_syncobj_signal_ioctl() 1388 u64_to_user_ptr(args->handles), in drm_syncobj_timeline_signal_ioctl() 1400 if (!u64_to_user_ptr(args->points)) { in drm_syncobj_timeline_signal_ioctl() 1402 } else if (copy_from_user(points, u64_to_user_ptr(args->points), in drm_syncobj_timeline_signal_ioctl() 1445 uint64_t __user *points = u64_to_user_ptr(args->points); in drm_syncobj_query_ioctl() 1459 u64_to_user_ptr(args->handles), in drm_syncobj_query_ioctl()
|
A D | drm_property.c | 480 values_ptr = u64_to_user_ptr(out_resp->values_ptr); in drm_mode_getproperty_ioctl() 491 enum_ptr = u64_to_user_ptr(out_resp->enum_blob_ptr); in drm_mode_getproperty_ioctl() 769 if (copy_to_user(u64_to_user_ptr(out_resp->data), in drm_mode_getblob_ioctl() 798 u64_to_user_ptr(out_resp->data), in drm_mode_createblob_ioctl()
|
A D | drm_mode_config.c | 110 fb_id = u64_to_user_ptr(card_res->fb_id_ptr); in drm_mode_getresources() 128 crtc_id = u64_to_user_ptr(card_res->crtc_id_ptr); in drm_mode_getresources() 140 encoder_id = u64_to_user_ptr(card_res->encoder_id_ptr); in drm_mode_getresources() 151 connector_id = u64_to_user_ptr(card_res->connector_id_ptr); in drm_mode_getresources()
|
/linux/drivers/infiniband/core/ |
A D | uverbs_ioctl.c | 141 return ib_is_buffer_cleared(u64_to_user_ptr(uattr->data) + len, in uverbs_is_attr_cleared() 201 ret = copy_from_user(idr_vals, u64_to_user_ptr(uattr->data), in uverbs_process_idrs_array() 297 if (copy_from_user(p, u64_to_user_ptr(uattr->data), in uverbs_process_attr() 708 udata->inbuf = u64_to_user_ptr(in->ptr_attr.data); in uverbs_fill_udata() 715 udata->outbuf = u64_to_user_ptr(out->ptr_attr.data); in uverbs_fill_udata() 733 if (copy_to_user(u64_to_user_ptr(attr->ptr_attr.data), from, min_size)) in uverbs_copy_to() 811 if (clear_user(u64_to_user_ptr(attr->ptr_attr.data) + size, in uverbs_copy_to_struct_or_zero()
|
A D | ucma.c | 399 if (copy_to_user(u64_to_user_ptr(cmd.response), in ucma_get_event() 469 if (copy_to_user(u64_to_user_ptr(cmd.response), in ucma_create_id() 615 if (copy_to_user(u64_to_user_ptr(cmd.response), in ucma_destroy_id() 859 if (copy_to_user(u64_to_user_ptr(cmd.response), &resp, in ucma_query_route() 1005 response = u64_to_user_ptr(cmd.response); in ucma_query() 1242 if (copy_to_user(u64_to_user_ptr(cmd.response), in ucma_init_qp_attr() 1394 optval = memdup_user(u64_to_user_ptr(cmd.optval), in ucma_set_option() 1486 if (copy_to_user(u64_to_user_ptr(cmd->response), in ucma_process_join() 1591 if (copy_to_user(u64_to_user_ptr(cmd.response), in ucma_leave_multicast() 1663 if (copy_to_user(u64_to_user_ptr(cmd.response), in ucma_migrate_id()
|
A D | uverbs_main.c | 525 if (!access_ok(u64_to_user_ptr(ex_hdr->response), in verify_hdr() 642 u64_to_user_ptr(response + in ib_uverbs_write() 653 &bundle.ucore, buf, u64_to_user_ptr(response), in ib_uverbs_write() 659 u64_to_user_ptr(ex_hdr.response), in ib_uverbs_write() 664 u64_to_user_ptr(ex_hdr.response) + bundle.ucore.outlen, in ib_uverbs_write()
|
/linux/kernel/bpf/ |
A D | syscall.c | 1060 void __user *ukey = u64_to_user_ptr(attr->key); in map_lookup_elem() 1061 void __user *uvalue = u64_to_user_ptr(attr->value); in map_lookup_elem() 1196 void __user *ukey = u64_to_user_ptr(attr->key); in map_delete_elem() 1251 void __user *ukey = u64_to_user_ptr(attr->key); in map_get_next_key() 1530 void __user *ukey = u64_to_user_ptr(attr->key); in map_lookup_and_delete_elem() 1531 void __user *uvalue = u64_to_user_ptr(attr->value); in map_lookup_and_delete_elem() 3696 uinsns = u64_to_user_ptr(info.xlated_prog_insns); in bpf_prog_get_info_by_fd() 3728 uinsns = u64_to_user_ptr(info.jited_prog_insns); in bpf_prog_get_info_by_fd() 3771 user_ksyms = u64_to_user_ptr(info.jited_ksyms); in bpf_prog_get_info_by_fd() 3825 user_finfo = u64_to_user_ptr(info.func_info); in bpf_prog_get_info_by_fd() [all …]
|
/linux/drivers/gpu/drm/qxl/ |
A D | qxl_ioctl.c | 168 if (!access_ok(u64_to_user_ptr(cmd->command), in qxl_process_single_command() 190 u64_to_user_ptr(cmd->command), cmd->command_size); in qxl_process_single_command() 209 struct drm_qxl_reloc __user *u = u64_to_user_ptr(cmd->relocs); in qxl_process_single_command() 287 u64_to_user_ptr(execbuffer->commands); in qxl_execbuffer_ioctl()
|
/linux/drivers/gpu/drm/virtio/ |
A D | virtgpu_ioctl.c | 193 user_bo_handles = u64_to_user_ptr(exbuf->bo_handles); in virtio_gpu_execbuffer_ioctl() 210 buf = vmemdup_user(u64_to_user_ptr(exbuf->command), exbuf->size); in virtio_gpu_execbuffer_ioctl() 298 if (copy_to_user(u64_to_user_ptr(param->value), &value, sizeof(int))) in virtio_gpu_getparam_ioctl() 597 if (copy_to_user(u64_to_user_ptr(args->addr), ptr, size)) in virtio_gpu_get_caps_ioctl() 685 buf = memdup_user(u64_to_user_ptr(rc_blob->cmd), in virtio_gpu_resource_create_blob_ioctl() 754 ctx_set_params = memdup_user(u64_to_user_ptr(args->ctx_set_params), in virtio_gpu_context_init_ioctl()
|
/linux/drivers/gpu/drm/etnaviv/ |
A D | etnaviv_gem_submit.c | 485 ret = copy_from_user(bos, u64_to_user_ptr(args->bos), in etnaviv_ioctl_gem_submit() 492 ret = copy_from_user(relocs, u64_to_user_ptr(args->relocs), in etnaviv_ioctl_gem_submit() 499 ret = copy_from_user(pmrs, u64_to_user_ptr(args->pmrs), in etnaviv_ioctl_gem_submit() 506 ret = copy_from_user(stream, u64_to_user_ptr(args->stream), in etnaviv_ioctl_gem_submit()
|
/linux/drivers/gpu/drm/i915/gem/ |
A D | i915_gem_phys.c | 142 char __user *user_data = u64_to_user_ptr(args->data_ptr); in i915_gem_object_pwrite_phys() 172 char __user *user_data = u64_to_user_ptr(args->data_ptr); in i915_gem_object_pread_phys()
|
A D | i915_gem_create.c | 271 u64_to_user_ptr(args->regions); in set_placements() 414 ret = i915_user_extensions(u64_to_user_ptr(args->extensions), in i915_gem_create_ext_ioctl()
|
/linux/drivers/gpu/drm/msm/ |
A D | msm_gem_submit.c | 102 u64_to_user_ptr(args->bos + (i * sizeof(submit_bo))); in submit_lookup_objects() 171 u64_to_user_ptr(args->cmds + (i * sizeof(submit_cmd))); in submit_lookup_cmds() 203 userptr = u64_to_user_ptr(submit_cmd.relocs); in submit_lookup_cmds() 562 u64_to_user_ptr(address), in msm_parse_deps() 640 u64_to_user_ptr(address), in msm_parse_post_deps()
|
/linux/fs/verity/ |
A D | enable.c | 219 copy_from_user(desc->salt, u64_to_user_ptr(arg->salt_ptr), in enable_verity() 228 copy_from_user(desc->signature, u64_to_user_ptr(arg->sig_ptr), in enable_verity()
|
/linux/include/linux/ |
A D | bpfptr.h | 31 return USER_BPFPTR(u64_to_user_ptr(addr)); in make_bpfptr()
|
/linux/drivers/gpu/drm/tegra/ |
A D | submit.c | 190 if (copy_from_user(bo->gather_data, u64_to_user_ptr(args->gather_data_ptr), copy_len)) { in submit_copy_gather_data() 242 bufs = alloc_copy_user_array(u64_to_user_ptr(args->bufs_ptr), args->num_bufs, in submit_process_bufs() 381 cmds = alloc_copy_user_array(u64_to_user_ptr(args->cmds_ptr), args->num_cmds, in submit_create_job()
|
/linux/lib/ |
A D | test_hmm.c | 375 if (copy_to_user(u64_to_user_ptr(cmd->ptr), bounce.ptr, in dmirror_read() 428 if (copy_from_user(bounce.ptr, u64_to_user_ptr(cmd->ptr), in dmirror_write() 767 if (copy_to_user(u64_to_user_ptr(cmd->ptr), bounce.ptr, in dmirror_exclusive() 838 if (copy_to_user(u64_to_user_ptr(cmd->ptr), bounce.ptr, in dmirror_migrate() 1001 uptr = u64_to_user_ptr(cmd->ptr); in dmirror_snapshot()
|
/linux/drivers/gpu/drm/v3d/ |
A D | v3d_gem.c | 485 struct drm_v3d_sem __user *handle = u64_to_user_ptr(se->in_syncs); in v3d_job_init() 607 post_deps = u64_to_user_ptr(handles); in v3d_get_multisync_post_deps() 679 user_ext = u64_to_user_ptr(ext_handles); in v3d_get_extensions() 699 user_ext = u64_to_user_ptr(ext.next); in v3d_get_extensions()
|
/linux/fs/ |
A D | io_uring.c | 3149 return u64_to_user_ptr(kbuf->addr); in io_rw_buffer_select() 3161 uiov = u64_to_user_ptr(req->rw.addr); in io_compat_import() 3241 buf = u64_to_user_ptr(req->rw.addr); in __io_import_iovec() 3757 oldf = u64_to_user_ptr(READ_ONCE(sqe->addr)); in io_renameat_prep() 3950 oldf = u64_to_user_ptr(READ_ONCE(sqe->addr)); in io_linkat_prep() 4251 how = u64_to_user_ptr(READ_ONCE(sqe->addr2)); in io_openat2_prep() 4513 ev = u64_to_user_ptr(READ_ONCE(sqe->addr)); in io_epoll_ctl_prep() 5109 buf = u64_to_user_ptr(kbuf->addr); in io_recv() 10012 *sig = u64_to_user_ptr(arg.sigmask); in io_get_ext_arg() 10014 *ts = u64_to_user_ptr(arg.ts); in io_get_ext_arg() [all …]
|
/linux/drivers/gpu/drm/vc4/ |
A D | vc4_gem.c | 130 if (copy_to_user(u64_to_user_ptr(get_state->bo), in vc4_get_hang_state_ioctl() 761 if (copy_from_user(handles, u64_to_user_ptr(args->bo_handles), in vc4_cl_lookup_bos() 869 u64_to_user_ptr(args->bin_cl), in vc4_get_bcl() 876 u64_to_user_ptr(args->shader_rec), in vc4_get_bcl() 883 u64_to_user_ptr(args->uniforms), in vc4_get_bcl()
|