Searched refs:u64_to_user_ptr (Results 1 - 25 of 118) sorted by relevance

12345

/linux-master/io_uring/
H A Dstatx.c35 path = u64_to_user_ptr(READ_ONCE(sqe->addr));
36 sx->buffer = u64_to_user_ptr(READ_ONCE(sqe->addr2));
H A Dfs.c61 oldf = u64_to_user_ptr(READ_ONCE(sqe->addr));
62 newf = u64_to_user_ptr(READ_ONCE(sqe->addr2));
120 fname = u64_to_user_ptr(READ_ONCE(sqe->addr));
167 fname = u64_to_user_ptr(READ_ONCE(sqe->addr));
209 oldpath = u64_to_user_ptr(READ_ONCE(sqe->addr));
210 newpath = u64_to_user_ptr(READ_ONCE(sqe->addr2));
253 oldf = u64_to_user_ptr(READ_ONCE(sqe->addr));
254 newf = u64_to_user_ptr(READ_ONCE(sqe->addr2));
H A Dxattr.c56 name = u64_to_user_ptr(READ_ONCE(sqe->addr));
57 ix->ctx.cvalue = u64_to_user_ptr(READ_ONCE(sqe->addr2));
97 path = u64_to_user_ptr(READ_ONCE(sqe->addr3));
159 name = u64_to_user_ptr(READ_ONCE(sqe->addr));
160 ix->ctx.cvalue = u64_to_user_ptr(READ_ONCE(sqe->addr2));
190 path = u64_to_user_ptr(READ_ONCE(sqe->addr3));
H A Depoll.c38 ev = u64_to_user_ptr(READ_ONCE(sqe->addr));
H A Drsrc.c100 dst->iov_base = u64_to_user_ptr((u64)ciov.iov_base);
358 u64 __user *tags = u64_to_user_ptr(up->tags);
359 __s32 __user *fds = u64_to_user_ptr(up->data);
423 u64 __user *tags = u64_to_user_ptr(up->tags);
424 struct iovec iov, __user *iovs = u64_to_user_ptr(up->data);
543 return io_sqe_files_register(ctx, u64_to_user_ptr(rr.data),
544 rr.nr, u64_to_user_ptr(rr.tags));
548 return io_sqe_buffers_register(ctx, u64_to_user_ptr(rr.data),
549 rr.nr, u64_to_user_ptr(rr.tags));
575 __s32 __user *fds = u64_to_user_ptr(u
[all...]
/linux-master/drivers/gpu/drm/i915/
H A Di915_query.c25 if (copy_from_user(query_hdr, u64_to_user_ptr(query_item->data_ptr),
68 if (copy_to_user(u64_to_user_ptr(query_item->data_ptr),
72 if (copy_to_user(u64_to_user_ptr(query_item->data_ptr + sizeof(topo)),
76 if (intel_sseu_copy_ssmask_to_user(u64_to_user_ptr(query_item->data_ptr +
81 if (intel_sseu_copy_eumask_to_user(u64_to_user_ptr(query_item->data_ptr +
132 u64_to_user_ptr(query_item->data_ptr);
200 u32 __user *p = u64_to_user_ptr(user_regs_ptr);
230 u64_to_user_ptr(query_item->data_ptr);
232 u64_to_user_ptr(query_item->data_ptr +
376 u64_to_user_ptr(query_ite
[all...]
H A Di915_user_extensions.c57 ext = u64_to_user_ptr(next);
/linux-master/drivers/gpu/drm/xe/
H A Dxe_query.c126 query_ptr = u64_to_user_ptr(query->data);
186 u64_to_user_ptr(query->data);
248 u64_to_user_ptr(query->data);
319 u64_to_user_ptr(query->data);
360 u64_to_user_ptr(query->data);
422 void __user *query_ptr = u64_to_user_ptr(query->data);
478 void __user *query_ptr = u64_to_user_ptr(query->data);
522 struct drm_xe_query_uc_fw_version __user *query_ptr = u64_to_user_ptr(query->data);
H A Dxe_exec.c114 struct drm_xe_sync __user *syncs_user = u64_to_user_ptr(args->syncs);
115 u64 __user *addresses_user = u64_to_user_ptr(args->address);
/linux-master/include/linux/
H A Dbpfptr.h31 return USER_BPFPTR(u64_to_user_ptr(addr));
/linux-master/net/bpf/
H A Dbpf_dummy_struct_ops.c43 ctx_in = u64_to_user_ptr(kattr->test.ctx_in);
48 u_state = u64_to_user_ptr(args->args[0]);
63 u_state = u64_to_user_ptr(args->args[0]);
/linux-master/drivers/gpu/drm/v3d/
H A Dv3d_submit.c177 struct drm_v3d_sem __user *handle = u64_to_user_ptr(se->in_syncs);
340 post_deps = u64_to_user_ptr(handles);
480 offsets = u64_to_user_ptr(timestamp.offsets);
481 syncs = u64_to_user_ptr(timestamp.syncs);
534 syncs = u64_to_user_ptr(reset.syncs);
587 offsets = u64_to_user_ptr(copy.offsets);
588 syncs = u64_to_user_ptr(copy.syncs);
648 syncs = u64_to_user_ptr(reset.syncs);
649 kperfmon_ids = u64_to_user_ptr(reset.kperfmon_ids);
669 ids_pointer = u64_to_user_ptr(id
[all...]
/linux-master/drivers/gpu/drm/imagination/
H A Dpvr_drv.c502 if (copy_to_user(u64_to_user_ptr(query.quirks), out,
585 if (copy_to_user(u64_to_user_ptr(query.enhancements), out,
1167 return copy_struct_from_user(out, obj_size, u64_to_user_ptr(usr_ptr), usr_stride);
1176 if (copy_to_user(u64_to_user_ptr(usr_ptr), in, min_t(u32, usr_stride, obj_size)))
1180 clear_user(u64_to_user_ptr(usr_ptr + obj_size), usr_stride - obj_size)) {
1204 if (copy_from_user(out_alloc, u64_to_user_ptr(in->array),
1208 void __user *in_ptr = u64_to_user_ptr(in->array);
1241 if (copy_to_user(u64_to_user_ptr(out->array), in,
1246 void __user *out_ptr = u64_to_user_ptr(out->array);
1258 clear_user(u64_to_user_ptr(ou
[all...]
/linux-master/kernel/bpf/
H A Dmprog.c420 uprog_id = u64_to_user_ptr(attr->query.prog_ids);
421 uprog_flags = u64_to_user_ptr(attr->query.prog_attach_flags);
422 ulink_id = u64_to_user_ptr(attr->query.link_ids);
423 ulink_flags = u64_to_user_ptr(attr->query.link_attach_flags);
H A Dsyscall.c1525 void __user *ukey = u64_to_user_ptr(attr->key);
1526 void __user *uvalue = u64_to_user_ptr(attr->value);
1714 void __user *ukey = u64_to_user_ptr(attr->key);
1715 void __user *unext_key = u64_to_user_ptr(attr->next_key);
1780 void __user *keys = u64_to_user_ptr(attr->batch.keys);
1836 void __user *values = u64_to_user_ptr(attr->batch.values);
1837 void __user *keys = u64_to_user_ptr(attr->batch.keys);
1899 void __user *uobatch = u64_to_user_ptr(attr->batch.out_batch);
1900 void __user *ubatch = u64_to_user_ptr(attr->batch.in_batch);
1901 void __user *values = u64_to_user_ptr(att
[all...]
/linux-master/drivers/gpu/drm/virtio/
H A Dvirtgpu_submit.c118 u64_to_user_ptr(address),
208 u64_to_user_ptr(address),
311 if (copy_from_user(bo_handles, u64_to_user_ptr(exbuf->bo_handles),
416 submit->buf = vmemdup_user(u64_to_user_ptr(exbuf->command), exbuf->size);
H A Dvirtgpu_ioctl.c123 if (copy_to_user(u64_to_user_ptr(param->value), &value, sizeof(int)))
433 if (copy_to_user(u64_to_user_ptr(args->addr), ptr, size))
520 buf = memdup_user(u64_to_user_ptr(rc_blob->cmd),
596 ctx_set_params = memdup_user(u64_to_user_ptr(args->ctx_set_params),
662 u64_to_user_ptr(value),
/linux-master/drivers/gpu/drm/qxl/
H A Dqxl_ioctl.c165 if (!access_ok(u64_to_user_ptr(cmd->command),
187 u64_to_user_ptr(cmd->command), cmd->command_size);
205 struct drm_qxl_reloc __user *u = u64_to_user_ptr(cmd->relocs);
281 u64_to_user_ptr(execbuffer->commands);
/linux-master/drivers/gpu/drm/i915/gem/
H A Di915_gem_phys.c143 char __user *user_data = u64_to_user_ptr(args->data_ptr);
174 char __user *user_data = u64_to_user_ptr(args->data_ptr);
/linux-master/drivers/gpu/drm/
H A Ddrm_syncobj.c1251 u64_to_user_ptr(timeline_wait->points),
1344 u64_to_user_ptr(args->handles),
1388 u64_to_user_ptr(args->handles),
1520 u64_to_user_ptr(args->handles),
1553 u64_to_user_ptr(args->handles),
1591 u64_to_user_ptr(args->handles),
1603 if (!u64_to_user_ptr(args->points)) {
1605 } else if (copy_from_user(points, u64_to_user_ptr(args->points),
1648 uint64_t __user *points = u64_to_user_ptr(args->points);
1662 u64_to_user_ptr(arg
[all...]
/linux-master/drivers/infiniband/core/
H A Duverbs_ioctl.c144 return ib_is_buffer_cleared(u64_to_user_ptr(uattr->data) + len,
206 ret = copy_from_user(idr_vals, u64_to_user_ptr(uattr->data),
304 if (copy_from_user(p, u64_to_user_ptr(uattr->data),
728 udata->inbuf = u64_to_user_ptr(in->ptr_attr.data);
735 udata->outbuf = u64_to_user_ptr(out->ptr_attr.data);
753 if (copy_to_user(u64_to_user_ptr(attr->ptr_attr.data), from, min_size))
831 if (clear_user(u64_to_user_ptr(attr->ptr_attr.data) + size,
/linux-master/drivers/iommu/iommufd/
H A Dioas.c80 ranges = u64_to_user_ptr(cmd->allowed_iovas);
160 u64_to_user_ptr(cmd->allowed_iovas),
223 u64_to_user_ptr(cmd->user_va), cmd->length,
/linux-master/drivers/misc/
H A Dnsm.c159 if (copy_from_user(req->data, u64_to_user_ptr(raw->request.addr),
176 if (copy_to_user(u64_to_user_ptr(raw->response.addr),
356 void __user *argp = u64_to_user_ptr((u64)arg);
/linux-master/fs/verity/
H A Dread_metadata.c180 buf = u64_to_user_ptr(arg.buf_ptr);
/linux-master/drivers/gpu/drm/etnaviv/
H A Detnaviv_gem_submit.c480 ret = copy_from_user(bos, u64_to_user_ptr(args->bos),
487 ret = copy_from_user(relocs, u64_to_user_ptr(args->relocs),
494 ret = copy_from_user(pmrs, u64_to_user_ptr(args->pmrs),
501 ret = copy_from_user(stream, u64_to_user_ptr(args->stream),

Completed in 227 milliseconds

12345