kdata 1592 arch/arm64/kernel/ptrace.c static int compat_ptrace_hbp_get_resource_info(u32 *kdata) kdata 1610 arch/arm64/kernel/ptrace.c *kdata = reg; kdata 1617 arch/arm64/kernel/ptrace.c u32 *kdata) kdata 1626 arch/arm64/kernel/ptrace.c *kdata = (u32)addr; kdata 1629 arch/arm64/kernel/ptrace.c *kdata = ctrl; kdata 1638 arch/arm64/kernel/ptrace.c u32 *kdata) kdata 1646 arch/arm64/kernel/ptrace.c addr = *kdata; kdata 1649 arch/arm64/kernel/ptrace.c ctrl = *kdata; kdata 1660 arch/arm64/kernel/ptrace.c u32 kdata; kdata 1664 arch/arm64/kernel/ptrace.c ret = compat_ptrace_hbp_get(NT_ARM_HW_WATCH, tsk, num, &kdata); kdata 1667 arch/arm64/kernel/ptrace.c ret = compat_ptrace_hbp_get_resource_info(&kdata); kdata 1670 arch/arm64/kernel/ptrace.c ret = compat_ptrace_hbp_get(NT_ARM_HW_BREAK, tsk, num, &kdata); kdata 1674 arch/arm64/kernel/ptrace.c ret = put_user(kdata, data); kdata 1683 arch/arm64/kernel/ptrace.c u32 kdata = 0; kdata 1688 arch/arm64/kernel/ptrace.c ret = get_user(kdata, data); kdata 1693 arch/arm64/kernel/ptrace.c ret = compat_ptrace_hbp_set(NT_ARM_HW_WATCH, tsk, num, &kdata); kdata 1695 arch/arm64/kernel/ptrace.c ret = compat_ptrace_hbp_set(NT_ARM_HW_BREAK, tsk, num, &kdata); kdata 447 drivers/gpu/drm/amd/amdgpu/amdgpu.h void *kdata; kdata 171 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c p->chunks[i].kdata = kvmalloc_array(size, sizeof(uint32_t), GFP_KERNEL); kdata 172 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c if (p->chunks[i].kdata == NULL) { kdata 178 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c if (copy_from_user(p->chunks[i].kdata, cdata, size)) { kdata 195 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c ret = amdgpu_cs_user_fence_chunk(p, p->chunks[i].kdata, kdata 209 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c ret = amdgpu_cs_bo_handles_chunk(p, p->chunks[i].kdata); kdata 251 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c kvfree(p->chunks[i].kdata); kdata 773 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c kvfree(parser->chunks[i].kdata); kdata 810 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c chunk_ib = chunk->kdata; kdata 957 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c chunk_ib = (struct drm_amdgpu_cs_chunk_ib *)chunk->kdata; kdata 1022 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c deps = (struct drm_amdgpu_cs_chunk_dep *)chunk->kdata; kdata 1095 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c deps = (struct drm_amdgpu_cs_chunk_sem *)chunk->kdata; kdata 1116 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c syncobj_deps = (struct drm_amdgpu_cs_chunk_syncobj *)chunk->kdata; kdata 1138 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c deps = (struct drm_amdgpu_cs_chunk_sem *)chunk->kdata; kdata 1174 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c syncobj_deps = (struct drm_amdgpu_cs_chunk_syncobj *)chunk->kdata; kdata 1783 drivers/gpu/drm/amd/amdkfd/kfd_chardev.c char *kdata = NULL; kdata 1823 drivers/gpu/drm/amd/amdkfd/kfd_chardev.c kdata = stack_kdata; kdata 1825 drivers/gpu/drm/amd/amdkfd/kfd_chardev.c kdata = kmalloc(asize, GFP_KERNEL); kdata 1826 drivers/gpu/drm/amd/amdkfd/kfd_chardev.c if (!kdata) { kdata 1832 drivers/gpu/drm/amd/amdkfd/kfd_chardev.c memset(kdata + usize, 0, asize - usize); kdata 1836 drivers/gpu/drm/amd/amdkfd/kfd_chardev.c if (copy_from_user(kdata, (void __user *)arg, usize) != 0) { kdata 1841 drivers/gpu/drm/amd/amdkfd/kfd_chardev.c memset(kdata, 0, usize); kdata 1844 drivers/gpu/drm/amd/amdkfd/kfd_chardev.c retcode = func(filep, process, kdata); kdata 1847 drivers/gpu/drm/amd/amdkfd/kfd_chardev.c if (copy_to_user((void __user *)arg, kdata, usize) != 0) kdata 1855 drivers/gpu/drm/amd/amdkfd/kfd_chardev.c if (kdata != stack_kdata) kdata 1856 drivers/gpu/drm/amd/amdkfd/kfd_chardev.c kfree(kdata); kdata 769 drivers/gpu/drm/drm_ioctl.c long drm_ioctl_kernel(struct file *file, drm_ioctl_t *func, void *kdata, kdata 786 drivers/gpu/drm/drm_ioctl.c retcode = func(dev, kdata, file_priv); kdata 789 drivers/gpu/drm/drm_ioctl.c retcode = func(dev, kdata, file_priv); kdata 819 drivers/gpu/drm/drm_ioctl.c char *kdata = NULL; kdata 869 drivers/gpu/drm/drm_ioctl.c kdata = stack_kdata; kdata 871 drivers/gpu/drm/drm_ioctl.c kdata = kmalloc(ksize, GFP_KERNEL); kdata 872 drivers/gpu/drm/drm_ioctl.c if (!kdata) { kdata 878 drivers/gpu/drm/drm_ioctl.c if (copy_from_user(kdata, (void __user *)arg, in_size) != 0) { kdata 884 drivers/gpu/drm/drm_ioctl.c memset(kdata + in_size, 0, ksize - in_size); kdata 886 drivers/gpu/drm/drm_ioctl.c retcode = drm_ioctl_kernel(filp, func, kdata, ioctl->flags); kdata 887 drivers/gpu/drm/drm_ioctl.c if (copy_to_user((void __user *)arg, kdata, out_size) != 0) kdata 897 drivers/gpu/drm/drm_ioctl.c if (kdata != stack_kdata) kdata 898 drivers/gpu/drm/drm_ioctl.c kfree(kdata); kdata 1053 drivers/gpu/drm/radeon/radeon.h uint32_t *kdata; kdata 1093 drivers/gpu/drm/radeon/radeon.h if (ibc->kdata) kdata 1094 drivers/gpu/drm/radeon/radeon.h return ibc->kdata[idx]; kdata 109 drivers/gpu/drm/radeon/radeon_cs.c r = (struct drm_radeon_cs_reloc *)&chunk->kdata[i*4]; kdata 351 drivers/gpu/drm/radeon/radeon_cs.c p->chunks[i].kdata = kvmalloc_array(size, sizeof(uint32_t), GFP_KERNEL); kdata 353 drivers/gpu/drm/radeon/radeon_cs.c if (p->chunks[i].kdata == NULL) { kdata 356 drivers/gpu/drm/radeon/radeon_cs.c if (copy_from_user(p->chunks[i].kdata, cdata, size)) { kdata 360 drivers/gpu/drm/radeon/radeon_cs.c p->cs_flags = p->chunks[i].kdata[0]; kdata 362 drivers/gpu/drm/radeon/radeon_cs.c ring = p->chunks[i].kdata[1]; kdata 364 drivers/gpu/drm/radeon/radeon_cs.c priority = (s32)p->chunks[i].kdata[2]; kdata 453 drivers/gpu/drm/radeon/radeon_cs.c kvfree(parser->chunks[i].kdata); kdata 653 drivers/gpu/drm/radeon/radeon_cs.c if (ib_chunk->kdata) kdata 654 drivers/gpu/drm/radeon/radeon_cs.c memcpy(parser->ib.ptr, ib_chunk->kdata, ib_chunk->length_dw * 4); kdata 878 drivers/gpu/drm/radeon/radeon_cs.c (u64)relocs_chunk->kdata[idx + 3] << 32; kdata 879 drivers/gpu/drm/radeon/radeon_cs.c (*cs_reloc)->gpu_offset |= relocs_chunk->kdata[idx + 0]; kdata 358 drivers/misc/habanalabs/habanalabs_ioctl.c char *kdata = NULL; kdata 388 drivers/misc/habanalabs/habanalabs_ioctl.c kdata = stack_kdata; kdata 390 drivers/misc/habanalabs/habanalabs_ioctl.c kdata = kzalloc(asize, GFP_KERNEL); kdata 391 drivers/misc/habanalabs/habanalabs_ioctl.c if (!kdata) { kdata 399 drivers/misc/habanalabs/habanalabs_ioctl.c if (copy_from_user(kdata, (void __user *)arg, usize)) { kdata 404 drivers/misc/habanalabs/habanalabs_ioctl.c memset(kdata, 0, usize); kdata 407 drivers/misc/habanalabs/habanalabs_ioctl.c retcode = func(hpriv, kdata); kdata 410 drivers/misc/habanalabs/habanalabs_ioctl.c if (copy_to_user((void __user *)arg, kdata, usize)) kdata 418 drivers/misc/habanalabs/habanalabs_ioctl.c if (kdata != stack_kdata) kdata 419 drivers/misc/habanalabs/habanalabs_ioctl.c kfree(kdata); kdata 628 drivers/xen/privcmd.c struct privcmd_dm_op kdata; kdata 636 drivers/xen/privcmd.c if (copy_from_user(&kdata, udata, sizeof(kdata))) kdata 640 drivers/xen/privcmd.c if (data->domid != DOMID_INVALID && data->domid != kdata.dom) kdata 643 drivers/xen/privcmd.c if (kdata.num == 0) kdata 646 drivers/xen/privcmd.c if (kdata.num > privcmd_dm_op_max_num) kdata 649 drivers/xen/privcmd.c kbufs = kcalloc(kdata.num, sizeof(*kbufs), GFP_KERNEL); kdata 653 drivers/xen/privcmd.c if (copy_from_user(kbufs, kdata.ubufs, kdata 654 drivers/xen/privcmd.c sizeof(*kbufs) * kdata.num)) { kdata 659 drivers/xen/privcmd.c for (i = 0; i < kdata.num; i++) { kdata 682 drivers/xen/privcmd.c xbufs = kcalloc(kdata.num, sizeof(*xbufs), GFP_KERNEL); kdata 688 drivers/xen/privcmd.c rc = lock_pages(kbufs, kdata.num, pages, nr_pages); kdata 692 drivers/xen/privcmd.c for (i = 0; i < kdata.num; i++) { kdata 698 drivers/xen/privcmd.c rc = HYPERVISOR_dm_op(kdata.dom, kdata.num, xbufs); kdata 732 drivers/xen/privcmd.c struct privcmd_mmap_resource kdata; kdata 737 drivers/xen/privcmd.c if (copy_from_user(&kdata, udata, sizeof(kdata))) kdata 741 drivers/xen/privcmd.c if (data->domid != DOMID_INVALID && data->domid != kdata.dom) kdata 746 drivers/xen/privcmd.c vma = find_vma(mm, kdata.addr); kdata 752 drivers/xen/privcmd.c pfns = kcalloc(kdata.num, sizeof(*pfns), GFP_KERNEL); kdata 760 drivers/xen/privcmd.c unsigned int nr = DIV_ROUND_UP(kdata.num, XEN_PFN_PER_PAGE); kdata 769 drivers/xen/privcmd.c for (i = 0; i < kdata.num; i++) { kdata 779 drivers/xen/privcmd.c xdata.domid = kdata.dom; kdata 780 drivers/xen/privcmd.c xdata.type = kdata.type; kdata 781 drivers/xen/privcmd.c xdata.id = kdata.id; kdata 782 drivers/xen/privcmd.c xdata.frame = kdata.idx; kdata 783 drivers/xen/privcmd.c xdata.nr_frames = kdata.num; kdata 795 drivers/xen/privcmd.c rc = xen_remap_vma_range(vma, kdata.addr, kdata.num << PAGE_SHIFT); kdata 799 drivers/xen/privcmd.c DOMID_SELF : kdata.dom; kdata 803 drivers/xen/privcmd.c kdata.addr & PAGE_MASK, kdata 804 drivers/xen/privcmd.c pfns, kdata.num, (int *)pfns, kdata 810 drivers/xen/privcmd.c else if (num != kdata.num) { kdata 167 kernel/capability.c struct __user_cap_data_struct kdata[_KERNEL_CAPABILITY_U32S]; kdata 171 kernel/capability.c kdata[i].effective = pE.cap[i]; kdata 172 kernel/capability.c kdata[i].permitted = pP.cap[i]; kdata 173 kernel/capability.c kdata[i].inheritable = pI.cap[i]; kdata 195 kernel/capability.c if (copy_to_user(dataptr, kdata, tocopy kdata 224 kernel/capability.c struct __user_cap_data_struct kdata[_KERNEL_CAPABILITY_U32S]; kdata 243 kernel/capability.c if (copybytes > sizeof(kdata)) kdata 246 kernel/capability.c if (copy_from_user(&kdata, data, copybytes)) kdata 250 kernel/capability.c effective.cap[i] = kdata[i].effective; kdata 251 kernel/capability.c permitted.cap[i] = kdata[i].permitted; kdata 252 kernel/capability.c inheritable.cap[i] = kdata[i].inheritable;