num_chunks 255 drivers/crypto/qat/qat_common/icp_qat_uclo.h unsigned short num_chunks; num_chunks 270 drivers/crypto/qat/qat_common/icp_qat_uclo.h short num_chunks; num_chunks 510 drivers/crypto/qat/qat_common/icp_qat_uclo.h unsigned short num_chunks; num_chunks 155 drivers/crypto/qat/qat_common/qat_uclo.c if (suof_hdr->num_chunks <= 0x1) { num_chunks 484 drivers/crypto/qat/qat_common/qat_uclo.c for (i = 0; i < obj_hdr->num_chunks; i++) { num_chunks 531 drivers/crypto/qat/qat_common/qat_uclo.c for (i = 0; i < file_hdr->num_chunks; i++) { num_chunks 1162 drivers/crypto/qat/qat_common/qat_uclo.c suof_handle->img_table.num_simgs = suof_ptr->num_chunks - 1; num_chunks 116 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c if (cs->in.num_chunks == 0) num_chunks 119 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c chunk_array = kmalloc_array(cs->in.num_chunks, sizeof(uint64_t), GFP_KERNEL); num_chunks 140 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c sizeof(uint64_t)*cs->in.num_chunks)) { num_chunks 145 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c p->nchunks = cs->in.num_chunks; num_chunks 280 drivers/gpu/drm/radeon/radeon_cs.c if (!cs->num_chunks) { num_chunks 292 drivers/gpu/drm/radeon/radeon_cs.c p->chunks_array = kcalloc(cs->num_chunks, sizeof(uint64_t), GFP_KERNEL); num_chunks 298 drivers/gpu/drm/radeon/radeon_cs.c sizeof(uint64_t)*cs->num_chunks)) { num_chunks 302 drivers/gpu/drm/radeon/radeon_cs.c p->nchunks = cs->num_chunks; num_chunks 375 drivers/infiniband/sw/siw/siw_mem.c int num_pages, num_chunks, i, rv = 0; num_chunks 385 drivers/infiniband/sw/siw/siw_mem.c num_chunks = (num_pages >> CHUNK_SHIFT) + 1; num_chunks 411 drivers/infiniband/sw/siw/siw_mem.c kcalloc(num_chunks, sizeof(struct siw_page_chunk), GFP_KERNEL); num_chunks 466 drivers/misc/habanalabs/command_submission.c u32 num_chunks, u64 *cs_seq) num_chunks 479 drivers/misc/habanalabs/command_submission.c if (num_chunks > HL_MAX_JOBS_PER_CS) { num_chunks 487 drivers/misc/habanalabs/command_submission.c cs_chunk_array = kmalloc_array(num_chunks, sizeof(*cs_chunk_array), num_chunks 494 drivers/misc/habanalabs/command_submission.c size_to_copy = num_chunks * sizeof(struct hl_cs_chunk); num_chunks 515 drivers/misc/habanalabs/command_submission.c for (i = 0, parse_cnt = 0 ; i < num_chunks ; i++, parse_cnt++) { num_chunks 616 drivers/misc/habanalabs/command_submission.c u32 num_chunks; num_chunks 635 drivers/misc/habanalabs/command_submission.c num_chunks = args->in.num_chunks_restore; num_chunks 663 drivers/misc/habanalabs/command_submission.c if (num_chunks == 0) { num_chunks 668 drivers/misc/habanalabs/command_submission.c rc = _hl_cs_ioctl(hpriv, chunks, num_chunks, num_chunks 682 drivers/misc/habanalabs/command_submission.c if (num_chunks > 0) { num_chunks 711 drivers/misc/habanalabs/command_submission.c num_chunks = args->in.num_chunks_execute; num_chunks 713 drivers/misc/habanalabs/command_submission.c if (num_chunks == 0) { num_chunks 721 drivers/misc/habanalabs/command_submission.c rc = _hl_cs_ioctl(hpriv, chunks, num_chunks, &cs_seq); num_chunks 255 drivers/mtd/nand/raw/mxc_nand.c u16 num_chunks = mtd->writesize / 512; num_chunks 262 drivers/mtd/nand/raw/mxc_nand.c oob_chunk_size = (host->used_oobsize / num_chunks) & ~1; num_chunks 265 drivers/mtd/nand/raw/mxc_nand.c for (i = 0; i < num_chunks - 1; i++) num_chunks 275 drivers/mtd/nand/raw/mxc_nand.c for (i = 0; i < num_chunks - 1; i++) num_chunks 3812 drivers/net/wireless/intel/ipw2x00/ipw2200.c if (le32_to_cpu(bd->u.data.num_chunks) > NUM_TFD_CHUNKS) { num_chunks 3814 drivers/net/wireless/intel/ipw2x00/ipw2200.c le32_to_cpu(bd->u.data.num_chunks)); num_chunks 3820 drivers/net/wireless/intel/ipw2x00/ipw2200.c for (i = 0; i < le32_to_cpu(bd->u.data.num_chunks); i++) { num_chunks 10213 drivers/net/wireless/intel/ipw2x00/ipw2200.c tfd->u.data.num_chunks = cpu_to_le32(min((u8) (NUM_TFD_CHUNKS - 2), num_chunks 10216 drivers/net/wireless/intel/ipw2x00/ipw2200.c txb->nr_frags, le32_to_cpu(tfd->u.data.num_chunks)); num_chunks 10217 drivers/net/wireless/intel/ipw2x00/ipw2200.c for (i = 0; i < le32_to_cpu(tfd->u.data.num_chunks); i++) { num_chunks 10219 drivers/net/wireless/intel/ipw2x00/ipw2200.c i, le32_to_cpu(tfd->u.data.num_chunks), num_chunks 10222 drivers/net/wireless/intel/ipw2x00/ipw2200.c i, tfd->u.data.num_chunks, num_chunks 10267 drivers/net/wireless/intel/ipw2x00/ipw2200.c le32_add_cpu(&tfd->u.data.num_chunks, 1); num_chunks 488 drivers/net/wireless/intel/ipw2x00/ipw2200.h __le32 num_chunks; num_chunks 1559 drivers/staging/media/ipu3/ipu3-abi.h u32 num_chunks; num_chunks 377 fs/btrfs/inode.c atomic_t num_chunks; num_chunks 1219 fs/btrfs/inode.c u64 num_chunks = DIV_ROUND_UP(end - start, SZ_512K); num_chunks 1228 fs/btrfs/inode.c num_chunks = 1; num_chunks 1235 fs/btrfs/inode.c ctx = kvmalloc(struct_size(ctx, chunks, num_chunks), GFP_KERNEL); num_chunks 1252 fs/btrfs/inode.c atomic_set(&ctx->num_chunks, num_chunks); num_chunks 1254 fs/btrfs/inode.c for (i = 0; i < num_chunks; i++) { num_chunks 1265 fs/btrfs/inode.c async_chunk[i].pending = &ctx->num_chunks; num_chunks 554 include/uapi/drm/amdgpu_drm.h __u32 num_chunks; num_chunks 984 include/uapi/drm/radeon_drm.h __u32 num_chunks; num_chunks 6989 net/sctp/socket.c u32 num_chunks = 0; num_chunks 7011 net/sctp/socket.c num_chunks = ntohs(ch->param_hdr.length) - sizeof(struct sctp_paramhdr); num_chunks 7012 net/sctp/socket.c if (len < num_chunks) num_chunks 7015 net/sctp/socket.c if (copy_to_user(to, ch->chunks, num_chunks)) num_chunks 7018 net/sctp/socket.c len = sizeof(struct sctp_authchunks) + num_chunks; num_chunks 7021 net/sctp/socket.c if (put_user(num_chunks, &p->gauth_number_of_chunks)) num_chunks 7034 net/sctp/socket.c u32 num_chunks = 0; num_chunks 7061 net/sctp/socket.c num_chunks = ntohs(ch->param_hdr.length) - sizeof(struct sctp_paramhdr); num_chunks 7062 net/sctp/socket.c if (len < sizeof(struct sctp_authchunks) + num_chunks) num_chunks 7065 net/sctp/socket.c if (copy_to_user(to, ch->chunks, num_chunks)) num_chunks 7068 net/sctp/socket.c len = sizeof(struct sctp_authchunks) + num_chunks; num_chunks 7071 net/sctp/socket.c if (put_user(num_chunks, &p->gauth_number_of_chunks))