/linux-4.1.27/include/linux/ceph/ |
H A D | buffer.h | 18 size_t alloc_len; member in struct:ceph_buffer
|
/linux-4.1.27/drivers/scsi/ |
H A D | scsi_trace.c | 197 u32 alloc_len = 0; scsi_trace_service_action_in() local 219 alloc_len |= (cdb[10] << 24); scsi_trace_service_action_in() 220 alloc_len |= (cdb[11] << 16); scsi_trace_service_action_in() 221 alloc_len |= (cdb[12] << 8); scsi_trace_service_action_in() 222 alloc_len |= cdb[13]; scsi_trace_service_action_in() 224 trace_seq_printf(p, "%s lba=%llu alloc_len=%u", cmd, scsi_trace_service_action_in() 225 (unsigned long long)lba, alloc_len); scsi_trace_service_action_in()
|
H A D | scsi_debug.c | 1284 int alloc_len, n, ret; resp_inquiry() local 1287 alloc_len = (cmd[3] << 8) + cmd[4]; resp_inquiry() 1388 len = min(((arr[2] << 8) + arr[3]) + 4, alloc_len); resp_inquiry() 1419 min(alloc_len, SDEBUG_LONG_INQ_SZ)); resp_inquiry() 1546 int k, alloc_len; resp_readcap16() local 1548 alloc_len = ((cmd[10] << 24) + (cmd[11] << 16) + (cmd[12] << 8) resp_readcap16() 1577 min(alloc_len, SDEBUG_READCAP16_ARR_SZ)); resp_readcap16() 1669 u32 alloc_len, a_len; resp_rsup_opcodes() local 1680 alloc_len = get_unaligned_be32(cmd + 6); resp_rsup_opcodes() 1681 if (alloc_len < 4 || alloc_len > 0xffff) { resp_rsup_opcodes() 1685 if (alloc_len > 8192) resp_rsup_opcodes() 1688 a_len = alloc_len; resp_rsup_opcodes() 1808 len = (offset < alloc_len) ? offset : alloc_len; resp_rsup_opcodes() 1818 u32 alloc_len, len; resp_rsup_tmfs() local 1824 alloc_len = get_unaligned_be32(cmd + 6); resp_rsup_tmfs() 1825 if (alloc_len < 4) { resp_rsup_tmfs() 1837 len = (len < alloc_len) ? len : alloc_len; resp_rsup_tmfs() 2006 int k, alloc_len, msense_6, offset, len, target_dev_id; resp_mode_sense() local 2022 alloc_len = msense_6 ? cmd[4] : ((cmd[7] << 8) | cmd[8]); resp_mode_sense() 2154 return fill_from_dev_buffer(scp, arr, min(alloc_len, offset)); resp_mode_sense() 2266 int ppc, sp, pcontrol, pcode, subpcode, alloc_len, len, n; resp_log_sense() local 2280 alloc_len = (cmd[7] << 8) + cmd[8]; resp_log_sense() 2337 len = min(((arr[2] << 8) + arr[3]) + 4, alloc_len); resp_log_sense() 3287 u32 alloc_len, mapped, num; resp_get_lba_status() local 3292 alloc_len = get_unaligned_be32(cmd + 10); resp_get_lba_status() 3294 if (alloc_len < 24) resp_get_lba_status() 3327 unsigned int alloc_len; resp_report_luns() local 3337 alloc_len = cmd[9] + (cmd[8] << 8) + (cmd[7] << 16) + (cmd[6] << 24); resp_report_luns() 3338 shortish = (alloc_len < 4); resp_report_luns() 3376 alloc_len = (unsigned char *)(one_lun + i) - arr; resp_report_luns() 3378 min((int)alloc_len, SDEBUG_RLUN_ARR_SZ)); resp_report_luns()
|
/linux-4.1.27/net/ceph/ |
H A D | buffer.c | 26 b->alloc_len = len; ceph_buffer_new()
|
H A D | auth_x.c | 305 if (au->buf && au->buf->alloc_len < maxlen) { ceph_x_build_authorizer()
|
/linux-4.1.27/drivers/staging/lustre/lustre/ptlrpc/ |
H A D | sec_plain.c | 544 int alloc_len; plain_alloc_reqbuf() local 557 alloc_len = lustre_msg_size_v2(PLAIN_PACK_SEGMENTS, buflens); plain_alloc_reqbuf() 562 alloc_len = size_roundup_power2(alloc_len); plain_alloc_reqbuf() 563 OBD_ALLOC_LARGE(req->rq_reqbuf, alloc_len); plain_alloc_reqbuf() 567 req->rq_reqbuf_len = alloc_len; plain_alloc_reqbuf() 570 LASSERT(req->rq_reqbuf_len >= alloc_len); plain_alloc_reqbuf() 571 memset(req->rq_reqbuf, 0, alloc_len); plain_alloc_reqbuf() 600 int alloc_len; plain_alloc_repbuf() local 610 alloc_len = lustre_msg_size_v2(PLAIN_PACK_SEGMENTS, buflens); plain_alloc_repbuf() 613 alloc_len += plain_at_offset; plain_alloc_repbuf() 615 alloc_len = size_roundup_power2(alloc_len); plain_alloc_repbuf() 617 OBD_ALLOC_LARGE(req->rq_repbuf, alloc_len); plain_alloc_repbuf() 621 req->rq_repbuf_len = alloc_len; plain_alloc_repbuf()
|
/linux-4.1.27/drivers/block/ |
H A D | nvme-scsi.c | 673 int alloc_len) nvme_trans_standard_inquiry_page() 727 xfer_len = min(alloc_len, STANDARD_INQUIRY_LENGTH); nvme_trans_standard_inquiry_page() 739 int alloc_len) nvme_trans_supported_vpd_pages() 754 xfer_len = min(alloc_len, STANDARD_INQUIRY_LENGTH); nvme_trans_supported_vpd_pages() 762 int alloc_len) nvme_trans_unit_serial_page() 773 xfer_len = min(alloc_len, STANDARD_INQUIRY_LENGTH); nvme_trans_unit_serial_page() 780 u8 *inq_response, int alloc_len) nvme_trans_device_id_page() 797 memset(inq_response, 0, alloc_len); nvme_trans_device_id_page() 831 if (alloc_len < 72) { nvme_trans_device_id_page() 850 xfer_len = alloc_len; nvme_trans_device_id_page() 861 int alloc_len) nvme_trans_ext_inq_page() 932 xfer_len = min(alloc_len, EXTENDED_INQUIRY_DATA_PAGE_LENGTH); nvme_trans_ext_inq_page() 945 u8 *inq_response, int alloc_len) nvme_trans_bdev_limits_page() 965 int alloc_len) nvme_trans_bdev_char_page() 984 xfer_len = min(alloc_len, EXTENDED_INQUIRY_DATA_PAGE_LENGTH); nvme_trans_bdev_char_page() 995 int alloc_len) nvme_trans_log_supp_pages() 1014 xfer_len = min(alloc_len, LOG_PAGE_SUPPORTED_LOG_PAGES_LENGTH); nvme_trans_log_supp_pages() 1023 struct sg_io_hdr *hdr, int alloc_len) nvme_trans_log_info_exceptions() 1078 xfer_len = min(alloc_len, LOG_INFO_EXCP_PAGE_LENGTH); nvme_trans_log_info_exceptions() 1090 int alloc_len) nvme_trans_log_temperature() 1159 xfer_len = min(alloc_len, LOG_TEMP_PAGE_LENGTH); nvme_trans_log_temperature() 1387 u16 alloc_len, u8 cdb10, nvme_trans_mode_page_create() 1436 xfer_len = min(alloc_len, resp_size); nvme_trans_mode_page_create() 2253 int alloc_len; nvme_trans_inquiry() local 2258 alloc_len = GET_INQ_ALLOC_LENGTH(cmd); nvme_trans_inquiry() 2260 inq_response = kmalloc(max(alloc_len, STANDARD_INQUIRY_LENGTH), nvme_trans_inquiry() 2270 inq_response, alloc_len); nvme_trans_inquiry() 2282 inq_response, alloc_len); nvme_trans_inquiry() 2286 alloc_len); nvme_trans_inquiry() 2290 alloc_len); nvme_trans_inquiry() 2293 res = nvme_trans_ext_inq_page(ns, hdr, alloc_len); nvme_trans_inquiry() 2297 alloc_len); nvme_trans_inquiry() 2300 res = nvme_trans_bdev_char_page(ns, hdr, alloc_len); nvme_trans_inquiry() 2320 u16 alloc_len; nvme_trans_log_sense() local 2341 alloc_len = GET_U16_FROM_CDB(cmd, LOG_SENSE_CDB_ALLOC_LENGTH_OFFSET); nvme_trans_log_sense() 2344 res = nvme_trans_log_supp_pages(ns, hdr, alloc_len); nvme_trans_log_sense() 2347 res = nvme_trans_log_info_exceptions(ns, hdr, alloc_len); nvme_trans_log_sense() 2350 res = nvme_trans_log_temperature(ns, hdr, alloc_len); nvme_trans_log_sense() 2403 u16 alloc_len; nvme_trans_mode_sense() local 2409 alloc_len = GET_U8_FROM_CDB(cmd, MODE_SENSE6_ALLOC_LEN_OFFSET); nvme_trans_mode_sense() 2411 alloc_len = GET_U16_FROM_CDB(cmd, nvme_trans_mode_sense() 2429 res = nvme_trans_mode_page_create(ns, hdr, cmd, alloc_len, nvme_trans_mode_sense() 2435 res = nvme_trans_mode_page_create(ns, hdr, cmd, alloc_len, nvme_trans_mode_sense() 2441 res = nvme_trans_mode_page_create(ns, hdr, cmd, alloc_len, nvme_trans_mode_sense() 2447 res = nvme_trans_mode_page_create(ns, hdr, cmd, alloc_len, nvme_trans_mode_sense() 2453 res = nvme_trans_mode_page_create(ns, hdr, cmd, alloc_len, nvme_trans_mode_sense() 2474 u32 alloc_len = READ_CAP_10_RESP_SIZE; nvme_trans_read_capacity() local 2486 alloc_len = GET_READ_CAP_16_ALLOC_LENGTH(cmd); nvme_trans_read_capacity() 2514 xfer_len = min(alloc_len, resp_size); nvme_trans_read_capacity() 2530 u32 alloc_len, xfer_len, resp_size; nvme_trans_report_luns() local 2541 alloc_len = GET_REPORT_LUNS_ALLOC_LENGTH(cmd); nvme_trans_report_luns() 2572 if (alloc_len < resp_size) { nvme_trans_report_luns() 2600 xfer_len = min(alloc_len, resp_size); nvme_trans_report_luns() 2615 u8 alloc_len, xfer_len, resp_size; nvme_trans_request_sense() local 2619 alloc_len = GET_REQUEST_SENSE_ALLOC_LENGTH(cmd); nvme_trans_request_sense() 2653 xfer_len = min(alloc_len, resp_size); nvme_trans_request_sense() 671 nvme_trans_standard_inquiry_page(struct nvme_ns *ns, struct sg_io_hdr *hdr, u8 *inq_response, int alloc_len) nvme_trans_standard_inquiry_page() argument 737 nvme_trans_supported_vpd_pages(struct nvme_ns *ns, struct sg_io_hdr *hdr, u8 *inq_response, int alloc_len) nvme_trans_supported_vpd_pages() argument 760 nvme_trans_unit_serial_page(struct nvme_ns *ns, struct sg_io_hdr *hdr, u8 *inq_response, int alloc_len) nvme_trans_unit_serial_page() argument 779 nvme_trans_device_id_page(struct nvme_ns *ns, struct sg_io_hdr *hdr, u8 *inq_response, int alloc_len) nvme_trans_device_id_page() argument 860 nvme_trans_ext_inq_page(struct nvme_ns *ns, struct sg_io_hdr *hdr, int alloc_len) nvme_trans_ext_inq_page() argument 944 nvme_trans_bdev_limits_page(struct nvme_ns *ns, struct sg_io_hdr *hdr, u8 *inq_response, int alloc_len) nvme_trans_bdev_limits_page() argument 964 nvme_trans_bdev_char_page(struct nvme_ns *ns, struct sg_io_hdr *hdr, int alloc_len) nvme_trans_bdev_char_page() argument 994 nvme_trans_log_supp_pages(struct nvme_ns *ns, struct sg_io_hdr *hdr, int alloc_len) nvme_trans_log_supp_pages() argument 1022 nvme_trans_log_info_exceptions(struct nvme_ns *ns, struct sg_io_hdr *hdr, int alloc_len) nvme_trans_log_info_exceptions() argument 1089 nvme_trans_log_temperature(struct nvme_ns *ns, struct sg_io_hdr *hdr, int alloc_len) nvme_trans_log_temperature() argument 1385 nvme_trans_mode_page_create(struct nvme_ns *ns, struct sg_io_hdr *hdr, u8 *cmd, u16 alloc_len, u8 cdb10, int (*mode_page_fill_func) (struct nvme_ns *, struct sg_io_hdr *hdr, u8 *, int), u16 mode_pages_tot_len) nvme_trans_mode_page_create() argument
|
/linux-4.1.27/tools/lib/traceevent/ |
H A D | plugin_scsi.c | 317 u32 alloc_len = 0; scsi_trace_service_action_in() local 339 alloc_len |= (cdb[10] << 24); scsi_trace_service_action_in() 340 alloc_len |= (cdb[11] << 16); scsi_trace_service_action_in() 341 alloc_len |= (cdb[12] << 8); scsi_trace_service_action_in() 342 alloc_len |= cdb[13]; scsi_trace_service_action_in() 344 trace_seq_printf(p, "%s lba=%llu alloc_len=%u", cmd, scsi_trace_service_action_in() 345 (unsigned long long)lba, alloc_len); scsi_trace_service_action_in()
|
/linux-4.1.27/fs/notify/inotify/ |
H A D | inotify_fsnotify.c | 77 int alloc_len = sizeof(struct inotify_event_info); inotify_handle_event() local 90 alloc_len += len + 1; inotify_handle_event() 99 event = kmalloc(alloc_len, GFP_KERNEL); inotify_handle_event()
|
/linux-4.1.27/drivers/net/wireless/cw1200/ |
H A D | bh.c | 252 size_t alloc_len; cw1200_bh_rx_helper() local 271 alloc_len = priv->hwbus_ops->align_size( cw1200_bh_rx_helper() 275 if (WARN_ON_ONCE(alloc_len > EFFECTIVE_BUF_SIZE)) { cw1200_bh_rx_helper() 277 alloc_len); cw1200_bh_rx_helper() 280 skb_rx = dev_alloc_skb(alloc_len); cw1200_bh_rx_helper() 290 if (WARN_ON(cw1200_data_read(priv, data, alloc_len))) { cw1200_bh_rx_helper() 291 pr_err("rx blew up, len %zu\n", alloc_len); cw1200_bh_rx_helper() 297 ((__le16 *)data)[alloc_len / 2 - 1]); cw1200_bh_rx_helper()
|
/linux-4.1.27/drivers/net/ethernet/ |
H A D | ec_bhf.c | 127 size_t alloc_len; member in struct:bhf_dma 348 buf->alloc_len = 2 * buf->len; ec_bhf_alloc_dma_mem() 350 buf->alloc = dma_alloc_coherent(dev, buf->alloc_len, &buf->alloc_phys, ec_bhf_alloc_dma_mem() 436 dma_free_coherent(dev, priv->rx_buf.alloc_len, priv->rx_buf.alloc, ec_bhf_open() 437 priv->rx_buf.alloc_len); ec_bhf_open() 453 dma_free_coherent(dev, priv->tx_buf.alloc_len, ec_bhf_stop() 455 dma_free_coherent(dev, priv->rx_buf.alloc_len, ec_bhf_stop()
|
/linux-4.1.27/drivers/net/ethernet/intel/i40e/ |
H A D | i40e_hmc.c | 53 u64 alloc_len; i40e_add_sd_table_entry() local 71 alloc_len = I40E_HMC_PAGED_BP_SIZE; i40e_add_sd_table_entry() 74 alloc_len = direct_mode_sz; i40e_add_sd_table_entry() 78 ret_code = i40e_allocate_dma_mem(hw, &mem, mem_type, alloc_len, i40e_add_sd_table_entry()
|
/linux-4.1.27/fs/jffs2/ |
H A D | file.c | 156 uint32_t alloc_len; jffs2_write_begin() local 161 ret = jffs2_reserve_space(c, sizeof(ri), &alloc_len, jffs2_write_begin()
|
/linux-4.1.27/drivers/scsi/lpfc/ |
H A D | lpfc_mbox.c | 1767 uint32_t alloc_len; lpfc_sli4_config() local 1808 for (pagen = 0, alloc_len = 0; pagen < pcount; pagen++) { lpfc_sli4_config() 1824 resid_len = length - alloc_len; lpfc_sli4_config() 1828 alloc_len += SLI4_PAGE_SIZE; lpfc_sli4_config() 1832 alloc_len = length; lpfc_sli4_config() 1837 sli4_config->header.cfg_mhdr.payload_length = alloc_len; lpfc_sli4_config() 1845 alloc_len - sizeof(union lpfc_sli4_cfg_shdr); lpfc_sli4_config() 1851 return alloc_len; lpfc_sli4_config() 2016 uint32_t alloc_len, req_len; lpfc_sli4_mbx_read_fcf_rec() local 2026 alloc_len = lpfc_sli4_config(phba, mboxq, LPFC_MBOX_SUBSYSTEM_FCOE, lpfc_sli4_mbx_read_fcf_rec() 2030 if (alloc_len < req_len) { lpfc_sli4_mbx_read_fcf_rec() 2034 "size (x%x)\n", alloc_len, req_len); lpfc_sli4_mbx_read_fcf_rec()
|
H A D | lpfc_bsg.c | 1889 uint32_t req_len, alloc_len; lpfc_sli4_bsg_set_link_diag_state() local 1898 alloc_len = lpfc_sli4_config(phba, pmboxq, LPFC_MBOX_SUBSYSTEM_FCOE, lpfc_sli4_bsg_set_link_diag_state() 1901 if (alloc_len != req_len) { lpfc_sli4_bsg_set_link_diag_state() 1949 uint32_t req_len, alloc_len; lpfc_sli4_bsg_set_internal_loopback() local 1958 alloc_len = lpfc_sli4_config(phba, pmboxq, LPFC_MBOX_SUBSYSTEM_FCOE, lpfc_sli4_bsg_set_internal_loopback() 1961 if (alloc_len != req_len) { lpfc_sli4_bsg_set_internal_loopback() 2290 uint32_t req_len, alloc_len; lpfc_sli4_bsg_link_diag_test() local 2355 alloc_len = lpfc_sli4_config(phba, pmboxq, LPFC_MBOX_SUBSYSTEM_FCOE, lpfc_sli4_bsg_link_diag_test() 2358 if (alloc_len != req_len) { lpfc_sli4_bsg_link_diag_test()
|
H A D | lpfc_sli.c | 5264 uint32_t alloc_len, mbox_tmo; lpfc_sli4_cfg_post_extnts() local 5288 alloc_len = lpfc_sli4_config(phba, mbox, LPFC_MBOX_SUBSYSTEM_COMMON, lpfc_sli4_cfg_post_extnts() 5291 if (alloc_len < req_len) { lpfc_sli4_cfg_post_extnts() 5295 "size (x%x)\n", alloc_len, req_len); lpfc_sli4_cfg_post_extnts() 5955 uint32_t alloc_len, mbox_tmo; lpfc_sli4_get_allocated_extnts() local 6021 alloc_len = lpfc_sli4_config(phba, mbox, LPFC_MBOX_SUBSYSTEM_COMMON, 6024 if (alloc_len < req_len) { 6028 "size (x%x)\n", alloc_len, req_len); 15943 uint32_t alloc_len, req_len; lpfc_sli4_add_fcf_record() local 15957 alloc_len = lpfc_sli4_config(phba, mboxq, LPFC_MBOX_SUBSYSTEM_FCOE, lpfc_sli4_add_fcf_record() 15960 if (alloc_len < req_len) { lpfc_sli4_add_fcf_record() 15964 "size (x%x)\n", alloc_len, req_len); lpfc_sli4_add_fcf_record()
|
/linux-4.1.27/drivers/usb/storage/ |
H A D | sddr09.c | 1203 int numblocks, alloc_len, alloc_blocks; sddr09_read_map() local 1222 alloc_len = (alloc_blocks << CONTROL_SHIFT); sddr09_read_map() 1223 buffer = kmalloc(alloc_len, GFP_NOIO); sddr09_read_map() 1229 buffer_end = buffer + alloc_len; sddr09_read_map()
|
/linux-4.1.27/fs/ceph/ |
H A D | xattr.c | 685 BUG_ON(need > ci->i_xattrs.prealloc_blob->alloc_len); __ceph_build_xattrs_blob() 962 required_blob_size > ci->i_xattrs.prealloc_blob->alloc_len) { __ceph_setxattr() 1076 required_blob_size > ci->i_xattrs.prealloc_blob->alloc_len) { __ceph_removexattr()
|
/linux-4.1.27/drivers/scsi/aacraid/ |
H A D | aachba.c | 2514 unsigned int alloc_len; aac_scsi_cmd() local 2532 alloc_len = ((scsicmd->cmnd[10] << 24) aac_scsi_cmd() 2536 alloc_len = min_t(size_t, alloc_len, sizeof(cp)); aac_scsi_cmd() 2537 scsi_sg_copy_from_buffer(scsicmd, cp, alloc_len); aac_scsi_cmd() 2538 if (alloc_len < scsi_bufflen(scsicmd)) aac_scsi_cmd() 2540 scsi_bufflen(scsicmd) - alloc_len); aac_scsi_cmd()
|
/linux-4.1.27/drivers/crypto/ |
H A D | talitos.c | 1127 int assoc_nents = 0, src_nents, dst_nents, alloc_len, dma_len; talitos_edesc_alloc() local 1175 alloc_len = sizeof(struct talitos_edesc); talitos_edesc_alloc() 1179 alloc_len += dma_len; talitos_edesc_alloc() 1182 alloc_len += icv_stashing ? authsize : 0; talitos_edesc_alloc() 1185 edesc = kmalloc(alloc_len, GFP_DMA | flags); talitos_edesc_alloc()
|
/linux-4.1.27/net/core/ |
H A D | dev.c | 1967 int alloc_len = XPS_MIN_MAP_ALLOC; expand_xps_map() local 1978 if (pos < map->alloc_len) expand_xps_map() 1981 alloc_len = map->alloc_len * 2; expand_xps_map() 1985 new_map = kzalloc_node(XPS_MAP_SIZE(alloc_len), GFP_KERNEL, expand_xps_map() 1992 new_map->alloc_len = alloc_len; expand_xps_map()
|
/linux-4.1.27/drivers/scsi/csiostor/ |
H A D | csio_scsi.c | 564 uint32_t alloc_len = 0; csio_setup_ddp() local 612 while (alloc_len < xfer_len) { csio_setup_ddp() 618 alloc_len += dma_buf->len; csio_setup_ddp()
|
/linux-4.1.27/include/linux/ |
H A D | netdevice.h | 715 unsigned int alloc_len; member in struct:xps_map
|