new_size 51 arch/arm/kernel/vdso.c unsigned long new_size = new_vma->vm_end - new_vma->vm_start; new_size 57 arch/arm/kernel/vdso.c if (vdso_size != new_size) new_size 88 arch/arm64/kernel/vdso.c unsigned long new_size = new_vma->vm_end - new_vma->vm_start; new_size 92 arch/arm64/kernel/vdso.c if (vdso_size != new_size) new_size 742 arch/powerpc/kernel/fadump.c u64 new_size; new_size 744 arch/powerpc/kernel/fadump.c new_size = mrange_info->mem_ranges_sz + PAGE_SIZE; new_size 746 arch/powerpc/kernel/fadump.c new_size, mrange_info->name); new_size 748 arch/powerpc/kernel/fadump.c new_array = krealloc(mrange_info->mem_ranges, new_size, GFP_KERNEL); new_size 757 arch/powerpc/kernel/fadump.c mrange_info->mem_ranges_sz = new_size; new_size 758 arch/powerpc/kernel/fadump.c mrange_info->max_mem_ranges = (new_size / new_size 100 arch/s390/appldata/appldata_os.c unsigned int new_size; new_size 136 arch/s390/appldata/appldata_os.c new_size = sizeof(struct appldata_os_data) + new_size 138 arch/s390/appldata/appldata_os.c if (ops.size != new_size) { new_size 142 arch/s390/appldata/appldata_os.c (unsigned long) ops.data, new_size, new_size 156 arch/s390/appldata/appldata_os.c ops.size = new_size; new_size 370 arch/sparc/mm/tsb.c static unsigned long tsb_size_to_rss_limit(unsigned long new_size) new_size 372 arch/sparc/mm/tsb.c unsigned long num_ents = (new_size / sizeof(struct tsb)); new_size 399 arch/sparc/mm/tsb.c unsigned long new_size, old_size, flags; new_size 409 arch/sparc/mm/tsb.c for (new_size = 8192; new_size < max_tsb_size; new_size <<= 1UL) { new_size 410 arch/sparc/mm/tsb.c new_rss_limit = tsb_size_to_rss_limit(new_size); new_size 416 arch/sparc/mm/tsb.c if (new_size == max_tsb_size) new_size 421 arch/sparc/mm/tsb.c if (new_size > (PAGE_SIZE * 2)) new_size 435 arch/sparc/mm/tsb.c new_size = 8192; new_size 449 arch/sparc/mm/tsb.c tsb_init(new_tsb, new_size); new_size 509 arch/sparc/mm/tsb.c copy_tsb(old_tsb_base, old_size, new_tsb_base, new_size, new_size 515 arch/sparc/mm/tsb.c setup_tsb_params(mm, tsb_index, new_size); new_size 75 arch/x86/entry/vdso/vma.c unsigned long new_size = new_vma->vm_end - new_vma->vm_start; new_size 78 arch/x86/entry/vdso/vma.c if (image->size != new_size) new_size 246 arch/x86/platform/efi/quirks.c phys_addr_t new_phys, new_size; new_size 274 arch/x86/platform/efi/quirks.c new_size = efi.memmap.desc_size * num_entries; new_size 282 arch/x86/platform/efi/quirks.c new = early_memremap(new_phys, new_size); new_size 289 arch/x86/platform/efi/quirks.c early_memunmap(new, new_size); new_size 407 arch/x86/platform/efi/quirks.c phys_addr_t new_phys, new_size; new_size 462 arch/x86/platform/efi/quirks.c new_size = efi.memmap.desc_size * num_entries; new_size 469 arch/x86/platform/efi/quirks.c new = memremap(new_phys, new_size, MEMREMAP_WB); new_size 551 block/bio.c void bio_truncate(struct bio *bio, unsigned new_size) new_size 558 block/bio.c if (new_size >= bio->bi_iter.bi_size) new_size 565 block/bio.c if (done + bv.bv_len > new_size) { new_size 569 block/bio.c offset = new_size - done; new_size 587 block/bio.c bio->bi_iter.bi_size = new_size; new_size 508 drivers/acpi/apei/erst.c int new_size; new_size 511 drivers/acpi/apei/erst.c new_size = erst_record_id_cache.size * 2; new_size 512 drivers/acpi/apei/erst.c new_size = clamp_val(new_size, ERST_RECORD_ID_CACHE_SIZE_MIN, new_size 514 drivers/acpi/apei/erst.c if (new_size <= erst_record_id_cache.size) { new_size 519 drivers/acpi/apei/erst.c new_entries = kvmalloc_array(new_size, sizeof(entries[0]), new_size 527 drivers/acpi/apei/erst.c erst_record_id_cache.size = new_size; new_size 363 drivers/base/component.c size_t new_size = match->alloc + 16; new_size 366 drivers/base/component.c ret = component_match_realloc(master, match, new_size); new_size 4111 drivers/block/drbd/drbd_receiver.c sector_t new_size, cur_size; new_size 4147 drivers/block/drbd/drbd_receiver.c new_size = drbd_new_dev_size(device, device->ldev, p_usize, 0); new_size 4148 drivers/block/drbd/drbd_receiver.c if (new_size < cur_size && new_size 4152 drivers/block/drbd/drbd_receiver.c (unsigned long long)new_size, (unsigned long long)cur_size); new_size 4216 drivers/block/drbd/drbd_receiver.c sector_t new_size = p_csize ?: p_usize ?: p_size; new_size 4218 drivers/block/drbd/drbd_receiver.c if (new_size == 0) { new_size 4220 drivers/block/drbd/drbd_receiver.c } else if (new_size == cur_size) { new_size 4224 drivers/block/drbd/drbd_receiver.c (unsigned long long)new_size, (unsigned long long)cur_size); new_size 4225 drivers/block/drbd/drbd_receiver.c } else if (new_size < cur_size && device->state.role == R_PRIMARY) { new_size 4227 drivers/block/drbd/drbd_receiver.c (unsigned long long)new_size, (unsigned long long)cur_size); new_size 4240 drivers/block/drbd/drbd_receiver.c drbd_set_my_capacity(device, new_size); new_size 537 drivers/block/xen-blkback/blkback.c unsigned long long new_size = vbd_sz(vbd); new_size 541 drivers/block/xen-blkback/blkback.c pr_info("VBD Resize: new size %llu\n", new_size); new_size 542 drivers/block/xen-blkback/blkback.c vbd->size = new_size; new_size 197 drivers/gpu/drm/vc4/vc4_bo.c uint32_t new_size = max(vc4->bo_cache.size_list_size * 2, new_size 202 drivers/gpu/drm/vc4/vc4_bo.c new_list = kmalloc_array(new_size, sizeof(struct list_head), new_size 220 drivers/gpu/drm/vc4/vc4_bo.c for (i = vc4->bo_cache.size_list_size; i < new_size; i++) new_size 225 drivers/gpu/drm/vc4/vc4_bo.c vc4->bo_cache.size_list_size = new_size; new_size 211 drivers/gpu/drm/vc4/vc4_plane.c u32 new_size = max(4u, vc4_state->dlist_count * 2); new_size 212 drivers/gpu/drm/vc4/vc4_plane.c u32 *new_dlist = kmalloc_array(new_size, 4, GFP_KERNEL); new_size 220 drivers/gpu/drm/vc4/vc4_plane.c vc4_state->dlist_size = new_size; new_size 381 drivers/gpu/drm/vmwgfx/vmwgfx_cotable.c static int vmw_cotable_resize(struct vmw_resource *res, size_t new_size) new_size 411 drivers/gpu/drm/vmwgfx/vmwgfx_cotable.c ret = vmw_bo_init(dev_priv, buf, new_size, &vmw_mob_ne_placement, new_size 460 drivers/gpu/drm/vmwgfx/vmwgfx_cotable.c res->backup_size = new_size; new_size 508 drivers/gpu/drm/vmwgfx/vmwgfx_cotable.c size_t new_size = res->backup_size; new_size 514 drivers/gpu/drm/vmwgfx/vmwgfx_cotable.c while (needed_size > new_size) new_size 515 drivers/gpu/drm/vmwgfx/vmwgfx_cotable.c new_size *= 2; new_size 517 drivers/gpu/drm/vmwgfx/vmwgfx_cotable.c if (likely(new_size <= res->backup_size)) { new_size 527 drivers/gpu/drm/vmwgfx/vmwgfx_cotable.c return vmw_cotable_resize(res, new_size); new_size 1011 drivers/hid/hid-asus.c size_t new_size = *rsize + sizeof(asus_g752_fixed_rdesc); new_size 1013 drivers/hid/hid-asus.c new_rdesc = devm_kzalloc(&hdev->dev, new_size, GFP_KERNEL); new_size 1025 drivers/hid/hid-asus.c *rsize = new_size; new_size 131 drivers/hid/hid-core.c unsigned int new_size = parser->collection_stack_size + new_size 135 drivers/hid/hid-core.c new_size * sizeof(unsigned int), new_size 141 drivers/hid/hid-core.c parser->collection_stack_size = new_size; new_size 67 drivers/hid/hid-gembird.c size_t new_size = *rsize + delta_size; new_size 72 drivers/hid/hid-gembird.c new_rdesc = devm_kzalloc(&hdev->dev, new_size, GFP_KERNEL); new_size 90 drivers/hid/hid-gembird.c *rsize = new_size; new_size 1983 drivers/infiniband/hw/hfi1/init.c u32 new_size, i, j; new_size 2000 drivers/infiniband/hw/hfi1/init.c new_size = rcd->egrbufs.rcvtid_size / 2; new_size 2008 drivers/infiniband/hw/hfi1/init.c rcd->egrbufs.rcvtid_size = new_size; new_size 2026 drivers/infiniband/hw/hfi1/init.c new_size) == new_size 2032 drivers/infiniband/hw/hfi1/init.c offset += new_size; new_size 2035 drivers/infiniband/hw/hfi1/init.c rcd->egrbufs.rcvtid_size = new_size; new_size 2994 drivers/md/dm-cache-target.c static bool can_resize(struct cache *cache, dm_cblock_t new_size) new_size 2996 drivers/md/dm-cache-target.c if (from_cblock(new_size) > from_cblock(cache->cache_size)) { new_size 3007 drivers/md/dm-cache-target.c while (from_cblock(new_size) < from_cblock(cache->cache_size)) { new_size 3008 drivers/md/dm-cache-target.c new_size = to_cblock(from_cblock(new_size) + 1); new_size 3009 drivers/md/dm-cache-target.c if (is_dirty(cache, new_size)) { new_size 3012 drivers/md/dm-cache-target.c (unsigned long long) from_cblock(new_size)); new_size 3020 drivers/md/dm-cache-target.c static int resize_cache_dev(struct cache *cache, dm_cblock_t new_size) new_size 3024 drivers/md/dm-cache-target.c r = dm_cache_resize(cache->cmd, new_size); new_size 3031 drivers/md/dm-cache-target.c set_cache_size(cache, new_size); new_size 821 drivers/md/dm-era-target.c dm_block_t *new_size = arg; new_size 824 drivers/md/dm-era-target.c if (!valid_nr_blocks(*new_size)) { new_size 826 drivers/md/dm-era-target.c (unsigned long long) *new_size); new_size 833 drivers/md/dm-era-target.c r = writeset_alloc(&md->writesets[0], *new_size); new_size 839 drivers/md/dm-era-target.c r = writeset_alloc(&md->writesets[1], *new_size); new_size 848 drivers/md/dm-era-target.c md->nr_blocks, *new_size, new_size 855 drivers/md/dm-era-target.c md->nr_blocks = *new_size; new_size 1571 drivers/md/dm-era-target.c dm_block_t new_size = calc_nr_blocks(era); new_size 1573 drivers/md/dm-era-target.c if (era->nr_blocks != new_size) { new_size 1574 drivers/md/dm-era-target.c r = in_worker1(era, metadata_resize, &new_size); new_size 1578 drivers/md/dm-era-target.c era->nr_blocks = new_size; new_size 551 drivers/md/dm-table.c unsigned new_size; new_size 555 drivers/md/dm-table.c new_size = *size * 2; new_size 558 drivers/md/dm-table.c new_size = 8; new_size 561 drivers/md/dm-table.c argv = kmalloc_array(new_size, sizeof(*argv), gfp); new_size 564 drivers/md/dm-table.c *size = new_size; new_size 207 drivers/md/dm-thin-metadata.h int dm_pool_resize_data_dev(struct dm_pool_metadata *pmd, dm_block_t new_size); new_size 208 drivers/md/dm-thin-metadata.h int dm_pool_resize_metadata_dev(struct dm_pool_metadata *pmd, dm_block_t new_size); new_size 656 drivers/md/persistent-data/dm-array.c uint32_t old_size, uint32_t new_size, new_size 662 drivers/md/persistent-data/dm-array.c if (old_size == new_size) { new_size 675 drivers/md/persistent-data/dm-array.c resize.new_nr_full_blocks = new_size / resize.max_entries; new_size 676 drivers/md/persistent-data/dm-array.c resize.new_nr_entries_in_last_block = new_size % resize.max_entries; new_size 679 drivers/md/persistent-data/dm-array.c r = ((new_size > old_size) ? grow : shrink)(&resize); new_size 688 drivers/md/persistent-data/dm-array.c uint32_t old_size, uint32_t new_size, new_size 692 drivers/md/persistent-data/dm-array.c int r = array_resize(info, root, old_size, new_size, value, new_root); new_size 110 drivers/md/persistent-data/dm-array.h uint32_t old_size, uint32_t new_size, new_size 2158 drivers/md/raid5-cache.c int new_size = conf->min_nr_stripes * 2; new_size 2161 drivers/md/raid5-cache.c new_size); new_size 2162 drivers/md/raid5-cache.c ret = raid5_set_cache_size(mddev, new_size); new_size 2163 drivers/md/raid5-cache.c if (conf->min_nr_stripes <= new_size / 2) { new_size 2167 drivers/md/raid5-cache.c new_size, new_size 54 drivers/misc/vmw_vmci/vmci_handle_array.c size_t new_size = handle_arr_calc_size(array->capacity + new_size 60 drivers/misc/vmw_vmci/vmci_handle_array.c new_array = krealloc(array, new_size, GFP_ATOMIC); new_size 2308 drivers/mtd/nand/raw/nandsim.c uint64_t new_size = (uint64_t)nsmtd->erasesize << overridesize; new_size 2314 drivers/mtd/nand/raw/nandsim.c if (new_size >> overridesize != nsmtd->erasesize) { new_size 2321 drivers/mtd/nand/raw/nandsim.c nsmtd->size = new_size; new_size 1048 drivers/net/ethernet/cortina/gemini.c unsigned int new_size = 0; new_size 1062 drivers/net/ethernet/cortina/gemini.c new_size = 1 << (port->rxq_order + 1); new_size 1065 drivers/net/ethernet/cortina/gemini.c new_size, new_size 1069 drivers/net/ethernet/cortina/gemini.c new_size += 1 << (other_port->rxq_order + 1); new_size 1076 drivers/net/ethernet/cortina/gemini.c new_order = min(15, ilog2(new_size - 1) + 1); new_size 1078 drivers/net/ethernet/cortina/gemini.c new_size, new_order); new_size 188 drivers/net/ethernet/mellanox/mlx4/en_rx.c int new_size; new_size 201 drivers/net/ethernet/mellanox/mlx4/en_rx.c new_size = rounddown_pow_of_two(ring->actual_size); new_size 203 drivers/net/ethernet/mellanox/mlx4/en_rx.c ring->actual_size, new_size); new_size 216 drivers/net/ethernet/mellanox/mlx4/en_rx.c while (ring->actual_size > new_size) { new_size 360 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_rule.c enum mlx5dr_icm_chunk_size new_size) new_size 383 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_rule.c new_size, new_size 479 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_rule.c enum mlx5dr_icm_chunk_size new_size; new_size 481 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_rule.c new_size = mlx5dr_icm_next_higher_chunk(cur_htbl->chunk_size); new_size 482 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_rule.c new_size = min_t(u32, new_size, dmn->info.max_log_sw_icm_sz); new_size 484 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_rule.c if (new_size == cur_htbl->chunk_size) new_size 488 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_rule.c update_list, new_size); new_size 16 drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_ctcam.c u16 new_size) new_size 21 drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_ctcam.c region->key_type, new_size, region->id, new_size 561 drivers/net/usb/cdc_ncm.c static void cdc_ncm_set_dgram_size(struct usbnet *dev, int new_size) new_size 570 drivers/net/usb/cdc_ncm.c ctx->max_datagram_size = clamp_t(u32, new_size, new_size 224 drivers/net/wimax/i2400m/fw.c new_size = el_size * new_count; new_size 225 drivers/net/wimax/i2400m/fw.c void *nptr = krealloc(*ptr, new_size, gfp_flags); new_size 230 drivers/net/wimax/i2400m/fw.c memset(nptr, 0, new_size); new_size 1592 drivers/of/base.c u32 list_size, new_size; new_size 1663 drivers/of/base.c ret = of_property_read_u32(new, cells_name, &new_size); new_size 1668 drivers/of/base.c if (WARN_ON(new_size > MAX_PHANDLE_ARGS)) new_size 1670 drivers/of/base.c if (map_len < new_size) new_size 1674 drivers/of/base.c map += new_size; new_size 1675 drivers/of/base.c map_len -= new_size; new_size 1690 drivers/of/base.c match_array = map - new_size; new_size 1691 drivers/of/base.c for (i = 0; i < new_size; i++) { new_size 1692 drivers/of/base.c __be32 val = *(map - new_size + i); new_size 1701 drivers/of/base.c out_args->args_count = list_size = new_size; new_size 359 drivers/pci/setup-res.c resource_size_t new_size; new_size 374 drivers/pci/setup-res.c new_size = resource_size(res) + addsize; new_size 375 drivers/pci/setup-res.c ret = _pci_assign_resource(dev, resno, new_size, min_align); new_size 505 drivers/scsi/cxlflash/vlun.c u64 *new_size) new_size 516 drivers/scsi/cxlflash/vlun.c u64 delta = *new_size - rhte->lxt_cnt; new_size 601 drivers/scsi/cxlflash/vlun.c *new_size = my_new_size; new_size 623 drivers/scsi/cxlflash/vlun.c u64 *new_size) new_size 636 drivers/scsi/cxlflash/vlun.c u64 delta = rhte->lxt_cnt - *new_size; new_size 709 drivers/scsi/cxlflash/vlun.c *new_size = my_new_size; new_size 741 drivers/scsi/cxlflash/vlun.c u64 new_size; new_size 755 drivers/scsi/cxlflash/vlun.c new_size = DIV_ROUND_UP(nsectors, MC_CHUNK_SIZE); new_size 759 drivers/scsi/cxlflash/vlun.c new_size); new_size 789 drivers/scsi/cxlflash/vlun.c if (new_size > rhte->lxt_cnt) new_size 790 drivers/scsi/cxlflash/vlun.c rc = grow_lxt(afu, sdev, ctxid, rhndl, rhte, &new_size); new_size 791 drivers/scsi/cxlflash/vlun.c else if (new_size < rhte->lxt_cnt) new_size 792 drivers/scsi/cxlflash/vlun.c rc = shrink_lxt(afu, sdev, rhndl, rhte, ctxi, &new_size); new_size 810 drivers/scsi/cxlflash/vlun.c resize->last_lba = (new_size * MC_CHUNK_SIZE * gli->blk_len); new_size 3906 drivers/scsi/st.c static int enlarge_buffer(struct st_buffer * STbuffer, int new_size, int need_dma) new_size 3911 drivers/scsi/st.c if (new_size <= STbuffer->buffer_size) new_size 3932 drivers/scsi/st.c max_segs * (PAGE_SIZE << order) < new_size; new_size 3937 drivers/scsi/st.c if (max_segs * (PAGE_SIZE << order) < new_size) { new_size 3941 drivers/scsi/st.c return enlarge_buffer(STbuffer, new_size, need_dma); new_size 3945 drivers/scsi/st.c segs < max_segs && got < new_size;) { new_size 258 drivers/staging/comedi/comedi_buf.c unsigned long new_size) new_size 265 drivers/staging/comedi/comedi_buf.c new_size = (new_size + PAGE_SIZE - 1) & PAGE_MASK; new_size 268 drivers/staging/comedi/comedi_buf.c if (async->prealloc_buf && async->prealloc_bufsz == new_size) new_size 275 drivers/staging/comedi/comedi_buf.c if (new_size) { new_size 276 drivers/staging/comedi/comedi_buf.c unsigned int n_pages = new_size >> PAGE_SHIFT; new_size 286 drivers/staging/comedi/comedi_buf.c async->prealloc_bufsz = new_size; new_size 337 drivers/staging/comedi/comedi_fops.c unsigned int new_size) new_size 344 drivers/staging/comedi/comedi_fops.c if (new_size > async->max_bufsize) new_size 359 drivers/staging/comedi/comedi_fops.c new_size = (new_size + PAGE_SIZE - 1) & PAGE_MASK; new_size 361 drivers/staging/comedi/comedi_fops.c retval = comedi_buf_alloc(dev, s, new_size); new_size 28 drivers/staging/comedi/comedi_internal.h unsigned long new_size); new_size 1066 drivers/staging/exfat/exfat_super.c static int ffsTruncateFile(struct inode *inode, u64 old_size, u64 new_size) new_size 1081 drivers/staging/exfat/exfat_super.c new_size); new_size 1097 drivers/staging/exfat/exfat_super.c if (old_size <= new_size) { new_size 1108 drivers/staging/exfat/exfat_super.c if (new_size > 0) { new_size 1109 drivers/staging/exfat/exfat_super.c num_clusters = (s32)((new_size - 1) >> new_size 1128 drivers/staging/exfat/exfat_super.c fid->size = new_size; new_size 1130 drivers/staging/exfat/exfat_super.c if (new_size == 0) { new_size 1156 drivers/staging/exfat/exfat_super.c p_fs->fs_func->set_entry_size(ep2, new_size); new_size 1157 drivers/staging/exfat/exfat_super.c if (new_size == 0) { new_size 446 drivers/staging/most/dim2/dim2.c u16 new_size; new_size 465 drivers/staging/most/dim2/dim2.c new_size = dim_norm_ctrl_async_buffer_size(buf_size); new_size 466 drivers/staging/most/dim2/dim2.c if (new_size == 0) { new_size 470 drivers/staging/most/dim2/dim2.c ccfg->buffer_size = new_size; new_size 471 drivers/staging/most/dim2/dim2.c if (new_size != buf_size) new_size 473 drivers/staging/most/dim2/dim2.c hdm_ch->name, buf_size, new_size); new_size 476 drivers/staging/most/dim2/dim2.c is_tx ? new_size * 2 : new_size); new_size 479 drivers/staging/most/dim2/dim2.c new_size = dim_norm_ctrl_async_buffer_size(buf_size); new_size 480 drivers/staging/most/dim2/dim2.c if (new_size == 0) { new_size 484 drivers/staging/most/dim2/dim2.c ccfg->buffer_size = new_size; new_size 485 drivers/staging/most/dim2/dim2.c if (new_size != buf_size) new_size 487 drivers/staging/most/dim2/dim2.c hdm_ch->name, buf_size, new_size); new_size 490 drivers/staging/most/dim2/dim2.c is_tx ? new_size * 2 : new_size); new_size 493 drivers/staging/most/dim2/dim2.c new_size = dim_norm_isoc_buffer_size(buf_size, sub_size); new_size 494 drivers/staging/most/dim2/dim2.c if (new_size == 0) { new_size 499 drivers/staging/most/dim2/dim2.c ccfg->buffer_size = new_size; new_size 500 drivers/staging/most/dim2/dim2.c if (new_size != buf_size) new_size 502 drivers/staging/most/dim2/dim2.c hdm_ch->name, buf_size, new_size); new_size 507 drivers/staging/most/dim2/dim2.c new_size = dim_norm_sync_buffer_size(buf_size, sub_size); new_size 508 drivers/staging/most/dim2/dim2.c if (new_size == 0) { new_size 513 drivers/staging/most/dim2/dim2.c ccfg->buffer_size = new_size; new_size 514 drivers/staging/most/dim2/dim2.c if (new_size != buf_size) new_size 516 drivers/staging/most/dim2/dim2.c hdm_ch->name, buf_size, new_size); new_size 103 drivers/staging/uwb/hwa-rc.c size_t *new_size) new_size 138 drivers/staging/uwb/hwa-rc.c *new_size = sizeof(*newbe) + ielength; new_size 161 drivers/staging/uwb/hwa-rc.c size_t *new_size) new_size 205 drivers/staging/uwb/hwa-rc.c *new_size = sizeof(*newda); new_size 236 drivers/staging/uwb/hwa-rc.c size_t *new_size) new_size 269 drivers/staging/uwb/hwa-rc.c *new_size = sizeof(*newdrpev) + ielength; new_size 310 drivers/xen/evtchn.c unsigned int new_size; new_size 321 drivers/xen/evtchn.c new_size = 64; new_size 323 drivers/xen/evtchn.c new_size = 2 * u->ring_size; new_size 325 drivers/xen/evtchn.c new_ring = kvmalloc_array(new_size, sizeof(*new_ring), GFP_KERNEL); new_size 354 drivers/xen/evtchn.c u->ring_size = new_size; new_size 4580 fs/btrfs/ctree.c void btrfs_truncate_item(struct btrfs_path *path, u32 new_size, int from_end) new_size 4597 fs/btrfs/ctree.c if (old_size == new_size) new_size 4605 fs/btrfs/ctree.c size_diff = old_size - new_size; new_size 4628 fs/btrfs/ctree.c data_end, old_data_start + new_size - data_end); new_size 4665 fs/btrfs/ctree.c btrfs_set_item_size(leaf, item, new_size); new_size 2540 fs/btrfs/ctree.h void btrfs_truncate_item(struct btrfs_path *path, u32 new_size, int from_end); new_size 2837 fs/btrfs/ctree.h struct inode *inode, u64 new_size, new_size 566 fs/btrfs/file-item.c u32 new_size = (bytenr - key->offset) >> blocksize_bits; new_size 567 fs/btrfs/file-item.c new_size *= csum_size; new_size 568 fs/btrfs/file-item.c btrfs_truncate_item(path, new_size, 1); new_size 577 fs/btrfs/file-item.c u32 new_size = (csum_end - end_byte) >> blocksize_bits; new_size 578 fs/btrfs/file-item.c new_size *= csum_size; new_size 580 fs/btrfs/file-item.c btrfs_truncate_item(path, new_size, 0); new_size 1752 fs/btrfs/inode.c u64 new_size; new_size 1758 fs/btrfs/inode.c new_size = orig->end - split + 1; new_size 1759 fs/btrfs/inode.c num_extents = count_max_extents(new_size); new_size 1760 fs/btrfs/inode.c new_size = split - orig->start; new_size 1761 fs/btrfs/inode.c num_extents += count_max_extents(new_size); new_size 1779 fs/btrfs/inode.c u64 new_size, old_size; new_size 1787 fs/btrfs/inode.c new_size = new->end - other->start + 1; new_size 1789 fs/btrfs/inode.c new_size = other->end - new->start + 1; new_size 1792 fs/btrfs/inode.c if (new_size <= BTRFS_MAX_EXTENT_SIZE) { new_size 1821 fs/btrfs/inode.c if (count_max_extents(new_size) >= num_extents) new_size 4713 fs/btrfs/inode.c u64 new_size, u32 min_type) new_size 4725 fs/btrfs/inode.c u64 last_size = new_size; new_size 4737 fs/btrfs/inode.c const u64 lock_start = ALIGN_DOWN(new_size, fs_info->sectorsize); new_size 4740 fs/btrfs/inode.c BUG_ON(new_size > 0 && min_type != BTRFS_EXTENT_DATA_KEY); new_size 4766 fs/btrfs/inode.c btrfs_drop_extent_cache(BTRFS_I(inode), ALIGN(new_size, new_size 4846 fs/btrfs/inode.c if (item_end < new_size) new_size 4848 fs/btrfs/inode.c if (found_key.offset >= new_size) new_size 4864 fs/btrfs/inode.c extent_num_bytes = ALIGN(new_size - new_size 4901 fs/btrfs/inode.c u32 size = (u32)(new_size - found_key.offset); new_size 4916 fs/btrfs/inode.c inode_sub_bytes(inode, item_end + 1 - new_size); new_size 4922 fs/btrfs/inode.c last_size = new_size; new_size 5017 fs/btrfs/inode.c ASSERT(last_size >= new_size); new_size 5018 fs/btrfs/inode.c if (!ret && last_size > new_size) new_size 5019 fs/btrfs/inode.c last_size = new_size; new_size 1601 fs/btrfs/ioctl.c u64 new_size; new_size 1667 fs/btrfs/ioctl.c new_size = device->bdev->bd_inode->i_size; new_size 1676 fs/btrfs/ioctl.c new_size = memparse(sizestr, &retptr); new_size 1677 fs/btrfs/ioctl.c if (*retptr != '\0' || new_size == 0) { new_size 1691 fs/btrfs/ioctl.c if (new_size > old_size) { new_size 1695 fs/btrfs/ioctl.c new_size = old_size - new_size; new_size 1697 fs/btrfs/ioctl.c if (new_size > ULLONG_MAX - old_size) { new_size 1701 fs/btrfs/ioctl.c new_size = old_size + new_size; new_size 1704 fs/btrfs/ioctl.c if (new_size < SZ_256M) { new_size 1708 fs/btrfs/ioctl.c if (new_size > device->bdev->bd_inode->i_size) { new_size 1713 fs/btrfs/ioctl.c new_size = round_down(new_size, fs_info->sectorsize); new_size 1716 fs/btrfs/ioctl.c rcu_str_deref(device->name), new_size); new_size 1718 fs/btrfs/ioctl.c if (new_size > old_size) { new_size 1724 fs/btrfs/ioctl.c ret = btrfs_grow_device(trans, device, new_size); new_size 1726 fs/btrfs/ioctl.c } else if (new_size < old_size) { new_size 1727 fs/btrfs/ioctl.c ret = btrfs_shrink_device(device, new_size); new_size 2857 fs/btrfs/volumes.c struct btrfs_device *device, u64 new_size) new_size 2867 fs/btrfs/volumes.c new_size = round_down(new_size, fs_info->sectorsize); new_size 2871 fs/btrfs/volumes.c diff = round_down(new_size - device->total_bytes, fs_info->sectorsize); new_size 2873 fs/btrfs/volumes.c if (new_size <= device->total_bytes || new_size 2883 fs/btrfs/volumes.c btrfs_device_set_total_bytes(device, new_size); new_size 2884 fs/btrfs/volumes.c btrfs_device_set_disk_total_bytes(device, new_size); new_size 4715 fs/btrfs/volumes.c int btrfs_shrink_device(struct btrfs_device *device, u64 new_size) new_size 4736 fs/btrfs/volumes.c new_size = round_down(new_size, fs_info->sectorsize); new_size 4737 fs/btrfs/volumes.c start = new_size; new_size 4738 fs/btrfs/volumes.c diff = round_down(old_size - new_size, fs_info->sectorsize); new_size 4757 fs/btrfs/volumes.c btrfs_device_set_total_bytes(device, new_size); new_size 4815 fs/btrfs/volumes.c if (key.offset + length <= new_size) { new_size 4867 fs/btrfs/volumes.c btrfs_device_set_disk_total_bytes(device, new_size); new_size 460 fs/btrfs/volumes.h struct btrfs_device *device, u64 new_size); new_size 463 fs/btrfs/volumes.h int btrfs_shrink_device(struct btrfs_device *device, u64 new_size); new_size 4610 fs/ext4/extents.c ext4_lblk_t len, loff_t new_size, new_size 4670 fs/ext4/extents.c if (new_size) { new_size 4671 fs/ext4/extents.c if (epos > new_size) new_size 4672 fs/ext4/extents.c epos = new_size; new_size 4701 fs/ext4/extents.c loff_t new_size = 0; new_size 4756 fs/ext4/extents.c new_size = offset + len; new_size 4757 fs/ext4/extents.c ret = inode_newsize_ok(inode, new_size); new_size 4775 fs/ext4/extents.c new_size, flags); new_size 4807 fs/ext4/extents.c ret = ext4_alloc_file_blocks(file, lblk, max_blocks, new_size, new_size 4831 fs/ext4/extents.c if (new_size) { new_size 4832 fs/ext4/extents.c ext4_update_inode_size(inode, new_size); new_size 4867 fs/ext4/extents.c loff_t new_size = 0; new_size 4932 fs/ext4/extents.c new_size = offset + len; new_size 4933 fs/ext4/extents.c ret = inode_newsize_ok(inode, new_size); new_size 4941 fs/ext4/extents.c ret = ext4_alloc_file_blocks(file, lblk, max_blocks, new_size, flags); new_size 5451 fs/ext4/extents.c loff_t new_size, ioffset; new_size 5565 fs/ext4/extents.c new_size = i_size_read(inode) - len; new_size 5566 fs/ext4/extents.c i_size_write(inode, new_size); new_size 5567 fs/ext4/extents.c EXT4_I(inode)->i_disksize = new_size; new_size 1071 fs/ext4/inline.c static void ext4_update_final_de(void *de_buf, int old_size, int new_size) new_size 1087 fs/ext4/inline.c prev_de->rec_len = ext4_rec_len_to_disk(de_len + new_size - new_size 1088 fs/ext4/inline.c old_size, new_size); new_size 1092 fs/ext4/inline.c de->rec_len = ext4_rec_len_to_disk(new_size, new_size); new_size 1101 fs/ext4/inline.c int new_size = get_max_inline_xattr_value_size(dir, iloc); new_size 1103 fs/ext4/inline.c if (new_size - old_size <= EXT4_DIR_REC_LEN(1)) new_size 1107 fs/ext4/inline.c new_size + EXT4_MIN_INLINE_DATA_SIZE); new_size 124 fs/ext4/inode.c loff_t new_size) new_size 126 fs/ext4/inode.c trace_ext4_begin_ordered_truncate(inode, new_size); new_size 137 fs/ext4/inode.c new_size); new_size 1571 fs/ext4/xattr.c size_t old_size, new_size; new_size 1577 fs/ext4/xattr.c new_size = (i->value && !in_inode) ? EXT4_XATTR_SIZE(i->value_len) : 0; new_size 1583 fs/ext4/xattr.c if (new_size && new_size == old_size) { new_size 1589 fs/ext4/xattr.c memset(val, 0, new_size); new_size 1593 fs/ext4/xattr.c memset(val + i->value_len, 0, new_size - i->value_len); new_size 1622 fs/ext4/xattr.c if (free < EXT4_XATTR_LEN(name_len) + new_size) { new_size 1634 fs/ext4/xattr.c new_size && is_block && new_size 1635 fs/ext4/xattr.c (min_offs + old_size - new_size) < new_size 1751 fs/ext4/xattr.c void *val = s->base + min_offs - new_size; new_size 1753 fs/ext4/xattr.c here->e_value_offs = cpu_to_le16(min_offs - new_size); new_size 1755 fs/ext4/xattr.c memset(val, 0, new_size); new_size 1760 fs/ext4/xattr.c new_size - i->value_len); new_size 1790 fs/ext4/xattr.c new_size >> 2); new_size 1103 fs/f2fs/file.c size_t new_size; new_size 1137 fs/f2fs/file.c new_size = (loff_t)(dst + i) << PAGE_SHIFT; new_size 1138 fs/f2fs/file.c if (dst_inode->i_size < new_size) new_size 1139 fs/f2fs/file.c f2fs_i_size_write(dst_inode, new_size); new_size 1250 fs/f2fs/file.c loff_t new_size; new_size 1278 fs/f2fs/file.c new_size = i_size_read(inode) - len; new_size 1279 fs/f2fs/file.c truncate_pagecache(inode, new_size); new_size 1281 fs/f2fs/file.c ret = f2fs_truncate_blocks(inode, new_size, true); new_size 1284 fs/f2fs/file.c f2fs_i_size_write(inode, new_size); new_size 1338 fs/f2fs/file.c loff_t new_size = i_size_read(inode); new_size 1366 fs/f2fs/file.c new_size = max_t(loff_t, new_size, offset + len); new_size 1374 fs/f2fs/file.c new_size = max_t(loff_t, new_size, new_size 1417 fs/f2fs/file.c new_size = max_t(loff_t, new_size, new_size 1426 fs/f2fs/file.c new_size = max_t(loff_t, new_size, offset + len); new_size 1431 fs/f2fs/file.c if (new_size > i_size_read(inode)) { new_size 1435 fs/f2fs/file.c f2fs_i_size_write(inode, new_size); new_size 1444 fs/f2fs/file.c loff_t new_size; new_size 1447 fs/f2fs/file.c new_size = i_size_read(inode) + len; new_size 1448 fs/f2fs/file.c ret = inode_newsize_ok(inode, new_size); new_size 1509 fs/f2fs/file.c f2fs_i_size_write(inode, new_size); new_size 1521 fs/f2fs/file.c loff_t new_size = i_size_read(inode); new_size 1558 fs/f2fs/file.c new_size = (last_off == pg_end) ? offset + len : new_size 1561 fs/f2fs/file.c new_size = ((loff_t)pg_end << PAGE_SHIFT) + off_end; new_size 1564 fs/f2fs/file.c if (new_size > i_size_read(inode)) { new_size 1568 fs/f2fs/file.c f2fs_i_size_write(inode, new_size); new_size 1022 fs/gfs2/lock_dlm.c uint32_t old_size, new_size; new_size 1038 fs/gfs2/lock_dlm.c new_size = old_size; new_size 1039 fs/gfs2/lock_dlm.c while (new_size < max_jid + 1) new_size 1040 fs/gfs2/lock_dlm.c new_size += RECOVER_SIZE_INC; new_size 1041 fs/gfs2/lock_dlm.c if (new_size == old_size) new_size 1044 fs/gfs2/lock_dlm.c submit = kcalloc(new_size, sizeof(uint32_t), GFP_NOFS); new_size 1045 fs/gfs2/lock_dlm.c result = kcalloc(new_size, sizeof(uint32_t), GFP_NOFS); new_size 1059 fs/gfs2/lock_dlm.c ls->ls_recover_size = new_size; new_size 814 fs/gfs2/xattr.c u32 new_size = GFS2_EA_REC_LEN(ea) - ea_size; new_size 820 fs/gfs2/xattr.c new->ea_rec_len = cpu_to_be32(new_size); new_size 2668 fs/jbd2/transaction.c loff_t new_size) new_size 2687 fs/jbd2/transaction.c new_size, LLONG_MAX); new_size 1134 fs/jffs2/readinode.c uint32_t crc, new_size; new_size 1232 fs/jffs2/readinode.c new_size = jffs2_truncate_fragtree(c, &f->fragtree, je32_to_cpu(latest_node->isize)); new_size 1233 fs/jffs2/readinode.c if (new_size != je32_to_cpu(latest_node->isize)) { new_size 1235 fs/jffs2/readinode.c f->inocache->ino, je32_to_cpu(latest_node->isize), new_size); new_size 1236 fs/jffs2/readinode.c latest_node->isize = cpu_to_je32(new_size); new_size 471 fs/jfs/namei.c s64 new_size = 0; new_size 526 fs/jfs/namei.c if ((new_size = commitZeroLink(tid, ip)) < 0) { new_size 532 fs/jfs/namei.c rc = new_size; new_size 545 fs/jfs/namei.c if (new_size) new_size 561 fs/jfs/namei.c while (new_size && (rc == 0)) { new_size 564 fs/jfs/namei.c new_size = xtTruncate_pmap(tid, ip, new_size); new_size 565 fs/jfs/namei.c if (new_size < 0) { new_size 567 fs/jfs/namei.c rc = new_size; new_size 1079 fs/jfs/namei.c s64 new_size = 0; new_size 1193 fs/jfs/namei.c if ((new_size = commitZeroLink(tid, new_ip)) < 0) { new_size 1195 fs/jfs/namei.c rc = new_size; new_size 1289 fs/jfs/namei.c if (new_size) new_size 1305 fs/jfs/namei.c while (new_size && (rc == 0)) { new_size 1308 fs/jfs/namei.c new_size = xtTruncate_pmap(tid, new_ip, new_size); new_size 1309 fs/jfs/namei.c if (new_size < 0) { new_size 1311 fs/jfs/namei.c rc = new_size; new_size 593 fs/jfs/xattr.c int new_size) new_size 599 fs/jfs/xattr.c if (new_size == 0) { new_size 603 fs/jfs/xattr.c assert(new_size <= sizeof (ji->i_inline_ea)); new_size 606 fs/jfs/xattr.c DXDsize(&ea_buf->new_ea, new_size); new_size 610 fs/jfs/xattr.c rc = ea_write(inode, ea_buf->xattr, new_size, &ea_buf->new_ea); new_size 617 fs/jfs/xattr.c rc = ea_write(inode, ea_buf->xattr, new_size, &ea_buf->new_ea); new_size 663 fs/jfs/xattr.c int new_size; new_size 679 fs/jfs/xattr.c new_size = sizeof (struct jfs_ea_list); new_size 695 fs/jfs/xattr.c new_size += EA_SIZE(ea); new_size 710 fs/jfs/xattr.c new_size += sizeof (struct jfs_ea) + namelen + 1 + value_len; new_size 712 fs/jfs/xattr.c if (new_size > ea_buf.max_size) { new_size 718 fs/jfs/xattr.c xattr_size = ea_get(inode, &ea_buf, new_size); new_size 766 fs/jfs/xattr.c if (xattr_size != new_size) { new_size 769 fs/jfs/xattr.c xattr_size, new_size); new_size 778 fs/jfs/xattr.c if (new_size == sizeof (struct jfs_ea_list)) new_size 779 fs/jfs/xattr.c new_size = 0; new_size 781 fs/jfs/xattr.c ealist->size = cpu_to_le32(new_size); new_size 783 fs/jfs/xattr.c rc = ea_put(tid, inode, &ea_buf, new_size); new_size 123 fs/nfsd/blocklayout.c loff_t new_size = lcp->lc_last_wr + 1; new_size 133 fs/nfsd/blocklayout.c if (new_size > i_size_read(inode)) { new_size 135 fs/nfsd/blocklayout.c iattr.ia_size = new_size; new_size 1691 fs/nfsd/nfs4proc.c loff_t new_size = lcp->lc_last_wr + 1; new_size 1707 fs/nfsd/nfs4proc.c if (new_size <= seg->offset) { new_size 1711 fs/nfsd/nfs4proc.c if (new_size > seg->offset + seg->length) { new_size 1715 fs/nfsd/nfs4proc.c if (!lcp->lc_newoffset && new_size > i_size_read(inode)) { new_size 1734 fs/nfsd/nfs4proc.c if (new_size > i_size_read(inode)) { new_size 1736 fs/nfsd/nfs4proc.c lcp->lc_newsize = new_size; new_size 1421 fs/ntfs/attrib.c int ntfs_attr_record_resize(MFT_RECORD *m, ATTR_RECORD *a, u32 new_size) new_size 1423 fs/ntfs/attrib.c ntfs_debug("Entering for new_size %u.", new_size); new_size 1425 fs/ntfs/attrib.c if (new_size & 7) new_size 1426 fs/ntfs/attrib.c new_size = (new_size + 7) & ~7; new_size 1428 fs/ntfs/attrib.c if (new_size != le32_to_cpu(a->length)) { new_size 1430 fs/ntfs/attrib.c le32_to_cpu(a->length) + new_size; new_size 1435 fs/ntfs/attrib.c memmove((u8*)a + new_size, (u8*)a + le32_to_cpu(a->length), new_size 1441 fs/ntfs/attrib.c if (new_size >= offsetof(ATTR_REC, length) + sizeof(a->length)) new_size 1442 fs/ntfs/attrib.c a->length = cpu_to_le32(new_size); new_size 1466 fs/ntfs/attrib.c const u32 new_size) new_size 1472 fs/ntfs/attrib.c le16_to_cpu(a->data.resident.value_offset) + new_size)) new_size 1479 fs/ntfs/attrib.c if (new_size > old_size) new_size 1481 fs/ntfs/attrib.c old_size, 0, new_size - old_size); new_size 1483 fs/ntfs/attrib.c a->data.resident.value_length = cpu_to_le32(new_size); new_size 1523 fs/ntfs/attrib.c s64 new_size; new_size 1559 fs/ntfs/attrib.c new_size = (data_size + vol->cluster_size - 1) & new_size 1561 fs/ntfs/attrib.c if (new_size > 0) { new_size 1571 fs/ntfs/attrib.c rl = ntfs_cluster_alloc(vol, 0, new_size >> new_size 1576 fs/ntfs/attrib.c "%i.", (new_size >> new_size 1672 fs/ntfs/attrib.c a->data.non_resident.highest_vcn = cpu_to_sle64((new_size - 1) >> new_size 1677 fs/ntfs/attrib.c a->data.non_resident.allocated_size = cpu_to_sle64(new_size); new_size 1700 fs/ntfs/attrib.c ni->allocated_size = new_size; new_size 88 fs/ntfs/attrib.h extern int ntfs_attr_record_resize(MFT_RECORD *m, ATTR_RECORD *a, u32 new_size); new_size 90 fs/ntfs/attrib.h const u32 new_size); new_size 2333 fs/ntfs/inode.c s64 new_size, old_size, nr_freed, new_alloc_size, old_alloc_size; new_size 2396 fs/ntfs/inode.c new_size = i_size_read(vi); new_size 2401 fs/ntfs/inode.c new_alloc_size = (new_size + vol->cluster_size - 1) & new_size 2404 fs/ntfs/inode.c new_alloc_size = (new_size + 7) & ~7; new_size 2414 fs/ntfs/inode.c if (new_size - old_size >= 0) { new_size 2416 fs/ntfs/inode.c if (new_size == old_size) new_size 2434 fs/ntfs/inode.c err = ntfs_attr_size_bounds_check(vol, ni->type, new_size); new_size 2442 fs/ntfs/inode.c new_size > old_size ? "exceed " new_size 2471 fs/ntfs/inode.c if (new_size < vol->mft_record_size && new_size 2472 fs/ntfs/inode.c !ntfs_resident_attr_value_resize(m, a, new_size)) { new_size 2508 fs/ntfs/inode.c ni->initialized_size = new_size; new_size 2618 fs/ntfs/inode.c if (new_size < ni->initialized_size) { new_size 2619 fs/ntfs/inode.c ni->initialized_size = new_size; new_size 2621 fs/ntfs/inode.c cpu_to_sle64(new_size); new_size 2623 fs/ntfs/inode.c a->data.non_resident.data_size = cpu_to_sle64(new_size); new_size 2659 fs/ntfs/inode.c err = ntfs_attr_extend_allocation(ni, new_size, new_size 2660 fs/ntfs/inode.c size_change > 0 ? new_size : -1, -1); new_size 2768 fs/ntfs/inode.c a->data.non_resident.data_size = cpu_to_sle64(new_size); new_size 322 fs/ntfs/mft.c int new_size = (base_ni->nr_extents + 4) * sizeof(ntfs_inode *); new_size 324 fs/ntfs/mft.c tmp = kmalloc(new_size, GFP_NOFS); new_size 334 fs/ntfs/mft.c memcpy(tmp, base_ni->ext.extent_ntfs_inos, new_size - new_size 2881 fs/ntfs/mft.c int new_size = (base_ni->nr_extents + 4) * sizeof(ntfs_inode*); new_size 2883 fs/ntfs/mft.c extent_nis = kmalloc(new_size, GFP_NOFS); new_size 2894 fs/ntfs/mft.c new_size - 4 * sizeof(ntfs_inode*)); new_size 61 fs/ntfs/runlist.c int old_size, int new_size) new_size 66 fs/ntfs/runlist.c new_size = PAGE_ALIGN(new_size * sizeof(*rl)); new_size 67 fs/ntfs/runlist.c if (old_size == new_size) new_size 70 fs/ntfs/runlist.c new_rl = ntfs_malloc_nofs(new_size); new_size 75 fs/ntfs/runlist.c if (unlikely(old_size > new_size)) new_size 76 fs/ntfs/runlist.c old_size = new_size; new_size 107 fs/ntfs/runlist.c int old_size, int new_size) new_size 112 fs/ntfs/runlist.c new_size = PAGE_ALIGN(new_size * sizeof(*rl)); new_size 113 fs/ntfs/runlist.c if (old_size == new_size) new_size 116 fs/ntfs/runlist.c new_rl = ntfs_malloc_nofs_nofail(new_size); new_size 120 fs/ntfs/runlist.c if (unlikely(old_size > new_size)) new_size 121 fs/ntfs/runlist.c old_size = new_size; new_size 1555 fs/ntfs/runlist.c int new_size = rl - runlist->rl + 1; new_size 1556 fs/ntfs/runlist.c rl = ntfs_rl_realloc(runlist->rl, old_size, new_size); new_size 1542 fs/ocfs2/aops.c int ocfs2_size_fits_inline_data(struct buffer_head *di_bh, u64 new_size) new_size 1546 fs/ocfs2/aops.c if (new_size <= le16_to_cpu(di->id2.i_data.id_count)) new_size 48 fs/ocfs2/aops.h int ocfs2_size_fits_inline_data(struct buffer_head *di_bh, u64 new_size); new_size 2723 fs/ocfs2/dir.c unsigned int new_size = sb->s_blocksize; new_size 2728 fs/ocfs2/dir.c new_size = ocfs2_dir_trailer_blk_off(sb); new_size 2730 fs/ocfs2/dir.c bytes = new_size - old_size; new_size 586 fs/ocfs2/journal.h loff_t new_size) new_size 591 fs/ocfs2/journal.h new_size); new_size 183 fs/reiserfs/objectid.c int new_size = (s->s_blocksize - SB_SIZE) / sizeof(__u32) / 2 * 2; new_size 193 fs/reiserfs/objectid.c if (cur_size > new_size) { new_size 198 fs/reiserfs/objectid.c objectid_map[new_size - 1] = objectid_map[cur_size - 1]; new_size 199 fs/reiserfs/objectid.c set_sb_oid_cursize(disk_sb, new_size); new_size 202 fs/reiserfs/objectid.c for (i = new_size - 1; i >= 0; i--) { new_size 203 fs/reiserfs/objectid.c objectid_map[i + (old_max - new_size)] = objectid_map[i]; new_size 207 fs/reiserfs/objectid.c set_sb_oid_maxsize(disk_sb, new_size); new_size 529 fs/reiserfs/xattr.c size_t new_size; new_size 596 fs/reiserfs/xattr.c new_size = buffer_size + sizeof(struct reiserfs_xattr_header); new_size 597 fs/reiserfs/xattr.c if (!err && new_size < i_size_read(d_inode(dentry))) { new_size 600 fs/reiserfs/xattr.c .ia_size = new_size, new_size 492 fs/ubifs/debug.c (unsigned long long)le64_to_cpu(trun->new_size)); new_size 1113 fs/ubifs/file.c loff_t old_size = inode->i_size, new_size = attr->ia_size; new_size 1114 fs/ubifs/file.c int offset = new_size & (UBIFS_BLOCK_SIZE - 1), budgeted = 1; new_size 1117 fs/ubifs/file.c dbg_gen("ino %lu, size %lld -> %lld", inode->i_ino, old_size, new_size); new_size 1125 fs/ubifs/file.c if (new_size & (UBIFS_BLOCK_SIZE - 1)) new_size 1137 fs/ubifs/file.c if (new_size || err != -ENOSPC) new_size 1142 fs/ubifs/file.c truncate_setsize(inode, new_size); new_size 1145 fs/ubifs/file.c pgoff_t index = new_size >> PAGE_SHIFT; new_size 1163 fs/ubifs/file.c offset = new_size & new_size 1191 fs/ubifs/file.c err = ubifs_jnl_truncate(c, inode, old_size, new_size); new_size 1218 fs/ubifs/file.c loff_t new_size = attr->ia_size; new_size 1228 fs/ubifs/file.c dbg_gen("size %lld -> %lld", inode->i_size, new_size); new_size 1229 fs/ubifs/file.c truncate_setsize(inode, new_size); new_size 1504 fs/ubifs/journal.c loff_t old_size, loff_t new_size) new_size 1518 fs/ubifs/journal.c (unsigned long)inum, old_size, new_size); new_size 1536 fs/ubifs/journal.c trun->new_size = cpu_to_le64(new_size); new_size 1539 fs/ubifs/journal.c dlen = new_size & (UBIFS_BLOCK_SIZE - 1); new_size 1543 fs/ubifs/journal.c blk = new_size >> UBIFS_BLOCK_SHIFT; new_size 1621 fs/ubifs/journal.c bit = new_size & (UBIFS_BLOCK_SIZE - 1); new_size 1622 fs/ubifs/journal.c blk = (new_size >> UBIFS_BLOCK_SHIFT) + (bit ? 1 : 0); new_size 1355 fs/ubifs/recovery.c int deletion, loff_t new_size) new_size 1368 fs/ubifs/recovery.c e->i_size = new_size; new_size 1371 fs/ubifs/recovery.c err = add_ino(c, inum, new_size, 0, 1); new_size 1380 fs/ubifs/recovery.c if (new_size > e->d_size) new_size 1381 fs/ubifs/recovery.c e->d_size = new_size; new_size 1383 fs/ubifs/recovery.c err = add_ino(c, inum, 0, new_size, 0); new_size 1391 fs/ubifs/recovery.c e->d_size = new_size; new_size 58 fs/ubifs/replay.c loff_t new_size; new_size 187 fs/ubifs/replay.c min_blk = r->new_size / UBIFS_BLOCK_SIZE; new_size 188 fs/ubifs/replay.c if (r->new_size & (UBIFS_BLOCK_SIZE - 1)) new_size 285 fs/ubifs/replay.c r->new_size); new_size 385 fs/ubifs/replay.c loff_t old_size, loff_t new_size) new_size 408 fs/ubifs/replay.c r->new_size = new_size; new_size 746 fs/ubifs/replay.c loff_t new_size = le64_to_cpu(ino->size); new_size 752 fs/ubifs/replay.c &used, 0, new_size); new_size 758 fs/ubifs/replay.c loff_t new_size = le32_to_cpu(dn->size) + new_size 764 fs/ubifs/replay.c &used, 0, new_size); new_size 786 fs/ubifs/replay.c loff_t new_size = le64_to_cpu(trun->new_size); new_size 791 fs/ubifs/replay.c new_size < 0 || new_size > c->max_inode_sz || new_size 792 fs/ubifs/replay.c old_size <= new_size) { new_size 804 fs/ubifs/replay.c old_size, new_size); new_size 603 fs/ubifs/ubifs-media.h __le64 new_size; new_size 1784 fs/ubifs/ubifs.h loff_t old_size, loff_t new_size); new_size 2047 fs/ubifs/ubifs.h int deletion, loff_t new_size); new_size 1613 fs/udf/super.c unsigned int new_size = ALIGN(partnum, PART_DESC_ALLOC_STEP); new_size 1615 fs/udf/super.c new_loc = kcalloc(new_size, sizeof(*new_loc), GFP_KERNEL); new_size 1622 fs/udf/super.c data->size_part_descs = new_size; new_size 229 fs/ufs/inode.c unsigned new_size; new_size 234 fs/ufs/inode.c new_size = (writes_to & uspi->s_fpbmask) + 1; new_size 236 fs/ufs/inode.c new_size = uspi->s_fpb; new_size 240 fs/ufs/inode.c new_size - (lastfrag & uspi->s_fpbmask), err, new_size 2822 fs/xfs/libxfs/xfs_ialloc.c int new_size = igeo->inode_cluster_size_raw; new_size 2824 fs/xfs/libxfs/xfs_ialloc.c new_size *= mp->m_sb.sb_inodesize / XFS_DINODE_MIN_SIZE; new_size 2825 fs/xfs/libxfs/xfs_ialloc.c if (mp->m_sb.sb_inoalignmt >= XFS_B_TO_FSBT(mp, new_size)) new_size 2826 fs/xfs/libxfs/xfs_ialloc.c igeo->inode_cluster_size_raw = new_size; new_size 599 fs/xfs/libxfs/xfs_iext_tree.c size_t new_size = ifp->if_bytes + sizeof(struct xfs_iext_rec); new_size 603 fs/xfs/libxfs/xfs_iext_tree.c if (new_size / sizeof(struct xfs_iext_rec) == RECS_PER_LEAF) new_size 604 fs/xfs/libxfs/xfs_iext_tree.c new_size = NODE_SIZE; new_size 606 fs/xfs/libxfs/xfs_iext_tree.c new = kmem_realloc(ifp->if_u1.if_root, new_size, KM_NOFS); new_size 607 fs/xfs/libxfs/xfs_iext_tree.c memset(new + ifp->if_bytes, 0, new_size - ifp->if_bytes); new_size 351 fs/xfs/libxfs/xfs_inode_fork.c size_t new_size; new_size 369 fs/xfs/libxfs/xfs_inode_fork.c new_size = XFS_BMAP_BROOT_SPACE_CALC(mp, rec_diff); new_size 370 fs/xfs/libxfs/xfs_inode_fork.c ifp->if_broot = kmem_alloc(new_size, KM_NOFS); new_size 371 fs/xfs/libxfs/xfs_inode_fork.c ifp->if_broot_bytes = (int)new_size; new_size 383 fs/xfs/libxfs/xfs_inode_fork.c new_size = XFS_BMAP_BROOT_SPACE_CALC(mp, new_max); new_size 384 fs/xfs/libxfs/xfs_inode_fork.c ifp->if_broot = kmem_realloc(ifp->if_broot, new_size, new_size 389 fs/xfs/libxfs/xfs_inode_fork.c (int)new_size); new_size 390 fs/xfs/libxfs/xfs_inode_fork.c ifp->if_broot_bytes = (int)new_size; new_size 407 fs/xfs/libxfs/xfs_inode_fork.c new_size = XFS_BMAP_BROOT_SPACE_CALC(mp, new_max); new_size 409 fs/xfs/libxfs/xfs_inode_fork.c new_size = 0; new_size 410 fs/xfs/libxfs/xfs_inode_fork.c if (new_size > 0) { new_size 411 fs/xfs/libxfs/xfs_inode_fork.c new_broot = kmem_alloc(new_size, KM_NOFS); new_size 439 fs/xfs/libxfs/xfs_inode_fork.c (int)new_size); new_size 444 fs/xfs/libxfs/xfs_inode_fork.c ifp->if_broot_bytes = (int)new_size; new_size 474 fs/xfs/libxfs/xfs_inode_fork.c int new_size = (int)ifp->if_bytes + byte_diff; new_size 476 fs/xfs/libxfs/xfs_inode_fork.c ASSERT(new_size >= 0); new_size 477 fs/xfs/libxfs/xfs_inode_fork.c ASSERT(new_size <= XFS_IFORK_SIZE(ip, whichfork)); new_size 482 fs/xfs/libxfs/xfs_inode_fork.c if (new_size == 0) { new_size 495 fs/xfs/libxfs/xfs_inode_fork.c roundup(new_size, 4), KM_NOFS); new_size 496 fs/xfs/libxfs/xfs_inode_fork.c ifp->if_bytes = new_size; new_size 808 fs/xfs/xfs_file.c loff_t new_size = 0; new_size 842 fs/xfs/xfs_file.c new_size = i_size_read(inode) - len; new_size 864 fs/xfs/xfs_file.c new_size = isize + len; new_size 877 fs/xfs/xfs_file.c new_size = offset + len; new_size 878 fs/xfs/xfs_file.c error = inode_newsize_ok(inode, new_size); new_size 919 fs/xfs/xfs_file.c if (new_size) { new_size 923 fs/xfs/xfs_file.c iattr.ia_size = new_size; new_size 1510 fs/xfs/xfs_inode.c xfs_fsize_t new_size, new_size 1524 fs/xfs/xfs_inode.c ASSERT(new_size <= XFS_ISIZE(ip)); new_size 1530 fs/xfs/xfs_inode.c trace_xfs_itruncate_extents_start(ip, new_size); new_size 1543 fs/xfs/xfs_inode.c first_unmap_block = XFS_B_TO_FSB(mp, (xfs_ufsize_t)new_size); new_size 1586 fs/xfs/xfs_inode.c trace_xfs_itruncate_extents_end(ip, new_size); new_size 104 fs/xfs/xfs_inode.h xfs_new_eof(struct xfs_inode *ip, xfs_fsize_t new_size) new_size 108 fs/xfs/xfs_inode.h if (new_size > i_size || new_size < 0) new_size 109 fs/xfs/xfs_inode.h new_size = i_size; new_size 110 fs/xfs/xfs_inode.h return new_size > ip->i_d.di_size ? new_size : 0; new_size 463 fs/xfs/xfs_inode.h xfs_fsize_t new_size) new_size 465 fs/xfs/xfs_inode.h return xfs_itruncate_extents_flags(tpp, ip, whichfork, new_size, 0); new_size 1317 fs/xfs/xfs_trace.h TP_PROTO(struct xfs_inode *ip, xfs_fsize_t new_size), new_size 1318 fs/xfs/xfs_trace.h TP_ARGS(ip, new_size), new_size 1323 fs/xfs/xfs_trace.h __field(xfs_fsize_t, new_size) new_size 1329 fs/xfs/xfs_trace.h __entry->new_size = new_size; new_size 1335 fs/xfs/xfs_trace.h __entry->new_size) new_size 1340 fs/xfs/xfs_trace.h TP_PROTO(struct xfs_inode *ip, xfs_fsize_t new_size), \ new_size 1341 fs/xfs/xfs_trace.h TP_ARGS(ip, new_size)) new_size 473 include/linux/bio.h void bio_truncate(struct bio *bio, unsigned new_size); new_size 270 include/linux/ceph/osdmap.h int new_size, new_size 1420 include/linux/jbd2.h struct jbd2_inode *inode, loff_t new_size); new_size 61 include/linux/kasan.h void * __must_check kasan_krealloc(const void *object, size_t new_size, new_size 128 include/linux/kasan.h static inline void *kasan_krealloc(const void *object, size_t new_size, new_size 340 include/linux/kexec.h int crash_shrink_memory(unsigned long new_size); new_size 75 include/linux/livepatch.h unsigned long old_size, new_size; new_size 1175 include/net/sctp/structs.h int new_size, __u8 addr_state, gfp_t gfp); new_size 272 include/trace/events/ext4.h TP_PROTO(struct inode *inode, loff_t new_size), new_size 274 include/trace/events/ext4.h TP_ARGS(inode, new_size), new_size 279 include/trace/events/ext4.h __field( loff_t, new_size ) new_size 285 include/trace/events/ext4.h __entry->new_size = new_size; new_size 291 include/trace/events/ext4.h __entry->new_size) new_size 829 kernel/bpf/btf.c u32 expand_by, new_size; new_size 837 kernel/bpf/btf.c new_size = min_t(u32, BTF_MAX_TYPE, new_size 840 kernel/bpf/btf.c new_types = kvcalloc(new_size, sizeof(*new_types), new_size 853 kernel/bpf/btf.c btf->types_size = new_size; new_size 1006 kernel/kexec_core.c int crash_shrink_memory(unsigned long new_size) new_size 1022 kernel/kexec_core.c if (new_size >= old_size) { new_size 1023 kernel/kexec_core.c ret = (new_size == old_size) ? 0 : -EINVAL; new_size 1034 kernel/kexec_core.c end = roundup(start + new_size, KEXEC_CRASH_MEM_ALIGN); new_size 756 kernel/livepatch/core.c &func->new_size, NULL); new_size 209 kernel/livepatch/transition.c func_size = func->new_size; new_size 227 kernel/livepatch/transition.c func_size = prev->new_size; new_size 402 mm/list_lru.c int old_size, int new_size) new_size 406 mm/list_lru.c BUG_ON(old_size > new_size); new_size 410 mm/list_lru.c new = kvmalloc(sizeof(*new) + new_size * sizeof(void *), GFP_KERNEL); new_size 414 mm/list_lru.c if (__memcg_init_list_lru_node(new, old_size, new_size)) { new_size 437 mm/list_lru.c int old_size, int new_size) new_size 445 mm/list_lru.c __memcg_destroy_list_lru_node(memcg_lrus, old_size, new_size); new_size 483 mm/list_lru.c int old_size, int new_size) new_size 492 mm/list_lru.c old_size, new_size)) new_size 502 mm/list_lru.c old_size, new_size); new_size 508 mm/list_lru.c int old_size, int new_size) new_size 517 mm/list_lru.c old_size, new_size); new_size 520 mm/list_lru.c int memcg_update_all_list_lrus(int new_size) new_size 528 mm/list_lru.c ret = memcg_update_list_lru(lru, old_size, new_size); new_size 537 mm/list_lru.c memcg_cancel_update_list_lru(lru, old_size, new_size); new_size 419 mm/memblock.c phys_addr_t old_size, new_size, addr, new_end; new_size 431 mm/memblock.c new_size = old_size << 1; new_size 437 mm/memblock.c new_alloc_size = PAGE_ALIGN(new_size); new_size 447 mm/memblock.c new_array = kmalloc(new_size, GFP_KERNEL); new_size 470 mm/memblock.c new_end = addr + new_size - 1; new_size 1651 mm/slab_common.c static __always_inline void *__do_krealloc(const void *p, size_t new_size, new_size 1660 mm/slab_common.c if (ks >= new_size) { new_size 1661 mm/slab_common.c p = kasan_krealloc((void *)p, new_size, flags); new_size 1665 mm/slab_common.c ret = kmalloc_track_caller(new_size, flags); new_size 1684 mm/slab_common.c void *__krealloc(const void *p, size_t new_size, gfp_t flags) new_size 1686 mm/slab_common.c if (unlikely(!new_size)) new_size 1689 mm/slab_common.c return __do_krealloc(p, new_size, flags); new_size 1707 mm/slab_common.c void *krealloc(const void *p, size_t new_size, gfp_t flags) new_size 1711 mm/slab_common.c if (unlikely(!new_size)) { new_size 1716 mm/slab_common.c ret = __do_krealloc(p, new_size, flags); new_size 121 net/batman-adv/tp_meter.c u32 new_size = base + increment; new_size 124 net/batman-adv/tp_meter.c if (new_size < base) new_size 125 net/batman-adv/tp_meter.c new_size = (u32)ULONG_MAX; new_size 127 net/batman-adv/tp_meter.c new_size = min_t(u32, new_size, BATADV_TP_AWND); new_size 129 net/batman-adv/tp_meter.c return max_t(u32, new_size, min); new_size 2109 net/ceph/osdmap.c int new_size, new_size 2122 net/ceph/osdmap.c old_size != new_size || new_size 1230 net/ipv4/fib_semantics.c unsigned int new_size) new_size 1239 net/ipv4/fib_semantics.c fib_info_hash_size = new_size; new_size 1389 net/ipv4/fib_semantics.c unsigned int new_size = fib_info_hash_size << 1; new_size 1394 net/ipv4/fib_semantics.c if (!new_size) new_size 1395 net/ipv4/fib_semantics.c new_size = 16; new_size 1396 net/ipv4/fib_semantics.c bytes = new_size * sizeof(struct hlist_head *); new_size 1403 net/ipv4/fib_semantics.c fib_info_hash_move(new_info_hash, new_laddrhash, new_size); new_size 141 net/openvswitch/flow_table.c static struct table_instance *table_instance_alloc(int new_size) new_size 149 net/openvswitch/flow_table.c ti->buckets = kvmalloc_array(new_size, sizeof(struct hlist_head), new_size 156 net/openvswitch/flow_table.c for (i = 0; i < new_size; i++) new_size 159 net/openvswitch/flow_table.c ti->n_buckets = new_size; new_size 139 net/sctp/bind_addr.c int new_size, __u8 addr_state, gfp_t gfp) new_size 148 net/sctp/bind_addr.c memcpy(&addr->a, new, min_t(size_t, sizeof(*new), new_size)); new_size 299 scripts/kconfig/confdata.c size_t new_size = slen + 1; new_size 300 scripts/kconfig/confdata.c if (new_size > *n) { new_size 301 scripts/kconfig/confdata.c new_size += LINE_GROWTH - 1; new_size 302 scripts/kconfig/confdata.c new_size *= 2; new_size 303 scripts/kconfig/confdata.c nline = xrealloc(*lineptr, new_size); new_size 308 scripts/kconfig/confdata.c *n = new_size; new_size 48 tools/lib/bpf/btf.c __u32 expand_by, new_size; new_size 54 tools/lib/bpf/btf.c new_size = min(BTF_MAX_NR_TYPES, btf->types_size + expand_by); new_size 56 tools/lib/bpf/btf.c new_types = realloc(btf->types, sizeof(*new_types) * new_size); new_size 64 tools/lib/bpf/btf.c btf->types_size = new_size; new_size 103 tools/perf/util/header.c size_t new_size = ff->size; new_size 109 tools/perf/util/header.c while (size > (new_size - ff->offset)) new_size 110 tools/perf/util/header.c new_size <<= 1; new_size 111 tools/perf/util/header.c new_size = min(max_size, new_size); new_size 113 tools/perf/util/header.c if (ff->size < new_size) { new_size 114 tools/perf/util/header.c addr = realloc(ff->buf, new_size); new_size 118 tools/perf/util/header.c ff->size = new_size; new_size 700 tools/testing/selftests/x86/protection_keys.c size_t new_size = new_nr_records * sizeof(struct pkey_malloc_record); new_size 702 tools/testing/selftests/x86/protection_keys.c dprintf2("new_size: %zd\n", new_size); new_size 703 tools/testing/selftests/x86/protection_keys.c pkey_malloc_records = realloc(pkey_malloc_records, new_size);