load 202 arch/alpha/boot/bootp.c load(initrd_start, KERNEL_ORIGIN+KERNEL_SIZE, INITRD_IMAGE_SIZE); load 204 arch/alpha/boot/bootp.c load(START_ADDR+(4*KERNEL_SIZE), KERNEL_ORIGIN, KERNEL_SIZE); load 205 arch/alpha/boot/bootp.c load(START_ADDR, START_ADDR+(4*KERNEL_SIZE), KERNEL_SIZE); load 172 arch/alpha/boot/main.c i = load(dev, START_ADDR, KERNEL_SIZE); load 554 arch/arm/mach-omap2/vc.c u8 load; load 564 arch/arm/mach-omap2/vc.c .load = 50, load 573 arch/arm/mach-omap2/vc.c .load = 25, load 582 arch/arm/mach-omap2/vc.c .load = 12, load 591 arch/arm/mach-omap2/vc.c .load = 0, load 634 arch/arm/mach-omap2/vc.c while (i2c_data->load > capacitance) load 393 arch/arm/mm/alignment.c int load; load 398 arch/arm/mm/alignment.c load = !!(LDST_L_BIT(instr)); load 402 arch/arm/mm/alignment.c load = ((instr & 0xf0) == 0xd0); load 411 arch/arm/mm/alignment.c if (load) { load 425 arch/arm/mm/alignment.c if (load) { load 126 arch/arm64/kernel/kexec_image.c .load = image_load, load 13 arch/mips/fw/arc/salone.c return ARC_CALL4(load, Path, TopAddr, ExecAddr, LowAddr); load 22 arch/mips/include/asm/ftrace.h #define safe_load(load, src, dst, error) \ load 25 arch/mips/include/asm/ftrace.h "1: " load " %[tmp_dst], 0(%[tmp_src])\n" \ load 36 arch/mips/include/asm/mach-cavium-octeon/kernel-entry-init.h # Disable unaligned load/store support but leave HW fixup enabled load 49 arch/mips/include/asm/mach-ip27/kernel-entry-init.h or t1, t1, t0 # Physical load address of kernel text load 50 arch/mips/include/asm/mach-ip27/kernel-entry-init.h or t2, t2, t0 # Physical load address of kernel data load 188 arch/mips/include/asm/sgiarcs.h LONG load; /* Load an executable image. */ load 53 arch/mips/kernel/cmpxchg.c u32 mask, old32, new32, load32, load; load 88 arch/mips/kernel/cmpxchg.c load = (load32 & mask) >> shift; load 89 arch/mips/kernel/cmpxchg.c if (load != old) load 90 arch/mips/kernel/cmpxchg.c return load; load 198 arch/nds32/mm/alignment.c int imm, regular, load, len, addr_mode, idx_mode; load 206 arch/nds32/mm/alignment.c load = 1; load 214 arch/nds32/mm/alignment.c load = 1; load 222 arch/nds32/mm/alignment.c load = 1; load 230 arch/nds32/mm/alignment.c load = 1; load 238 arch/nds32/mm/alignment.c load = 0; load 246 arch/nds32/mm/alignment.c load = 0; load 254 arch/nds32/mm/alignment.c load = 0; load 262 arch/nds32/mm/alignment.c load = 0; load 291 arch/nds32/mm/alignment.c if (load) { load 315 arch/nds32/mm/alignment.c int imm, regular, load, len, sign_ext; load 325 arch/nds32/mm/alignment.c load = 1; load 332 arch/nds32/mm/alignment.c load = 1; load 339 arch/nds32/mm/alignment.c load = 1; load 346 arch/nds32/mm/alignment.c load = 1; load 353 arch/nds32/mm/alignment.c load = 1; load 360 arch/nds32/mm/alignment.c load = 1; load 367 arch/nds32/mm/alignment.c load = 0; load 374 arch/nds32/mm/alignment.c load = 0; load 381 arch/nds32/mm/alignment.c load = 0; load 388 arch/nds32/mm/alignment.c load = 0; load 399 arch/nds32/mm/alignment.c load = 1; load 406 arch/nds32/mm/alignment.c load = 1; load 413 arch/nds32/mm/alignment.c load = 1; load 420 arch/nds32/mm/alignment.c load = 1; load 427 arch/nds32/mm/alignment.c load = 1; load 434 arch/nds32/mm/alignment.c load = 1; load 441 arch/nds32/mm/alignment.c load = 0; load 448 arch/nds32/mm/alignment.c load = 0; load 455 arch/nds32/mm/alignment.c load = 0; load 462 arch/nds32/mm/alignment.c load = 0; load 480 arch/nds32/mm/alignment.c if (load) { load 80 arch/parisc/kernel/kexec_file.c .load = elf_load, load 286 arch/powerpc/include/asm/fsl_hcalls.h uint32_t entry_point, int load) load 296 arch/powerpc/include/asm/fsl_hcalls.h r5 = load; load 124 arch/powerpc/kernel/kexec_elf_64.c .load = elf64_load, load 133 arch/powerpc/perf/power8-pmu.c CACHE_EVENT_ATTR(L1-dcache-load-misses, PM_LD_MISS_L1); load 138 arch/powerpc/perf/power8-pmu.c CACHE_EVENT_ATTR(L1-icache-load-misses, PM_L1_ICACHE_MISS); load 142 arch/powerpc/perf/power8-pmu.c CACHE_EVENT_ATTR(LLC-load-misses, PM_DATA_FROM_L3MISS); load 148 arch/powerpc/perf/power8-pmu.c CACHE_EVENT_ATTR(branch-load-misses, PM_BR_MPRED_CMPL); load 150 arch/powerpc/perf/power8-pmu.c CACHE_EVENT_ATTR(dTLB-load-misses, PM_DTLB_MISS); load 151 arch/powerpc/perf/power8-pmu.c CACHE_EVENT_ATTR(iTLB-load-misses, PM_ITLB_MISS); load 163 arch/powerpc/perf/power9-pmu.c CACHE_EVENT_ATTR(L1-dcache-load-misses, PM_LD_MISS_L1_FIN); load 167 arch/powerpc/perf/power9-pmu.c CACHE_EVENT_ATTR(L1-icache-load-misses, PM_L1_ICACHE_MISS); load 170 arch/powerpc/perf/power9-pmu.c CACHE_EVENT_ATTR(LLC-load-misses, PM_DATA_FROM_L3MISS); load 173 arch/powerpc/perf/power9-pmu.c CACHE_EVENT_ATTR(branch-load-misses, PM_BR_MPRED_CMPL); load 175 arch/powerpc/perf/power9-pmu.c CACHE_EVENT_ATTR(dTLB-load-misses, PM_DTLB_MISS); load 176 arch/powerpc/perf/power9-pmu.c CACHE_EVENT_ATTR(iTLB-load-misses, PM_ITLB_MISS); load 132 arch/s390/kernel/kexec_elf.c .load = s390_elf_load, load 61 arch/s390/kernel/kexec_image.c .load = s390_image_load, load 137 arch/sparc/include/asm/cmpxchg_64.h unsigned int old32, new32, load; load 146 arch/sparc/include/asm/cmpxchg_64.h load = (load32 & mask) >> bit_shift; load 147 arch/sparc/include/asm/cmpxchg_64.h } while (load == old); load 149 arch/sparc/include/asm/cmpxchg_64.h return load; load 38 arch/sparc/kernel/unaligned_32.c return load; load 255 arch/sparc/kernel/unaligned_32.c case load: load 340 arch/sparc/kernel/unaligned_32.c case load: load 47 arch/sparc/kernel/unaligned_64.c return load; load 353 arch/sparc/kernel/unaligned_64.c case load: load 561 arch/x86/kernel/kexec-bzimage64.c .load = bzImage64_load, load 475 arch/x86/kernel/machine_kexec_64.c if (!image->fops || !image->fops->load) load 478 arch/x86/kernel/machine_kexec_64.c return image->fops->load(image, image->kernel_buf, load 1659 arch/x86/kvm/emulate.c goto load; load 1668 arch/x86/kvm/emulate.c goto load; load 1699 arch/x86/kvm/emulate.c goto load; load 1802 arch/x86/kvm/emulate.c load: load 90 drivers/ata/libata-zpodd.c if (desc->mech_type == 0 && desc->load == 0 && desc->eject == 1) { load 93 drivers/ata/libata-zpodd.c } else if (desc->mech_type == 1 && desc->load == 0 && load 37 drivers/clk/mvebu/dove-divider.c static void dove_load_divider(void __iomem *base, u32 val, u32 mask, u32 load) load 44 drivers/clk/mvebu/dove-divider.c v = (readl_relaxed(base + DIV_CTRL0) & ~(mask | load)) | val; load 46 drivers/clk/mvebu/dove-divider.c writel_relaxed(v | load, base + DIV_CTRL0); load 134 drivers/clk/mvebu/dove-divider.c u32 mask, load, div; load 146 drivers/clk/mvebu/dove-divider.c load = BIT(dc->div_bit_load); load 149 drivers/clk/mvebu/dove-divider.c dove_load_divider(dc->base, div, mask, load); load 564 drivers/clocksource/timer-ti-dm.c unsigned int load) load 578 drivers/clocksource/timer-ti-dm.c omap_dm_timer_write_reg(timer, OMAP_TIMER_LOAD_REG, load); load 583 drivers/clocksource/timer-ti-dm.c timer->context.tldr = load; load 65 drivers/cpufreq/cpufreq_conservative.c unsigned int load = dbs_update(policy); load 103 drivers/cpufreq/cpufreq_conservative.c if (load > dbs_data->up_threshold) { load 125 drivers/cpufreq/cpufreq_conservative.c if (load < cs_tuners->down_threshold) { load 141 drivers/cpufreq/cpufreq_governor.c unsigned int load; load 164 drivers/cpufreq/cpufreq_governor.c load = j_cdbs->prev_load; load 189 drivers/cpufreq/cpufreq_governor.c load = j_cdbs->prev_load; load 193 drivers/cpufreq/cpufreq_governor.c load = 100 * (time_elapsed - idle_time) / time_elapsed; load 210 drivers/cpufreq/cpufreq_governor.c load = (int)idle_time < 0 ? 100 : 0; load 212 drivers/cpufreq/cpufreq_governor.c j_cdbs->prev_load = load; load 222 drivers/cpufreq/cpufreq_governor.c if (load > max_load) load 223 drivers/cpufreq/cpufreq_governor.c max_load = load; load 140 drivers/cpufreq/cpufreq_ondemand.c unsigned int load = dbs_update(policy); load 145 drivers/cpufreq/cpufreq_ondemand.c if (load > dbs_data->up_threshold) { load 156 drivers/cpufreq/cpufreq_ondemand.c freq_next = min_f + load * (max_f - min_f) / 100; load 288 drivers/crypto/caam/desc_constr.h APPEND_CMD_PTR(load, LOAD) load 331 drivers/crypto/caam/desc_constr.h APPEND_CMD_PTR_TO_IMM(load, LOAD); load 388 drivers/crypto/caam/desc_constr.h APPEND_CMD_RAW_IMM(load, LOAD, u32); load 405 drivers/crypto/caam/desc_constr.h APPEND_CMD_RAW_IMM2(load, LOAD, be, 32); load 552 drivers/crypto/marvell/cesa.c atomic_set(&engine->load, 0); load 457 drivers/crypto/marvell/cesa.h atomic_t load; load 736 drivers/crypto/marvell/cesa.h u32 load = atomic_read(&engine->load); load 737 drivers/crypto/marvell/cesa.h if (load < min_load) { load 738 drivers/crypto/marvell/cesa.h min_load = load; load 743 drivers/crypto/marvell/cesa.h atomic_add(weight, &selected->load); load 205 drivers/crypto/marvell/cipher.c atomic_sub(skreq->cryptlen, &engine->load); load 367 drivers/crypto/marvell/hash.c atomic_sub(ahashreq->nbytes, &engine->load); load 1427 drivers/gpu/drm/amd/amdgpu/amdgpu_drv.c .load = amdgpu_driver_load_kms, load 212 drivers/gpu/drm/ast/ast_drv.c .load = ast_driver_load, load 489 drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c unsigned int pixels, load = 0; load 503 drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c load += pixels * plane_state->bpp[i]; load 510 drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c ahb_load[plane_state->ahb_id] += load; load 981 drivers/gpu/drm/drm_drv.c if (dev->driver->load) { load 982 drivers/gpu/drm/drm_drv.c ret = dev->driver->load(dev, flags); load 45 drivers/gpu/drm/drm_mode_object.c WARN_ON(!dev->driver->load && dev->registered && !obj_free_cb); load 107 drivers/gpu/drm/drm_mode_object.c WARN_ON(!dev->driver->load && dev->registered && !object->free_cb); load 139 drivers/gpu/drm/fsl-dcu/fsl_dcu_drm_drv.c .load = fsl_dcu_load, load 469 drivers/gpu/drm/gma500/psb_drv.c .load = psb_driver_load, load 61 drivers/gpu/drm/i810/i810_drv.c .load = i810_driver_load, load 61 drivers/gpu/drm/mga/mga_drv.c .load = mga_driver_load, load 98 drivers/gpu/drm/mgag200/mgag200_drv.c .load = mgag200_driver_load, load 60 drivers/gpu/drm/nouveau/dispnv50/atom.h void (*load)(struct drm_color_lut *, int size, void __iomem *); load 183 drivers/gpu/drm/nouveau/dispnv50/atom.h void (*load)(struct drm_color_lut *, int size, load 83 drivers/gpu/drm/nouveau/dispnv50/base907c.c asyw->xlut.i.load = head907d_olut_load; load 409 drivers/gpu/drm/nouveau/dispnv50/disp.c struct nv50_disp_dac_load_v0 load; load 418 drivers/gpu/drm/nouveau/dispnv50/disp.c args.load.data = nouveau_drm(encoder->dev)->vbios.dactestval; load 419 drivers/gpu/drm/nouveau/dispnv50/disp.c if (args.load.data == 0) load 420 drivers/gpu/drm/nouveau/dispnv50/disp.c args.load.data = 340; load 423 drivers/gpu/drm/nouveau/dispnv50/disp.c if (ret || !args.load.load) load 55 drivers/gpu/drm/nouveau/dispnv50/head.c asyh->olut.load); load 282 drivers/gpu/drm/nouveau/dispnv50/head507d.c asyh->olut.load = head507d_olut_load; load 237 drivers/gpu/drm/nouveau/dispnv50/head907d.c asyh->olut.load = head907d_olut_load; load 158 drivers/gpu/drm/nouveau/dispnv50/headc37d.c asyh->olut.load = head907d_olut_load; load 162 drivers/gpu/drm/nouveau/dispnv50/headc57d.c asyh->olut.load = headc57d_olut_load_8; load 164 drivers/gpu/drm/nouveau/dispnv50/headc57d.c asyh->olut.load = headc57d_olut_load; load 33 drivers/gpu/drm/nouveau/dispnv50/lut.c void (*load)(struct drm_color_lut *, int, void __iomem *)) load 48 drivers/gpu/drm/nouveau/dispnv50/lut.c load(in, 1024, mem); load 52 drivers/gpu/drm/nouveau/dispnv50/lut.c load(in, blob->length / sizeof(*in), mem); load 146 drivers/gpu/drm/nouveau/dispnv50/wndw.c asyw->ilut, asyw->xlut.i.load); load 81 drivers/gpu/drm/nouveau/dispnv50/wndwc37e.c asyw->xlut.i.load = head907d_olut_load; load 171 drivers/gpu/drm/nouveau/dispnv50/wndwc57e.c asyw->xlut.i.load = wndwc57e_ilut_load; load 54 drivers/gpu/drm/nouveau/include/nvif/cl5070.h __u8 load; load 129 drivers/gpu/drm/nouveau/nvkm/engine/disp/rootnv50.c args->v0.load = ret; load 44 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c bool load; load 66 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c status->load = !!(stat & 0x00002000); load 72 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c if (status->load && status->save) { load 78 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c if (status->load) { load 84 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c if (status->load) { load 91 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c status->chsw, status->save, status->load, load 68 drivers/gpu/drm/nouveau/nvkm/subdev/pmu/gk20a.c int *state, int load) load 77 drivers/gpu/drm/nouveau/nvkm/subdev/pmu/gk20a.c if (load > data->p_load_max) { load 80 drivers/gpu/drm/nouveau/nvkm/subdev/pmu/gk20a.c level += ((load - data->p_load_target) * 10 / load 40 drivers/gpu/drm/nouveau/nvkm/subdev/secboot/acr.h int (*load)(struct nvkm_acr *, struct nvkm_falcon *, load 258 drivers/gpu/drm/nouveau/nvkm/subdev/secboot/acr_r352.c ret = func->load(sb, func->version_max, &img->base); load 1076 drivers/gpu/drm/nouveau/nvkm/subdev/secboot/acr_r352.c .load = acr_ls_ucode_load_fecs, load 1093 drivers/gpu/drm/nouveau/nvkm/subdev/secboot/acr_r352.c .load = acr_ls_ucode_load_gpccs, load 1179 drivers/gpu/drm/nouveau/nvkm/subdev/secboot/acr_r352.c .load = acr_ls_ucode_load_pmu, load 1206 drivers/gpu/drm/nouveau/nvkm/subdev/secboot/acr_r352.c .load = acr_r352_load, load 71 drivers/gpu/drm/nouveau/nvkm/subdev/secboot/acr_r352.h int (*load)(const struct nvkm_secboot *, int maxver, load 77 drivers/gpu/drm/nouveau/nvkm/subdev/secboot/acr_r361.c .load = acr_ls_ucode_load_fecs, load 94 drivers/gpu/drm/nouveau/nvkm/subdev/secboot/acr_r361.c .load = acr_ls_ucode_load_gpccs, load 152 drivers/gpu/drm/nouveau/nvkm/subdev/secboot/acr_r361.c .load = acr_ls_ucode_load_pmu, load 199 drivers/gpu/drm/nouveau/nvkm/subdev/secboot/acr_r361.c .load = acr_ls_ucode_load_sec2, load 127 drivers/gpu/drm/nouveau/nvkm/subdev/secboot/acr_r367.c ret = func->load(sb, func->version_max, &img->base); load 387 drivers/gpu/drm/nouveau/nvkm/subdev/secboot/acr_r367.c .load = acr_ls_ucode_load_sec2, load 60 drivers/gpu/drm/nouveau/nvkm/subdev/secboot/acr_r370.c .load = acr_ls_ucode_load_fecs, load 77 drivers/gpu/drm/nouveau/nvkm/subdev/secboot/acr_r370.c .load = acr_ls_ucode_load_gpccs, load 122 drivers/gpu/drm/nouveau/nvkm/subdev/secboot/acr_r370.c .load = acr_ls_ucode_load_sec2, load 65 drivers/gpu/drm/nouveau/nvkm/subdev/secboot/acr_r375.c .load = acr_ls_ucode_load_pmu, load 68 drivers/gpu/drm/nouveau/nvkm/subdev/secboot/gm200.c ret = sb->acr->func->load(sb->acr, falcon, blob, vma->addr); load 65 drivers/gpu/drm/r128/r128_drv.c .load = r128_driver_load, load 621 drivers/gpu/drm/radeon/radeon_drv.c .load = radeon_driver_load_kms, load 54 drivers/gpu/drm/savage/savage_drv.c .load = savage_driver_load, load 104 drivers/gpu/drm/sis/sis_drv.c .load = sis_driver_load, load 1017 drivers/gpu/drm/tegra/drm.c .load = tegra_drm_load, load 77 drivers/gpu/drm/via/via_drv.c .load = via_driver_load, load 1390 drivers/gpu/drm/vmwgfx/vmwgfx_drv.c .load = vmw_driver_load, load 68 drivers/gpu/host1x/dev.h u32 (*load)(struct host1x_syncpt *syncpt); load 173 drivers/gpu/host1x/dev.h return host->syncpt_op->load(sp); load 140 drivers/gpu/host1x/hw/syncpt_hw.c .load = syncpt_load, load 92 drivers/infiniband/hw/mlx5/ib_rep.c .load = mlx5_ib_vport_rep_load, load 220 drivers/macintosh/rack-meter.c int i, offset, load, cumm, pause; load 234 drivers/macintosh/rack-meter.c load = div64_u64(9 * (total_nsecs - idle_nsecs), total_nsecs); load 239 drivers/macintosh/rack-meter.c u8 ub = (load > i) ? 0xff : 0; load 104 drivers/media/dvb-frontends/dib0090.c u16 load; load 2488 drivers/media/dvb-frontends/dib0090.c dib0090_write_reg(state, 0x0e, tune->load); load 1261 drivers/media/i2c/max2175.c static int max2175_refout_load_to_bits(struct i2c_client *client, u32 load, load 1264 drivers/media/i2c/max2175.c if (load <= 40) load 1265 drivers/media/i2c/max2175.c *bits = load / 10; load 1266 drivers/media/i2c/max2175.c else if (load >= 60 && load <= 70) load 1267 drivers/media/i2c/max2175.c *bits = load / 10 - 1; load 20 drivers/media/platform/qcom/venus/core.h unsigned int load; load 415 drivers/media/platform/qcom/venus/helpers.c if (mbs_per_sec > table[i].load) load 498 drivers/misc/genwqe/card_dev.c struct genwqe_bitstream *load) load 513 drivers/misc/genwqe/card_dev.c if ((load->size & 0x3) != 0) load 516 drivers/misc/genwqe/card_dev.c if (((unsigned long)(load->data_addr) & ~PAGE_MASK) != 0) load 520 drivers/misc/genwqe/card_dev.c switch ((char)load->partition) { load 534 drivers/misc/genwqe/card_dev.c buf = (u8 __user *)load->data_addr; load 539 drivers/misc/genwqe/card_dev.c blocks_to_flash = load->size / FLASH_BLOCK; load 540 drivers/misc/genwqe/card_dev.c while (load->size) { load 547 drivers/misc/genwqe/card_dev.c tocopy = min_t(size_t, load->size, FLASH_BLOCK); load 577 drivers/misc/genwqe/card_dev.c req->__asiv[24] = load->uid; load 581 drivers/misc/genwqe/card_dev.c *(__be64 *)&req->__asiv[88] = cpu_to_be64(load->slu_id); load 582 drivers/misc/genwqe/card_dev.c *(__be64 *)&req->__asiv[96] = cpu_to_be64(load->app_id); load 589 drivers/misc/genwqe/card_dev.c *(__be32 *)&req->asiv[24] = cpu_to_be32(load->uid<<24); load 593 drivers/misc/genwqe/card_dev.c *(__be64 *)&req->asiv[80] = cpu_to_be64(load->slu_id); load 594 drivers/misc/genwqe/card_dev.c *(__be64 *)&req->asiv[88] = cpu_to_be64(load->app_id); load 607 drivers/misc/genwqe/card_dev.c load->retc = req->retc; load 608 drivers/misc/genwqe/card_dev.c load->attn = req->attn; load 609 drivers/misc/genwqe/card_dev.c load->progress = req->progress; load 622 drivers/misc/genwqe/card_dev.c load->size -= tocopy; load 635 drivers/misc/genwqe/card_dev.c struct genwqe_bitstream *load) load 649 drivers/misc/genwqe/card_dev.c if ((load->size & 0x3) != 0) load 652 drivers/misc/genwqe/card_dev.c if (((unsigned long)(load->data_addr) & ~PAGE_MASK) != 0) load 656 drivers/misc/genwqe/card_dev.c switch ((char)load->partition) { load 670 drivers/misc/genwqe/card_dev.c buf = (u8 __user *)load->data_addr; load 675 drivers/misc/genwqe/card_dev.c blocks_to_flash = load->size / FLASH_BLOCK; load 676 drivers/misc/genwqe/card_dev.c while (load->size) { load 681 drivers/misc/genwqe/card_dev.c tocopy = min_t(size_t, load->size, FLASH_BLOCK); load 703 drivers/misc/genwqe/card_dev.c cmd->__asiv[24] = load->uid; load 711 drivers/misc/genwqe/card_dev.c *(__be32 *)&cmd->asiv[24] = cpu_to_be32(load->uid<<24); load 725 drivers/misc/genwqe/card_dev.c load->retc = cmd->retc; load 726 drivers/misc/genwqe/card_dev.c load->attn = cmd->attn; load 727 drivers/misc/genwqe/card_dev.c load->progress = cmd->progress; load 751 drivers/misc/genwqe/card_dev.c load->size -= tocopy; load 1140 drivers/misc/genwqe/card_dev.c struct genwqe_bitstream load; load 1148 drivers/misc/genwqe/card_dev.c if (copy_from_user(&load, (void __user *)arg, load 1149 drivers/misc/genwqe/card_dev.c sizeof(load))) load 1152 drivers/misc/genwqe/card_dev.c rc = do_flash_update(cfile, &load); load 1154 drivers/misc/genwqe/card_dev.c if (copy_to_user((void __user *)arg, &load, sizeof(load))) load 1161 drivers/misc/genwqe/card_dev.c struct genwqe_bitstream load; load 1169 drivers/misc/genwqe/card_dev.c if (copy_from_user(&load, (void __user *)arg, sizeof(load))) load 1172 drivers/misc/genwqe/card_dev.c rc = do_flash_read(cfile, &load); load 1174 drivers/misc/genwqe/card_dev.c if (copy_to_user((void __user *)arg, &load, sizeof(load))) load 89 drivers/net/bonding/bond_alb.c SLAVE_TLB_INFO(slave).load = 0; load 163 drivers/net/bonding/bond_alb.c (s64) (SLAVE_TLB_INFO(slave).load << 3); /* Bytes to bits */ load 215 drivers/net/bonding/bond_alb.c slave_info->load += load 1550 drivers/net/bonding/bond_alb.c SLAVE_TLB_INFO(slave).load = load 36 drivers/net/can/softing/softing_cs.c .load = {0x0120, 0x00f600, fw_dir "ldcard.bin",}, load 48 drivers/net/can/softing/softing_cs.c .load = {0x0120, 0x00f600, fw_dir "ldcard.bin",}, load 60 drivers/net/can/softing/softing_cs.c .load = {0x0120, 0x00f600, fw_dir "ldcard.bin",}, load 72 drivers/net/can/softing/softing_cs.c .load = {0x0120, 0x00f600, fw_dir "ldcard2.bin",}, load 84 drivers/net/can/softing/softing_cs.c .load = {0x0120, 0x00f600, fw_dir "ldcard.bin",}, load 96 drivers/net/can/softing/softing_cs.c .load = {0x0120, 0x00f600, fw_dir "ldcard.bin",}, load 108 drivers/net/can/softing/softing_cs.c .load = {0x0120, 0x00f600, fw_dir "ldcard2.bin",}, load 120 drivers/net/can/softing/softing_cs.c .load = {0x0120, 0x00f600, fw_dir "ldcard.bin",}, load 132 drivers/net/can/softing/softing_cs.c .load = {0x0120, 0x00f600, fw_dir "ldcard2.bin",}, load 512 drivers/net/can/softing/softing_main.c ret = softing_load_fw(card->pdat->load.fw, card, card->dpram, load 514 drivers/net/can/softing/softing_main.c card->pdat->load.offs - card->pdat->load.addr); load 32 drivers/net/can/softing/softing_platform.h } boot, load, app; load 2698 drivers/net/ethernet/brocade/bna/bna_tx_rx.c u32 load, bias; load 2714 drivers/net/ethernet/brocade/bna/bna_tx_rx.c load = BNA_LOAD_T_LOW_4; load 2716 drivers/net/ethernet/brocade/bna/bna_tx_rx.c load = BNA_LOAD_T_LOW_3; load 2718 drivers/net/ethernet/brocade/bna/bna_tx_rx.c load = BNA_LOAD_T_LOW_2; load 2720 drivers/net/ethernet/brocade/bna/bna_tx_rx.c load = BNA_LOAD_T_LOW_1; load 2722 drivers/net/ethernet/brocade/bna/bna_tx_rx.c load = BNA_LOAD_T_HIGH_1; load 2724 drivers/net/ethernet/brocade/bna/bna_tx_rx.c load = BNA_LOAD_T_HIGH_2; load 2726 drivers/net/ethernet/brocade/bna/bna_tx_rx.c load = BNA_LOAD_T_HIGH_3; load 2728 drivers/net/ethernet/brocade/bna/bna_tx_rx.c load = BNA_LOAD_T_HIGH_4; load 2738 drivers/net/ethernet/brocade/bna/bna_tx_rx.c coalescing_timeo = bna->rx_mod.dim_vector[load][bias]; load 1973 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c .load = mlx5e_vport_rep_load, load 1492 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c err = esw->offloads.rep_ops[rep_type]->load(esw->dev, rep); load 586 drivers/net/wireless/ath/wil6210/fw_inc.c const void *data, size_t size, bool load) load 592 drivers/net/wireless/ath/wil6210/fw_inc.c return load ? load 611 drivers/net/wireless/ath/wil6210/fw_inc.c size_t size, bool load) load 630 drivers/net/wireless/ath/wil6210/fw_inc.c &hdr[1], hdr_sz, load); load 658 drivers/net/wireless/ath/wil6210/fw_inc.c bool load) load 683 drivers/net/wireless/ath/wil6210/fw_inc.c rc = wil_fw_process(wil, d, rc1, load); load 1399 drivers/net/wireless/ath/wil6210/wil6210.h bool load); load 692 drivers/net/wireless/cisco/airo.c __le16 load; load 273 drivers/net/wireless/intel/iwlegacy/4965-rs.c tl = &lq_data->load[tid]; load 319 drivers/net/wireless/intel/iwlegacy/4965-rs.c tl = &(lq_data->load[tid]); load 342 drivers/net/wireless/intel/iwlegacy/4965-rs.c u32 load; load 344 drivers/net/wireless/intel/iwlegacy/4965-rs.c load = il4965_rs_tl_get_load(lq_data, tid); load 346 drivers/net/wireless/intel/iwlegacy/4965-rs.c if (load > IL_AGG_LOAD_THRESHOLD) { load 360 drivers/net/wireless/intel/iwlegacy/4965-rs.c tid, load); load 2807 drivers/net/wireless/intel/iwlegacy/common.h struct il_traffic_load load[TID_MAX_LOAD_COUNT]; load 287 drivers/net/wireless/intel/iwlwifi/dvm/rs.c tl = &lq_data->load[tid]; load 362 drivers/net/wireless/intel/iwlwifi/dvm/rs.c tl = &(lq_data->load[tid]); load 356 drivers/net/wireless/intel/iwlwifi/dvm/rs.h struct iwl_traffic_load load[IWL_MAX_TID_COUNT]; load 413 drivers/net/wireless/intel/iwlwifi/mvm/coex.c data->primary_load = mvm->tcm.result.load[mvmvif->id]; load 415 drivers/net/wireless/intel/iwlwifi/mvm/coex.c data->secondary_load = mvm->tcm.result.load[mvmvif->id]; load 430 drivers/net/wireless/intel/iwlwifi/mvm/coex.c data->primary_load = mvm->tcm.result.load[mvmvif->id]; load 432 drivers/net/wireless/intel/iwlwifi/mvm/coex.c data->secondary_load = mvm->tcm.result.load[mvmvif->id]; load 267 drivers/net/wireless/intel/iwlwifi/mvm/debugfs-vif.c mvm->tcm.result.load[mvmvif->id]); load 641 drivers/net/wireless/intel/iwlwifi/mvm/mvm.h enum iwl_mvm_traffic_load load[NUM_MAC_INDEX_DRIVER]; load 271 drivers/net/wireless/intel/iwlwifi/mvm/scan.c enum iwl_mvm_traffic_load load, load 285 drivers/net/wireless/intel/iwlwifi/mvm/scan.c if ((load == IWL_MVM_TRAFFIC_HIGH || low_latency) && load 308 drivers/net/wireless/intel/iwlwifi/mvm/scan.c if (load >= IWL_MVM_TRAFFIC_MEDIUM || low_latency) load 318 drivers/net/wireless/intel/iwlwifi/mvm/scan.c enum iwl_mvm_traffic_load load; load 321 drivers/net/wireless/intel/iwlwifi/mvm/scan.c load = iwl_mvm_get_traffic_load(mvm); load 324 drivers/net/wireless/intel/iwlwifi/mvm/scan.c return _iwl_mvm_get_scan_type(mvm, vif, load, low_latency); load 332 drivers/net/wireless/intel/iwlwifi/mvm/scan.c enum iwl_mvm_traffic_load load; load 335 drivers/net/wireless/intel/iwlwifi/mvm/scan.c load = iwl_mvm_get_traffic_load_band(mvm, band); load 338 drivers/net/wireless/intel/iwlwifi/mvm/scan.c return _iwl_mvm_get_scan_type(mvm, vif, load, low_latency); load 1049 drivers/net/wireless/intel/iwlwifi/mvm/utils.c u8 load = iwl_mvm_tcm_load_percentage(airtime, elapsed); load 1051 drivers/net/wireless/intel/iwlwifi/mvm/utils.c if (load > IWL_MVM_TCM_LOAD_HIGH_THRESH) load 1053 drivers/net/wireless/intel/iwlwifi/mvm/utils.c if (load > IWL_MVM_TCM_LOAD_MEDIUM_THRESH) load 1230 drivers/net/wireless/intel/iwlwifi/mvm/utils.c enum iwl_mvm_traffic_load load, band_load; load 1253 drivers/net/wireless/intel/iwlwifi/mvm/utils.c load = iwl_mvm_tcm_load(mvm, airtime, elapsed); load 1254 drivers/net/wireless/intel/iwlwifi/mvm/utils.c mvm->tcm.result.change[mac] = load != mvm->tcm.result.load[mac]; load 1255 drivers/net/wireless/intel/iwlwifi/mvm/utils.c mvm->tcm.result.load[mac] = load; load 1285 drivers/net/wireless/intel/iwlwifi/mvm/utils.c load = iwl_mvm_tcm_load(mvm, total_airtime, elapsed); load 1286 drivers/net/wireless/intel/iwlwifi/mvm/utils.c mvm->tcm.result.global_change = load != mvm->tcm.result.global_load; load 1287 drivers/net/wireless/intel/iwlwifi/mvm/utils.c mvm->tcm.result.global_load = load; load 1300 drivers/net/wireless/intel/iwlwifi/mvm/utils.c if (load != IWL_MVM_TRAFFIC_LOW) load 142 drivers/perf/fsl_imx8_ddr_perf.c IMX8_DDR_PMU_EVENT_ATTR(load-mode, 0x33), load 26 drivers/regulator/qcom_smd-regulator.c u32 load; load 66 drivers/regulator/qcom_smd-regulator.c req[reqlen].value = cpu_to_le32(vreg->load / 1000); load 151 drivers/regulator/qcom_smd-regulator.c u32 old_load = vreg->load; load 154 drivers/regulator/qcom_smd-regulator.c vreg->load = load_uA; load 158 drivers/regulator/qcom_smd-regulator.c vreg->load = old_load; load 290 drivers/remoteproc/qcom_q6v5_adsp.c .load = adsp_load, load 1228 drivers/remoteproc/qcom_q6v5_mss.c .load = q6v5_load, load 179 drivers/remoteproc/qcom_q6v5_pas.c .load = adsp_load, load 434 drivers/remoteproc/qcom_q6v5_wcss.c .load = q6v5_wcss_load, load 307 drivers/remoteproc/qcom_wcnss.c .load = wcnss_load, load 2053 drivers/remoteproc/remoteproc_core.c if (!rproc->ops->load) { load 2054 drivers/remoteproc/remoteproc_core.c rproc->ops->load = rproc_elf_load_segments; load 87 drivers/remoteproc/remoteproc_internal.h if (rproc->ops->load) load 88 drivers/remoteproc/remoteproc_internal.h return rproc->ops->load(rproc, fw); load 234 drivers/remoteproc/st_remoteproc.c .load = rproc_elf_load_segments, load 205 drivers/remoteproc/st_slim_rproc.c .load = rproc_elf_load_segments, load 474 drivers/remoteproc/stm32_rproc.c .load = rproc_elf_load_segments, load 348 drivers/rtc/rtc-pcf85063.c u32 load = 7000; load 352 drivers/rtc/rtc-pcf85063.c load = force_cap; load 354 drivers/rtc/rtc-pcf85063.c of_property_read_u32(np, "quartz-load-femtofarads", &load); load 356 drivers/rtc/rtc-pcf85063.c switch (load) { load 359 drivers/rtc/rtc-pcf85063.c load); load 99 drivers/rtc/rtc-pcf8523.c u32 load; load 107 drivers/rtc/rtc-pcf8523.c load = 12500; load 109 drivers/rtc/rtc-pcf8523.c &load); load 111 drivers/rtc/rtc-pcf8523.c switch (load) { load 114 drivers/rtc/rtc-pcf8523.c load); load 578 drivers/s390/crypto/zcrypt_api.c atomic_add(weight, &zc->load); load 579 drivers/s390/crypto/zcrypt_api.c atomic_add(weight, &zq->load); load 591 drivers/s390/crypto/zcrypt_api.c atomic_sub(weight, &zc->load); load 592 drivers/s390/crypto/zcrypt_api.c atomic_sub(weight, &zq->load); load 605 drivers/s390/crypto/zcrypt_api.c weight += atomic_read(&zc->load); load 606 drivers/s390/crypto/zcrypt_api.c pref_weight += atomic_read(&pref_zc->load); load 620 drivers/s390/crypto/zcrypt_api.c weight += atomic_read(&zq->load); load 621 drivers/s390/crypto/zcrypt_api.c pref_weight += atomic_read(&pref_zq->load); load 86 drivers/s390/crypto/zcrypt_api.h atomic_t load; /* Utilization of the crypto device */ load 99 drivers/s390/crypto/zcrypt_api.h atomic_t load; /* Utilization of the crypto device */ load 89 drivers/s390/crypto/zcrypt_card.c return snprintf(buf, PAGE_SIZE, "%d\n", atomic_read(&zc->load)); load 92 drivers/s390/crypto/zcrypt_card.c static DEVICE_ATTR_RO(load); load 177 drivers/s390/crypto/zcrypt_cex2a.c atomic_set(&zq->load, 0); load 222 drivers/s390/crypto/zcrypt_cex2c.c atomic_set(&zq->load, 0); load 383 drivers/s390/crypto/zcrypt_cex4.c atomic_set(&zq->load, 0); load 81 drivers/s390/crypto/zcrypt_queue.c return snprintf(buf, PAGE_SIZE, "%d\n", atomic_read(&zq->load)); load 84 drivers/s390/crypto/zcrypt_queue.c static DEVICE_ATTR_RO(load); load 1660 drivers/scsi/be2iscsi/be_cmds.c int beiscsi_cmd_special_wrb(struct be_ctrl_info *ctrl, u32 load) load 1671 drivers/scsi/be2iscsi/be_cmds.c if (load) { load 790 drivers/scsi/be2iscsi/be_cmds.h int beiscsi_cmd_special_wrb(struct be_ctrl_info *ctrl, u32 load); load 53 drivers/staging/fbtft/fb_uc1611.c static unsigned int load = 1; load 54 drivers/staging/fbtft/fb_uc1611.c module_param(load, uint, 0000); load 55 drivers/staging/fbtft/fb_uc1611.c MODULE_PARM_DESC(load, "PC[1:0] Panel Loading: 0-3 (default: 1)"); load 88 drivers/staging/fbtft/fb_uc1611.c write_reg(par, 0x28 | (load & 0x03)); load 96 drivers/staging/olpc_dcon/olpc_dcon.h void (*set_dconload)(int load); load 232 drivers/thermal/cpu_cooling.c u32 load; load 241 drivers/thermal/cpu_cooling.c load = 0; load 243 drivers/thermal/cpu_cooling.c load = div64_u64(100 * (delta_time - delta_idle), delta_time); load 248 drivers/thermal/cpu_cooling.c return load; load 381 drivers/thermal/cpu_cooling.c u32 load; load 384 drivers/thermal/cpu_cooling.c load = get_load(cpufreq_cdev, cpu, i); load 386 drivers/thermal/cpu_cooling.c load = 0; load 388 drivers/thermal/cpu_cooling.c total_load += load; load 390 drivers/thermal/cpu_cooling.c load_cpu[i] = load; load 332 drivers/usb/host/isp116x-hcd.c isp116x->load[i] -= ep->load; load 336 drivers/usb/host/isp116x-hcd.c ep->load / ep->period; load 484 drivers/usb/host/isp116x-hcd.c u16 load = 0; load 504 drivers/usb/host/isp116x-hcd.c load = isp116x->load[index]; load 505 drivers/usb/host/isp116x-hcd.c if (load) { load 530 drivers/usb/host/isp116x-hcd.c len = (MAX_LOAD_LIMIT - load) / byte_time; load 551 drivers/usb/host/isp116x-hcd.c load += len * byte_time; load 552 drivers/usb/host/isp116x-hcd.c if (load > MAX_LOAD_LIMIT) load 652 drivers/usb/host/isp116x-hcd.c static int balance(struct isp116x *isp116x, u16 period, u16 load) load 659 drivers/usb/host/isp116x-hcd.c if (branch < 0 || isp116x->load[branch] > isp116x->load[i]) { load 663 drivers/usb/host/isp116x-hcd.c if ((isp116x->load[j] + load) load 757 drivers/usb/host/isp116x-hcd.c ep->load = usb_calc_bus_time(udev->speed, load 784 drivers/usb/host/isp116x-hcd.c ep->branch = ret = balance(isp116x, ep->period, ep->load); load 810 drivers/usb/host/isp116x-hcd.c isp116x->load[i] += ep->load; load 812 drivers/usb/host/isp116x-hcd.c hcd->self.bandwidth_allocated += ep->load / ep->period; load 279 drivers/usb/host/isp116x.h u16 load[PERIODIC_SIZE]; load 320 drivers/usb/host/isp116x.h u16 load; load 488 drivers/usb/host/isp1362-hcd.c ep, ep->branch, ep->load, load 489 drivers/usb/host/isp1362-hcd.c isp1362_hcd->load[ep->branch], load 490 drivers/usb/host/isp1362-hcd.c isp1362_hcd->load[ep->branch] - ep->load); load 491 drivers/usb/host/isp1362-hcd.c isp1362_hcd->load[ep->branch] -= ep->load; load 1184 drivers/usb/host/isp1362-hcd.c static int balance(struct isp1362_hcd *isp1362_hcd, u16 interval, u16 load) load 1192 drivers/usb/host/isp1362-hcd.c if (branch < 0 || isp1362_hcd->load[branch] > isp1362_hcd->load[i]) { load 1196 drivers/usb/host/isp1362-hcd.c if ((isp1362_hcd->load[j] + load) > MAX_PERIODIC_LOAD) { load 1198 drivers/usb/host/isp1362-hcd.c load, j, isp1362_hcd->load[j], MAX_PERIODIC_LOAD); load 1301 drivers/usb/host/isp1362-hcd.c ep->load = usb_calc_bus_time(udev->speed, !is_out, load 1328 drivers/usb/host/isp1362-hcd.c retval = balance(isp1362_hcd, ep->interval, ep->load); load 1363 drivers/usb/host/isp1362-hcd.c ep->load / ep->interval, isp1362_hcd->load[ep->branch], load 1364 drivers/usb/host/isp1362-hcd.c isp1362_hcd->load[ep->branch] + ep->load); load 1365 drivers/usb/host/isp1362-hcd.c isp1362_hcd->load[ep->branch] += ep->load; load 2133 drivers/usb/host/isp1362-hcd.c isp1362_hcd->load[ep->branch], ep->ptd_index, ep->ptd_offset); load 404 drivers/usb/host/isp1362.h u16 load; load 459 drivers/usb/host/isp1362.h u16 load[PERIODIC_SIZE]; load 515 drivers/usb/host/ohci-dbg.c temp = scnprintf (next, size, "%2d [%3d]:", i, ohci->load [i]); load 1071 drivers/usb/host/ohci-hcd.c for (i = 0; i < NUM_INTS; i++) ohci->load [i] = 0; load 109 drivers/usb/host/ohci-q.c static int balance (struct ohci_hcd *ohci, int interval, int load) load 121 drivers/usb/host/ohci-q.c if (branch < 0 || ohci->load [branch] > ohci->load [i]) { load 126 drivers/usb/host/ohci-q.c if ((ohci->load [j] + load) > 900) load 149 drivers/usb/host/ohci-q.c ed, ed->branch, ed->load, ed->interval); load 176 drivers/usb/host/ohci-q.c ohci->load [i] += ed->load; load 178 drivers/usb/host/ohci-q.c ohci_to_hcd(ohci)->self.bandwidth_allocated += ed->load / ed->interval; load 247 drivers/usb/host/ohci-q.c branch = balance (ohci, ed->interval, ed->load); load 251 drivers/usb/host/ohci-q.c branch, ed->interval, ed->load); load 287 drivers/usb/host/ohci-q.c ohci->load [i] -= ed->load; load 289 drivers/usb/host/ohci-q.c ohci_to_hcd(ohci)->self.bandwidth_allocated -= ed->load / ed->interval; load 293 drivers/usb/host/ohci-q.c ed, ed->branch, ed->load, ed->interval); load 459 drivers/usb/host/ohci-q.c ed->load = usb_calc_bus_time ( load 66 drivers/usb/host/ohci.h u16 load; load 403 drivers/usb/host/ohci.h int load [NUM_INTS]; load 460 drivers/usb/host/sl811-hcd.c sl811->load[i] -= ep->load; load 465 drivers/usb/host/sl811-hcd.c -= ep->load / ep->period; load 764 drivers/usb/host/sl811-hcd.c static int balance(struct sl811 *sl811, u16 period, u16 load) load 772 drivers/usb/host/sl811-hcd.c if (branch < 0 || sl811->load[branch] > sl811->load[i]) { load 776 drivers/usb/host/sl811-hcd.c if ((sl811->load[j] + load) load 880 drivers/usb/host/sl811-hcd.c ep->load = usb_calc_bus_time(udev->speed, !is_out, load 912 drivers/usb/host/sl811-hcd.c retval = balance(sl811, ep->period, ep->load); load 940 drivers/usb/host/sl811-hcd.c sl811->load[i] += ep->load; load 943 drivers/usb/host/sl811-hcd.c hcd->self.bandwidth_allocated += ep->load / ep->period; load 1467 drivers/usb/host/sl811-hcd.c seq_printf(s, "%2d [%3d]:\n", i, sl811->load[i]); load 154 drivers/usb/host/sl811.h u16 load[PERIODIC_SIZE]; load 185 drivers/usb/host/sl811.h u16 load; load 188 drivers/usb/host/uhci-debug.c space, "", qh->period, qh->phase, qh->load, load 192 drivers/usb/host/uhci-debug.c space, "", qh->period, qh->phase, qh->load); load 400 drivers/usb/host/uhci-debug.c out += sprintf(out, "\t%d", uhci->load[i]); load 172 drivers/usb/host/uhci-hcd.h short load; /* Periodic time requirement, in us */ load 449 drivers/usb/host/uhci-hcd.h short load[MAX_PHASE]; /* Periodic allocations */ load 279 drivers/usb/host/uhci-q.c qh->load = usb_calc_bus_time(udev->speed, load 612 drivers/usb/host/uhci-q.c int highest_load = uhci->load[phase]; load 615 drivers/usb/host/uhci-q.c highest_load = max_t(int, highest_load, uhci->load[phase]); load 632 drivers/usb/host/uhci-q.c int phase, load; load 638 drivers/usb/host/uhci-q.c load = uhci_highest_load(uhci, phase, qh->period); load 639 drivers/usb/host/uhci-q.c if (load < minimax_load) { load 640 drivers/usb/host/uhci-q.c minimax_load = load; load 647 drivers/usb/host/uhci-q.c if (minimax_load + qh->load > 900) { load 650 drivers/usb/host/uhci-q.c qh->period, qh->phase, minimax_load, qh->load); load 662 drivers/usb/host/uhci-q.c int load = qh->load; load 666 drivers/usb/host/uhci-q.c uhci->load[i] += load; load 667 drivers/usb/host/uhci-q.c uhci->total_load += load; load 686 drivers/usb/host/uhci-q.c qh->period, qh->phase, load); load 695 drivers/usb/host/uhci-q.c int load = qh->load; load 699 drivers/usb/host/uhci-q.c uhci->load[i] -= load; load 700 drivers/usb/host/uhci-q.c uhci->total_load -= load; load 719 drivers/usb/host/uhci-q.c qh->period, qh->phase, load); load 109 drivers/virt/fsl_hypervisor.c param.load); load 94 drivers/watchdog/sp805_wdt.c u64 load, rate; load 104 drivers/watchdog/sp805_wdt.c load = div_u64(rate, 2) * timeout - 1; load 106 drivers/watchdog/sp805_wdt.c load = (load > LOAD_MAX) ? LOAD_MAX : load; load 107 drivers/watchdog/sp805_wdt.c load = (load < LOAD_MIN) ? LOAD_MIN : load; load 110 drivers/watchdog/sp805_wdt.c wdt->load_val = load; load 112 drivers/watchdog/sp805_wdt.c wdd->timeout = div_u64((load + 1) * 2 + (rate / 2), rate); load 122 drivers/watchdog/sp805_wdt.c u64 load; load 125 drivers/watchdog/sp805_wdt.c load = readl_relaxed(wdt->base + WDTVALUE); load 129 drivers/watchdog/sp805_wdt.c load += wdt->load_val + 1; load 132 drivers/watchdog/sp805_wdt.c return div_u64(load, wdt->rate); load 368 include/clocksource/timer-ti-dm.h u32 ctrl, unsigned int load, load 371 include/clocksource/timer-ti-dm.h __omap_dm_timer_write(timer, OMAP_TIMER_COUNTER_REG, load, posted); load 180 include/drm/drm_drv.h int (*load) (struct drm_device *, unsigned long flags); load 20 include/linux/frontswap.h int (*load)(unsigned, pgoff_t, struct page *); /* load a page */ load 135 include/linux/kexec.h kexec_load_t *load; load 34 include/linux/mlx5/eswitch.h int (*load)(struct mlx5_core_dev *dev, struct mlx5_eswitch_rep *rep); load 383 include/linux/remoteproc.h int (*load)(struct rproc *rproc, const struct firmware *fw); load 446 include/linux/sched.h struct load_weight load; load 29 include/linux/sched/loadavg.h calc_load(unsigned long load, unsigned long exp, unsigned long active) load 33 include/linux/sched/loadavg.h newload = load * exp + active * (FIXED_1 - exp); load 34 include/linux/sched/loadavg.h if (active >= load) load 40 include/linux/sched/loadavg.h extern unsigned long calc_load_n(unsigned long load, unsigned long exp, load 121 include/net/bond_alb.h u32 load; /* Each slave sums the loadHistory of all clients load 1088 include/net/tcp.h int tcp_set_congestion_control(struct sock *sk, const char *name, bool load, load 244 include/trace/events/kvm.h TP_PROTO(int load), load 245 include/trace/events/kvm.h TP_ARGS(load), load 248 include/trace/events/kvm.h __field( u32, load ) load 252 include/trace/events/kvm.h __entry->load = load; load 255 include/trace/events/kvm.h TP_printk("%s", __print_symbolic(__entry->load, kvm_fpu_load_symbol)) load 96 include/trace/events/thermal.h TP_PROTO(const struct cpumask *cpus, unsigned long freq, u32 *load, load 99 include/trace/events/thermal.h TP_ARGS(cpus, freq, load, load_len, dynamic_power), load 104 include/trace/events/thermal.h __dynamic_array(u32, load, load_len) load 113 include/trace/events/thermal.h memcpy(__get_dynamic_array(load), load, load 114 include/trace/events/thermal.h load_len * sizeof(*load)); load 121 include/trace/events/thermal.h __print_array(__get_dynamic_array(load), __entry->load_len, 4), load 163 include/trace/events/thermal.h __field(u32, load ) load 172 include/trace/events/thermal.h __entry->load = (100 * status->busy_time) / status->total_time; load 180 include/trace/events/thermal.h __entry->load, __entry->dynamic_power, __entry->static_power, load 929 include/uapi/linux/cdrom.h __u8 load:1; load 939 include/uapi/linux/cdrom.h __u8 load:1; load 93 include/uapi/linux/fsl_hypervisor.h __u32 load; load 64 kernel/kexec_file.c if (!image->fops || !image->fops->load) load 67 kernel/kexec_file.c return image->fops->load(image, image->kernel_buf, load 750 kernel/sched/core.c struct load_weight *load = &p->se.load; load 756 kernel/sched/core.c load->weight = scale_load(WEIGHT_IDLEPRIO); load 757 kernel/sched/core.c load->inv_weight = WMULT_IDLEPRIO; load 758 kernel/sched/core.c p->se.runnable_weight = load->weight; load 769 kernel/sched/core.c load->weight = scale_load(sched_prio_to_weight[prio]); load 770 kernel/sched/core.c load->inv_weight = sched_prio_to_wmult[prio]; load 771 kernel/sched/core.c p->se.runnable_weight = load->weight; load 404 kernel/sched/debug.c P(se->load.weight); load 525 kernel/sched/debug.c SEQ_printf(m, " .%-30s: %ld\n", "load", cfs_rq->load.weight); load 942 kernel/sched/debug.c P(se.load.weight); load 664 kernel/sched/fair.c if (unlikely(se->load.weight != NICE_0_LOAD)) load 665 kernel/sched/fair.c delta = __calc_delta(delta, NICE_0_LOAD, &se->load); load 697 kernel/sched/fair.c struct load_weight *load; load 701 kernel/sched/fair.c load = &cfs_rq->load; load 704 kernel/sched/fair.c lw = cfs_rq->load; load 706 kernel/sched/fair.c update_load_add(&lw, se->load.weight); load 707 kernel/sched/fair.c load = &lw; load 709 kernel/sched/fair.c slice = __calc_delta(slice, se->load.weight, load); load 745 kernel/sched/fair.c sa->runnable_load_avg = sa->load_avg = scale_load_down(se->load.weight); load 747 kernel/sched/fair.c se->runnable_weight = se->load.weight; load 790 kernel/sched/fair.c sa->util_avg = cfs_rq->avg.util_avg * se->load.weight; load 1481 kernel/sched/fair.c unsigned long load; load 1498 kernel/sched/fair.c ns->load += cpu_runnable_load(rq); load 1567 kernel/sched/fair.c orig_src_load = env->src_stats.load; load 1568 kernel/sched/fair.c orig_dst_load = env->dst_stats.load; load 1599 kernel/sched/fair.c long load; load 1679 kernel/sched/fair.c load = task_h_load(env->p) - task_h_load(cur); load 1680 kernel/sched/fair.c if (!load) load 1683 kernel/sched/fair.c dst_load = env->dst_stats.load + load; load 1684 kernel/sched/fair.c src_load = env->src_stats.load - load; load 1713 kernel/sched/fair.c long src_load, dst_load, load; load 1717 kernel/sched/fair.c load = task_h_load(env->p); load 1718 kernel/sched/fair.c dst_load = env->dst_stats.load + load; load 1719 kernel/sched/fair.c src_load = env->src_stats.load - load; load 2759 kernel/sched/fair.c update_load_add(&cfs_rq->load, se->load.weight); load 2774 kernel/sched/fair.c update_load_sub(&cfs_rq->load, se->load.weight); load 2889 kernel/sched/fair.c update_load_set(&se->load, weight); load 2912 kernel/sched/fair.c struct load_weight *load = &se->load; load 2916 kernel/sched/fair.c load->inv_weight = sched_prio_to_wmult[prio]; load 2996 kernel/sched/fair.c long tg_weight, tg_shares, load, shares; load 3001 kernel/sched/fair.c load = max(scale_load_down(cfs_rq->load.weight), cfs_rq->avg.load_avg); load 3007 kernel/sched/fair.c tg_weight += load; load 3009 kernel/sched/fair.c shares = (tg_shares * load); load 3060 kernel/sched/fair.c scale_load_down(cfs_rq->load.weight)); load 3093 kernel/sched/fair.c if (likely(se->load.weight == shares)) load 3334 kernel/sched/fair.c if (scale_load_down(gcfs_rq->load.weight)) { load 3336 kernel/sched/fair.c scale_load_down(gcfs_rq->load.weight)); load 4184 kernel/sched/fair.c rq_of(cfs_rq)->cfs.load.weight >= 2*se->load.weight) { load 4526 kernel/sched/fair.c if (qcfs_rq->load.weight) load 4580 kernel/sched/fair.c if (!cfs_rq->load.weight) load 5334 kernel/sched/fair.c if (cfs_rq->load.weight) { load 5594 kernel/sched/fair.c unsigned long load, avg_load, runnable_load; load 5616 kernel/sched/fair.c load = cpu_runnable_load(cpu_rq(i)); load 5617 kernel/sched/fair.c runnable_load += load; load 5716 kernel/sched/fair.c unsigned long load, min_load = ULONG_MAX; load 5757 kernel/sched/fair.c load = cpu_runnable_load(cpu_rq(i)); load 5758 kernel/sched/fair.c if (load < min_load) { load 5759 kernel/sched/fair.c min_load = load; load 7377 kernel/sched/fair.c unsigned long load; load 7410 kernel/sched/fair.c load = task_h_load(p); load 7412 kernel/sched/fair.c if (sched_feat(LB_MIN) && load < 16 && !env->sd->nr_balance_failed) load 7415 kernel/sched/fair.c if ((load / 2) > env->imbalance) load 7422 kernel/sched/fair.c env->imbalance -= load; load 7572 kernel/sched/fair.c if (cfs_rq->load.weight) load 7637 kernel/sched/fair.c unsigned long load; load 7656 kernel/sched/fair.c load = cfs_rq->h_load; load 7657 kernel/sched/fair.c load = div64_ul(load * se->avg.load_avg, load 7660 kernel/sched/fair.c cfs_rq->h_load = load; load 8644 kernel/sched/fair.c unsigned long capacity, load; load 8698 kernel/sched/fair.c load = cpu_runnable_load(rq); load 8705 kernel/sched/fair.c if (rq->nr_running == 1 && load > env->imbalance && load 8720 kernel/sched/fair.c if (load * busiest_capacity > busiest_load * capacity) { load 8721 kernel/sched/fair.c busiest_load = load; load 10389 kernel/sched/fair.c update_load_set(&se->load, NICE_0_LOAD); load 10456 kernel/sched/fair.c if (rq->cfs.load.weight) load 157 kernel/sched/loadavg.c calc_load_n(unsigned long load, unsigned long exp, load 160 kernel/sched/loadavg.c return calc_load(load, fixed_power_int(exp, FSHIFT, n), active); load 111 kernel/sched/pelt.c unsigned long load, unsigned long runnable, int running) load 137 kernel/sched/pelt.c if (load) load 138 kernel/sched/pelt.c sa->load_sum += load * contrib; load 177 kernel/sched/pelt.c unsigned long load, unsigned long runnable, int running) load 210 kernel/sched/pelt.c if (!load) load 220 kernel/sched/pelt.c if (!accumulate_sum(delta, sa, load, runnable, running)) load 227 kernel/sched/pelt.c ___update_load_avg(struct sched_avg *sa, unsigned long load, unsigned long runnable) load 234 kernel/sched/pelt.c sa->load_avg = div_u64(load * sa->load_sum, divider); load 294 kernel/sched/pelt.c scale_load_down(cfs_rq->load.weight), load 497 kernel/sched/sched.h struct load_weight load; load 707 kernel/sched/sched.h return scale_load_down(se->load.weight); load 312 mm/frontswap.c ret = ops->load(type, offset, page); load 5632 mm/page_alloc.c int node, load, nr_nodes = 0; load 5638 mm/page_alloc.c load = nr_online_nodes; load 5651 mm/page_alloc.c node_load[node] = load; load 5655 mm/page_alloc.c load--; load 1237 mm/zswap.c .load = zswap_frontswap_load, load 336 net/ipv4/tcp_cong.c int tcp_set_congestion_control(struct sock *sk, const char *name, bool load, load 347 net/ipv4/tcp_cong.c if (!load) load 360 net/ipv4/tcp_cong.c } else if (!load) { load 112 net/sched/em_meta.c static inline unsigned long fixed_loadavg(int load) load 114 net/sched/em_meta.c int rnd_load = load + (FIXED_1/200); load 367 scripts/kconfig/confdata.c goto load; load 382 scripts/kconfig/confdata.c goto load; load 389 scripts/kconfig/confdata.c load: load 25 sound/soc/intel/skylake/skl-sst-utils.c u32 load : 1; load 55 tools/perf/builtin-c2c.c struct stats load; load 153 tools/perf/builtin-c2c.c init_stats(&c2c_he->cstats.load); load 252 tools/perf/builtin-c2c.c else if (stats->load) load 253 tools/perf/builtin-c2c.c update_stats(&cstats->load, weight); load 1202 tools/perf/builtin-c2c.c MEAN_ENTRY(mean_load_entry, load); load 2137 tools/perf/builtin-c2c.c fprintf(out, " Load Operations : %10d\n", stats->load); load 2176 tools/perf/builtin-c2c.c fprintf(out, " Load HITs on shared lines : %10d\n", stats->load); load 328 tools/perf/util/jitdump.c jr->load.pid = bswap_32(jr->load.pid); load 329 tools/perf/util/jitdump.c jr->load.tid = bswap_32(jr->load.tid); load 330 tools/perf/util/jitdump.c jr->load.vma = bswap_64(jr->load.vma); load 331 tools/perf/util/jitdump.c jr->load.code_addr = bswap_64(jr->load.code_addr); load 332 tools/perf/util/jitdump.c jr->load.code_size = bswap_64(jr->load.code_size); load 333 tools/perf/util/jitdump.c jr->load.code_index= bswap_64(jr->load.code_index); load 406 tools/perf/util/jitdump.c pid = jr->load.pid; load 407 tools/perf/util/jitdump.c tid = jr->load.tid; load 408 tools/perf/util/jitdump.c csize = jr->load.code_size; load 410 tools/perf/util/jitdump.c addr = jr->load.code_addr; load 411 tools/perf/util/jitdump.c sym = (void *)((unsigned long)jr + sizeof(jr->load)); load 412 tools/perf/util/jitdump.c code = (unsigned long)jr + jr->load.p.total_size - csize; load 413 tools/perf/util/jitdump.c count = jr->load.code_index; load 475 tools/perf/util/jitdump.c id->time = convert_timestamp(jd, jr->load.p.timestamp); load 567 tools/perf/util/jitdump.c id->time = convert_timestamp(jd, jr->load.p.timestamp); load 119 tools/perf/util/jitdump.h struct jr_code_load load; load 341 tools/perf/util/mem-events.c stats->load++; load 433 tools/perf/util/mem-events.c stats->load += add->load; load 59 tools/perf/util/mem-events.h u32 load; /* count of all loads in trace */ load 32 tools/power/cpupower/bench/benchmark.c unsigned int calculate_timespace(long load, struct config *config) load 41 tools/power/cpupower/bench/benchmark.c printf("calibrating load of %lius, please wait...\n", load); load 53 tools/power/cpupower/bench/benchmark.c rounds = (unsigned int)(load * estimated / timed); load 88 tools/power/cpupower/bench/benchmark.c load_time = config->load; load 92 tools/power/cpupower/bench/benchmark.c total_time += _round * (config->sleep + config->load); load 97 tools/power/cpupower/bench/main.c sscanf(optarg, "%li", &config->load); load 169 tools/power/cpupower/bench/main.c config->load, load 127 tools/power/cpupower/bench/parse.c config->load = 500000; load 187 tools/power/cpupower/bench/parse.c sscanf(val, "%li", &config->load); load 11 tools/power/cpupower/bench/parse.h long load; /* load time in µs */ load 136 tools/power/cpupower/bench/system.c (config->load + config->load_step * round) + load 137 tools/power/cpupower/bench/system.c (config->load + config->load_step * round * 4); load 96 tools/testing/selftests/bpf/test_verifier_log.c ret = load(log, log_len, 1); load 128 tools/testing/selftests/bpf/test_verifier_log.c ret = load(log, log_len, log_level); load 13 tools/testing/selftests/powerpc/ptrace/ptrace-vsx.h int validate_vsx(unsigned long *vsx, unsigned long *load) load 18 tools/testing/selftests/powerpc/ptrace/ptrace-vsx.h if (vsx[i] != load[2 * i + 1]) { load 20 tools/testing/selftests/powerpc/ptrace/ptrace-vsx.h i, vsx[i], 2 * i + 1, load[2 * i + 1]); load 31 tools/testing/selftests/powerpc/ptrace/ptrace-vsx.h int validate_vmx(unsigned long vmx[][2], unsigned long *load) load 37 tools/testing/selftests/powerpc/ptrace/ptrace-vsx.h if ((vmx[i][0] != load[64 + 2 * i]) || load 38 tools/testing/selftests/powerpc/ptrace/ptrace-vsx.h (vmx[i][1] != load[65 + 2 * i])) { load 41 tools/testing/selftests/powerpc/ptrace/ptrace-vsx.h load[64 + 2 * i]); load 44 tools/testing/selftests/powerpc/ptrace/ptrace-vsx.h load[65 + 2 * i]); load 51 tools/testing/selftests/powerpc/ptrace/ptrace-vsx.h if ((vmx[i][0] != load[65 + 2 * i]) || load 52 tools/testing/selftests/powerpc/ptrace/ptrace-vsx.h (vmx[i][1] != load[64 + 2 * i])) { load 55 tools/testing/selftests/powerpc/ptrace/ptrace-vsx.h load[65 + 2 * i]); load 58 tools/testing/selftests/powerpc/ptrace/ptrace-vsx.h load[64 + 2 * i]); load 70 tools/testing/selftests/powerpc/ptrace/ptrace-vsx.h int compare_vsx_vmx(unsigned long *store, unsigned long *load) load 75 tools/testing/selftests/powerpc/ptrace/ptrace-vsx.h if (store[1 + 2 * i] != load[1 + 2 * i]) { load 78 tools/testing/selftests/powerpc/ptrace/ptrace-vsx.h 1 + 2 * i, load[i]); load 85 tools/testing/selftests/powerpc/ptrace/ptrace-vsx.h if (store[i] != load[i]) { load 87 tools/testing/selftests/powerpc/ptrace/ptrace-vsx.h i, store[i], i, load[i]); load 93 tools/testing/selftests/powerpc/ptrace/ptrace-vsx.h if (!(i % 2) && (store[i] != load[i+1])) { load 95 tools/testing/selftests/powerpc/ptrace/ptrace-vsx.h i, store[i], i+1, load[i+1]); load 98 tools/testing/selftests/powerpc/ptrace/ptrace-vsx.h if ((i % 2) && (store[i] != load[i-1])) { load 100 tools/testing/selftests/powerpc/ptrace/ptrace-vsx.h i, store[i], i-1, load[i-1]); load 108 tools/testing/selftests/powerpc/ptrace/ptrace-vsx.h void load_vsx_vmx(unsigned long *load, unsigned long *vsx, load 114 tools/testing/selftests/powerpc/ptrace/ptrace-vsx.h vsx[i] = load[1 + 2 * i]; load 117 tools/testing/selftests/powerpc/ptrace/ptrace-vsx.h vmx[i][0] = load[64 + 2 * i]; load 118 tools/testing/selftests/powerpc/ptrace/ptrace-vsx.h vmx[i][1] = load[65 + 2 * i]; load 39 tools/testing/selftests/powerpc/security/rfi_flush.c load(p + j); load 170 tools/testing/selftests/rseq/basic_percpu_ops_test.c intptr_t *targetptr, expectnot, *load; load 178 tools/testing/selftests/rseq/basic_percpu_ops_test.c load = (intptr_t *)&head; load 180 tools/testing/selftests/rseq/basic_percpu_ops_test.c offset, load, cpu); load 551 tools/testing/selftests/rseq/param_test.c intptr_t *targetptr, expectnot, *load; load 559 tools/testing/selftests/rseq/param_test.c load = (intptr_t *)&head; load 561 tools/testing/selftests/rseq/param_test.c offset, load, cpu); load 220 tools/testing/selftests/rseq/rseq-arm.h off_t voffp, intptr_t *load, int cpu) load 264 tools/testing/selftests/rseq/rseq-arm.h [load] "m" (*load) load 262 tools/testing/selftests/rseq/rseq-arm64.h off_t voffp, intptr_t *load, int cpu) load 283 tools/testing/selftests/rseq/rseq-arm64.h RSEQ_ASM_OP_R_STORE(load) load 294 tools/testing/selftests/rseq/rseq-arm64.h [load] "Qo" (*load), load 225 tools/testing/selftests/rseq/rseq-mips.h off_t voffp, intptr_t *load, int cpu) load 267 tools/testing/selftests/rseq/rseq-mips.h [load] "m" (*load) load 267 tools/testing/selftests/rseq/rseq-ppc.h off_t voffp, intptr_t *load, int cpu) load 295 tools/testing/selftests/rseq/rseq-ppc.h RSEQ_ASM_OP_R_STORE(load) load 310 tools/testing/selftests/rseq/rseq-ppc.h [load] "m" (*load) load 201 tools/testing/selftests/rseq/rseq-s390.h off_t voffp, intptr_t *load, int cpu) load 242 tools/testing/selftests/rseq/rseq-s390.h [load] "m" (*load) load 16 tools/testing/selftests/rseq/rseq-skip.h off_t voffp, intptr_t *load, int cpu) load 175 tools/testing/selftests/rseq/rseq-x86.h off_t voffp, intptr_t *load, int cpu) load 215 tools/testing/selftests/rseq/rseq-x86.h [load] "m" (*load) load 696 tools/testing/selftests/rseq/rseq-x86.h off_t voffp, intptr_t *load, int cpu) load 736 tools/testing/selftests/rseq/rseq-x86.h [load] "m" (*load)