tmpp 121 arch/x86/kernel/cpu/mce/genpool.c struct gen_pool *tmpp; tmpp 124 arch/x86/kernel/cpu/mce/genpool.c tmpp = gen_pool_create(ilog2(sizeof(struct mce_evt_llist)), -1); tmpp 125 arch/x86/kernel/cpu/mce/genpool.c if (!tmpp) tmpp 128 arch/x86/kernel/cpu/mce/genpool.c ret = gen_pool_add(tmpp, (unsigned long)gen_pool_buf, MCE_POOLSZ, -1); tmpp 130 arch/x86/kernel/cpu/mce/genpool.c gen_pool_destroy(tmpp); tmpp 134 arch/x86/kernel/cpu/mce/genpool.c mce_evt_pool = tmpp; tmpp 1327 drivers/infiniband/hw/qib/qib_init.c struct page **tmpp = dd->pageshadow; tmpp 1336 drivers/infiniband/hw/qib/qib_init.c if (!tmpp[i]) tmpp 1340 drivers/infiniband/hw/qib/qib_init.c qib_release_user_pages(&tmpp[i], 1); tmpp 1341 drivers/infiniband/hw/qib/qib_init.c tmpp[i] = NULL; tmpp 1346 drivers/infiniband/hw/qib/qib_init.c vfree(tmpp); tmpp 459 drivers/soc/fsl/qbman/qman_ccsr.c void __iomem *tmpp = ioremap_cache(addr, sz); tmpp 461 drivers/soc/fsl/qbman/qman_ccsr.c if (!tmpp) tmpp 464 drivers/soc/fsl/qbman/qman_ccsr.c memset_io(tmpp, 0, sz); tmpp 465 drivers/soc/fsl/qbman/qman_ccsr.c flush_dcache_range((unsigned long)tmpp, tmpp 466 drivers/soc/fsl/qbman/qman_ccsr.c (unsigned long)tmpp + sz); tmpp 467 drivers/soc/fsl/qbman/qman_ccsr.c iounmap(tmpp); tmpp 95 fs/erofs/decompressor.c u8 *tmpp = tmp; tmpp 99 fs/erofs/decompressor.c while (tmpp < tmp + inlen) { tmpp 102 fs/erofs/decompressor.c memcpy(tmpp, src + pageofs_in, count); tmpp 105 fs/erofs/decompressor.c tmpp += count; tmpp 116 fs/nfsd/nfs4xdr.c if (!(x = (p==argp->tmp || p == argp->tmpp) ? \ tmpp 189 fs/nfsd/nfs4xdr.c kfree(argp->tmpp); tmpp 190 fs/nfsd/nfs4xdr.c p = argp->tmpp = kmalloc(nbytes, GFP_KERNEL); tmpp 4541 fs/nfsd/nfs4xdr.c kfree(args->tmpp); tmpp 4542 fs/nfsd/nfs4xdr.c args->tmpp = NULL; tmpp 4566 fs/nfsd/nfs4xdr.c args->tmpp = NULL; tmpp 653 fs/nfsd/xdr4.h __be32 * tmpp;