src_paddr        6735 arch/x86/kvm/svm.c static int __sev_dbg_decrypt(struct kvm *kvm, unsigned long src_paddr,
src_paddr        6744 arch/x86/kvm/svm.c 	src_paddr = round_down(src_paddr, 16);
src_paddr        6745 arch/x86/kvm/svm.c 	offset = src_paddr & 15;
src_paddr        6748 arch/x86/kvm/svm.c 	return __sev_issue_dbg_cmd(kvm, src_paddr, dst_paddr, sz, err, false);
src_paddr         108 arch/xtensa/mm/cache.c 	unsigned long dst_paddr, src_paddr;
src_paddr         112 arch/xtensa/mm/cache.c 					  &src_paddr);
src_paddr         117 arch/xtensa/mm/cache.c 	copy_page_alias(dst_vaddr, src_vaddr, dst_paddr, src_paddr);
src_paddr         680 drivers/crypto/n2_core.c 		unsigned long	src_paddr : 44;
src_paddr         848 drivers/crypto/n2_core.c 	in_place = (cp->dest_paddr == cp->arr[0].src_paddr);
src_paddr         855 drivers/crypto/n2_core.c 	ent->src_addr = cp->arr[0].src_paddr;
src_paddr         867 drivers/crypto/n2_core.c 		ent->src_addr = cp->arr[i].src_paddr;
src_paddr         909 drivers/crypto/n2_core.c 		unsigned long dest_paddr, src_paddr;
src_paddr         913 drivers/crypto/n2_core.c 		src_paddr = (page_to_phys(walk->src.page) +
src_paddr         917 drivers/crypto/n2_core.c 		in_place = (src_paddr == dest_paddr);
src_paddr         941 drivers/crypto/n2_core.c 		chunk->arr[chunk->arr_len].src_paddr = src_paddr;
src_paddr        1072 drivers/crypto/n2_core.c 				iv_paddr = (tmp->arr[tmp->arr_len-1].src_paddr +
src_paddr        1079 drivers/crypto/n2_core.c 				pa = (c->arr[c->arr_len-1].src_paddr +
src_paddr         894 drivers/crypto/ux500/cryp/cryp_core.c 	unsigned long src_paddr;
src_paddr         916 drivers/crypto/ux500/cryp/cryp_core.c 		src_paddr = (page_to_phys(walk.src.page) + walk.src.offset);
src_paddr         917 drivers/crypto/ux500/cryp/cryp_core.c 		ctx->indata = phys_to_virt(src_paddr);