src_paddr 6735 arch/x86/kvm/svm.c static int __sev_dbg_decrypt(struct kvm *kvm, unsigned long src_paddr, src_paddr 6744 arch/x86/kvm/svm.c src_paddr = round_down(src_paddr, 16); src_paddr 6745 arch/x86/kvm/svm.c offset = src_paddr & 15; src_paddr 6748 arch/x86/kvm/svm.c return __sev_issue_dbg_cmd(kvm, src_paddr, dst_paddr, sz, err, false); src_paddr 108 arch/xtensa/mm/cache.c unsigned long dst_paddr, src_paddr; src_paddr 112 arch/xtensa/mm/cache.c &src_paddr); src_paddr 117 arch/xtensa/mm/cache.c copy_page_alias(dst_vaddr, src_vaddr, dst_paddr, src_paddr); src_paddr 680 drivers/crypto/n2_core.c unsigned long src_paddr : 44; src_paddr 848 drivers/crypto/n2_core.c in_place = (cp->dest_paddr == cp->arr[0].src_paddr); src_paddr 855 drivers/crypto/n2_core.c ent->src_addr = cp->arr[0].src_paddr; src_paddr 867 drivers/crypto/n2_core.c ent->src_addr = cp->arr[i].src_paddr; src_paddr 909 drivers/crypto/n2_core.c unsigned long dest_paddr, src_paddr; src_paddr 913 drivers/crypto/n2_core.c src_paddr = (page_to_phys(walk->src.page) + src_paddr 917 drivers/crypto/n2_core.c in_place = (src_paddr == dest_paddr); src_paddr 941 drivers/crypto/n2_core.c chunk->arr[chunk->arr_len].src_paddr = src_paddr; src_paddr 1072 drivers/crypto/n2_core.c iv_paddr = (tmp->arr[tmp->arr_len-1].src_paddr + src_paddr 1079 drivers/crypto/n2_core.c pa = (c->arr[c->arr_len-1].src_paddr + src_paddr 894 drivers/crypto/ux500/cryp/cryp_core.c unsigned long src_paddr; src_paddr 916 drivers/crypto/ux500/cryp/cryp_core.c src_paddr = (page_to_phys(walk.src.page) + walk.src.offset); src_paddr 917 drivers/crypto/ux500/cryp/cryp_core.c ctx->indata = phys_to_virt(src_paddr);