mop 4257 arch/s390/kvm/kvm-s390.c struct kvm_s390_mem_op *mop) mop 4259 arch/s390/kvm/kvm-s390.c void __user *uaddr = (void __user *)mop->buf; mop 4265 arch/s390/kvm/kvm-s390.c if (mop->flags & ~supported_flags || mop->ar >= NUM_ACRS || !mop->size) mop 4268 arch/s390/kvm/kvm-s390.c if (mop->size > MEM_OP_MAX_SIZE) mop 4271 arch/s390/kvm/kvm-s390.c if (!(mop->flags & KVM_S390_MEMOP_F_CHECK_ONLY)) { mop 4272 arch/s390/kvm/kvm-s390.c tmpbuf = vmalloc(mop->size); mop 4279 arch/s390/kvm/kvm-s390.c switch (mop->op) { mop 4281 arch/s390/kvm/kvm-s390.c if (mop->flags & KVM_S390_MEMOP_F_CHECK_ONLY) { mop 4282 arch/s390/kvm/kvm-s390.c r = check_gva_range(vcpu, mop->gaddr, mop->ar, mop 4283 arch/s390/kvm/kvm-s390.c mop->size, GACC_FETCH); mop 4286 arch/s390/kvm/kvm-s390.c r = read_guest(vcpu, mop->gaddr, mop->ar, tmpbuf, mop->size); mop 4288 arch/s390/kvm/kvm-s390.c if (copy_to_user(uaddr, tmpbuf, mop->size)) mop 4293 arch/s390/kvm/kvm-s390.c if (mop->flags & KVM_S390_MEMOP_F_CHECK_ONLY) { mop 4294 arch/s390/kvm/kvm-s390.c r = check_gva_range(vcpu, mop->gaddr, mop->ar, mop 4295 arch/s390/kvm/kvm-s390.c mop->size, GACC_STORE); mop 4298 arch/s390/kvm/kvm-s390.c if (copy_from_user(tmpbuf, uaddr, mop->size)) { mop 4302 arch/s390/kvm/kvm-s390.c r = write_guest(vcpu, mop->gaddr, mop->ar, tmpbuf, mop->size); mop 4310 arch/s390/kvm/kvm-s390.c if (r > 0 && (mop->flags & KVM_S390_MEMOP_F_INJECT_EXCEPTION) != 0) mop 331 drivers/net/xen-netback/netback.c struct gnttab_map_grant_ref *mop) mop 333 drivers/net/xen-netback/netback.c queue->pages_to_map[mop-queue->tx_map_ops] = queue->mmap_pages[pending_idx]; mop 334 drivers/net/xen-netback/netback.c gnttab_set_map_op(mop, idx_to_kaddr(queue, pending_idx), mop 429 drivers/pinctrl/pinctrl-tb10x.c DEFPINFUNCGRP(mop, 5, 0, 0),