Lines Matching refs:arch
93 vcpu->kvm->arch.vgic.vm_ops.add_sgi_source(vcpu, irq, source); in add_sgi_source()
98 return vcpu->kvm->arch.vgic.vm_ops.queue_sgi(vcpu, irq); in queue_sgi()
103 return kvm->arch.vgic.vm_ops.map_resources(kvm, vgic); in kvm_vgic_map_resources()
248 struct vgic_dist *dist = &vcpu->kvm->arch.vgic; in vgic_irq_is_edge()
257 struct vgic_dist *dist = &vcpu->kvm->arch.vgic; in vgic_irq_is_enabled()
264 struct vgic_dist *dist = &vcpu->kvm->arch.vgic; in vgic_irq_is_queued()
271 struct vgic_dist *dist = &vcpu->kvm->arch.vgic; in vgic_irq_is_active()
278 struct vgic_dist *dist = &vcpu->kvm->arch.vgic; in vgic_irq_set_queued()
285 struct vgic_dist *dist = &vcpu->kvm->arch.vgic; in vgic_irq_clear_queued()
292 struct vgic_dist *dist = &vcpu->kvm->arch.vgic; in vgic_irq_set_active()
299 struct vgic_dist *dist = &vcpu->kvm->arch.vgic; in vgic_irq_clear_active()
306 struct vgic_dist *dist = &vcpu->kvm->arch.vgic; in vgic_dist_irq_get_level()
313 struct vgic_dist *dist = &vcpu->kvm->arch.vgic; in vgic_dist_irq_set_level()
320 struct vgic_dist *dist = &vcpu->kvm->arch.vgic; in vgic_dist_irq_clear_level()
327 struct vgic_dist *dist = &vcpu->kvm->arch.vgic; in vgic_dist_irq_soft_pend()
334 struct vgic_dist *dist = &vcpu->kvm->arch.vgic; in vgic_dist_irq_clear_soft_pend()
341 struct vgic_dist *dist = &vcpu->kvm->arch.vgic; in vgic_dist_irq_is_pending()
348 struct vgic_dist *dist = &vcpu->kvm->arch.vgic; in vgic_dist_irq_set_pending()
355 struct vgic_dist *dist = &vcpu->kvm->arch.vgic; in vgic_dist_irq_clear_pending()
363 set_bit(irq, vcpu->arch.vgic_cpu.pending_percpu); in vgic_cpu_irq_set()
366 vcpu->arch.vgic_cpu.pending_shared); in vgic_cpu_irq_set()
372 clear_bit(irq, vcpu->arch.vgic_cpu.pending_percpu); in vgic_cpu_irq_clear()
375 vcpu->arch.vgic_cpu.pending_shared); in vgic_cpu_irq_clear()
459 reg = vgic_bitmap_get_reg(&kvm->arch.vgic.irq_enabled, vcpu_id, offset); in vgic_handle_enable_reg()
481 struct vgic_dist *dist = &kvm->arch.vgic; in vgic_handle_set_pending_reg()
518 struct vgic_dist *dist = &kvm->arch.vgic; in vgic_handle_clear_pending_reg()
552 struct vgic_dist *dist = &kvm->arch.vgic; in vgic_handle_set_active_reg()
571 struct vgic_dist *dist = &kvm->arch.vgic; in vgic_handle_clear_active_reg()
662 struct vgic_cpu *vgic_cpu = &vcpu->arch.vgic_cpu; in vgic_unqueue_irqs()
803 struct vgic_dist *dist = &vcpu->kvm->arch.vgic; in vgic_handle_mmio_access()
922 struct vgic_dist *dist = &vcpu->kvm->arch.vgic; in compute_active_for_cpu()
929 act_percpu = vcpu->arch.vgic_cpu.active_percpu; in compute_active_for_cpu()
930 act_shared = vcpu->arch.vgic_cpu.active_shared; in compute_active_for_cpu()
952 struct vgic_dist *dist = &vcpu->kvm->arch.vgic; in compute_pending_for_cpu()
959 pend_percpu = vcpu->arch.vgic_cpu.pending_percpu; in compute_pending_for_cpu()
960 pend_shared = vcpu->arch.vgic_cpu.pending_shared; in compute_pending_for_cpu()
985 struct vgic_dist *dist = &kvm->arch.vgic; in vgic_update_state()
1069 struct vgic_cpu *vgic_cpu = &vcpu->arch.vgic_cpu; in vgic_retire_lr()
1090 struct vgic_cpu *vgic_cpu = &vcpu->arch.vgic_cpu; in vgic_retire_disabled_irqs()
1131 struct vgic_cpu *vgic_cpu = &vcpu->arch.vgic_cpu; in vgic_queue_irq()
1132 struct vgic_dist *dist = &vcpu->kvm->arch.vgic; in vgic_queue_irq()
1199 struct vgic_cpu *vgic_cpu = &vcpu->arch.vgic_cpu; in __kvm_vgic_flush_hwstate()
1200 struct vgic_dist *dist = &vcpu->kvm->arch.vgic; in __kvm_vgic_flush_hwstate()
1208 pa_percpu = vcpu->arch.vgic_cpu.pend_act_percpu; in __kvm_vgic_flush_hwstate()
1209 pa_shared = vcpu->arch.vgic_cpu.pend_act_shared; in __kvm_vgic_flush_hwstate()
1262 struct vgic_dist *dist = &vcpu->kvm->arch.vgic; in vgic_process_maintenance()
1347 struct vgic_cpu *vgic_cpu = &vcpu->arch.vgic_cpu; in __kvm_vgic_sync_hwstate()
1348 struct vgic_dist *dist = &vcpu->kvm->arch.vgic; in __kvm_vgic_sync_hwstate()
1379 struct vgic_dist *dist = &vcpu->kvm->arch.vgic; in kvm_vgic_flush_hwstate()
1399 struct vgic_dist *dist = &vcpu->kvm->arch.vgic; in kvm_vgic_vcpu_pending_irq()
1409 struct vgic_dist *dist = &vcpu->kvm->arch.vgic; in kvm_vgic_vcpu_active_irq()
1454 struct vgic_dist *dist = &kvm->arch.vgic; in vgic_update_irq_pending()
1552 if (kvm->arch.vgic.vgic_model != KVM_DEV_TYPE_ARM_VGIC_V2) { in kvm_vgic_inject_irq()
1564 if (irq_num >= min(kvm->arch.vgic.nr_irqs, 1020)) in kvm_vgic_inject_irq()
1590 struct vgic_cpu *vgic_cpu = &vcpu->arch.vgic_cpu; in kvm_vgic_vcpu_destroy()
1604 struct vgic_cpu *vgic_cpu = &vcpu->arch.vgic_cpu; in vgic_vcpu_init_maps()
1645 struct vgic_dist *dist = &kvm->arch.vgic; in kvm_vgic_destroy()
1683 struct vgic_dist *dist = &kvm->arch.vgic; in vgic_init()
1740 ret = kvm->arch.vgic.vm_ops.init_model(kvm); in vgic_init()
1786 if (atomic_read(&kvm->online_vcpus) > kvm->arch.max_vcpus) in init_vgic_model()
1828 if (vcpu->arch.has_run_once) in kvm_vgic_create()
1837 spin_lock_init(&kvm->arch.vgic.lock); in kvm_vgic_create()
1838 kvm->arch.vgic.in_kernel = true; in kvm_vgic_create()
1839 kvm->arch.vgic.vgic_model = type; in kvm_vgic_create()
1840 kvm->arch.vgic.vctrl_base = vgic->vctrl_base; in kvm_vgic_create()
1841 kvm->arch.vgic.vgic_dist_base = VGIC_ADDR_UNDEF; in kvm_vgic_create()
1842 kvm->arch.vgic.vgic_cpu_base = VGIC_ADDR_UNDEF; in kvm_vgic_create()
1843 kvm->arch.vgic.vgic_redist_base = VGIC_ADDR_UNDEF; in kvm_vgic_create()
1858 phys_addr_t dist = kvm->arch.vgic.vgic_dist_base; in vgic_ioaddr_overlap()
1859 phys_addr_t cpu = kvm->arch.vgic.vgic_cpu_base; in vgic_ioaddr_overlap()
1909 struct vgic_dist *vgic = &kvm->arch.vgic; in kvm_vgic_addr()
2004 if (vgic_ready(dev->kvm) || dev->kvm->arch.vgic.nr_irqs) in vgic_set_common_attr()
2007 dev->kvm->arch.vgic.nr_irqs = val; in vgic_set_common_attr()
2047 r = put_user(dev->kvm->arch.vgic.nr_irqs, uaddr); in vgic_get_common_attr()