/xen/xen/arch/arm/ |
A D | gic-vgic.c | 53 ASSERT(spin_is_locked(&v->arch.vgic.lock)); in gic_add_to_lr_pending() 71 ASSERT(spin_is_locked(&v->arch.vgic.lock)); in gic_remove_from_lr_pending() 84 ASSERT(spin_is_locked(&v->arch.vgic.lock)); in gic_raise_inflight_irq() 117 ASSERT(spin_is_locked(&v->arch.vgic.lock)); in gic_find_unused_lr() 143 ASSERT(spin_is_locked(&v->arch.vgic.lock)); in gic_raise_guest_irq() 169 ASSERT(spin_is_locked(&v->arch.vgic.lock)); in gic_update_one_lr() 338 ASSERT(spin_is_locked(&v->arch.vgic.lock)); in gic_clear_pending_irqs()
|
A D | gic.c | 106 ASSERT(spin_is_locked(&desc->lock)); in gic_set_irq_type() 125 ASSERT(spin_is_locked(&desc->lock)); in gic_route_irq_to_xen() 141 ASSERT(spin_is_locked(&desc->lock)); in gic_route_irq_to_guest() 175 ASSERT(spin_is_locked(&desc->lock)); in gic_remove_irq_from_guest()
|
A D | vgic-v3-its.c | 130 ASSERT(spin_is_locked(&its->its_lock)); in its_set_collection() 148 ASSERT(spin_is_locked(&its->its_lock)); in get_vcpu_from_collection() 234 ASSERT(spin_is_locked(&its->its_lock)); in read_itte() 265 ASSERT(spin_is_locked(&its->its_lock)); in write_itte() 444 ASSERT(spin_is_locked(&v->arch.vgic.lock)); in update_lpi_vgic_status() 595 ASSERT(spin_is_locked(&its->its_lock)); in its_discard_event() 915 ASSERT(spin_is_locked(&its->vcmd_lock)); in vgic_its_handle_cmds() 1167 ASSERT(spin_is_locked(&its->its_lock)); in vgic_v3_verify_its_status()
|
A D | gic-v2.c | 264 ASSERT(spin_is_locked(&irqd->lock)); in gicv2_set_active_state() 280 ASSERT(spin_is_locked(&irqd->lock)); in gicv2_set_pending_state() 617 ASSERT(spin_is_locked(&desc->lock)); in gicv2_irq_enable() 631 ASSERT(spin_is_locked(&desc->lock)); in gicv2_irq_disable()
|
A D | vgic-v2.c | 76 ASSERT(spin_is_locked(&rank->lock)); in vgic_fetch_itargetsr() 99 ASSERT(spin_is_locked(&rank->lock)); in vgic_store_itargetsr()
|
A D | vgic.c | 378 ASSERT(spin_is_locked(&r->lock)); in vgic_get_virq_type() 521 ASSERT(spin_is_locked(&v->arch.vgic.lock)); in vgic_remove_irq_from_queues()
|
A D | gic-v3.c | 500 ASSERT(spin_is_locked(&irqd->lock)); in gicv3_set_active_state() 516 ASSERT(spin_is_locked(&irqd->lock)); in gicv3_set_pending_state() 1187 ASSERT(spin_is_locked(&desc->lock)); in gicv3_irq_enable() 1201 ASSERT(spin_is_locked(&desc->lock)); in gicv3_irq_disable()
|
A D | gic-v3-lpi.c | 424 ASSERT(spin_is_locked(&lpi_data.host_lpis_lock)); in find_unused_host_lpi()
|
/xen/xen/arch/x86/ |
A D | emul-i8254.c | 66 ASSERT(spin_is_locked(&pit->lock)); in pit_get_count() 97 ASSERT(spin_is_locked(&pit->lock)); in pit_get_out() 131 ASSERT(spin_is_locked(&pit->lock)); in pit_set_gate() 155 ASSERT(spin_is_locked(&pit->lock)); in pit_get_gate() 172 ASSERT(spin_is_locked(&pit->lock)); in pit_load_count() 214 ASSERT(spin_is_locked(&pit->lock)); in pit_latch_count() 227 ASSERT(spin_is_locked(&pit->lock)); in pit_latch_status()
|
A D | irq.c | 873 ASSERT(spin_is_locked(&desc->lock)); in irq_set_affinity() 1547 WARN_ON(!spin_is_locked(&v->domain->event_lock)); in pirq_guest_bind() 1767 WARN_ON(!spin_is_locked(&d->event_lock)); in pirq_guest_unbind() 1804 WARN_ON(!spin_is_locked(&d->event_lock)); in pirq_guest_force_unbind() 2048 ASSERT(spin_is_locked(&d->event_lock)); in get_free_pirq() 2073 ASSERT(spin_is_locked(&d->event_lock)); in get_free_pirqs() 2101 ASSERT(spin_is_locked(&d->event_lock)); in map_domain_pirq() 2320 ASSERT(spin_is_locked(&d->event_lock)); in unmap_domain_pirq() 2694 ASSERT(spin_is_locked(&d->event_lock)); in map_domain_emuirq_pirq() 2760 ASSERT(spin_is_locked(&d->event_lock)); in unmap_domain_pirq_emuirq() [all …]
|
/xen/xen/arch/arm/vgic/ |
A D | vgic.c | 169 ASSERT(spin_is_locked(&irq->irq_lock)); in vgic_target_oracle() 250 ASSERT(spin_is_locked(&vgic_cpu->ap_list_lock)); in vgic_sort_ap_list() 286 ASSERT(spin_is_locked(&irq->irq_lock)); in vgic_queue_irq_unlock() 530 ASSERT(spin_is_locked(&irq->irq_lock)); in vgic_populate_lr() 549 ASSERT(spin_is_locked(&vgic_cpu->ap_list_lock)); in compute_ap_list_depth() 564 ASSERT(spin_is_locked(&vgic_cpu->ap_list_lock)); in vgic_flush_lr_state()
|
/xen/xen/drivers/passthrough/amd/ |
A D | iommu_cmd.c | 366 ASSERT( spin_is_locked(&iommu->lock) ); in amd_iommu_flush_device() 374 ASSERT( spin_is_locked(&iommu->lock) ); in amd_iommu_flush_intremap() 382 ASSERT( spin_is_locked(&iommu->lock) ); in amd_iommu_flush_all_caches()
|
/xen/xen/arch/x86/hvm/ |
A D | rtc.c | 65 ASSERT(spin_is_locked(&s->lock)); in rtc_update_irq() 130 ASSERT(spin_is_locked(&s->lock)); in rtc_timer_update() 185 ASSERT(spin_is_locked(&s->lock)); in check_update_timer() 257 ASSERT(spin_is_locked(&s->lock)); in alarm_timer_update() 580 ASSERT(spin_is_locked(&s->lock)); in rtc_set_time() 608 ASSERT(spin_is_locked(&s->lock)); in rtc_copy_date()
|
A D | pmtimer.c | 61 ASSERT(spin_is_locked(&s->lock)); in pmt_update_sci() 103 ASSERT(spin_is_locked(&s->lock)); in pmt_update_time()
|
A D | vioapic.c | 411 ASSERT(spin_is_locked(&d->arch.hvm.irq_lock)); in vioapic_deliver() 499 ASSERT(spin_is_locked(&d->arch.hvm.irq_lock)); in vioapic_irq_positive_edge()
|
A D | vmsi.c | 468 ASSERT(spin_is_locked(&d->event_lock)); in msixtbl_pt_register() 538 ASSERT(spin_is_locked(&d->event_lock)); in msixtbl_pt_unregister()
|
A D | vpic.c | 43 #define vpic_is_locked(v) spin_is_locked(__vpic_lock(v))
|
/xen/xen/include/xen/ |
A D | spinlock.h | 197 #define spin_is_locked(l) _spin_is_locked(l) macro
|
/xen/xen/drivers/passthrough/vtd/ |
A D | intremap.c | 182 ASSERT(spin_is_locked(&iommu->intremap.lock)); in update_irte() 227 ASSERT(spin_is_locked(&iommu->intremap.lock)); in free_remap_entry() 249 ASSERT(spin_is_locked(&iommu->intremap.lock)); in alloc_remap_entry()
|
/xen/xen/common/sched/ |
A D | credit2.c | 1430 ASSERT(spin_is_locked(get_sched_res(cpu)->schedule_lock)); in runq_insert() 1909 ASSERT(spin_is_locked(&svc->sdom->budget_lock)); in park_unit() 1939 ASSERT(spin_is_locked(get_sched_res(cpu)->schedule_lock)); in unit_grab_budget() 1986 ASSERT(spin_is_locked(get_sched_res(cpu)->schedule_lock)); in unit_return_budget() 2240 ASSERT(spin_is_locked(get_sched_res(cpu)->schedule_lock)); in csched2_unit_wake() 2366 ASSERT(spin_is_locked(get_sched_res(cpu)->schedule_lock)); in csched2_res_pick() 2703 ASSERT(spin_is_locked(get_sched_res(cpu)->schedule_lock)); in balance_load() 3560 ASSERT(spin_is_locked(get_sched_res(sched_cpu)->schedule_lock)); in csched2_schedule()
|
A D | credit.c | 254 ASSERT(spin_is_locked(get_sched_res(cpu)->schedule_lock)); in is_runq_idle() 263 ASSERT(spin_is_locked(get_sched_res(cpu)->schedule_lock)); in inc_nr_runnable() 271 ASSERT(spin_is_locked(get_sched_res(cpu)->schedule_lock)); in dec_nr_runnable() 585 ASSERT(spin_is_locked(&prv->lock)); in init_pdata()
|
/xen/xen/common/ |
A D | livepatch.c | 142 ASSERT(spin_is_locked(&payload_lock)); in livepatch_symbols_lookup_by_name() 237 ASSERT(spin_is_locked(&payload_lock)); in find_payload() 1003 ASSERT(spin_is_locked(&payload_lock)); in free_payload() 1520 ASSERT(spin_is_locked(&payload_lock)); in is_work_scheduled() 1560 ASSERT(spin_is_locked(&payload_lock)); in schedule_work()
|
/xen/tools/tests/vhpet/ |
A D | emul.h | 337 #define spin_is_locked(a) 1 macro
|
/xen/xen/drivers/char/ |
A D | console.c | 330 ASSERT(spin_is_locked(&console_lock)); in conring_puts() 741 ASSERT(spin_is_locked(&console_lock)); in __putstr()
|
/xen/xen/drivers/passthrough/ |
A D | io.c | 108 ASSERT(spin_is_locked(&d->event_lock)); in pt_pirq_softirq_reset() 802 ASSERT(spin_is_locked(&d->event_lock)); in pt_pirq_iterate()
|