/xen/xen/include/xen/ |
A D | rwlock.h | 14 spinlock_t lock; member 59 cnts = atomic_read(&lock->cnts); in _read_trylock() 93 queue_read_lock_slowpath(lock); in _read_lock() 104 _read_lock(lock); in _read_lock_irq() 111 _read_lock(lock); in _read_lock_irqsave() 131 _read_unlock(lock); in _read_unlock_irq() 137 _read_unlock(lock); in _read_unlock_irqrestore() 178 _write_lock(lock); in _write_lock_irq() 186 _write_lock(lock); in _write_lock_irqsave() 225 _write_unlock(lock); in _write_unlock_irq() [all …]
|
A D | spinlock.h | 167 void _spin_lock(spinlock_t *lock); 169 void _spin_lock_irq(spinlock_t *lock); 172 void _spin_unlock(spinlock_t *lock); 173 void _spin_unlock_irq(spinlock_t *lock); 176 int _spin_is_locked(spinlock_t *lock); 177 int _spin_trylock(spinlock_t *lock); 178 void _spin_barrier(spinlock_t *lock); 180 int _spin_trylock_recursive(spinlock_t *lock); 181 void _spin_lock_recursive(spinlock_t *lock); 182 void _spin_unlock_recursive(spinlock_t *lock); [all …]
|
/xen/xen/common/ |
A D | spinlock.c | 117 lock->profile->time_hold += NOW() - lock->profile->time_locked; \ 128 lock->profile->time_block += lock->profile->time_locked - block; \ 187 _spin_lock(lock); in _spin_lock_irq() 195 _spin_lock(lock); in _spin_lock_irqsave() 211 _spin_unlock(lock); in _spin_unlock_irq() 217 _spin_unlock(lock); in _spin_unlock_irqrestore() 231 ? lock->tickets.head != lock->tickets.tail in _spin_is_locked() 254 if (lock->profile) in _spin_trylock() 308 lock->recurse_cnt++; in _spin_trylock_recursive() 325 lock->recurse_cnt++; in _spin_lock_recursive() [all …]
|
A D | rwlock.c | 18 cnts = atomic_read(&lock->cnts); in rspin_until_writer_unlock() 26 void queue_read_lock_slowpath(rwlock_t *lock) in queue_read_lock_slowpath() argument 33 atomic_sub(_QR_BIAS, &lock->cnts); in queue_read_lock_slowpath() 38 spin_lock(&lock->lock); in queue_read_lock_slowpath() 51 rspin_until_writer_unlock(lock, cnts); in queue_read_lock_slowpath() 56 spin_unlock(&lock->lock); in queue_read_lock_slowpath() 68 spin_lock(&lock->lock); in queue_write_lock_slowpath() 71 if ( !atomic_read(&lock->cnts) && in queue_write_lock_slowpath() 81 cnts = atomic_read(&lock->cnts); in queue_write_lock_slowpath() 93 cnts = atomic_read(&lock->cnts); in queue_write_lock_slowpath() [all …]
|
A D | rangeset.c | 31 rwlock_t lock; member 123 write_lock(&r->lock); in rangeset_add_range() 191 write_lock(&r->lock); in rangeset_remove_range() 262 read_lock(&r->lock); in rangeset_contains_range() 265 read_unlock(&r->lock); in rangeset_contains_range() 281 read_lock(&r->lock); in rangeset_overlaps_range() 284 read_unlock(&r->lock); in rangeset_overlaps_range() 296 read_lock(&r->lock); in rangeset_report_ranges() 313 write_lock(&r->lock); in rangeset_claim_range() 366 write_lock(&r->lock); in rangeset_consume_ranges() [all …]
|
A D | timer.c | 33 spinlock_t lock; member 446 spin_unlock_irq(&ts->lock); in execute_timer() 448 spin_lock_irq(&ts->lock); in execute_timer() 479 spin_lock_irq(&ts->lock); in timer_softirq_action() 490 spin_lock_irq(&ts->lock); in timer_softirq_action() 532 spin_unlock_irq(&ts->lock); in timer_softirq_action() 587 spin_lock_irq(&old_ts->lock); in migrate_timers_from_cpu() 588 spin_lock(&new_ts->lock); in migrate_timers_from_cpu() 593 spin_lock(&old_ts->lock); in migrate_timers_from_cpu() 612 spin_unlock(&old_ts->lock); in migrate_timers_from_cpu() [all …]
|
A D | wait.c | 84 spin_lock_init(&wq->lock); in init_waitqueue_head() 97 spin_lock(&wq->lock); in wake_up_nr() 107 spin_unlock(&wq->lock); in wake_up_nr() 236 spin_lock(&wq->lock); in prepare_to_wait() 240 spin_unlock(&wq->lock); in prepare_to_wait() 253 spin_lock(&wq->lock); in finish_wait() 260 spin_unlock(&wq->lock); in finish_wait()
|
/xen/tools/libxl/ |
A D | libxl_internal.c | 457 return lock; in libxl__lock_file() 460 if (lock) libxl__unlock_file(lock); in libxl__lock_file() 480 if (lock->path) unlink(lock->path); in libxl__unlock_file() 481 if (lock->carefd) libxl__carefd_close(lock->carefd); in libxl__unlock_file() 483 free(lock); in libxl__unlock_file() 636 lock->callback(egc, lock, ERROR_LOCK_FAIL); in libxl__ev_slowlock_lock() 707 lock->callback(egc, lock, ERROR_LOCK_FAIL); in ev_lock_prepare_fork() 766 lock->callback(egc, lock, rc); in ev_lock_child_callback() 778 if (lock->path && lock->held) in libxl__ev_slowlock_unlock() 787 lock->fd, lock->path); in libxl__ev_slowlock_unlock() [all …]
|
/xen/xen/arch/x86/mm/ |
A D | mm-locks.h | 37 spin_lock_init(&l->lock); in mm_lock_init() 45 return (l->lock.recurse_cpu == current->processor); in mm_locked_by_me() 96 spin_lock_recursive(&l->lock); in _mm_lock() 97 if ( l->lock.recurse_cnt == 1 ) in _mm_lock() 171 percpu_read_lock(p2m_percpu_rwlock, &l->lock); in _mm_read_lock() 178 percpu_read_unlock(p2m_percpu_rwlock, &l->lock); in mm_read_unlock() 212 if ( l->lock.recurse_cnt == 1 ) in mm_unlock() 217 spin_unlock_recursive(&l->lock); in mm_unlock() 327 mm_write_lock(altp2m, p->domain, &p->lock); in p2m_lock() 329 mm_write_lock(p2m, p->domain, &p->lock); in p2m_lock() [all …]
|
/xen/xen/arch/x86/hvm/ |
A D | pmtimer.c | 76 spin_lock(&s->lock); in hvm_acpi_power_button() 79 spin_unlock(&s->lock); in hvm_acpi_power_button() 89 spin_lock(&s->lock); in hvm_acpi_sleep_button() 92 spin_unlock(&s->lock); in hvm_acpi_sleep_button() 133 spin_lock(&s->lock); in pmt_timer_callback() 151 spin_unlock(&s->lock); in pmt_timer_callback() 169 spin_lock(&s->lock); in handle_evt_io() 212 spin_unlock(&s->lock); in handle_evt_io() 263 spin_lock(&s->lock); in acpi_save() 279 spin_unlock(&s->lock); in acpi_save() [all …]
|
A D | rtc.c | 88 spin_lock(&s->lock); in rtc_pf_callback() 102 spin_unlock(&s->lock); in rtc_pf_callback() 224 spin_lock(&s->lock); in rtc_update_timer() 237 spin_lock(&s->lock); in rtc_update_timer2() 428 spin_lock(&s->lock); in rtc_alarm_cb() 444 spin_lock(&s->lock); in rtc_ioport_write() 655 spin_lock(&s->lock); in rtc_ioport_read() 750 spin_lock(&s->lock); in rtc_save() 766 spin_lock(&s->lock); in rtc_load() 854 spin_lock(&s->lock); in rtc_init() [all …]
|
A D | ioreq.c | 508 spin_lock(&s->lock); in hvm_ioreq_server_add_vcpu() 534 spin_unlock(&s->lock); in hvm_ioreq_server_add_vcpu() 541 spin_unlock(&s->lock); in hvm_ioreq_server_add_vcpu() 553 spin_lock(&s->lock); in hvm_ioreq_server_remove_vcpu() 573 spin_unlock(&s->lock); in hvm_ioreq_server_remove_vcpu() 580 spin_lock(&s->lock); in hvm_ioreq_server_remove_all_vcpus() 599 spin_unlock(&s->lock); in hvm_ioreq_server_remove_all_vcpus() 694 spin_lock(&s->lock); in hvm_ioreq_server_enable() 710 spin_unlock(&s->lock); in hvm_ioreq_server_enable() 715 spin_lock(&s->lock); in hvm_ioreq_server_disable() [all …]
|
/xen/xen/arch/x86/ |
A D | emul-i8254.c | 248 spin_lock(&pit->lock); in pit_ioport_write() 314 spin_unlock(&pit->lock); in pit_ioport_write() 325 spin_lock(&pit->lock); in pit_ioport_read() 377 spin_unlock(&pit->lock); in pit_ioport_read() 389 spin_lock(&pit->lock); in pit_stop_channel0_irq() 391 spin_unlock(&pit->lock); in pit_stop_channel0_irq() 403 spin_lock(&pit->lock); in pit_save() 407 spin_unlock(&pit->lock); in pit_save() 420 spin_lock(&pit->lock); in pit_load() 461 spin_lock(&pit->lock); in pit_reset() [all …]
|
A D | irq.c | 758 spin_lock(&desc->lock); in irq_move_cleanup_interrupt() 916 spin_lock(&lock); in alloc_direct_apic_vector() 921 spin_unlock(&lock); in alloc_direct_apic_vector() 933 spin_lock(&desc->lock); in irq_ratelimit_timer_fn() 1120 spin_lock_irq(&desc->lock); in irq_guest_eoi_timer_fn() 1353 spin_lock(&desc->lock); in flush_ready_eoi() 1392 spin_lock(&desc->lock); in set_eoi_ready() 1394 spin_unlock(&desc->lock); in set_eoi_ready() 1949 spin_lock(&desc->lock); in do_IRQ() 2031 spin_unlock(&desc->lock); in do_IRQ() [all …]
|
/xen/xen/arch/arm/ |
A D | irq.c | 146 ASSERT(spin_is_locked(&desc->lock)); in irq_get_guest_info() 217 spin_lock(&desc->lock); in do_IRQ() 253 spin_unlock_irq(&desc->lock); in do_IRQ() 261 spin_lock_irq(&desc->lock); in do_IRQ() 268 spin_unlock(&desc->lock); in do_IRQ() 280 spin_lock_irqsave(&desc->lock,flags); in release_irq() 309 spin_unlock_irqrestore(&desc->lock,flags); in release_irq() 354 spin_lock_irqsave(&desc->lock, flags); in setup_irq() 463 spin_lock_irqsave(&desc->lock, flags); in route_irq_to_guest() 548 spin_lock_irqsave(&desc->lock, flags); in release_guest_irq() [all …]
|
A D | vgic.c | 89 spin_lock_init(&rank->lock); in vgic_rank_init() 140 spin_lock_init(&d->arch.vgic.lock); in domain_vgic_init() 224 spin_lock_init(&v->arch.vgic.lock); in vcpu_vgic_init() 262 spin_lock_irqsave(&old->arch.vgic.lock, flags); in vgic_migrate_irq() 362 spin_lock_irqsave(&desc->lock, flags); in vgic_disable_irqs() 378 ASSERT(spin_is_locked(&r->lock)); in vgic_get_virq_type() 411 spin_lock_irqsave(&p->desc->lock, flags); in vgic_enable_irqs() 512 spin_lock_irqsave(&v->arch.vgic.lock, flags); in vgic_clear_pending_irqs() 516 spin_unlock_irqrestore(&v->arch.vgic.lock, flags); in vgic_clear_pending_irqs() 521 ASSERT(spin_is_locked(&v->arch.vgic.lock)); in vgic_remove_irq_from_queues() [all …]
|
A D | gic-vgic.c | 53 ASSERT(spin_is_locked(&v->arch.vgic.lock)); in gic_add_to_lr_pending() 71 ASSERT(spin_is_locked(&v->arch.vgic.lock)); in gic_remove_from_lr_pending() 84 ASSERT(spin_is_locked(&v->arch.vgic.lock)); in gic_raise_inflight_irq() 117 ASSERT(spin_is_locked(&v->arch.vgic.lock)); in gic_find_unused_lr() 143 ASSERT(spin_is_locked(&v->arch.vgic.lock)); in gic_raise_guest_irq() 169 ASSERT(spin_is_locked(&v->arch.vgic.lock)); in gic_update_one_lr() 266 spin_lock_irqsave(&v->arch.vgic.lock, flags); in vgic_sync_from_lrs() 287 spin_lock(&v->arch.vgic.lock); in gic_restore_pending_irqs() 331 spin_unlock(&v->arch.vgic.lock); in gic_restore_pending_irqs() 338 ASSERT(spin_is_locked(&v->arch.vgic.lock)); in gic_clear_pending_irqs() [all …]
|
/xen/xen/common/sched/ |
A D | null.c | 145 spin_lock_init(&prv->lock); in null_init() 233 spin_lock_irqsave(&prv->lock, flags); in null_alloc_domdata() 460 spinlock_t *lock; in null_unit_insert() local 477 spin_unlock(lock); in null_unit_insert() 479 lock = unit_schedule_lock(unit); in null_unit_insert() 514 spin_unlock_irq(lock); in null_unit_insert() 524 spinlock_t *lock; in null_unit_remove() local 878 spinlock_t *lock; in null_schedule() local 897 if ( lock ) { in null_schedule() 947 spinlock_t *lock; in null_dump_pcpu() local [all …]
|
A D | rt.c | 358 spin_lock_irqsave(&prv->lock, flags); in rt_dump_pcpu() 378 spin_lock_irqsave(&prv->lock, flags); in rt_dump() 688 spin_lock_init(&prv->lock); in rt_init() 750 return &prv->lock; in rt_switch_sched() 873 spinlock_t *lock; in rt_unit_insert() local 879 lock = pcpu_schedule_lock_irq(cpu); in rt_unit_insert() 881 pcpu_schedule_unlock_irq(lock, cpu); in rt_unit_insert() 883 lock = unit_schedule_lock_irq(unit); in rt_unit_insert() 909 spinlock_t *lock; in rt_unit_remove() local 1465 spin_lock_irq(&prv->lock); in repl_timer_handler() [all …]
|
A D | arinc653.c | 109 spinlock_t lock; member 239 spin_lock_irqsave(&sched_priv->lock, flags); in arinc653_sched_set() 291 spin_unlock_irqrestore(&sched_priv->lock, flags); in arinc653_sched_set() 314 spin_lock_irqsave(&sched_priv->lock, flags); in arinc653_sched_get() 327 spin_unlock_irqrestore(&sched_priv->lock, flags); in arinc653_sched_get() 358 spin_lock_init(&prv->lock); in a653sched_init() 401 spin_lock_irqsave(&sched_priv->lock, flags); in a653sched_alloc_udata() 435 spin_unlock_irqrestore(&sched_priv->lock, flags); in a653sched_alloc_udata() 455 spin_lock_irqsave(&sched_priv->lock, flags); in a653sched_free_udata() 463 spin_unlock_irqrestore(&sched_priv->lock, flags); in a653sched_free_udata() [all …]
|
A D | credit.c | 211 spinlock_t lock; member 636 spin_lock(&prv->lock); in csched_switch_sched() 638 spin_unlock(&prv->lock); in csched_switch_sched() 1014 spinlock_t *lock; in csched_unit_insert() local 1023 spin_unlock_irq(lock); in csched_unit_insert() 1062 spin_lock_irq(&prv->lock); in csched_unit_remove() 1321 spinlock_t *lock; in csched_runq_sort() local 1767 if ( !lock ) in csched_load_balance() 2039 spinlock_t *lock; in csched_dump_pcpu() local 2133 spinlock_t *lock; in csched_dump() local [all …]
|
/xen/xen/drivers/passthrough/amd/ |
A D | iommu_intr.c | 277 spinlock_t *lock; in update_intremap_entry_from_ioapic() local 289 spin_lock_irqsave(lock, flags); in update_intremap_entry_from_ioapic() 311 spin_unlock(lock); in update_intremap_entry_from_ioapic() 313 spin_lock(&iommu->lock); in update_intremap_entry_from_ioapic() 315 spin_unlock(&iommu->lock); in update_intremap_entry_from_ioapic() 317 spin_lock(lock); in update_intremap_entry_from_ioapic() 513 spinlock_t *lock; in update_intremap_entry_from_msi_msg() local 568 spin_unlock(lock); in update_intremap_entry_from_msi_msg() 570 spin_lock(&iommu->lock); in update_intremap_entry_from_msi_msg() 576 spin_lock(lock); in update_intremap_entry_from_msi_msg() [all …]
|
/xen/tools/libs/toolcore/ |
A D | handlereg.c | 36 static void lock(void) { in lock() function 47 lock(); in xentoolcore__register_active_handle() 53 lock(); in xentoolcore__deregister_active_handle() 62 lock(); in xentoolcore_restrict_all()
|
/xen/xen/drivers/vpci/ |
A D | vpci.c | 40 spin_lock(&pdev->vpci->lock); in vpci_remove_device() 50 spin_unlock(&pdev->vpci->lock); in vpci_remove_device() 70 spin_lock_init(&pdev->vpci->lock); in vpci_add_handlers() 149 spin_lock(&vpci->lock); in vpci_add_register() 169 spin_unlock(&vpci->lock); in vpci_add_register() 180 spin_lock(&vpci->lock); in vpci_remove_register() 199 spin_unlock(&vpci->lock); in vpci_remove_register() 324 spin_lock(&pdev->vpci->lock); in vpci_read() 376 spin_unlock(&pdev->vpci->lock); in vpci_read() 437 spin_lock(&pdev->vpci->lock); in vpci_write() [all …]
|
/xen/docs/designs/ |
A D | argo.pandoc | 149 1) Acquire a Read lock on L1. 150 2) Acquire a Read lock on L2. 155 1) Acquire a Read lock on L1. 156 2) Acquire a Write lock on L2. 165 1) Acquire a Write lock on L1. 176 ### The global singleton lock: 185 The majority of operations take a read-lock on this lock, allowing concurrent 216 ### Per-Domain: Ring hash lock 247 ### Send hash lock 261 ### Wildcard pending list lock [all …]
|