Lines Matching refs:eoi
557 struct lateeoi_work *eoi = &per_cpu(lateeoi, info->eoi_cpu);
560 spin_lock_irqsave(&eoi->eoi_list_lock, flags);
562 spin_unlock_irqrestore(&eoi->eoi_list_lock, flags);
567 struct lateeoi_work *eoi = &per_cpu(lateeoi, info->eoi_cpu);
578 spin_lock_irqsave(&eoi->eoi_list_lock, flags);
580 elem = list_first_entry_or_null(&eoi->eoi_list, struct irq_info,
583 list_add(&info->eoi_list, &eoi->eoi_list);
585 &eoi->delayed, delay);
587 list_for_each_entry_reverse(elem, &eoi->eoi_list, eoi_list) {
594 spin_unlock_irqrestore(&eoi->eoi_list_lock, flags);
650 struct lateeoi_work *eoi;
655 eoi = container_of(to_delayed_work(work), struct lateeoi_work, delayed);
660 spin_lock_irqsave(&eoi->eoi_list_lock, flags);
662 info = list_first_entry_or_null(&eoi->eoi_list, struct irq_info,
670 &eoi->delayed,
677 spin_unlock_irqrestore(&eoi->eoi_list_lock, flags);
684 spin_unlock_irqrestore(&eoi->eoi_list_lock, flags);
691 struct lateeoi_work *eoi = &per_cpu(lateeoi, cpu);
693 INIT_DELAYED_WORK(&eoi->delayed, xen_irq_lateeoi_worker);
694 spin_lock_init(&eoi->eoi_list_lock);
695 INIT_LIST_HEAD(&eoi->eoi_list);
816 struct physdev_eoi eoi = { .irq = pirq_from_irq(info) };
825 rc = HYPERVISOR_physdev_op(PHYSDEVOP_eoi, &eoi);
1068 * interrupts too. In any case Xen always honors the eoi mechanism,
1070 * hasn't received an eoi yet. Therefore using the fasteoi handler