Lines Matching refs:lock
141 spinlock_t lock;
308 __releases(&ffs->ev.waitq.lock)
314 spin_unlock_irq(&ffs->ev.waitq.lock);
320 spin_unlock_irq(&ffs->ev.waitq.lock);
434 spin_lock_irq(&ffs->ev.waitq.lock);
450 spin_unlock_irq(&ffs->ev.waitq.lock);
458 spin_unlock_irq(&ffs->ev.waitq.lock);
466 spin_lock_irq(&ffs->ev.waitq.lock);
484 spin_unlock_irq(&ffs->ev.waitq.lock);
501 /* Called with ffs->ev.waitq.lock and ffs->mutex held, both released on exit. */
504 __releases(&ffs->ev.waitq.lock)
530 spin_unlock_irq(&ffs->ev.waitq.lock);
563 spin_lock_irq(&ffs->ev.waitq.lock);
594 spin_unlock_irq(&ffs->ev.waitq.lock);
601 spin_unlock_irq(&ffs->ev.waitq.lock);
611 spin_lock_irq(&ffs->ev.waitq.lock);
631 spin_unlock_irq(&ffs->ev.waitq.lock);
1389 * It can't be done here, as the unref functions might try to lock
1515 spin_lock_init(&priv->lock);
1668 &priv->lock, priv->context, seqno);
3163 * of ffs->setup_state without holding the lock because when
3224 spin_lock_irqsave(&ffs->ev.waitq.lock, flags);
3226 spin_unlock_irqrestore(&ffs->ev.waitq.lock, flags);
3485 * which already uses locking; taking the same lock here would
3803 spin_lock_irqsave(&ffs->ev.waitq.lock, flags);
3807 spin_unlock_irqrestore(&ffs->ev.waitq.lock, flags);