Searched refs:lock (Results 101 - 125 of 6515) sorted by last modified time

1234567891011>>

/linux-master/drivers/net/ethernet/mellanox/mlx5/core/en_accel/
H A Dmacsec.c117 struct mutex lock; /* Protects mlx5e_macsec internal contexts */ member in struct:mlx5e_macsec
512 mutex_lock(&priv->macsec->lock);
557 mutex_unlock(&macsec->lock);
567 mutex_unlock(&macsec->lock);
584 mutex_lock(&priv->macsec->lock);
629 mutex_unlock(&macsec->lock);
643 mutex_lock(&priv->macsec->lock);
665 mutex_unlock(&macsec->lock);
681 mutex_lock(&priv->macsec->lock);
734 mutex_unlock(&macsec->lock);
[all...]
/linux-master/drivers/net/ethernet/marvell/octeontx2/af/
H A Drvu_npc.c957 mutex_lock(&mcam->lock);
987 mutex_unlock(&mcam->lock);
1171 mutex_lock(&mcam->lock);
1181 mutex_unlock(&mcam->lock);
1189 mutex_unlock(&mcam->lock);
1215 mutex_lock(&mcam->lock);
1234 mutex_unlock(&mcam->lock);
1251 mutex_lock(&mcam->lock);
1267 mutex_unlock(&mcam->lock);
1957 mutex_init(&mcam->lock);
[all...]
/linux-master/drivers/net/ethernet/broadcom/bnxt/
H A Dbnxt.c9017 mutex_init(&bp->fw_health->lock);
/linux-master/drivers/net/ethernet/broadcom/
H A Db44.c580 spin_lock_irq(&bp->lock);
586 spin_unlock_irq(&bp->lock);
846 spin_lock_irqsave(&bp->lock, flags);
862 spin_unlock_irqrestore(&bp->lock, flags);
869 spin_lock_irqsave(&bp->lock, flags);
874 spin_unlock_irqrestore(&bp->lock, flags);
893 spin_lock(&bp->lock);
924 spin_unlock(&bp->lock);
934 spin_lock_irq(&bp->lock);
940 spin_unlock_irq(&bp->lock);
[all...]
/linux-master/drivers/mmc/host/
H A Dsdhci-msm.c437 spin_lock_irqsave(&host->lock, flags);
477 spin_unlock_irqrestore(&host->lock, flags);
638 spin_lock_irqsave(&host->lock, flags);
769 spin_unlock_irqrestore(&host->lock, flags);
775 spin_unlock_irqrestore(&host->lock, flags);
1784 /* sdhci_msm_set_clock - Called with (host->lock) spinlock held. */
1952 spin_lock_irqsave(&host->lock, flags);
1965 spin_unlock_irqrestore(&host->lock, flags);
2699 spin_lock_irqsave(&host->lock, flags);
2701 spin_unlock_irqrestore(&host->lock, flag
[all...]
H A Dmoxart-mmc.c123 spinlock_t lock; member in struct:moxart_host
388 spin_lock_irqsave(&host->lock, flags);
408 spin_unlock_irqrestore(&host->lock, flags);
412 spin_lock_irqsave(&host->lock, flags);
417 spin_unlock_irqrestore(&host->lock, flags);
423 spin_lock_irqsave(&host->lock, flags);
449 spin_unlock_irqrestore(&host->lock, flags);
458 spin_lock(&host->lock);
475 spin_unlock(&host->lock);
487 spin_lock_irqsave(&host->lock, flag
[all...]
/linux-master/drivers/md/
H A Ddm.c556 spin_lock_irqsave(&io->lock, flags);
558 spin_unlock_irqrestore(&io->lock, flags);
562 spin_unlock_irqrestore(&io->lock, flags);
596 spin_lock_init(&io->lock);
998 /* reuse deferred lock to simplify dm_handle_requeue */
1059 spin_lock_irqsave(&io->lock, flags);
1064 spin_unlock_irqrestore(&io->lock, flags);
2034 * Hold lock to make sure del_gendisk() won't concurrent
2391 * Hold lock to make sure add_disk() and del_gendisk() won't concurrent
2679 * Functions to lock an
[all...]
/linux-master/drivers/gpu/drm/xe/
H A Dxe_guc_ct.c134 might_lock(&ct->lock);
148 drmm_mutex_init(&xe->drm, &ct->lock);
293 mutex_lock(&ct->lock); /* Serialise dequeue_one_g2h() */
305 * Lockdep doesn't like this under the fast lock and he destroy only
306 * needs to be serialized with the send path which ct lock provides.
310 mutex_unlock(&ct->lock);
383 lockdep_assert_held(&ct->lock);
409 lockdep_assert_held(&ct->lock);
419 lockdep_assert_held(&ct->lock);
467 lockdep_assert_held(&ct->lock);
[all...]
H A Dxe_gt_ccs_mode.c128 spin_lock(&xe->clients.lock);
130 spin_unlock(&xe->clients.lock);
140 spin_unlock(&xe->clients.lock);
/linux-master/drivers/gpu/drm/etnaviv/
H A Detnaviv_gpu.h111 struct mutex lock; member in struct:etnaviv_gpu
H A Detnaviv_gpu.c883 mutex_lock(&gpu->lock);
885 mutex_unlock(&gpu->lock);
1137 * GPU lock must already be held, otherwise fence completion order might
1140 lockdep_assert_held(&gpu->lock);
1277 * Although the retirement happens under the gpu lock, we don't want to hold
1278 * that lock in this function while waiting.
1387 mutex_lock(&gpu->lock);
1424 mutex_unlock(&gpu->lock);
1469 mutex_lock(&gpu->lock);
1481 mutex_unlock(&gpu->lock);
[all...]
/linux-master/drivers/gpu/drm/amd/pm/
H A Damdgpu_pm.c1706 raw_spin_lock_irqsave(&adev->throttling_logging_rs.lock, flags);
1716 raw_spin_unlock_irqrestore(&adev->throttling_logging_rs.lock, flags);
/linux-master/drivers/gpu/drm/amd/amdkfd/
H A Dkfd_svm.c90 * Context: The caller must hold svms->lock
126 * Context: The caller must hold svms->lock
306 mutex_destroy(&prange->lock);
355 mutex_init(&prange->lock);
398 mutex_lock(&prange->lock);
402 mutex_unlock(&prange->lock);
447 mutex_lock(&prange->lock);
449 mutex_unlock(&prange->lock);
454 mutex_unlock(&prange->lock);
464 mutex_unlock(&prange->lock);
[all...]
H A Dkfd_migrate.c476 * Context: Process context, caller hold mmap read lock, svms lock, prange lock
673 * Context: Process context, caller hold mmap read lock, prange->migrate_mutex
782 * Context: Process context, caller hold mmap read lock, prange->migrate_mutex
870 * Context: Process context, caller hold mmap read lock, svms lock, prange lock
924 * Context: vm fault handler, caller holds the mmap read lock
969 mutex_lock(&p->svms.lock);
[all...]
/linux-master/drivers/gpio/
H A Dgpio-tangier.c101 guard(raw_spinlock_irqsave)(&priv->lock);
115 guard(raw_spinlock_irqsave)(&priv->lock);
134 guard(raw_spinlock_irqsave)(&priv->lock);
166 guard(raw_spinlock_irqsave)(&priv->lock);
206 guard(raw_spinlock_irqsave)(&priv->lock);
219 guard(raw_spinlock_irqsave)(&priv->lock);
261 guard(raw_spinlock_irqsave)(&priv->lock);
317 guard(raw_spinlock_irqsave)(&priv->lock);
439 raw_spin_lock_init(&gpio->lock);
470 guard(raw_spinlock_irqsave)(&priv->lock);
[all...]
/linux-master/drivers/cxl/core/
H A Dmbox.c1300 rc = mutex_lock_interruptible(&mds->poison.lock);
1336 mutex_unlock(&mds->poison.lock);
1346 /* Get Poison List output buffer is protected by mds->poison.lock */
1370 mutex_init(&mds->poison.lock);
/linux-master/fs/btrfs/tests/
H A Dextent-map-tests.c19 write_lock(&em_tree->lock);
37 write_unlock(&em_tree->lock);
75 write_lock(&em_tree->lock);
77 write_unlock(&em_tree->lock);
96 write_lock(&em_tree->lock);
98 write_unlock(&em_tree->lock);
117 write_lock(&em_tree->lock);
119 write_unlock(&em_tree->lock);
168 write_lock(&em_tree->lock);
170 write_unlock(&em_tree->lock);
[all...]
/linux-master/fs/btrfs/
H A Dscrub.c2125 spin_lock(&bg->lock);
2127 spin_unlock(&bg->lock);
2131 spin_unlock(&bg->lock);
2410 spin_lock(&bg->lock);
2413 spin_unlock(&bg->lock);
2582 spin_lock(&cache->lock);
2584 spin_unlock(&cache->lock);
2589 spin_unlock(&cache->lock);
2716 spin_lock(&cache->lock);
2719 spin_unlock(&cache->lock);
[all...]
H A Dinode.c116 * For the file_extent_tree, we want to hold the inode lock when we lookup and
118 * the tree lock and get the inode lock when setting delalloc. These two things
363 * BTRFS_ILOCK_SHARED - acquire a shared lock on the inode
364 * BTRFS_ILOCK_TRY - try to acquire the lock, if fails on first attempt
366 * BTRFS_ILOCK_MMAP - acquire a write lock on the i_mmap_lock
396 * to decide whether the lock acquired is shared or exclusive.
771 * We lock the full page then run each delalloc range in the page, thus
780 * unlocked already, triggering various page lock requirement BUG_ON()s.
789 * TODO: Make btrfs_run_delalloc_range() to lock al
[all...]
H A Dextent_map.c37 rwlock_init(&tree->lock);
305 write_lock(&tree->lock);
344 write_unlock(&tree->lock);
352 lockdep_assert_held_write(&tree->lock);
393 lockdep_assert_held_write(&tree->lock);
477 lockdep_assert_held_write(&tree->lock);
491 lockdep_assert_held_write(&tree->lock);
603 * while we had the lock dropped. It is also possible that
655 write_lock(&tree->lock);
665 cond_resched_rwlock_write(&tree->lock);
[all...]
H A Dbackref.c838 struct preftrees *preftrees, bool lock)
869 if (lock)
875 if (lock)
898 spin_lock(&head->lock);
1004 spin_unlock(&head->lock);
1246 * The caller has joined a transaction or is holding a read lock on the
1316 * The caller has joined a transaction or is holding a read lock on the
1445 * means we have the path lock, we need to grab the ref head and
1446 * lock it so we have a consistent view of the refs at the given
1450 spin_lock(&delayed_refs->lock);
837 add_missing_keys(struct btrfs_fs_info *fs_info, struct preftrees *preftrees, bool lock) argument
[all...]
/linux-master/drivers/hid/
H A Dhid-nintendo.c575 spinlock_t lock; member in struct:joycon_ctlr
832 spin_lock_irqsave(&ctlr->lock, flags);
834 spin_unlock_irqrestore(&ctlr->lock, flags);
873 spin_lock_irqsave(&ctlr->lock, flags);
875 spin_unlock_irqrestore(&ctlr->lock, flags);
960 spin_lock_irqsave(&ctlr->lock, flags);
966 spin_unlock_irqrestore(&ctlr->lock, flags);
971 spin_unlock_irqrestore(&ctlr->lock, flags);
1593 spin_lock_irqsave(&ctlr->lock, flags);
1610 spin_unlock_irqrestore(&ctlr->lock, flag
[all...]
H A Dhid-mcp2221.c91 * lock is used for this purpose mainly.
96 struct mutex lock; member in struct:mcp2221
361 mutex_lock(&mcp->lock);
403 mutex_unlock(&mcp->lock);
463 mutex_lock(&mcp->lock);
591 mutex_unlock(&mcp->lock);
620 mutex_lock(&mcp->lock);
622 mutex_unlock(&mcp->lock);
640 mutex_lock(&mcp->lock);
642 mutex_unlock(&mcp->lock);
[all...]
H A Dhid-logitech-dj.c153 spinlock_t lock; member in struct:dj_receiver_dev
654 spin_lock_init(&djrcv_dev->lock);
688 spin_lock_irqsave(&djrcv_dev->lock, flags);
691 spin_unlock_irqrestore(&djrcv_dev->lock, flags);
717 /* We are the only one ever adding a device, no need to lock */
780 spin_lock_irqsave(&djrcv_dev->lock, flags);
782 spin_unlock_irqrestore(&djrcv_dev->lock, flags);
792 spin_lock_irqsave(&djrcv_dev->lock, flags);
794 spin_unlock_irqrestore(&djrcv_dev->lock, flags);
812 spin_lock_irqsave(&djrcv_dev->lock, flag
[all...]
/linux-master/virt/kvm/
H A Dpfncache.c32 read_lock_irq(&gpc->lock);
37 read_unlock_irq(&gpc->lock);
43 * is still necessary once the write lock has been
47 write_lock_irq(&gpc->lock);
51 write_unlock_irq(&gpc->lock);
55 read_unlock_irq(&gpc->lock);
137 * is not protected by gpc->lock. It is guaranteed to
138 * be elevated before the mmu_notifier acquires gpc->lock, and
165 lockdep_assert_held_write(&gpc->lock);
168 * Invalidate the cache prior to dropping gpc->lock, th
[all...]

Completed in 358 milliseconds

1234567891011>>