/linux-master/drivers/net/ethernet/mellanox/mlx5/core/en_accel/ |
H A D | macsec.c | 117 struct mutex lock; /* Protects mlx5e_macsec internal contexts */ member in struct:mlx5e_macsec 512 mutex_lock(&priv->macsec->lock); 557 mutex_unlock(&macsec->lock); 567 mutex_unlock(&macsec->lock); 584 mutex_lock(&priv->macsec->lock); 629 mutex_unlock(&macsec->lock); 643 mutex_lock(&priv->macsec->lock); 665 mutex_unlock(&macsec->lock); 681 mutex_lock(&priv->macsec->lock); 734 mutex_unlock(&macsec->lock); [all...] |
/linux-master/drivers/net/ethernet/marvell/octeontx2/af/ |
H A D | rvu_npc.c | 957 mutex_lock(&mcam->lock); 987 mutex_unlock(&mcam->lock); 1171 mutex_lock(&mcam->lock); 1181 mutex_unlock(&mcam->lock); 1189 mutex_unlock(&mcam->lock); 1215 mutex_lock(&mcam->lock); 1234 mutex_unlock(&mcam->lock); 1251 mutex_lock(&mcam->lock); 1267 mutex_unlock(&mcam->lock); 1957 mutex_init(&mcam->lock); [all...] |
/linux-master/drivers/net/ethernet/broadcom/bnxt/ |
H A D | bnxt.c | 9017 mutex_init(&bp->fw_health->lock);
|
/linux-master/drivers/net/ethernet/broadcom/ |
H A D | b44.c | 580 spin_lock_irq(&bp->lock); 586 spin_unlock_irq(&bp->lock); 846 spin_lock_irqsave(&bp->lock, flags); 862 spin_unlock_irqrestore(&bp->lock, flags); 869 spin_lock_irqsave(&bp->lock, flags); 874 spin_unlock_irqrestore(&bp->lock, flags); 893 spin_lock(&bp->lock); 924 spin_unlock(&bp->lock); 934 spin_lock_irq(&bp->lock); 940 spin_unlock_irq(&bp->lock); [all...] |
/linux-master/drivers/mmc/host/ |
H A D | sdhci-msm.c | 437 spin_lock_irqsave(&host->lock, flags); 477 spin_unlock_irqrestore(&host->lock, flags); 638 spin_lock_irqsave(&host->lock, flags); 769 spin_unlock_irqrestore(&host->lock, flags); 775 spin_unlock_irqrestore(&host->lock, flags); 1784 /* sdhci_msm_set_clock - Called with (host->lock) spinlock held. */ 1952 spin_lock_irqsave(&host->lock, flags); 1965 spin_unlock_irqrestore(&host->lock, flags); 2699 spin_lock_irqsave(&host->lock, flags); 2701 spin_unlock_irqrestore(&host->lock, flag [all...] |
H A D | moxart-mmc.c | 123 spinlock_t lock; member in struct:moxart_host 388 spin_lock_irqsave(&host->lock, flags); 408 spin_unlock_irqrestore(&host->lock, flags); 412 spin_lock_irqsave(&host->lock, flags); 417 spin_unlock_irqrestore(&host->lock, flags); 423 spin_lock_irqsave(&host->lock, flags); 449 spin_unlock_irqrestore(&host->lock, flags); 458 spin_lock(&host->lock); 475 spin_unlock(&host->lock); 487 spin_lock_irqsave(&host->lock, flag [all...] |
/linux-master/drivers/md/ |
H A D | dm.c | 556 spin_lock_irqsave(&io->lock, flags); 558 spin_unlock_irqrestore(&io->lock, flags); 562 spin_unlock_irqrestore(&io->lock, flags); 596 spin_lock_init(&io->lock); 998 /* reuse deferred lock to simplify dm_handle_requeue */ 1059 spin_lock_irqsave(&io->lock, flags); 1064 spin_unlock_irqrestore(&io->lock, flags); 2034 * Hold lock to make sure del_gendisk() won't concurrent 2391 * Hold lock to make sure add_disk() and del_gendisk() won't concurrent 2679 * Functions to lock an [all...] |
/linux-master/drivers/gpu/drm/xe/ |
H A D | xe_guc_ct.c | 134 might_lock(&ct->lock); 148 drmm_mutex_init(&xe->drm, &ct->lock); 293 mutex_lock(&ct->lock); /* Serialise dequeue_one_g2h() */ 305 * Lockdep doesn't like this under the fast lock and he destroy only 306 * needs to be serialized with the send path which ct lock provides. 310 mutex_unlock(&ct->lock); 383 lockdep_assert_held(&ct->lock); 409 lockdep_assert_held(&ct->lock); 419 lockdep_assert_held(&ct->lock); 467 lockdep_assert_held(&ct->lock); [all...] |
H A D | xe_gt_ccs_mode.c | 128 spin_lock(&xe->clients.lock); 130 spin_unlock(&xe->clients.lock); 140 spin_unlock(&xe->clients.lock);
|
/linux-master/drivers/gpu/drm/etnaviv/ |
H A D | etnaviv_gpu.h | 111 struct mutex lock; member in struct:etnaviv_gpu
|
H A D | etnaviv_gpu.c | 883 mutex_lock(&gpu->lock); 885 mutex_unlock(&gpu->lock); 1137 * GPU lock must already be held, otherwise fence completion order might 1140 lockdep_assert_held(&gpu->lock); 1277 * Although the retirement happens under the gpu lock, we don't want to hold 1278 * that lock in this function while waiting. 1387 mutex_lock(&gpu->lock); 1424 mutex_unlock(&gpu->lock); 1469 mutex_lock(&gpu->lock); 1481 mutex_unlock(&gpu->lock); [all...] |
/linux-master/drivers/gpu/drm/amd/pm/ |
H A D | amdgpu_pm.c | 1706 raw_spin_lock_irqsave(&adev->throttling_logging_rs.lock, flags); 1716 raw_spin_unlock_irqrestore(&adev->throttling_logging_rs.lock, flags);
|
/linux-master/drivers/gpu/drm/amd/amdkfd/ |
H A D | kfd_svm.c | 90 * Context: The caller must hold svms->lock 126 * Context: The caller must hold svms->lock 306 mutex_destroy(&prange->lock); 355 mutex_init(&prange->lock); 398 mutex_lock(&prange->lock); 402 mutex_unlock(&prange->lock); 447 mutex_lock(&prange->lock); 449 mutex_unlock(&prange->lock); 454 mutex_unlock(&prange->lock); 464 mutex_unlock(&prange->lock); [all...] |
H A D | kfd_migrate.c | 476 * Context: Process context, caller hold mmap read lock, svms lock, prange lock 673 * Context: Process context, caller hold mmap read lock, prange->migrate_mutex 782 * Context: Process context, caller hold mmap read lock, prange->migrate_mutex 870 * Context: Process context, caller hold mmap read lock, svms lock, prange lock 924 * Context: vm fault handler, caller holds the mmap read lock 969 mutex_lock(&p->svms.lock); [all...] |
/linux-master/drivers/gpio/ |
H A D | gpio-tangier.c | 101 guard(raw_spinlock_irqsave)(&priv->lock); 115 guard(raw_spinlock_irqsave)(&priv->lock); 134 guard(raw_spinlock_irqsave)(&priv->lock); 166 guard(raw_spinlock_irqsave)(&priv->lock); 206 guard(raw_spinlock_irqsave)(&priv->lock); 219 guard(raw_spinlock_irqsave)(&priv->lock); 261 guard(raw_spinlock_irqsave)(&priv->lock); 317 guard(raw_spinlock_irqsave)(&priv->lock); 439 raw_spin_lock_init(&gpio->lock); 470 guard(raw_spinlock_irqsave)(&priv->lock); [all...] |
/linux-master/drivers/cxl/core/ |
H A D | mbox.c | 1300 rc = mutex_lock_interruptible(&mds->poison.lock); 1336 mutex_unlock(&mds->poison.lock); 1346 /* Get Poison List output buffer is protected by mds->poison.lock */ 1370 mutex_init(&mds->poison.lock);
|
/linux-master/fs/btrfs/tests/ |
H A D | extent-map-tests.c | 19 write_lock(&em_tree->lock); 37 write_unlock(&em_tree->lock); 75 write_lock(&em_tree->lock); 77 write_unlock(&em_tree->lock); 96 write_lock(&em_tree->lock); 98 write_unlock(&em_tree->lock); 117 write_lock(&em_tree->lock); 119 write_unlock(&em_tree->lock); 168 write_lock(&em_tree->lock); 170 write_unlock(&em_tree->lock); [all...] |
/linux-master/fs/btrfs/ |
H A D | scrub.c | 2125 spin_lock(&bg->lock); 2127 spin_unlock(&bg->lock); 2131 spin_unlock(&bg->lock); 2410 spin_lock(&bg->lock); 2413 spin_unlock(&bg->lock); 2582 spin_lock(&cache->lock); 2584 spin_unlock(&cache->lock); 2589 spin_unlock(&cache->lock); 2716 spin_lock(&cache->lock); 2719 spin_unlock(&cache->lock); [all...] |
H A D | inode.c | 116 * For the file_extent_tree, we want to hold the inode lock when we lookup and 118 * the tree lock and get the inode lock when setting delalloc. These two things 363 * BTRFS_ILOCK_SHARED - acquire a shared lock on the inode 364 * BTRFS_ILOCK_TRY - try to acquire the lock, if fails on first attempt 366 * BTRFS_ILOCK_MMAP - acquire a write lock on the i_mmap_lock 396 * to decide whether the lock acquired is shared or exclusive. 771 * We lock the full page then run each delalloc range in the page, thus 780 * unlocked already, triggering various page lock requirement BUG_ON()s. 789 * TODO: Make btrfs_run_delalloc_range() to lock al [all...] |
H A D | extent_map.c | 37 rwlock_init(&tree->lock); 305 write_lock(&tree->lock); 344 write_unlock(&tree->lock); 352 lockdep_assert_held_write(&tree->lock); 393 lockdep_assert_held_write(&tree->lock); 477 lockdep_assert_held_write(&tree->lock); 491 lockdep_assert_held_write(&tree->lock); 603 * while we had the lock dropped. It is also possible that 655 write_lock(&tree->lock); 665 cond_resched_rwlock_write(&tree->lock); [all...] |
H A D | backref.c | 838 struct preftrees *preftrees, bool lock) 869 if (lock) 875 if (lock) 898 spin_lock(&head->lock); 1004 spin_unlock(&head->lock); 1246 * The caller has joined a transaction or is holding a read lock on the 1316 * The caller has joined a transaction or is holding a read lock on the 1445 * means we have the path lock, we need to grab the ref head and 1446 * lock it so we have a consistent view of the refs at the given 1450 spin_lock(&delayed_refs->lock); 837 add_missing_keys(struct btrfs_fs_info *fs_info, struct preftrees *preftrees, bool lock) argument [all...] |
/linux-master/drivers/hid/ |
H A D | hid-nintendo.c | 575 spinlock_t lock; member in struct:joycon_ctlr 832 spin_lock_irqsave(&ctlr->lock, flags); 834 spin_unlock_irqrestore(&ctlr->lock, flags); 873 spin_lock_irqsave(&ctlr->lock, flags); 875 spin_unlock_irqrestore(&ctlr->lock, flags); 960 spin_lock_irqsave(&ctlr->lock, flags); 966 spin_unlock_irqrestore(&ctlr->lock, flags); 971 spin_unlock_irqrestore(&ctlr->lock, flags); 1593 spin_lock_irqsave(&ctlr->lock, flags); 1610 spin_unlock_irqrestore(&ctlr->lock, flag [all...] |
H A D | hid-mcp2221.c | 91 * lock is used for this purpose mainly. 96 struct mutex lock; member in struct:mcp2221 361 mutex_lock(&mcp->lock); 403 mutex_unlock(&mcp->lock); 463 mutex_lock(&mcp->lock); 591 mutex_unlock(&mcp->lock); 620 mutex_lock(&mcp->lock); 622 mutex_unlock(&mcp->lock); 640 mutex_lock(&mcp->lock); 642 mutex_unlock(&mcp->lock); [all...] |
H A D | hid-logitech-dj.c | 153 spinlock_t lock; member in struct:dj_receiver_dev 654 spin_lock_init(&djrcv_dev->lock); 688 spin_lock_irqsave(&djrcv_dev->lock, flags); 691 spin_unlock_irqrestore(&djrcv_dev->lock, flags); 717 /* We are the only one ever adding a device, no need to lock */ 780 spin_lock_irqsave(&djrcv_dev->lock, flags); 782 spin_unlock_irqrestore(&djrcv_dev->lock, flags); 792 spin_lock_irqsave(&djrcv_dev->lock, flags); 794 spin_unlock_irqrestore(&djrcv_dev->lock, flags); 812 spin_lock_irqsave(&djrcv_dev->lock, flag [all...] |
/linux-master/virt/kvm/ |
H A D | pfncache.c | 32 read_lock_irq(&gpc->lock); 37 read_unlock_irq(&gpc->lock); 43 * is still necessary once the write lock has been 47 write_lock_irq(&gpc->lock); 51 write_unlock_irq(&gpc->lock); 55 read_unlock_irq(&gpc->lock); 137 * is not protected by gpc->lock. It is guaranteed to 138 * be elevated before the mmu_notifier acquires gpc->lock, and 165 lockdep_assert_held_write(&gpc->lock); 168 * Invalidate the cache prior to dropping gpc->lock, th [all...] |