/linux-master/drivers/crypto/intel/qat/qat_common/ |
H A D | adf_heartbeat_dbgfs.c | 198 struct adf_heartbeat *hb = accel_dev->heartbeat; local 200 if (!hb) 203 hb->dbgfs.base_dir = debugfs_create_dir("heartbeat", accel_dev->debugfs_dir); 204 hb->dbgfs.status = debugfs_create_file("status", 0400, hb->dbgfs.base_dir, 206 hb->dbgfs.sent = debugfs_create_file("queries_sent", 0400, hb->dbgfs.base_dir, 207 &hb->hb_sent_counter, &adf_hb_stats_fops); 208 hb->dbgfs.failed = debugfs_create_file("queries_failed", 0400, hb 228 struct adf_heartbeat *hb = accel_dev->heartbeat; local [all...] |
H A D | adf_heartbeat.c | 87 /* fill hb stats memory with pattern */ 224 struct adf_heartbeat *hb; local 237 hb = accel_dev->heartbeat; 238 hb->hb_sent_counter++; 244 hb->hb_failed_counter++; 282 struct adf_heartbeat *hb; local 284 hb = kzalloc(sizeof(*hb), GFP_KERNEL); 285 if (!hb) 288 hb 334 struct adf_heartbeat *hb = accel_dev->heartbeat; local [all...] |
/linux-master/include/linux/ |
H A D | hdlcdrv.h | 158 static inline int hdlcdrv_hbuf_full(struct hdlcdrv_hdlcbuffer *hb) argument 163 spin_lock_irqsave(&hb->lock, flags); 164 ret = !((HDLCDRV_HDLCBUFFER - 1 + hb->rd - hb->wr) % HDLCDRV_HDLCBUFFER); 165 spin_unlock_irqrestore(&hb->lock, flags); 171 static inline int hdlcdrv_hbuf_empty(struct hdlcdrv_hdlcbuffer *hb) argument 176 spin_lock_irqsave(&hb->lock, flags); 177 ret = (hb->rd == hb->wr); 178 spin_unlock_irqrestore(&hb 184 hdlcdrv_hbuf_get(struct hdlcdrv_hdlcbuffer *hb) argument 204 hdlcdrv_hbuf_put(struct hdlcdrv_hdlcbuffer *hb, unsigned short val) argument [all...] |
/linux-master/kernel/futex/ |
H A D | futex.h | 221 struct futex_q *q, struct futex_hash_bucket **hb); 222 extern void futex_wait_queue(struct futex_hash_bucket *hb, struct futex_q *q, 230 extern struct futex_q *futex_top_waiter(struct futex_hash_bucket *hb, union futex_key *key); 233 extern void __futex_queue(struct futex_q *q, struct futex_hash_bucket *hb); 239 * @hb: The destination hash bucket 241 * The hb->lock must be held by the caller, and is released here. A call to 248 static inline void futex_queue(struct futex_q *q, struct futex_hash_bucket *hb) 249 __releases(&hb->lock) 251 __futex_queue(q, hb); 252 spin_unlock(&hb 262 futex_hb_waiters_inc(struct futex_hash_bucket *hb) argument 277 futex_hb_waiters_dec(struct futex_hash_bucket *hb) argument 284 futex_hb_waiters_pending(struct futex_hash_bucket *hb) argument [all...] |
H A D | waitwake.c | 29 * the hb spinlock can be avoided and simply return. In order for this 147 * the hb->lock. 157 struct futex_hash_bucket *hb; local 173 hb = futex_hash(&key); 176 if (!futex_hb_waiters_pending(hb)) 179 spin_lock(&hb->lock); 181 plist_for_each_entry_safe(this, next, &hb->chain, list) { 198 spin_unlock(&hb->lock); 339 * @hb: the futex hash bucket, must be locked by the caller 343 void futex_wait_queue(struct futex_hash_bucket *hb, struc argument 416 struct futex_hash_bucket *hb; local 592 futex_wait_setup(u32 __user *uaddr, u32 val, unsigned int flags, struct futex_q *q, struct futex_hash_bucket **hb) argument 651 struct futex_hash_bucket *hb; local [all...] |
H A D | core.c | 437 * @hb: the hash bucket the futex_q's reside in 440 * Must be called with the hb lock held. 442 struct futex_q *futex_top_waiter(struct futex_hash_bucket *hb, union futex_key *key) argument 446 plist_for_each_entry(this, &hb->chain, list) { 514 struct futex_hash_bucket *hb; local 520 hb = container_of(q->lock_ptr, struct futex_hash_bucket, lock); 521 plist_del(&q->list, &hb->chain); 522 futex_hb_waiters_dec(hb); 527 __acquires(&hb->lock) 529 struct futex_hash_bucket *hb; variable in typeref:struct:futex_hash_bucket 541 futex_hb_waiters_inc(hb); /* implies smp_mb(); (A) */ variable 553 futex_hb_waiters_dec(hb); variable 556 __futex_queue(struct futex_q *q, struct futex_hash_bucket *hb) argument 972 struct futex_hash_bucket *hb; local [all...] |
H A D | requeue.c | 69 * requeue_futex() - Requeue a futex_q from one hb to another 200 * @hb: the hash_bucket of the requeue target futex 213 * 4) Set the q->lock_ptr to the requeue target hb->lock for the case that 222 * Must be called with both q->lock_ptr and hb->lock held. 226 struct futex_hash_bucket *hb) 235 q->lock_ptr = &hb->lock; 685 * @hb: the hash_bucket futex_q was original enqueued on 695 int handle_early_requeue_pi_wakeup(struct futex_hash_bucket *hb, argument 702 * With the hb lock held, we avoid races while we process the wakeup. 703 * We only need to hold hb (an 225 requeue_pi_wake_futex(struct futex_q *q, union futex_key *key, struct futex_hash_bucket *hb) argument 772 struct futex_hash_bucket *hb; local [all...] |
H A D | pi.c | 166 * hb->lock: 168 * hb -> futex_q, relation 171 * (cannot be raw because hb can contain arbitrary amount 192 * hb->lock 218 * We get here with hb->lock held, and having found a 220 * has dropped the hb->lock in between futex_queue() and futex_unqueue_pi(), 383 * This creates pi_state, we have hb->lock held, this means nothing can 494 * @hb: the pi futex hash bucket 495 * @key: the futex key associated with uaddr and hb 509 * The hb 515 futex_lock_pi_atomic(u32 __user *uaddr, struct futex_hash_bucket *hb, union futex_key *key, struct futex_pi_state **ps, struct task_struct *task, struct task_struct **exiting, int set_waiters) argument 923 struct futex_hash_bucket *hb; local 1116 struct futex_hash_bucket *hb; local [all...] |
/linux-master/kernel/power/ |
H A D | swap.c | 233 static void hib_init_batch(struct hib_bio_batch *hb) argument 235 atomic_set(&hb->count, 0); 236 init_waitqueue_head(&hb->wait); 237 hb->error = BLK_STS_OK; 238 blk_start_plug(&hb->plug); 241 static void hib_finish_batch(struct hib_bio_batch *hb) argument 243 blk_finish_plug(&hb->plug); 248 struct hib_bio_batch *hb = bio->bi_private; local 263 if (bio->bi_status && !hb->error) 264 hb 271 hib_submit_io(blk_opf_t opf, pgoff_t page_off, void *addr, struct hib_bio_batch *hb) argument 302 hib_wait_io(struct hib_bio_batch *hb) argument 385 write_page(void *buf, sector_t offset, struct hib_bio_batch *hb) argument 457 swap_write_page(struct swap_map_handle *handle, void *buf, struct hib_bio_batch *hb) argument 560 struct hib_bio_batch hb; local 705 struct hib_bio_batch hb; local 1051 swap_read_page(struct swap_map_handle *handle, void *buf, struct hib_bio_batch *hb) argument 1101 struct hib_bio_batch hb; local 1210 struct hib_bio_batch hb; local [all...] |
/linux-master/io_uring/ |
H A D | poll.c | 124 struct io_hash_bucket *hb = &table->hbs[index]; local 126 spin_lock(&hb->lock); 127 hlist_add_head(&req->hash_node, &hb->list); 128 spin_unlock(&hb->lock); 776 struct io_hash_bucket *hb = &table->hbs[i]; local 778 spin_lock(&hb->lock); 779 hlist_for_each_entry_safe(req, tmp, &hb->list, hash_node) { 786 spin_unlock(&hb->lock); 812 struct io_hash_bucket *hb = &table->hbs[index]; local 816 spin_lock(&hb 845 struct io_hash_bucket *hb = &table->hbs[i]; local [all...] |
H A D | fdinfo.c | 195 struct io_hash_bucket *hb = &ctx->cancel_table.hbs[i]; local 199 spin_lock(&hb->lock); 200 hlist_for_each_entry(req, &hb->list, hash_node) 203 spin_unlock(&hb->lock);
|
H A D | futex.c | 331 struct futex_hash_bucket *hb; local 353 &ifd->q, &hb); 358 futex_queue(&ifd->q, hb);
|
/linux-master/tools/perf/ui/browsers/ |
H A D | hists.c | 51 static void hist_browser__update_nr_entries(struct hist_browser *hb); 56 static bool hist_browser__has_filter(struct hist_browser *hb) argument 58 return hists__has_filter(hb->hists) || hb->min_pcnt || symbol_conf.has_filter || hb->c2c_filter; 79 static void hist_browser__set_title_space(struct hist_browser *hb) argument 81 struct ui_browser *browser = &hb->b; 82 struct hists *hists = hb->hists; 85 browser->extra_title_lines = hb->show_headers ? hpp_list->nr_header_lines : 0; 88 static u32 hist_browser__nr_entries(struct hist_browser *hb) argument 103 hist_browser__update_rows(struct hist_browser *hb) argument 129 struct hist_browser *hb = container_of(browser, struct hist_browser, b); local 272 hierarchy_count_rows(struct hist_browser *hb, struct hist_entry *he, bool include_children) argument 564 hierarchy_set_folding(struct hist_browser *hb, struct hist_entry *he, bool unfold __maybe_unused) argument 582 hist_entry__set_folding(struct hist_entry *he, struct hist_browser *hb, bool unfold) argument 1807 struct hist_browser *hb; local 1818 struct hist_browser *hb = container_of(browser, struct hist_browser, b); local 1911 struct hist_browser *hb; local 2880 hist_browser__update_nr_entries(struct hist_browser *hb) argument 2899 hist_browser__update_percent_limit(struct hist_browser *hb, double percent) argument [all...] |
/linux-master/drivers/misc/bcm-vk/ |
H A D | bcm_vk_msg.c | 143 struct bcm_vk_hb_ctrl *hb = container_of(to_delayed_work(work), struct bcm_vk_hb_ctrl, local 145 struct bcm_vk *vk = container_of(hb, struct bcm_vk, hb_ctrl); 151 if (uptime_s == hb->last_uptime) 152 hb->lost_cnt++; 154 hb->lost_cnt = 0; 157 hb->last_uptime, uptime_s, hb->lost_cnt); 164 hb->last_uptime = uptime_s; 167 hb->lost_cnt = 0; 171 if (hb 185 struct bcm_vk_hb_ctrl *hb = &vk->hb_ctrl; local 193 struct bcm_vk_hb_ctrl *hb = &vk->hb_ctrl; local [all...] |
/linux-master/drivers/scsi/ |
H A D | constants.c | 411 enum scsi_host_status hb = host_byte(result); local 414 if (hb < ARRAY_SIZE(hostbyte_table)) 415 hb_string = hostbyte_table[hb];
|
/linux-master/drivers/cxl/ |
H A D | acpi.c | 499 static int cxl_get_chbs(struct device *dev, struct acpi_device *hb, argument 505 rc = acpi_evaluate_integer(hb->handle, METHOD_NAME__UID, NULL, &uid); 526 struct acpi_device *hb = to_cxl_host_bridge(NULL, dev); local 529 if (kstrtou32(acpi_device_uid(hb), 0, &uid)) 545 struct acpi_device *hb = to_cxl_host_bridge(host, match); local 547 if (!hb) 550 rc = cxl_get_chbs(match, hb, &ctx); 566 pci_root = acpi_pci_find_root(hb->handle); 602 struct acpi_device *hb = to_cxl_host_bridge(host, match); local 611 if (!hb) [all...] |
/linux-master/drivers/atm/ |
H A D | nicstar.c | 203 struct sk_buff *hb; local 228 while ((hb = skb_dequeue(&card->hbpool.queue)) != NULL) { 229 dev_kfree_skb_any(hb); 658 struct sk_buff *hb; local 659 hb = __dev_alloc_skb(NS_HBUFSIZE, GFP_KERNEL); 660 if (hb == NULL) { 668 NS_PRV_BUFTYPE(hb) = BUF_NONE; 669 skb_queue_tail(&card->hbpool.queue, hb); 838 struct sk_buff *hb; local 839 while ((hb 2226 struct sk_buff *hb, *sb, *lb; local 2612 struct sk_buff *hb; local 2627 struct sk_buff *hb; local [all...] |
/linux-master/drivers/net/phy/ |
H A D | bcm-phy-ptp.c | 180 u16 hb[4]; local 184 hb[0] = bcm_phy_read_exp(phydev, HB_REG_0); 185 hb[1] = bcm_phy_read_exp(phydev, HB_REG_1); 186 hb[2] = bcm_phy_read_exp(phydev, HB_REG_2); 187 hb[3] = bcm_phy_read_exp(phydev, HB_REG_3); 192 ts->tv_sec = (hb[3] << 16) | hb[2]; 193 ts->tv_nsec = (hb[1] << 16) | hb[0];
|
/linux-master/arch/mips/kernel/ |
H A D | head.S | 138 jr.hb v0
|
H A D | cps-vec.S | 208 jr.hb t1 363 jr.hb t1 477 1: jr.hb t0
|
/linux-master/net/ipv4/ |
H A D | tcp_metrics.c | 772 struct tcpm_hash_bucket *hb = tcp_metrics_hash + row; local 775 for (col = 0, tm = rcu_dereference(hb->chain); tm; 896 struct tcpm_hash_bucket *hb = tcp_metrics_hash; local 900 for (row = 0; row < max_rows; row++, hb++) { 901 struct tcp_metrics_block __rcu **pp = &hb->chain; 925 struct tcpm_hash_bucket *hb; local 947 hb = tcp_metrics_hash + hash; 948 pp = &hb->chain;
|
/linux-master/drivers/hwmon/ |
H A D | w83773g.c | 56 static inline long temp_of_remote(s8 hb, u8 lb) argument 58 return (hb << 3 | lb >> 5) * 125;
|
/linux-master/drivers/media/i2c/ |
H A D | mt9v111.c | 459 unsigned int hb; local 490 best_fps = vb = hb = 0; 498 for (hb = MT9V111_CORE_R05_MIN_HBLANK; 499 hb < MT9V111_CORE_R05_MAX_HBLANK; hb += 10) { 500 unsigned int t_frame = (row_pclk + hb) * 518 ret = v4l2_ctrl_s_ctrl_int64(mt9v111->hblank, hb);
|
/linux-master/drivers/comedi/drivers/ |
H A D | dt2801.c | 248 int hb = 0; local 254 ret = dt2801_readdata(dev, &hb); 258 *data = (hb << 8) + lb;
|
/linux-master/drivers/net/ethernet/pensando/ionic/ |
H A D | ionic_dev.c | 19 int hb; local 27 hb = ionic_heartbeat_check(ionic); 28 dev_dbg(ionic->dev, "%s: hb %d running %d UP %d\n", 29 __func__, hb, netif_running(lif->netdev), 32 if (hb >= 0 &&
|