/linux-master/arch/x86/kvm/svm/ |
H A D | sev.c | 2967 int count; local 2974 count = svm->vmcb->control.exit_info_2; 2975 if (unlikely(check_mul_overflow(count, size, &bytes))) 2983 count, in);
|
/linux-master/arch/x86/kvm/mmu/ |
H A D | mmu.c | 372 /* Ensure the spte is completely set before we increase the count */ 445 * non-present->present), hence it is enough to count the number of 453 int count; local 456 count = sp->clear_spte_count; 466 count != sp->clear_spte_count)) 803 gfn_t gfn, int count) 812 linfo->disallow_lpage += count; 937 int count = 0; local 948 ++count; 951 count 802 update_gfn_disallow_lpage_count(const struct kvm_memory_slot *slot, gfn_t gfn, int count) argument [all...] |
/linux-master/arch/x86/kvm/ |
H A D | lapic.c | 647 u8 count = 0; local 651 count += hweight32(*reg); 654 return count; 838 int i, count = 0; local 848 count += kvm_apic_set_irq(vcpu, irq, NULL); 852 return count; 862 int count; local 875 count = -EOPNOTSUPP; 877 count = __pv_send_ipi(&ipi_bitmap_low, map, &irq, min); 879 count [all...] |
/linux-master/arch/x86/include/asm/ |
H A D | kvm_host.h | 409 unsigned long count; member in struct:kvm_pio_request 505 * Base value of the PMC counter, relative to the *consumed* count in 511 * The count is relative to the associated perf_event so that KVM 623 u64 count; member in struct:kvm_vcpu_hv_stimer 1183 u8 count; member in struct:kvm_x86_msr_filter 1445 * count to zero should removed the root from the list and clean
|
/linux-master/arch/powerpc/kernel/ |
H A D | iommu.c | 147 size_t count) 151 if (count > 0 && sscanf(buf, "%d", &i) > 0) 154 return count; 145 fail_iommu_store(struct device *dev, struct device_attribute *attr, const char *buf, size_t count) argument
|
/linux-master/tools/testing/selftests/ |
H A D | kselftest.h | 70 #define __cpuid_count(level, count, a, b, c, d) \ 73 : "0" (level), "2" (count))
|
/linux-master/tools/perf/util/bpf_skel/ |
H A D | lock_contention.bpf.c | 491 .count = 1, 512 __sync_fetch_and_add(&data->count, 1);
|
/linux-master/tools/perf/util/ |
H A D | annotate.c | 601 int count = 0; local 613 count++; 617 return count > 1; 1939 int pc = 0, count, sub_id; local 2023 count = disassemble(pc, &info); 2062 pc += count; 2063 } while (count > 0 && pc < len);
|
/linux-master/tools/include/uapi/sound/ |
H A D | asound.h | 59 unsigned char db1_ct_cc; /* coding type and channel count */ 193 * available bit count in most significant bit. It's for the case of so-called 'left-justified' or 490 snd_pcm_uframes_t overrange; /* count of ADC (capture) overrange detections from last status */ 785 size_t xruns; /* count of overruns since last status (in bytes) */ 970 unsigned int overrun; /* count of read queue overruns */ 1112 unsigned int space; /* W: count of element IDs to get */ 1113 unsigned int used; /* R: count of element IDs set */ 1114 unsigned int count; /* R: count of all elements */ member in struct:snd_ctl_elem_list 1123 unsigned int count; /* coun member in struct:snd_ctl_elem_info [all...] |
/linux-master/tools/include/uapi/linux/ |
H A D | kvm.h | 239 __u32 count; member in struct:kvm_run::__anon162::__anon166
|
/linux-master/tools/arch/x86/include/uapi/asm/ |
H A D | kvm.h | 54 /* Architectural interrupt line count. */ 269 __u32 count; /* can be 65536 */ member in struct:kvm_pit_channel_state
|
/linux-master/tools/arch/s390/include/uapi/asm/ |
H A D | kvm.h | 18 __u64 count; member in struct:kvm_s390_skeys 32 * @count: Size of the result buffer. 44 __u32 count; member in struct:kvm_s390_cmma_log
|
/linux-master/sound/pci/hda/ |
H A D | tas2781_hda_i2c.c | 174 uinfo->count = 1; 231 uinfo->count = 1; 245 uinfo->count = 1;
|
/linux-master/net/unix/ |
H A D | af_unix.c | 1805 for (i = scm->fp->count - 1; i >= 0; i--) 1818 for (i = scm->fp->count - 1; i >= 0; i--) 1832 * protected by unix_gc_lock, total_refs (file count) is not, hence this 1841 * Any operations that changes the file count through file descriptors 1846 * that takes unix_gc_lock to decrement the inflight count, so it's 1849 * MSG_PEEK is special in that it does not change the inflight count, 1852 * done between incrementing the file count and installing the file into 1858 * before the file count was incremented, then the lock/unlock pair will 1940 if (unlikely(fp && fp->count)) 1941 atomic_add(fp->count, 3265 unsigned long count = 0; local [all...] |
/linux-master/net/netfilter/ |
H A D | nft_set_pipapo.c | 149 * this information as pairs of first rule index, rule count. 2147 if (iter->count < iter->skip) 2157 iter->count++; 2179 * @desc: Set description, element count and field description used
|
H A D | nft_set_rbtree.c | 21 seqcount_rwlock_t count; member in struct:nft_rbtree 67 if (read_seqcount_retry(&priv->count, seq)) 121 unsigned int seq = read_seqcount_begin(&priv->count); 125 if (ret || !read_seqcount_retry(&priv->count, seq)) 129 seq = read_seqcount_begin(&priv->count); 148 if (read_seqcount_retry(&priv->count, seq)) 205 unsigned int seq = read_seqcount_begin(&priv->count); 212 if (ret || !read_seqcount_retry(&priv->count, seq)) 216 seq = read_seqcount_begin(&priv->count); 501 write_seqcount_begin(&priv->count); [all...] |
H A D | nft_set_hash.c | 287 if (iter->count < iter->skip) 295 iter->count++; 651 if (iter->count < iter->skip) 658 iter->count++;
|
H A D | nft_set_bitmap.c | 223 if (iter->count < iter->skip) 231 iter->count++;
|
H A D | nf_tables_api.c | 5469 iter.count = 0; 5929 args.iter.count = 0; 5934 if (!args.iter.err && args.iter.count == cb->args[0]) 5944 if (args.iter.count == cb->args[0]) 5947 cb->args[0] = args.iter.count; 9738 for (i = 0; i < trans->count; i++) { 9761 for (i = 0; i < trans->count; i++) { 9844 trans->priv[trans->count++] = priv; 9858 return NFT_TRANS_GC_BATCHCOUNT - trans->count; 9877 if (trans->count [all...] |
H A D | nft_lookup.c | 221 iter.count = 0;
|
/linux-master/net/core/ |
H A D | dev.c | 789 * the usage count incremented and the caller must use netdev_put() to 953 * The returned device has not had its ref count increased 1289 * away, e.g. rcu read lock or own a reference count to device. 2312 * offset/count pair it will no longer be used. In the worst case TC0 2323 if (tc->offset + tc->count > txq) { 2334 if (tc->offset + tc->count > txq) { 2350 if ((txq - tc->offset) < tc->count) 2400 int cpu, u16 offset, u16 count) 2409 for (i = count, j = offset; i--; j++) { 2434 u16 offset, u16 count) 2398 remove_xps_queue_cpu(struct net_device *dev, struct xps_dev_maps *dev_maps, int cpu, u16 offset, u16 count) argument 2433 clean_xps_maps(struct net_device *dev, enum xps_map_type type, u16 offset, u16 count) argument 2456 netif_reset_xps_queues(struct net_device *dev, u16 offset, u16 count) argument 2782 netdev_set_tc_queue(struct net_device *dev, u8 tc, u16 count, u16 offset) argument 2829 netdev_bind_sb_channel_queue(struct net_device *dev, struct net_device *sb_dev, u8 tc, u16 count, u16 offset) argument 3054 int cpu, count = 0; local 9213 u8 count = 0; local 10021 unsigned int i, count = dev->num_rx_queues; local 10055 unsigned int i, count = dev->num_rx_queues; local 10088 unsigned int count = dev->num_tx_queues; local [all...] |
/linux-master/net/bridge/ |
H A D | br_private.h | 847 int br_fdb_fillbuf(struct net_bridge *br, void *buf, unsigned long count, 2049 int br_cfm_mep_count(struct net_bridge *br, u32 *count); 2050 int br_cfm_peer_mep_count(struct net_bridge *br, u32 *count); 2081 static inline int br_cfm_mep_count(struct net_bridge *br, u32 *count) argument 2083 *count = 0; 2087 static inline int br_cfm_peer_mep_count(struct net_bridge *br, u32 *count) argument 2089 *count = 0;
|
/linux-master/mm/ |
H A D | shmem.c | 1312 * Move the swapped pages for an inode to page cache. Returns the count 3289 dget(dentry); /* Extra count - pin the dentry in core */ 3378 dget(dentry); /* Extra pinning count for the created dentry */ 3398 dput(dentry); /* Undo the count from "create" - does all the work */ 3441 * we just have to decrement the usage count for the target if 4730 struct kobj_attribute *attr, const char *buf, size_t count) 4735 if (count + 1 > sizeof(tmp)) 4737 memcpy(tmp, buf, count); 4738 tmp[count] = '\0'; 4739 if (count 4729 shmem_enabled_store(struct kobject *kobj, struct kobj_attribute *attr, const char *buf, size_t count) argument [all...] |
H A D | page_owner.c | 126 refcount_set(&dummy_stack.stack_record->count, 1); 128 refcount_set(&failure_stack.stack_record->count, 1); 214 if (refcount_read(&stack_record->count) == REFCOUNT_SATURATED) { 217 if (atomic_try_cmpxchg_relaxed(&stack_record->count.refs, &old, 1)) 221 refcount_add(nr_base_pages, &stack_record->count); 232 if (refcount_sub_and_test(nr_base_pages, &stack_record->count)) 426 unsigned long count[MIGRATE_TYPES] = { 0, }; local 436 * not matter as the mixed block count will still be correct 480 count[MIGRATE_MOVABLE]++; 482 count[pageblock_m 504 print_page_owner_memcg(char *kbuf, size_t count, int ret, struct page *page) argument 541 print_page_owner(char __user *buf, size_t count, unsigned long pfn, struct page *page, struct page_owner *page_owner, depot_stack_handle_t handle) argument 653 read_page_owner(struct file *file, char __user *buf, size_t count, loff_t *ppos) argument 769 unsigned long count = 0; local [all...] |
H A D | memory-failure.c | 920 * The page count will stop it from being freed by unpoison. 934 * Poisoned page might never drop its ref count to 0 so we have 994 int count = page_count(p) - 1; local 997 count -= folio_nr_pages(page_folio(p)); 999 if (count > 0) { 1001 page_to_pfn(p), action_page_types[ps->type], count); 1808 * @count: length of the range, in unit of PAGE_SIZE 1812 unsigned long count, int mf_flags) 1817 size_t end = index + count; 1903 unsigned long count local 1811 mf_dax_kill_procs(struct address_space *mapping, pgoff_t index, unsigned long count, int mf_flags) argument 2533 unsigned long count = 1; local [all...] |