/linux-master/net/ipv4/ |
H A D | udp.c | 2489 unsigned int slot; local 2491 slot = udp_hashfn(net, hnum, udptable->mask); 2492 hslot = &udptable->hash[slot]; 3419 21, /* one slot per 2 MB */
|
/linux-master/kernel/bpf/ |
H A D | verifier.c | 597 * takes the size of two stack slots; the first slot will be at 598 * spi and the second slot will be at spi - 1. 778 /* Why do we need to set REG_LIVE_WRITTEN for STACK_INVALID slot? 785 * the slot conservatively. Hence we need to prevent those liveness 835 * matches then the stack slot also belongs to a 875 /* Reposition spi to first slot */ 887 /* Writing partially to one dynptr stack slot destroys both. */ 949 /* This already represents first slot of initialized bpf_dynptr. 1018 struct bpf_stack_state *slot = &state->stack[spi - i]; local 1019 struct bpf_reg_state *st = &slot 1056 struct bpf_stack_state *slot = &state->stack[spi - i]; local 1093 struct bpf_stack_state *slot = &state->stack[spi - i]; local 1114 struct bpf_stack_state *slot = &state->stack[spi - i]; local 3485 bt_set_frame_slot(struct backtrack_state *bt, u32 frame, u32 slot) argument 3490 bt_clear_frame_slot(struct backtrack_state *bt, u32 frame, u32 slot) argument 3520 bt_is_frame_slot_set(struct backtrack_state *bt, u32 frame, u32 slot) argument 4486 int i, slot = -off - 1, spi = slot / BPF_REG_SIZE, err; local 4660 int slot, spi; local 4745 int i, slot, spi; local 4786 int i, slot = -off - 1, spi = slot / BPF_REG_SIZE; local 7096 int err, min_off, max_off, i, j, slot, spi; local 17141 struct bpf_reg_state *slot, *cur_slot; local [all...] |
H A D | core.c | 834 u32 slot = prog->aux->size_poke_tab; local 835 u32 size = slot + 1; 856 memcpy(&tab[slot], poke, sizeof(*poke)); 860 return slot;
|
/linux-master/fs/btrfs/ |
H A D | volumes.c | 1628 int slot; local 1663 slot = path->slots[0]; 1664 if (slot >= btrfs_header_nritems(l)) { 1673 btrfs_item_key_to_cpu(l, &key, slot); 1712 dev_extent = btrfs_item_ptr(l, slot, struct btrfs_dev_extent); 4015 int slot; local 4082 slot = path->slots[0]; 4083 btrfs_item_key_to_cpu(leaf, &found_key, slot); 4090 chunk = btrfs_item_ptr(leaf, slot, struct btrfs_chunk); 4791 int slot; local 4975 int slot; local 7532 int slot; local 7706 int i, ret, slot; local 8173 int slot = path->slots[0]; local [all...] |
H A D | qgroup.c | 373 struct extent_buffer *leaf, int slot, 377 ASSERT(btrfs_item_size(leaf, slot) >= sizeof(*ptr)); 392 int slot; local 431 slot = path->slots[0]; 433 btrfs_item_key_to_cpu(l, &found_key, slot); 438 ptr = btrfs_item_ptr(l, slot, 449 qgroup_read_enable_gen(fs_info, l, slot, ptr); 487 ptr = btrfs_item_ptr(l, slot, 499 ptr = btrfs_item_ptr(l, slot, 530 slot 372 qgroup_read_enable_gen(struct btrfs_fs_info *fs_info, struct extent_buffer *leaf, int slot, struct btrfs_qgroup_status_item *ptr) argument 828 int slot; local 871 int slot; local 916 int slot; local 1011 int slot; local 2110 int nr, slot; local 3604 int slot; local [all...] |
H A D | ioctl.c | 1585 int slot; local 1589 slot = path->slots[0]; 1598 for (i = slot; i < nritems; i++) { 1845 int slot; local 1884 slot = path->slots[0]; 1886 iref = btrfs_item_ptr(l, slot, struct btrfs_inode_ref); 1934 int slot; local 1969 slot = path->slots[0]; 1971 iref = btrfs_item_ptr(leaf, slot, struct btrfs_inode_ref); 1995 slot 2174 int slot; local 2291 int slot; local [all...] |
H A D | scrub.c | 1362 * slot, to avoid duplicated btrfs_search_slot() calls. 1891 * There should always be one slot left, as caller filling the last 1892 * slot should flush them all. 1917 /* Last slot used, flush them all. */ 2466 int slot; local 2507 slot = path->slots[0]; 2509 btrfs_item_key_to_cpu(l, &found_key, slot); 2523 dev_extent = btrfs_item_ptr(l, slot, struct btrfs_dev_extent);
|
H A D | inode.c | 2024 * search, then go back to the previous slot as it will be the 3695 * slot is the slot the inode is in, objectid is the objectid of the inode 3698 int slot, u64 objectid, 3714 slot++; 3716 while (slot < nritems) { 3717 btrfs_item_key_to_cpu(leaf, &found_key, slot); 3726 *first_xattr_slot = slot; 3739 slot++; 3756 *first_xattr_slot = slot; 3697 acls_after_inode_item(struct extent_buffer *leaf, int slot, u64 objectid, int *first_xattr_slot) argument [all...] |
H A D | backref.c | 111 int slot; local 122 for (slot = 0; slot < nritems; ++slot) { 123 btrfs_item_key_to_cpu(eb, &key, slot); 126 fi = btrfs_item_ptr(eb, slot, struct btrfs_file_extent_item); 475 int slot; local 498 * slot == nritems. 517 slot = path->slots[0]; 519 btrfs_item_key_to_cpu(eb, &key, slot); 1019 int slot; local 1154 int slot; local 1481 int slot; local 2026 int slot = ctx->prev_extents_cache_slot; local 2054 int ret, slot; local 2136 int slot; local 2590 int slot; local 2651 int slot; local [all...] |
/linux-master/drivers/nvme/target/ |
H A D | core.c | 1357 void __rcu **slot; local 1359 radix_tree_for_each_slot(slot, &ctrl->p2p_ns_map, &iter, 0) 1360 pci_dev_put(radix_tree_deref_slot(slot));
|
/linux-master/drivers/gpu/drm/nouveau/nvkm/subdev/gsp/ |
H A D | r535.c | 1501 u32 slot = cmd->payload.regStore.index; local 1503 seq->regSaveArea[slot] = nvkm_rd32(device, addr); 1504 nvkm_trace(subdev, "seq save %08x -> %d: %08x\n", addr, slot, 1505 seq->regSaveArea[slot]);
|
/linux-master/arch/s390/mm/ |
H A D | gmap.c | 142 void __rcu **slot; local 149 radix_tree_for_each_slot(slot, root, &iter, index) { 167 void __rcu **slot; local 174 radix_tree_for_each_slot(slot, root, &iter, index) { 1200 void __rcu **slot; local 1203 slot = radix_tree_lookup_slot(&sg->host_to_rmap, vmaddr >> PAGE_SHIFT); 1204 if (slot) { 1205 rmap->next = radix_tree_deref_slot_protected(slot, 1213 radix_tree_replace_slot(&sg->host_to_rmap, slot, rmap);
|
/linux-master/kernel/ |
H A D | workqueue.c | 5081 struct pool_workqueue __rcu **slot = unbound_pwq_slot(wq, cpu); local 5090 old_pwq = rcu_access_pointer(*slot); 5091 rcu_assign_pointer(*slot, pwq);
|
/linux-master/fs/nfsd/ |
H A D | nfs4xdr.c | 5642 struct nfsd4_slot *slot = resp->cstate.slot; local 5648 if (slot->sl_flags & NFSD4_SLOT_CACHETHIS) { 5693 struct nfsd4_slot *slot = resp->cstate.slot; local 5695 if (slot->sl_flags & NFSD4_SLOT_CACHETHIS)
|
/linux-master/drivers/pinctrl/ |
H A D | core.c | 713 void __rcu **slot; local 715 radix_tree_for_each_slot(slot, &pctldev->pin_group_tree, &iter, 0)
|
/linux-master/kernel/sched/ |
H A D | fair.c | 2656 int slot = ps_ratio - NUMA_PERIOD_THRESHOLD; local 2657 if (!slot) 2658 slot = 1; 2659 diff = slot * period_slot; 2666 int slot = lr_ratio - NUMA_PERIOD_THRESHOLD; local 2667 if (!slot) 2668 slot = 1; 2669 diff = slot * period_slot;
|
/linux-master/drivers/mmc/host/ |
H A D | sdhci-msm.c | 1863 int slot) 1881 cfg->data_unit_size, slot); 1883 return qcom_ice_evict_key(msm_host->ice, slot); 1861 sdhci_msm_program_key(struct cqhci_host *cq_host, const union cqhci_crypto_cfg_entry *cfg, int slot) argument
|
/linux-master/arch/arc/kernel/ |
H A D | entry-arcv2.S | 224 ;####### Return from Intr to insn in delay slot ####### 235 ; and return from pure kernel mode which does right things for delay slot
|
/linux-master/virt/kvm/ |
H A D | kvm_mm.h | 41 int kvm_gmem_bind(struct kvm *kvm, struct kvm_memory_slot *slot, 43 void kvm_gmem_unbind(struct kvm_memory_slot *slot); 51 struct kvm_memory_slot *slot, 58 static inline void kvm_gmem_unbind(struct kvm_memory_slot *slot) argument 50 kvm_gmem_bind(struct kvm *kvm, struct kvm_memory_slot *slot, unsigned int fd, loff_t offset) argument
|
H A D | kvm_main.c | 602 struct kvm_memory_slot *slot; local 624 slot = container_of(node, struct kvm_memory_slot, hva_node[slots->node_idx]); 625 hva_start = max_t(unsigned long, range->start, slot->userspace_addr); 627 slot->userspace_addr + (slot->npages << PAGE_SHIFT)); 642 gfn_range.start = hva_to_gfn_memslot(hva_start, slot); 643 gfn_range.end = hva_to_gfn_memslot(hva_end + PAGE_SIZE - 1, slot); 644 gfn_range.slot = slot; 719 if (range->slot 1027 kvm_free_memslot(struct kvm *kvm, struct kvm_memory_slot *slot) argument 1489 kvm_insert_gfn_node(struct kvm_memslots *slots, struct kvm_memory_slot *slot) argument 1514 kvm_erase_gfn_node(struct kvm_memslots *slots, struct kvm_memory_slot *slot) argument 2482 struct kvm_memory_slot *slot; local 2637 struct kvm_memory_slot *slot; local 2706 memslot_is_readonly(const struct kvm_memory_slot *slot) argument 2711 __gfn_to_hva_many(const struct kvm_memory_slot *slot, gfn_t gfn, gfn_t *nr_pages, bool write) argument 2726 gfn_to_hva_many(struct kvm_memory_slot *slot, gfn_t gfn, gfn_t *nr_pages) argument 2732 gfn_to_hva_memslot(struct kvm_memory_slot *slot, gfn_t gfn) argument 2759 gfn_to_hva_memslot_prot(struct kvm_memory_slot *slot, gfn_t gfn, bool *writable) argument 2772 struct kvm_memory_slot *slot = gfn_to_memslot(kvm, gfn); local 2779 struct kvm_memory_slot *slot = kvm_vcpu_gfn_to_memslot(vcpu, gfn); local 3028 __gfn_to_pfn_memslot(const struct kvm_memory_slot *slot, gfn_t gfn, bool atomic, bool interruptible, bool *async, bool write_fault, bool *writable, hva_t *hva) argument 3068 gfn_to_pfn_memslot(const struct kvm_memory_slot *slot, gfn_t gfn) argument 3075 gfn_to_pfn_memslot_atomic(const struct kvm_memory_slot *slot, gfn_t gfn) argument 3100 gfn_to_page_many_atomic(struct kvm_memory_slot *slot, gfn_t gfn, struct page **pages, int nr_pages) argument 3310 __kvm_read_guest_page(struct kvm_memory_slot *slot, gfn_t gfn, void *data, int offset, int len) argument 3328 struct kvm_memory_slot *slot = gfn_to_memslot(kvm, gfn); local 3337 struct kvm_memory_slot *slot = kvm_vcpu_gfn_to_memslot(vcpu, gfn); local 3383 __kvm_read_guest_atomic(struct kvm_memory_slot *slot, gfn_t gfn, void *data, int offset, unsigned long len) argument 3404 struct kvm_memory_slot *slot = kvm_vcpu_gfn_to_memslot(vcpu, gfn); local 3431 struct kvm_memory_slot *slot = gfn_to_memslot(kvm, gfn); local 3440 struct kvm_memory_slot *slot = kvm_vcpu_gfn_to_memslot(vcpu, gfn); local 3647 u32 slot = (memslot->as_id << 16) | memslot->id; local 5355 __u32 slot; member in struct:compat_kvm_dirty_log 5364 __u32 slot; member in struct:compat_kvm_clear_dirty_log [all...] |
/linux-master/tools/testing/selftests/kvm/ |
H A D | set_memory_region_test.c | 29 * Somewhat arbitrary location and slot, intended to not overlap anything. 397 uint32_t slot; local 425 for (slot = 0; slot < max_mem_slots; slot++) 426 vm_set_user_memory_region(vm, slot, 0, 427 ((uint64_t)slot * MEM_REGION_SIZE), 429 mem_aligned + (uint64_t)slot * MEM_REGION_SIZE); 440 "Adding one more memory slot should fail with EINVAL"); 520 /* Overlap the front half of the other slot [all...] |
H A D | max_guest_memory_test.c | 168 int max_slots, slot, opt, fd; local 230 for (slot = first_slot; slot < max_slots; slot++) { 231 gpa = start_gpa + ((slot - first_slot) * slot_size); 238 vm_set_user_memory_region(vm, slot, 0, gpa, slot_size, mem); 278 for (slot = (slot - 1) & ~1ull; slot >= first_slot; slot [all...] |
/linux-master/drivers/usb/host/ |
H A D | xhci-trace.h | 186 TP_printk("vdev %p slot %d ctx %llx | %llx current_mel %d", 222 TP_printk("vdev %p ctx %llx | %llx num %d state %d speed %d port %d level %d slot %d", 282 TP_printk("ep%d%s-%s: urb %p pipe %u slot %d length %d/%d sgs %d/%d stream %d flags %08x", 545 TP_PROTO(u32 slot, u32 doorbell), 546 TP_ARGS(slot, doorbell), 548 __field(u32, slot) 552 __entry->slot = slot; 556 xhci_decode_doorbell(__get_buf(XHCI_MSG_MAX), __entry->slot, __entry->doorbell) 561 TP_PROTO(u32 slot, u3 [all...] |
/linux-master/arch/x86/kvm/ |
H A D | x86.c | 364 unsigned slot; local 380 for (slot = 0; slot < kvm_nr_uret_msrs; ++slot) { 381 values = &msrs->values[slot]; 383 wrmsrl(kvm_uret_msrs_list[slot], values->host); 442 int kvm_set_user_return_msr(unsigned slot, u64 value, u64 mask) argument 448 value = (value & mask) | (msrs->values[slot].host & ~mask); 449 if (value == msrs->values[slot].curr) 451 err = wrmsrl_safe(kvm_uret_msrs_list[slot], valu 12647 struct kvm_memory_slot *slot; local 12729 memslot_rmap_free(struct kvm_memory_slot *slot) argument 12739 kvm_arch_free_memslot(struct kvm *kvm, struct kvm_memory_slot *slot) argument 12753 memslot_rmap_alloc(struct kvm_memory_slot *slot, unsigned long npages) argument 12775 kvm_alloc_memslot_metadata(struct kvm *kvm, struct kvm_memory_slot *slot) argument [all...] |
/linux-master/arch/x86/kvm/vmx/ |
H A D | vmx.c | 722 unsigned int slot = msr - vmx->guest_uret_msrs; local 727 ret = kvm_set_user_return_msr(slot, data, msr->mask); 6767 struct kvm_memory_slot *slot; local 6782 * Explicitly grab the memslot using KVM's internal slot ID to ensure 6787 slot = id_to_memslot(slots, APIC_ACCESS_PAGE_PRIVATE_MEMSLOT); 6788 if (!slot || slot->flags & KVM_MEMSLOT_INVALID) 6805 pfn = gfn_to_pfn_memslot(slot, gfn);
|