/linux-master/drivers/gpu/drm/xe/ |
H A D | xe_pt.c | 24 /** @children: Array of page-table child nodes */ 472 struct xe_ptw **child, 535 if (covers || !*child) { 548 *child = &xe_child->base; 725 struct xe_ptw **child, 731 struct xe_pt *xe_child = container_of(*child, typeof(*xe_child), base); 734 XE_WARN_ON(!*child); 739 * with the child of that entry rather than the parent, so need to 1373 const struct xe_pt *child, 1383 ((next - addr) >> shift) == child 470 xe_pt_stage_bind_entry(struct xe_ptw *parent, pgoff_t offset, unsigned int level, u64 addr, u64 next, struct xe_ptw **child, enum page_walk_action *action, struct xe_pt_walk *walk) argument 723 xe_pt_zap_ptes_entry(struct xe_ptw *parent, pgoff_t offset, unsigned int level, u64 addr, u64 next, struct xe_ptw **child, enum page_walk_action *action, struct xe_pt_walk *walk) argument 1372 xe_pt_check_kill(u64 addr, u64 next, unsigned int level, const struct xe_pt *child, enum page_walk_action *action, struct xe_pt_walk *walk) argument 1399 xe_pt_stage_unbind_entry(struct xe_ptw *parent, pgoff_t offset, unsigned int level, u64 addr, u64 next, struct xe_ptw **child, enum page_walk_action *action, struct xe_pt_walk *walk) argument 1416 xe_pt_stage_unbind_post_descend(struct xe_ptw *parent, pgoff_t offset, unsigned int level, u64 addr, u64 next, struct xe_ptw **child, enum page_walk_action *action, struct xe_pt_walk *walk) argument [all...] |
/linux-master/drivers/gpu/drm/i915/gt/uc/ |
H A D | intel_guc_submission.c | 417 * + sizeof(struct guc_process_desc) child go 418 * + CACHELINE_BYTES child join[0] 420 * + CACHELINE_BYTES child join[n - 1] 853 struct intel_context *child; local 885 for_each_child(ce, child) 886 *wqi++ = child->ring->tail / sizeof(u64); 2318 struct intel_context *child; local 2321 for_each_child(ce, child) 2322 child->guc_id.id = ce->guc_id.id + i++; 2404 struct intel_context *child; local 2427 struct intel_context *child; local 2742 struct intel_context *child; local 5500 struct intel_context *child; local [all...] |
/linux-master/drivers/gpu/drm/i915/gem/ |
H A D | i915_gem_execbuffer.c | 1975 * the last child. Requests are added in the reverse order, from the last child 2567 struct intel_context *ce = eb->context, *child; local 2584 for_each_child(ce, child) { 2585 err = intel_context_pin_ww(child, &eb->ww); 2589 for_each_child(ce, child) { 2590 err = eb_pin_timeline(eb, child, throttle); 2603 for_each_child(ce, child) { 2605 mutex_lock(&child->timeline->mutex); 2606 intel_context_exit(child); 2618 struct intel_context *ce = eb->context, *child; local 2687 struct intel_context *ce, *child; local 2763 struct intel_context *child; local 3276 struct intel_context *child; local [all...] |
H A D | i915_gem_context.c | 1012 struct intel_context *ce = e->engines[count], *child; local 1017 for_each_child(ce, child) 1018 intel_context_unpin(child); 1165 struct intel_context *child; local 1174 for_each_child(ce, child) { 1175 ret = intel_context_pin(child); 1187 for_each_child(ce, child) { 1189 intel_context_unpin(child); 1210 struct intel_context *ce, *child; local 1247 for_each_child(ce, child) { [all...] |
/linux-master/drivers/gpu/drm/i915/display/ |
H A D | intel_bios.c | 65 /* Wrapper for VBT child device config */ 69 struct child_device_config child; member in struct:intel_bios_encoder_data 643 if (devdata && devdata->child.handle == DEVICE_HANDLE_LFP2) 646 drm_WARN_ON(&i915->drm, devdata && devdata->child.handle != DEVICE_HANDLE_LFP1); 1215 const struct child_device_config *child = &devdata->child; local 1218 if (child->slave_addr != SLAVE_ADDR1 && 1219 child->slave_addr != SLAVE_ADDR2) { 1226 if (child->dvo_port != DEVICE_PORT_DVOB && 1227 child 2132 const struct child_device_config *child = &devdata->child; local 2407 const struct child_device_config *child = &devdata->child; local 2610 const struct child_device_config *child = &devdata->child; local 2773 const struct child_device_config *child; local 2883 struct child_device_config *child; local 3340 const struct child_device_config *child = &devdata->child; local 3379 const struct child_device_config *child = &devdata->child; local 3429 const struct child_device_config *child = &devdata->child; local 3440 const struct child_device_config *child = &devdata->child; local 3473 const struct child_device_config *child = &devdata->child; local 3568 const struct child_device_config *child = &devdata->child; local [all...] |
/linux-master/drivers/cxl/core/ |
H A D | hdm.c | 204 for (p1 = cxlds->dpa_res.child; p1; p1 = p1->sibling) { 206 for (p2 = p1->child; p2; p2 = p2->sibling) 479 for (p = cxlds->ram_res.child, last = NULL; p; p = p->sibling) 486 for (p = cxlds->pmem_res.child, last = NULL; p; p = p->sibling) 508 if (cxlds->pmem_res.child && 509 skip_start == cxlds->pmem_res.child->start)
|
/linux-master/drivers/cxl/ |
H A D | acpi.c | 715 for (res = cxl->child; res; res = next) { 753 for (res = cxl_res->child; res; res = next) { 794 for (p = cxl_res->child; p; p = p->sibling) {
|
/linux-master/arch/x86/kvm/mmu/ |
H A D | mmu.c | 836 * child PTEs in FNAME(gpte_changed), i.e. guarantee either in-flight 1794 struct kvm_mmu_page *child; local 1802 child = spte_to_child_sp(ent); 1804 if (child->unsync_children) { 1805 if (mmu_pages_add(pvec, child, i)) 1808 ret = __mmu_unsync_walk(child, pvec); 1816 } else if (child->unsync) { 1818 if (mmu_pages_add(pvec, child, i)) 2297 * bit 21 in the PTE (the child here), KVM propagates that bit to the 2431 struct kvm_mmu_page *child; local 2454 struct kvm_mmu_page *child; local 2892 struct kvm_mmu_page *child; local [all...] |
/linux-master/arch/powerpc/platforms/maple/ |
H A D | pci.c | 52 higher = fixup_one_level_bus_range(node->child, higher); 77 bus_range[1] = fixup_one_level_bus_range(bridge->child, bus_range[1]); 574 struct device_node *np, *child; local 585 for_each_child_of_node(np, child) 586 PCI_DN(child)->busno = 0xf0;
|
/linux-master/arch/powerpc/kernel/ |
H A D | prom_init.c | 647 && (*nodep = call_prom("child", 1, 1, node)) != 0) 2545 phandle child; local 2580 child = call_prom("child", 1, 1, node); 2581 while (child != 0) { 2582 scan_dt_build_strings(child, mem_start, mem_end); 2583 child = call_prom("peer", 1, 1, child); 2590 phandle child; local 2692 child [all...] |
H A D | eeh_driver.c | 733 list_for_each_entry_safe(child_pe, tmp, &pe->child_list, child) 743 list_del(&pe->child); 932 * to accomplish the reset. Each child gets a report of the 933 * status ... if any child can't handle the reset, then the entire
|
H A D | eeh_pe.c | 92 list_add_tail(&pe->child, &eeh_phb_pe); 162 list_for_each_entry(pe, &eeh_phb_pe, child) { 191 next = pe->child.next; 198 return list_entry(next, struct eeh_pe, child); 208 * child PEs. The traversing is to be terminated once the 233 * PE and its child PEs. 301 * a new PE is created and inserted into the PE tree as a child of 380 list_add_tail(&pe->child, &new_pe_parent->child_list); 383 * Put the newly created PE into the child list and 405 struct eeh_pe *pe, *parent, *child; local [all...] |
/linux-master/arch/powerpc/include/asm/ |
H A D | pmac_feature.h | 365 extern struct macio_chip* macio_find(struct device_node* child, int type);
|
H A D | eeh.h | 87 struct list_head child; /* Memb. child_list/eeh_phb_pe */ member in struct:eeh_pe
|
/linux-master/arch/loongarch/kvm/ |
H A D | mmu.c | 154 kvm_pte_t *entry, *child; local 157 child = kvm->arch.pgd; 159 entry = kvm_pgtable_offset(&ctx, child, addr); 164 child = kvm_mmu_memory_cache_alloc(cache); 165 _kvm_pte_init(child, ctx.invalid_ptes[ctx.level - 1]); 166 kvm_set_pte(entry, __pa(child)); 170 child = (kvm_pte_t *)__va(PHYSADDR(*entry)); 174 entry = kvm_pgtable_offset(&ctx, child, addr); 188 kvm_pte_t *entry, *child; local 192 child 222 kvm_pte_t *entry, *child; local 728 kvm_pte_t val, *child; local [all...] |
/linux-master/arch/arm64/kvm/hyp/nvhe/ |
H A D | mem_protect.c | 436 static inline bool range_included(struct kvm_mem_range *child, argument 439 return parent->start <= child->start && child->end <= parent->end;
|
/linux-master/tools/testing/selftests/x86/ |
H A D | test_mremap_vdso.c | 60 pid_t child; local 65 child = fork(); 66 if (child == -1) 69 if (child == 0) { 96 if (waitpid(child, &status, 0) != child ||
|
H A D | amx.c | 498 pid_t parent, child, grandchild; local 513 /* fork() succeeded. Now in the child . */ 517 printf("\tFork a child.\n"); 518 child = fork(); 519 if (child < 0) { 521 } else if (child > 0) { 526 fatal_error("arch_prctl test child exit"); 534 printf("\tTest XCOMP_PERM at child.\n"); 541 printf("\tTest dynamic sigaltstack at child:\n"); 545 printf("\tTest XCOMP_PERM again at child 626 pid_t child, grandchild; local 887 pid_t child; local [all...] |
/linux-master/tools/testing/selftests/pidfd/ |
H A D | pidfd_fdinfo_test.c | 86 struct child { struct 92 static struct child clone_newns(int (*fn)(void *), void *args, 96 struct child ret; 120 ksft_print_msg("New child: %d, fd: %d\n", ret.pid, ret.fd); 125 static inline void child_close(struct child *child) argument 127 close(child->fd); 130 static inline int child_join(struct child *child, struct error *err) argument 134 r = wait_for_pid(child 150 child_join_close(struct child *child, struct error *err) argument [all...] |
/linux-master/tools/testing/selftests/perf_events/ |
H A D | watermark_signal.c | 48 pid_t child = -1; local 66 child = fork(); 67 EXPECT_GE(child, 0); 68 if (child == 0) 70 else if (child < 0) { 75 if (waitpid(child, &child_status, WSTOPPED) != child || 78 "failed to sycnhronize with child errno=%d status=%x\n", 84 fd = syscall(__NR_perf_event_open, &attr, child, -1, -1, 117 if (kill(child, SIGCON [all...] |
/linux-master/tools/testing/selftests/capabilities/ |
H A D | test_execve.c | 172 pid_t child = fork(); local 173 if (child == 0) { 176 } else if (child > 0) { 178 if (waitpid(child, &status, 0) != child ||
|
/linux-master/tools/testing/selftests/bpf/progs/ |
H A D | bench_local_storage_create.c | 43 int BPF_PROG(sched_process_fork, struct task_struct *parent, struct task_struct *child) argument 50 stg = bpf_task_storage_get(&task_storage_map, child, NULL,
|
/linux-master/tools/perf/util/ |
H A D | probe-finder.c | 501 /* Search child die for local variables and parameters. */ 1410 bool externs = !af->child; 1444 if (af->child && dwarf_haspc(die_mem, af->pf.addr)) 1504 af->child = true; 1510 /* Don't need to search child DIE for external vars. */ 1511 af->child = false;
|
/linux-master/tools/net/ynl/ |
H A D | ynl-gen-c.py | 1082 child = self.pure_nested_structs.get(child_name) 1083 if child: 1084 if not child.recursive: 1085 struct.child_nests.update(child.child_nests) 1086 child.request |= struct.request 1087 child.reply |= struct.reply
|
/linux-master/security/yama/ |
H A D | yama_lsm.c | 268 * @parent: the process to compare against while walking up from child 269 * @child: the process to start from while looking upwards for parent 271 * Returns 1 if child is a descendant of parent, 0 if not. 274 struct task_struct *child) 277 struct task_struct *walker = child; 279 if (!parent || !child) 350 * @child: task that current task is attempting to ptrace 355 static int yama_ptrace_access_check(struct task_struct *child, argument 360 /* require ptrace target be a child of ptracer on attach */ 368 if (!pid_alive(child)) 273 task_is_descendant(struct task_struct *parent, struct task_struct *child) argument [all...] |