/linux-master/drivers/gpu/drm/xe/ |
H A D | xe_vm.c | 1440 migrate_vm = xe_migrate_get_vm(tile->migrate);
|
H A D | xe_gt.c | 452 tile->migrate = xe_migrate_init(tile); 453 if (IS_ERR(tile->migrate)) { 454 err = PTR_ERR(tile->migrate);
|
H A D | xe_migrate.c | 39 * struct xe_migrate - migrate context. 87 * xe_tile_migrate_engine() - Get this tile's migrate engine. 90 * Returns the default migrate engine of this tile. 93 * Return: The default migrate engine 97 return tile->migrate->q; 128 * migrate offset 305 * reserved copy engine to be used as logical mask for migrate engine. 307 * migrate jobs servicing the faults gets stuck behind the job that faulted. 328 * xe_migrate_init() - Initialize a migrate context 331 * Return: Pointer to a migrate contex [all...] |
H A D | xe_pt.c | 1278 fence = xe_migrate_update_pgtables(tile->migrate, 1626 fence = xe_migrate_update_pgtables(tile->migrate,
|
H A D | xe_device_types.h | 202 /** @migrate: Migration helper for vram blits and clearing */ 203 struct xe_migrate *migrate; member in struct:xe_tile
|
H A D | xe_exec_queue.c | 125 * some engines we use the kernels migrate vm underneath which offers no 604 migrate_vm = xe_migrate_get_vm(gt_to_tile(gt)->migrate); 757 * Typically vm->resv, but user-created timeline locks use the migrate vm 758 * and never grabs the migrate vm->resv so we have a race there.
|
H A D | xe_bo.c | 124 return tile->migrate; 622 struct xe_migrate *migrate = NULL; local 708 migrate = bo->tile->migrate; 710 migrate = mem_type_to_migrate(xe, new_mem->mem_type); 712 migrate = mem_type_to_migrate(xe, old_mem_type); 714 migrate = xe->tiles[0].migrate; 716 xe_assert(xe, migrate); 754 fence = xe_migrate_clear(migrate, b [all...] |
/linux-master/kernel/sched/ |
H A D | fair.c | 1840 * Cannot migrate to memoryless nodes. 1884 * Allow first faults or private faults to migrate immediately early in 1914 /* Always allow migrate on private faults */ 2110 * Clear previous best_cpu/rq numa-migrate flag, since task now 2405 /* Skip this CPU if the source task cannot migrate */ 2550 /* Attempt to migrate a task to a CPU on the preferred node. */ 2567 /* Otherwise, try migrate to a CPU on the preferred node */ 2900 /* Cannot migrate task to CPU-less node */ 3140 * Retry to migrate task to preferred node periodically, in case it 3208 unsigned long migrate, next_sca local [all...] |
/linux-master/mm/ |
H A D | hugetlb.c | 33 #include <linux/migrate.h>
|
H A D | shmem.c | 72 #include <linux/migrate.h>
|
H A D | page_owner.c | 10 #include <linux/migrate.h> 577 "Page has been migrated, last migrate reason: %s\n", 647 pr_alert("page has been migrated, last migrate reason: %s\n",
|
H A D | memory-failure.c | 51 #include <linux/migrate.h> 1395 /* Soft offline could migrate non-LRU movable pages */
|
H A D | huge_memory.c | 30 #include <linux/migrate.h>
|
H A D | gup.c | 18 #include <linux/migrate.h> 2133 * We can't migrate pages with unexpected references, so drop 2176 * will migrate those pages away, unpin all the pages in the range and return
|
H A D | memory.c | 69 #include <linux/migrate.h>
|
H A D | Makefile | 88 obj-$(CONFIG_MIGRATION) += migrate.o
|
H A D | filemap.c | 45 #include <linux/migrate.h>
|
/linux-master/drivers/gpu/drm/amd/amdkfd/ |
H A D | kfd_migrate.c | 27 #include <linux/migrate.h> 181 * @mfence: migrate fence 263 static unsigned long svm_migrate_unsuccessful_pages(struct migrate_vma *migrate) argument 268 for (i = 0; i < migrate->npages; i++) { 269 if (migrate->src[i] & MIGRATE_PFN_VALID && 270 !(migrate->src[i] & MIGRATE_PFN_MIGRATE)) 278 struct migrate_vma *migrate, struct dma_fence **mfence, 281 uint64_t npages = migrate->cpages; 302 migrate->dst[i] = svm_migrate_addr_to_pfn(adev, dst[i]); 303 svm_migrate_get_vram_page(prange, migrate 277 svm_migrate_copy_to_vram(struct kfd_node *node, struct svm_range *prange, struct migrate_vma *migrate, struct dma_fence **mfence, dma_addr_t *scratch, uint64_t ttm_res_offset) argument 390 struct migrate_vma migrate = { 0 }; local 576 svm_migrate_copy_to_ram(struct amdgpu_device *adev, struct svm_range *prange, struct migrate_vma *migrate, struct dma_fence **mfence, dma_addr_t *scratch, uint64_t npages) argument 692 struct migrate_vma migrate = { 0 }; local [all...] |
/linux-master/fs/btrfs/ |
H A D | inode.c | 32 #include <linux/migrate.h> 4211 * plenty of slack room in the global reserve to migrate, otherwise we cannot
|
/linux-master/fs/bcachefs/ |
H A D | fs-io.c | 29 #include <linux/migrate.h>
|
H A D | Makefile | 63 migrate.o \
|
/linux-master/fs/ |
H A D | aio.c | 40 #include <linux/migrate.h>
|
/linux-master/drivers/gpu/drm/i915/gt/ |
H A D | intel_gt.c | 762 intel_migrate_init(>->migrate, gt); 787 intel_migrate_fini(>->migrate);
|
/linux-master/arch/arm64/kvm/ |
H A D | arm.c | 2137 init_psci_0_1_impl_state(kvm_host_psci_config, migrate);
|
/linux-master/drivers/gpu/drm/nouveau/ |
H A D | nouveau_dmem.c | 42 #include <linux/migrate.h> 82 struct nouveau_dmem_migrate migrate; member in struct:nouveau_dmem 152 if (drm->dmem->migrate.copy_func(drm, 1, NOUVEAU_APER_HOST, *dma_addr, 183 * FIXME what we really want is to find some heuristic to migrate more 212 nouveau_fence_new(&fence, dmem->migrate.chan); 405 nouveau_fence_new(&fence, chunk->drm->dmem->migrate.chan); 446 struct nvif_push *push = drm->dmem->migrate.chan->chan.push; 519 struct nvif_push *push = drm->dmem->migrate.chan->chan.push; 579 drm->dmem->migrate.copy_func = nvc0b5_migrate_copy; 580 drm->dmem->migrate [all...] |