Searched refs:migrate (Results 1 - 25 of 87) sorted by last modified time

1234

/linux-master/drivers/gpu/drm/xe/
H A Dxe_vm.c1440 migrate_vm = xe_migrate_get_vm(tile->migrate);
H A Dxe_gt.c452 tile->migrate = xe_migrate_init(tile);
453 if (IS_ERR(tile->migrate)) {
454 err = PTR_ERR(tile->migrate);
H A Dxe_migrate.c39 * struct xe_migrate - migrate context.
87 * xe_tile_migrate_engine() - Get this tile's migrate engine.
90 * Returns the default migrate engine of this tile.
93 * Return: The default migrate engine
97 return tile->migrate->q;
128 * migrate offset
305 * reserved copy engine to be used as logical mask for migrate engine.
307 * migrate jobs servicing the faults gets stuck behind the job that faulted.
328 * xe_migrate_init() - Initialize a migrate context
331 * Return: Pointer to a migrate contex
[all...]
H A Dxe_pt.c1278 fence = xe_migrate_update_pgtables(tile->migrate,
1626 fence = xe_migrate_update_pgtables(tile->migrate,
H A Dxe_device_types.h202 /** @migrate: Migration helper for vram blits and clearing */
203 struct xe_migrate *migrate; member in struct:xe_tile
H A Dxe_exec_queue.c125 * some engines we use the kernels migrate vm underneath which offers no
604 migrate_vm = xe_migrate_get_vm(gt_to_tile(gt)->migrate);
757 * Typically vm->resv, but user-created timeline locks use the migrate vm
758 * and never grabs the migrate vm->resv so we have a race there.
H A Dxe_bo.c124 return tile->migrate;
622 struct xe_migrate *migrate = NULL; local
708 migrate = bo->tile->migrate;
710 migrate = mem_type_to_migrate(xe, new_mem->mem_type);
712 migrate = mem_type_to_migrate(xe, old_mem_type);
714 migrate = xe->tiles[0].migrate;
716 xe_assert(xe, migrate);
754 fence = xe_migrate_clear(migrate, b
[all...]
/linux-master/kernel/sched/
H A Dfair.c1840 * Cannot migrate to memoryless nodes.
1884 * Allow first faults or private faults to migrate immediately early in
1914 /* Always allow migrate on private faults */
2110 * Clear previous best_cpu/rq numa-migrate flag, since task now
2405 /* Skip this CPU if the source task cannot migrate */
2550 /* Attempt to migrate a task to a CPU on the preferred node. */
2567 /* Otherwise, try migrate to a CPU on the preferred node */
2900 /* Cannot migrate task to CPU-less node */
3140 * Retry to migrate task to preferred node periodically, in case it
3208 unsigned long migrate, next_sca local
[all...]
/linux-master/mm/
H A Dhugetlb.c33 #include <linux/migrate.h>
H A Dshmem.c72 #include <linux/migrate.h>
H A Dpage_owner.c10 #include <linux/migrate.h>
577 "Page has been migrated, last migrate reason: %s\n",
647 pr_alert("page has been migrated, last migrate reason: %s\n",
H A Dmemory-failure.c51 #include <linux/migrate.h>
1395 /* Soft offline could migrate non-LRU movable pages */
H A Dhuge_memory.c30 #include <linux/migrate.h>
H A Dgup.c18 #include <linux/migrate.h>
2133 * We can't migrate pages with unexpected references, so drop
2176 * will migrate those pages away, unpin all the pages in the range and return
H A Dmemory.c69 #include <linux/migrate.h>
H A DMakefile88 obj-$(CONFIG_MIGRATION) += migrate.o
H A Dfilemap.c45 #include <linux/migrate.h>
/linux-master/drivers/gpu/drm/amd/amdkfd/
H A Dkfd_migrate.c27 #include <linux/migrate.h>
181 * @mfence: migrate fence
263 static unsigned long svm_migrate_unsuccessful_pages(struct migrate_vma *migrate) argument
268 for (i = 0; i < migrate->npages; i++) {
269 if (migrate->src[i] & MIGRATE_PFN_VALID &&
270 !(migrate->src[i] & MIGRATE_PFN_MIGRATE))
278 struct migrate_vma *migrate, struct dma_fence **mfence,
281 uint64_t npages = migrate->cpages;
302 migrate->dst[i] = svm_migrate_addr_to_pfn(adev, dst[i]);
303 svm_migrate_get_vram_page(prange, migrate
277 svm_migrate_copy_to_vram(struct kfd_node *node, struct svm_range *prange, struct migrate_vma *migrate, struct dma_fence **mfence, dma_addr_t *scratch, uint64_t ttm_res_offset) argument
390 struct migrate_vma migrate = { 0 }; local
576 svm_migrate_copy_to_ram(struct amdgpu_device *adev, struct svm_range *prange, struct migrate_vma *migrate, struct dma_fence **mfence, dma_addr_t *scratch, uint64_t npages) argument
692 struct migrate_vma migrate = { 0 }; local
[all...]
/linux-master/fs/btrfs/
H A Dinode.c32 #include <linux/migrate.h>
4211 * plenty of slack room in the global reserve to migrate, otherwise we cannot
/linux-master/fs/bcachefs/
H A Dfs-io.c29 #include <linux/migrate.h>
H A DMakefile63 migrate.o \
/linux-master/fs/
H A Daio.c40 #include <linux/migrate.h>
/linux-master/drivers/gpu/drm/i915/gt/
H A Dintel_gt.c762 intel_migrate_init(&gt->migrate, gt);
787 intel_migrate_fini(&gt->migrate);
/linux-master/arch/arm64/kvm/
H A Darm.c2137 init_psci_0_1_impl_state(kvm_host_psci_config, migrate);
/linux-master/drivers/gpu/drm/nouveau/
H A Dnouveau_dmem.c42 #include <linux/migrate.h>
82 struct nouveau_dmem_migrate migrate; member in struct:nouveau_dmem
152 if (drm->dmem->migrate.copy_func(drm, 1, NOUVEAU_APER_HOST, *dma_addr,
183 * FIXME what we really want is to find some heuristic to migrate more
212 nouveau_fence_new(&fence, dmem->migrate.chan);
405 nouveau_fence_new(&fence, chunk->drm->dmem->migrate.chan);
446 struct nvif_push *push = drm->dmem->migrate.chan->chan.push;
519 struct nvif_push *push = drm->dmem->migrate.chan->chan.push;
579 drm->dmem->migrate.copy_func = nvc0b5_migrate_copy;
580 drm->dmem->migrate
[all...]

Completed in 525 milliseconds

1234