/linux/drivers/gpu/drm/amd/amdkfd/ |
A D | kfd_migrate.c | 332 migrate->dst[i] = migrate_pfn(migrate->dst[i]); in svm_migrate_copy_to_vram() 419 memset(&migrate, 0, sizeof(migrate)); in svm_migrate_vma_to_vram() 420 migrate.vma = vma; in svm_migrate_vma_to_vram() 422 migrate.end = end; in svm_migrate_vma_to_vram() 432 migrate.src = buf; in svm_migrate_vma_to_vram() 433 migrate.dst = migrate.src + npages; in svm_migrate_vma_to_vram() 459 svm_migrate_successful_pages(&migrate), cpages, migrate.npages); in svm_migrate_vma_to_vram() 658 memset(&migrate, 0, sizeof(migrate)); in svm_migrate_vma_to_ram() 659 migrate.vma = vma; in svm_migrate_vma_to_ram() 661 migrate.end = end; in svm_migrate_vma_to_ram() [all …]
|
/linux/drivers/gpu/drm/i915/gt/ |
A D | selftest_migrate.c | 32 static int copy(struct intel_migrate *migrate, in copy() argument 33 int (*fn)(struct intel_migrate *migrate, in copy() argument 85 err = fn(migrate, &ww, src, dst, &rq); in copy() 133 int (*fn)(struct intel_migrate *migrate, in clear() argument 167 err = fn(migrate, &ww, obj, sz, &rq); in clear() 288 struct intel_migrate *migrate = arg; in live_migrate_copy() local 309 struct intel_migrate *migrate = arg; in live_migrate_clear() local 330 struct intel_migrate *migrate; member 352 thread[i].migrate = migrate; in threaded_migrate() 446 if (!gt->migrate.context) in intel_migrate_live_selftests() [all …]
|
A D | intel_gt_types.h | 179 struct intel_migrate migrate; member
|
/linux/mm/ |
A D | migrate.c | 2212 migrate->dst[migrate->npages] = 0; in migrate_vma_collect_skip() 2213 migrate->src[migrate->npages++] = 0; in migrate_vma_collect_skip() 2232 migrate->src[migrate->npages] = MIGRATE_PFN_MIGRATE; in migrate_vma_collect_hole() 2233 migrate->dst[migrate->npages] = 0; in migrate_vma_collect_hole() 2234 migrate->npages++; in migrate_vma_collect_hole() 2414 migrate->dst[migrate->npages] = 0; in migrate_vma_collect_pmd() 2415 migrate->src[migrate->npages++] = mpfn; in migrate_vma_collect_pmd() 2450 migrate->vma, migrate->vma->vm_mm, migrate->start, migrate->end, in migrate_vma_collect() 2454 walk_page_range(migrate->vma->vm_mm, migrate->start, migrate->end, in migrate_vma_collect() 2458 migrate->end = migrate->start + (migrate->npages << PAGE_SHIFT); in migrate_vma_collect() [all …]
|
A D | .migrate.o.cmd | 1 …migrate.o := /usr/bin/ccache /home/test/workspace/code/optee_3.16/build/../toolchains/aarch64/bin/… 3 source_mm/migrate.o := mm/migrate.c 5 deps_mm/migrate.o := \ 31 include/linux/migrate.h \ 1287 include/trace/events/migrate.h \ 1292 mm/migrate.o: $(deps_mm/migrate.o) 1294 $(deps_mm/migrate.o):
|
A D | built-in.a | 62 migrate.o/
|
/linux/drivers/gpu/drm/nouveau/ |
A D | nouveau_dmem.c | 81 struct nouveau_dmem_migrate migrate; member 164 if (drm->dmem->migrate.copy_func(drm, 1, NOUVEAU_APER_HOST, *dma_addr, in nouveau_dmem_fault_copy_one() 212 nouveau_fence_new(dmem->migrate.chan, false, &fence); in nouveau_dmem_migrate_to_ram() 399 struct nvif_push *push = drm->dmem->migrate.chan->chan.push; in nvc0b5_migrate_copy() 472 struct nvif_push *push = drm->dmem->migrate.chan->chan.push; in nvc0b5_migrate_clear() 532 drm->dmem->migrate.copy_func = nvc0b5_migrate_copy; in nouveau_dmem_migrate_init() 533 drm->dmem->migrate.clear_func = nvc0b5_migrate_clear; in nouveau_dmem_migrate_init() 534 drm->dmem->migrate.chan = drm->ttm.chan; in nouveau_dmem_migrate_init() 590 if (drm->dmem->migrate.copy_func(drm, 1, in nouveau_dmem_migrate_copy_one() 595 if (drm->dmem->migrate.clear_func(drm, page_size(dpage), in nouveau_dmem_migrate_copy_one() [all …]
|
/linux/include/linux/ |
A D | psci.h | 29 int (*migrate)(unsigned long cpuid); member 41 u32 migrate; member
|
A D | migrate.h | 163 void migrate_vma_pages(struct migrate_vma *migrate); 164 void migrate_vma_finalize(struct migrate_vma *migrate);
|
/linux/tools/testing/selftests/kvm/x86_64/ |
A D | xapic_ipi_test.c | 368 bool *migrate, int *delay_usecs) in get_cmdline_args() argument 380 *migrate = true; in get_cmdline_args() 406 bool migrate = false; in main() local 412 get_cmdline_args(argc, argv, &run_secs, &migrate, &delay_usecs); in main() 473 if (!migrate) in main()
|
/linux/drivers/gpu/drm/i915/gem/ |
A D | i915_gem_ttm.c | 459 if (!i915->gt.migrate.context || intel_gt_is_wedged(&i915->gt)) in i915_ttm_accel_move() 467 intel_engine_pm_get(i915->gt.migrate.context->engine); in i915_ttm_accel_move() 468 ret = intel_context_migrate_clear(i915->gt.migrate.context, NULL, in i915_ttm_accel_move() 477 intel_engine_pm_put(i915->gt.migrate.context->engine); in i915_ttm_accel_move() 483 intel_engine_pm_get(i915->gt.migrate.context->engine); in i915_ttm_accel_move() 484 ret = intel_context_migrate_copy(i915->gt.migrate.context, in i915_ttm_accel_move() 494 intel_engine_pm_put(i915->gt.migrate.context->engine); in i915_ttm_accel_move() 891 .migrate = i915_ttm_migrate,
|
A D | i915_gem_object.c | 608 if (!obj->ops->migrate) in i915_gem_object_can_migrate() 667 if (!obj->ops->migrate) { in i915_gem_object_migrate() 673 return obj->ops->migrate(obj, mr); in i915_gem_object_migrate()
|
/linux/drivers/firmware/psci/ |
A D | psci.c | 235 return __psci_migrate(psci_0_1_function_ids.migrate, cpuid); in psci_0_1_migrate() 483 .migrate = psci_0_2_migrate, in psci_0_2_set_functions() 581 psci_0_1_function_ids.migrate = id; in psci_0_1_init() 582 psci_ops.migrate = psci_0_1_migrate; in psci_0_1_init()
|
/linux/arch/um/drivers/ |
A D | Kconfig | 171 migrate to UML_NET_VECTOR. 188 migrate to UML_NET_VECTOR. 209 migrate to UML_NET_VECTOR. 230 migrate to UML_NET_VECTOR. 267 migrate to UML_NET_VECTOR. 290 migrate to UML_NET_VECTOR. 309 migrate to UML_NET_VECTOR. 335 migrate to UML_NET_VECTOR.
|
/linux/drivers/gpu/drm/i915/selftests/ |
A D | i915_perf_selftests.h | 20 selftest(migrate, intel_migrate_perf_selftests)
|
A D | i915_live_selftests.h | 29 selftest(migrate, intel_migrate_live_selftests)
|
/linux/tools/perf/scripts/python/ |
A D | sched-migration.py | 191 def migrate(self, ts_list, new, old_cpu, new_cpu): member in TimeSlice 351 def migrate(self, headers, pid, prio, orig_cpu, dest_cpu): member in SchedEventProxy 353 ts.migrate(self.timeslices, pid, orig_cpu, dest_cpu) 418 parser.migrate(headers, pid, prio, orig_cpu, dest_cpu)
|
/linux/fs/ocfs2/dlm/ |
A D | dlmmaster.c | 3027 struct dlm_migrate_request migrate; in dlm_do_migrate_request() local 3031 memset(&migrate, 0, sizeof(migrate)); in dlm_do_migrate_request() 3033 memcpy(migrate.name, res->lockname.name, migrate.namelen); in dlm_do_migrate_request() 3034 migrate.new_master = new_master; in dlm_do_migrate_request() 3035 migrate.master = master; in dlm_do_migrate_request() 3055 &migrate, sizeof(migrate), nodenum, in dlm_do_migrate_request() 3060 migrate.namelen, migrate.name, ret, nodenum); in dlm_do_migrate_request() 3113 name = migrate->name; in dlm_migrate_request_handler() 3114 namelen = migrate->namelen; in dlm_migrate_request_handler() 3149 migrate->new_master, in dlm_migrate_request_handler() [all …]
|
/linux/arch/arm64/kvm/hyp/nvhe/ |
A D | psci-relay.c | 53 is_psci_0_1(migrate, func_id)); in is_psci_0_1_call() 226 if (is_psci_0_1(cpu_off, func_id) || is_psci_0_1(migrate, func_id)) in psci_0_1_handler()
|
/linux/fs/ext4/ |
A D | built-in.a | 20 migrate.o/
|
A D | Makefile | 10 indirect.o inline.o inode.o ioctl.o mballoc.o migrate.o \
|
/linux/net/xfrm/ |
A D | Kconfig | 61 bool "Transformation migrate database" 133 <draft-sugimoto-mip6-pfkey-migrate>.
|
/linux/drivers/staging/r8188eu/ |
A D | TODO | 10 * Remove the HAL layer and migrate its functionality into the relevant parts of
|
/linux/Documentation/ABI/testing/ |
A D | sysfs-kernel-mm-numa | 17 systems to migrate pages from fast tiers to slow tiers
|
/linux/Documentation/trace/ |
A D | hwlat_detector.rst | 81 By default, one hwlat detector's kernel thread will migrate across each CPU 87 - round-robin: migrate across each CPU specified in cpumask [default]
|