/linux/mm/ |
A D | memremap.c | 102 return (range->start + range_len(range)) >> PAGE_SHIFT; in pfn_end() 150 PHYS_PFN(range_len(range))); in pageunmap_range() 153 PHYS_PFN(range_len(range)), NULL); in pageunmap_range() 155 arch_remove_memory(range->start, range_len(range), in pageunmap_range() 161 untrack_pfn(NULL, PHYS_PFN(range->start), range_len(range)); in pageunmap_range() 223 is_ram = region_intersects(range->start, range_len(range), in pagemap_range() 242 range_len(range)); in pagemap_range() 266 PHYS_PFN(range_len(range)), params); in pagemap_range() 283 PHYS_PFN(range_len(range)), params->altmap, in pagemap_range() 297 PHYS_PFN(range_len(range)), pgmap); in pagemap_range() [all …]
|
/linux/drivers/comedi/ |
A D | range.c | 110 int chan, range_len, i; in comedi_check_chanlist() local 116 range_len = s->range_table->length; in comedi_check_chanlist() 118 range_len = s->range_table_list[chan]->length; in comedi_check_chanlist() 120 range_len = 0; in comedi_check_chanlist() 122 CR_RANGE(chanspec) >= range_len) { in comedi_check_chanlist() 125 i, chanspec, chan, range_len); in comedi_check_chanlist()
|
/linux/drivers/dax/ |
A D | kmem.c | 74 total_len += range_len(&range); in dev_dax_kmem_probe() 105 res = request_mem_region(range.start, range_len(&range), data->res_name); in dev_dax_kmem_probe() 133 range_len(&range), kmem_name, MHP_NID_IS_MGID); in dev_dax_kmem_probe() 182 rc = remove_memory(range.start, range_len(&range)); in dev_dax_kmem_remove()
|
A D | bus.c | 145 size += range_len(&dev_dax->ranges[i].range); in dev_dax_size() 380 __release_region(&dax_region->res, range->start, range_len(range)); in trim_dev_dax_range() 559 || !IS_ALIGNED(range_len(range), align)) in alloc_dax_region() 786 pgoff += PHYS_PFN(range_len(&ranges[i].range)); in alloc_dev_dax_range() 879 shrink = min_t(u64, to_shrink, range_len(range)); in dev_dax_shrink() 880 if (shrink >= range_len(range)) { in dev_dax_shrink() 900 return adjust_dev_dax_range(dev_dax, adjust, range_len(range) in dev_dax_shrink() 1087 to_alloc = range_len(&r); in mapping_store() 1111 size_t len = range_len(&dev_dax->ranges[i].range); in dev_dax_validate_align() 1365 if (dev_dax->nr_range && range_len(&dev_dax->ranges[0].range)) { in devm_create_dev_dax()
|
A D | device.c | 65 pgoff_end = dax_range->pgoff + PHYS_PFN(range_len(range)) - 1; in dax_pgoff_to_phys() 418 range_len(range), dev_name(dev))) { in dev_dax_probe()
|
/linux/fs/btrfs/ |
A D | ordered-data.h | 207 const u64 range_start, const u64 range_len); 209 const u64 range_start, const u64 range_len);
|
A D | ordered-data.c | 626 const u64 range_start, const u64 range_len) in btrfs_wait_ordered_extents() argument 634 const u64 range_end = range_start + range_len; in btrfs_wait_ordered_extents() 682 const u64 range_start, const u64 range_len) in btrfs_wait_ordered_roots() argument 703 range_start, range_len); in btrfs_wait_ordered_roots()
|
A D | ioctl.c | 1175 u64 range_len; in defrag_collect_targets() local 1219 range_len = min(extent_map_end(em), start + len) - cur; in defrag_collect_targets() 1232 last->len += range_len; in defrag_collect_targets() 1246 new->len = range_len; in defrag_collect_targets() 1417 u32 range_len = entry->len; in defrag_one_cluster() local 1424 range_len = min_t(u32, range_len, in defrag_one_cluster() 1430 ((entry->start + range_len - 1) >> PAGE_SHIFT) - in defrag_one_cluster() 1438 ret = defrag_one_range(inode, entry->start, range_len, in defrag_one_cluster() 1442 *sectors_defragged += range_len; in defrag_one_cluster()
|
A D | qgroup.c | 3703 u64 range_len = unode->aux - range_start + 1; in qgroup_free_reserved_data() local 3711 range_start + range_len <= start) in qgroup_free_reserved_data() 3714 free_len = min(start + len, range_start + range_len) - in qgroup_free_reserved_data()
|
/linux/include/linux/ |
A D | range.h | 11 static inline u64 range_len(const struct range *range) in range_len() function
|
/linux/drivers/nvdimm/ |
A D | badrange.c | 237 len = range->start + range_len(range) in badblocks_populate() 252 len = range_len(range); in badblocks_populate()
|
A D | pmem.c | 455 range_len(&pmem->pgmap.range); in pmem_attach_disk()
|
A D | pfn_devs.c | 702 nd_pfn->npfns = PHYS_PFN((range_len(range) - offset)); in __nvdimm_setup_pfn()
|
/linux/tools/testing/nvdimm/ |
A D | dax-dev.c | 20 pgoff_end = dax_range->pgoff + PHYS_PFN(range_len(range)) - 1; in dax_pgoff_to_phys()
|
/linux/drivers/cxl/core/ |
A D | memdev.c | 71 unsigned long long len = range_len(&cxlm->ram_range); in ram_size_show() 84 unsigned long long len = range_len(&cxlm->pmem_range); in pmem_size_show()
|
A D | bus.c | 59 return sysfs_emit(buf, "%#llx\n", range_len(&cxld->range)); in size_show()
|
/linux/drivers/net/wireless/intel/iwlwifi/fw/ |
A D | dbg.c | 525 u32 range_len, void *ptr) in iwl_dump_prph() argument 541 for (i = 0; i < range_len; i++) { in iwl_dump_prph() 604 u32 range_len, void *ptr) in iwl_fw_get_prph_len() argument 612 for (i = 0; i < range_len; i++) { in iwl_fw_get_prph_len() 629 u32 range_len; in iwl_fw_prph_handler() local 632 range_len = ARRAY_SIZE(iwl_prph_dump_addr_ax210); in iwl_fw_prph_handler() 633 handler(fwrt, iwl_prph_dump_addr_ax210, range_len, ptr); in iwl_fw_prph_handler() 636 range_len = ARRAY_SIZE(iwl_prph_dump_addr_22000); in iwl_fw_prph_handler() 639 range_len = ARRAY_SIZE(iwl_prph_dump_addr_comm); in iwl_fw_prph_handler() 640 handler(fwrt, iwl_prph_dump_addr_comm, range_len, ptr); in iwl_fw_prph_handler() [all …]
|
/linux/drivers/dax/pmem/ |
A D | core.c | 67 .size = range_len(&range), in __dax_pmem_probe()
|
/linux/tools/testing/cxl/test/ |
A D | mem.c | 234 if (range_len(&cxlm->pmem_range) && IS_ENABLED(CONFIG_CXL_PMEM)) in cxl_mock_mem_probe()
|
A D | cxl.c | 201 range_len(&res->range)); in depopulate_all_mock_resources()
|
/linux/lib/ |
A D | test_hmm.c | 503 pfn_last = pfn_first + (range_len(&devmem->pagemap.range) >> PAGE_SHIFT); in dmirror_allocate_chunk() 532 release_mem_region(devmem->pagemap.range.start, range_len(&devmem->pagemap.range)); in dmirror_allocate_chunk() 1225 range_len(&devmem->pagemap.range)); in dmirror_device_remove()
|
/linux/drivers/gpu/drm/nouveau/ |
A D | nouveau_dmem.c | 299 release_mem_region(chunk->pagemap.range.start, range_len(&chunk->pagemap.range)); in nouveau_dmem_chunk_alloc() 387 range_len(&chunk->pagemap.range)); in nouveau_dmem_fini()
|
/linux/arch/powerpc/kvm/ |
A D | book3s_hv_uvmem.c | 693 (range_len(&kvmppc_uvmem_pgmap.range) >> PAGE_SHIFT); in kvmppc_uvmem_get_page() 1212 range_len(&kvmppc_uvmem_pgmap.range)); in kvmppc_uvmem_free()
|
/linux/drivers/cxl/ |
A D | pci.c | 524 if (range_len(&cxlm->pmem_range) && IS_ENABLED(CONFIG_CXL_PMEM)) in cxl_pci_probe()
|
/linux/drivers/pci/ |
A D | p2pdma.c | 221 range_len(&pgmap->range), dev_to_node(&pdev->dev), in pci_p2pdma_add_resource()
|