/linux/drivers/gpu/drm/ |
A D | drm_gem_shmem_helper.c | 95 return shmem; in __drm_gem_shmem_create() 138 if (shmem->sgt) { in drm_gem_shmem_free_object() 144 if (shmem->pages) in drm_gem_shmem_free_object() 153 kfree(shmem); in drm_gem_shmem_free_object() 178 if (shmem->map_wc) in drm_gem_shmem_get_pages_locked() 224 if (shmem->map_wc) in drm_gem_shmem_put_pages_locked() 314 shmem->vaddr = vmap(shmem->pages, obj->size >> PAGE_SHIFT, in drm_gem_shmem_vmap_locked() 425 return shmem; in drm_gem_shmem_create_with_handle() 437 return shmem; in drm_gem_shmem_create_with_handle() 475 shmem->madv = -1; in drm_gem_shmem_purge_locked() [all …]
|
/linux/drivers/firmware/arm_scmi/ |
A D | shmem.c | 32 void shmem_tx_prepare(struct scmi_shared_mem __iomem *shmem, in shmem_tx_prepare() argument 41 spin_until_cond(ioread32(&shmem->channel_status) & in shmem_tx_prepare() 44 iowrite32(0x0, &shmem->channel_status); in shmem_tx_prepare() 46 &shmem->flags); in shmem_tx_prepare() 47 iowrite32(sizeof(shmem->msg_header) + xfer->tx.len, &shmem->length); in shmem_tx_prepare() 53 u32 shmem_read_header(struct scmi_shared_mem __iomem *shmem) in shmem_read_header() argument 55 return ioread32(&shmem->msg_header); in shmem_read_header() 61 xfer->hdr.status = ioread32(shmem->msg_payload); in shmem_fetch_response() 64 ioread32(&shmem->length) - 8); in shmem_fetch_response() 90 xfer_id = MSG_XTRACT_TOKEN(ioread32(&shmem->msg_header)); in shmem_poll_done() [all …]
|
A D | mailbox.c | 30 struct scmi_shared_mem __iomem *shmem; member 39 shmem_tx_prepare(smbox->shmem, m); in tx_prepare() 61 struct device_node *shmem; in mailbox_chan_setup() local 71 shmem = of_parse_phandle(cdev->of_node, "shmem", idx); in mailbox_chan_setup() 75 ret = of_address_to_resource(shmem, 0, &res); in mailbox_chan_setup() 76 of_node_put(shmem); in mailbox_chan_setup() 83 smbox->shmem = devm_ioremap(dev, res.start, size); in mailbox_chan_setup() 84 if (!smbox->shmem) { in mailbox_chan_setup() 161 shmem_fetch_response(smbox->shmem, xfer); in mailbox_fetch_response() 176 shmem_clear_channel(smbox->shmem); in mailbox_clear_channel() [all …]
|
A D | smc.c | 34 struct scmi_shared_mem __iomem *shmem; member 90 scmi_info->shmem = devm_ioremap(dev, res.start, size); in smc_chan_setup() 91 if (!scmi_info->shmem) { in smc_chan_setup() 147 shmem_tx_prepare(scmi_info->shmem, xfer); in smc_send_message() 158 shmem_read_header(scmi_info->shmem), NULL); in smc_send_message() 173 shmem_fetch_response(scmi_info->shmem, xfer); in smc_fetch_response() 181 return shmem_poll_done(scmi_info->shmem, xfer); in smc_poll_done()
|
A D | common.h | 431 void shmem_tx_prepare(struct scmi_shared_mem __iomem *shmem, 433 u32 shmem_read_header(struct scmi_shared_mem __iomem *shmem); 434 void shmem_fetch_response(struct scmi_shared_mem __iomem *shmem, 436 void shmem_fetch_notification(struct scmi_shared_mem __iomem *shmem, 438 void shmem_clear_channel(struct scmi_shared_mem __iomem *shmem); 439 bool shmem_poll_done(struct scmi_shared_mem __iomem *shmem,
|
A D | built-in.a | 14 shmem.o/
|
/linux/drivers/gpu/drm/virtio/ |
A D | virtgpu_object.c | 72 if (shmem->pages) { in virtio_gpu_cleanup_object() 73 if (shmem->mapped) { in virtio_gpu_cleanup_object() 76 shmem->mapped = 0; in virtio_gpu_cleanup_object() 79 sg_free_table(shmem->pages); in virtio_gpu_cleanup_object() 80 kfree(shmem->pages); in virtio_gpu_cleanup_object() 81 shmem->pages = NULL; in virtio_gpu_cleanup_object() 141 shmem = kzalloc(sizeof(*shmem), GFP_KERNEL); in virtio_gpu_create_object() 142 if (!shmem) in virtio_gpu_create_object() 145 dshmem = &shmem->base.base; in virtio_gpu_create_object() 171 if (!shmem->pages) { in virtio_gpu_object_shmem_init() [all …]
|
/linux/drivers/gpu/drm/panfrost/ |
A D | panfrost_gem_shrinker.c | 23 struct drm_gem_shmem_object *shmem; in panfrost_gem_shrinker_count() local 29 list_for_each_entry(shmem, &pfdev->shrinker_list, madv_list) { in panfrost_gem_shrinker_count() 30 if (drm_gem_shmem_is_purgeable(shmem)) in panfrost_gem_shrinker_count() 31 count += shmem->base.size >> PAGE_SHIFT; in panfrost_gem_shrinker_count() 51 if (!mutex_trylock(&shmem->pages_lock)) in panfrost_gem_purge() 58 mutex_unlock(&shmem->pages_lock); in panfrost_gem_purge() 70 struct drm_gem_shmem_object *shmem, *tmp; in panfrost_gem_shrinker_scan() local 79 if (drm_gem_shmem_is_purgeable(shmem) && in panfrost_gem_shrinker_scan() 80 panfrost_gem_purge(&shmem->base)) { in panfrost_gem_shrinker_scan() 81 freed += shmem->base.size >> PAGE_SHIFT; in panfrost_gem_shrinker_scan() [all …]
|
A D | panfrost_gem.c | 241 struct drm_gem_shmem_object *shmem; in panfrost_gem_create_with_handle() local 248 shmem = drm_gem_shmem_create(dev, size); in panfrost_gem_create_with_handle() 249 if (IS_ERR(shmem)) in panfrost_gem_create_with_handle() 250 return ERR_CAST(shmem); in panfrost_gem_create_with_handle() 252 bo = to_panfrost_bo(&shmem->base); in panfrost_gem_create_with_handle() 260 ret = drm_gem_handle_create(file_priv, &shmem->base, handle); in panfrost_gem_create_with_handle() 262 drm_gem_object_put(&shmem->base); in panfrost_gem_create_with_handle()
|
/linux/include/drm/ |
A D | drm_gem_shmem_helper.h | 112 int drm_gem_shmem_get_pages(struct drm_gem_shmem_object *shmem); 113 void drm_gem_shmem_put_pages(struct drm_gem_shmem_object *shmem); 121 static inline bool drm_gem_shmem_is_purgeable(struct drm_gem_shmem_object *shmem) in drm_gem_shmem_is_purgeable() argument 123 return (shmem->madv > 0) && in drm_gem_shmem_is_purgeable() 124 !shmem->vmap_use_count && shmem->sgt && in drm_gem_shmem_is_purgeable() 125 !shmem->base.dma_buf && !shmem->base.import_attach; in drm_gem_shmem_is_purgeable()
|
/linux/drivers/net/ethernet/microsoft/mana/ |
A D | shm_channel.c | 145 u64 *shmem; in mana_smc_setup_hwc() local 185 shmem = (u64 *)ptr; in mana_smc_setup_hwc() 187 *shmem = frame_addr & PAGE_FRAME_L48_MASK; in mana_smc_setup_hwc() 193 shmem = (u64 *)ptr; in mana_smc_setup_hwc() 195 *shmem = frame_addr & PAGE_FRAME_L48_MASK; in mana_smc_setup_hwc() 201 shmem = (u64 *)ptr; in mana_smc_setup_hwc() 203 *shmem = frame_addr & PAGE_FRAME_L48_MASK; in mana_smc_setup_hwc() 209 shmem = (u64 *)ptr; in mana_smc_setup_hwc() 211 *shmem = frame_addr & PAGE_FRAME_L48_MASK; in mana_smc_setup_hwc()
|
/linux/Documentation/devicetree/bindings/firmware/ |
A D | arm,scmi.yaml | 68 shmem: 194 shmem: 216 - shmem 227 - shmem 237 shmem = <&cpu_scp_lpri0>, 255 shmem = <&cpu_scp_hpri0>, 308 compatible = "arm,scmi-shmem"; 313 compatible = "arm,scmi-shmem"; 318 compatible = "arm,scmi-shmem"; 323 compatible = "arm,scmi-shmem"; [all …]
|
A D | nvidia,tegra186-bpmp.txt | 16 - shmem : List of the phandle of the TX and RX shared memory area that 80 cpu_bpmp_tx: shmem@4e000 { 81 compatible = "nvidia,tegra186-bpmp-shmem"; 87 cpu_bpmp_rx: shmem@4f000 { 88 compatible = "nvidia,tegra186-bpmp-shmem"; 98 shmem = <&cpu_bpmp_tx &cpu_bpmp_rx>;
|
A D | arm,scpi.yaml | 47 shmem: 168 - shmem 176 shmem = <&cpu_scp_hpri>; /* HP-NonSecure */ 223 compatible = "arm,scp-shmem"; 228 compatible = "arm,scp-shmem"; 239 shmem = <&cpu_scp_lpri &cpu_scp_hpri>;
|
/linux/Documentation/devicetree/bindings/mailbox/ |
A D | mailbox.txt | 26 - shmem : List of phandle pointing to the shared memory(SHM) area between the 39 Example with shared memory(shmem): 49 cl_shmem: shmem@0 { 50 compatible = "client-shmem"; 58 shmem = <&cl_shmem>;
|
A D | arm,mhu.yaml | 110 shmem = <&cpu_scp_hpri>; /* HP-NonSecure */ 144 shmem = <&cpu_scp_lpri0>, 162 shmem = <&cpu_scp_hpri0>,
|
/linux/drivers/net/arcnet/ |
A D | com90xx.c | 88 static int shmem; variable 93 module_param(shmem, int, 0); 107 if (!io && !irq && !shmem && !*device && com90xx_skip_probe) in com90xx_probe() 131 if (shmem) in com90xx_probe() 132 shmems[numshmems++] = shmem; in com90xx_probe() 473 release_mem_region(shmem, MIRROR_SIZE); in com90xx_found() 485 check_mirror(shmem - MIRROR_SIZE, MIRROR_SIZE) == 0 && in com90xx_found() 489 first_mirror = shmem - mirror_size; in com90xx_found() 494 last_mirror = shmem + mirror_size; in com90xx_found() 503 release_mem_region(shmem, MIRROR_SIZE); in com90xx_found() [all …]
|
A D | arc-rimi.c | 126 unsigned long first_mirror, last_mirror, shmem; in arcrimi_found() local 146 shmem = dev->mem_start; in arcrimi_found() 159 check_mirror(shmem - MIRROR_SIZE, MIRROR_SIZE) == 0 && in arcrimi_found() 160 check_mirror(shmem - 2 * MIRROR_SIZE, MIRROR_SIZE) == 1) in arcrimi_found() 163 first_mirror = shmem - mirror_size; in arcrimi_found() 168 last_mirror = shmem + mirror_size; in arcrimi_found() 194 release_mem_region(shmem, MIRROR_SIZE); in arcrimi_found()
|
/linux/Documentation/admin-guide/cgroup-v1/ |
A D | memcg_test.rst | 111 The best way to understand shmem's page state transition is to read 112 mm/shmem.c. 114 But brief explanation of the behavior of memcg around shmem will be 119 - radix-tree of shmem's inode. 126 - A new page is added to shmem's radix-tree. 158 Historically, memcg's shmem handling was poor and we saw some amount 159 of troubles here. This is because shmem is page-cache but can be 160 SwapCache. Test with shmem/tmpfs is always good test.
|
/linux/drivers/gpu/drm/lima/ |
A D | lima_gem.c | 107 struct drm_gem_shmem_object *shmem; in lima_gem_create_handle() local 112 shmem = drm_gem_shmem_create(dev, size); in lima_gem_create_handle() 113 if (IS_ERR(shmem)) in lima_gem_create_handle() 114 return PTR_ERR(shmem); in lima_gem_create_handle() 116 obj = &shmem->base; in lima_gem_create_handle()
|
/linux/drivers/net/ethernet/8390/ |
A D | mac8390.c | 738 long shmem = (start_page - WD_START_PG)<<8; in sane_block_output() local 740 memcpy_toio((void __iomem *)dev->mem_start + shmem, buf, count); in sane_block_output() 780 long shmem = (start_page - WD_START_PG)<<8; in dayna_block_output() local 782 dayna_memcpy_tocard(dev, shmem, buf, count); in dayna_block_output() 819 long shmem = (start_page - WD_START_PG)<<8; in slow_sane_block_output() local 821 word_memcpy_tocard(dev->mem_start + shmem, buf, count); in slow_sane_block_output()
|
/linux/arch/arm64/boot/dts/nvidia/ |
A D | tegra234.dtsi | 132 cpu_bpmp_tx: shmem@4e000 { 138 cpu_bpmp_rx: shmem@4f000 { 149 shmem = <&cpu_bpmp_tx &cpu_bpmp_rx>;
|
/linux/Documentation/devicetree/bindings/sram/ |
A D | sram.yaml | 84 - amlogic,meson-gxbb-scp-shmem 85 - amlogic,meson-axg-scp-shmem 86 - arm,juno-scp-shmem 87 - arm,scmi-shmem 88 - arm,scp-shmem
|
/linux/drivers/tee/amdtee/ |
A D | core.c | 344 struct shmem_desc shmem; in amdtee_map_shmem() local 356 shmem.kaddr = shm->kaddr; in amdtee_map_shmem() 357 shmem.size = shm->size; in amdtee_map_shmem() 363 rc = handle_map_shmem(count, &shmem, &buf_id); in amdtee_map_shmem()
|
/linux/drivers/gpu/drm/i915/selftests/ |
A D | i915_mock_selftests.h | 19 selftest(shmem, shmem_utils_mock_selftests)
|