/linux/drivers/gpu/drm/ |
A D | drm_gem_cma_helper.c | 94 return cma_obj; in __drm_gem_cma_create() 97 kfree(cma_obj); in __drm_gem_cma_create() 123 return cma_obj; in drm_gem_cma_create() 131 cma_obj->vaddr = dma_alloc_wc(drm->dev, size, &cma_obj->paddr, in drm_gem_cma_create() 141 return cma_obj; in drm_gem_cma_create() 176 return cma_obj; in drm_gem_cma_create_with_handle() 190 return cma_obj; in drm_gem_cma_create_with_handle() 215 cma_obj->vaddr, cma_obj->paddr, in drm_gem_cma_free_object() 219 cma_obj->vaddr, cma_obj->paddr); in drm_gem_cma_free_object() 224 kfree(cma_obj); in drm_gem_cma_free_object() [all …]
|
A D | drm_fb_cma_helper.c | 121 const struct drm_gem_cma_object *cma_obj; in drm_fb_cma_sync_non_coherent() local 128 cma_obj = drm_fb_cma_get_gem_obj(state->fb, i); in drm_fb_cma_sync_non_coherent() 129 if (!cma_obj->map_noncoherent) in drm_fb_cma_sync_non_coherent()
|
A D | drm_mipi_dbi.c | 203 struct drm_gem_cma_object *cma_obj = to_drm_gem_cma_obj(gem); in mipi_dbi_buf_copy() local 204 void *src = cma_obj->vaddr; in mipi_dbi_buf_copy() 253 struct drm_gem_cma_object *cma_obj = to_drm_gem_cma_obj(gem); in mipi_dbi_fb_dirty() local 280 tr = cma_obj->vaddr; in mipi_dbi_fb_dirty()
|
/linux/drivers/gpu/drm/imx/ |
A D | ipuv3-plane.c | 127 struct drm_gem_cma_object *cma_obj; in drm_plane_state_to_eba() local 131 cma_obj = drm_fb_cma_get_gem_obj(fb, plane); in drm_plane_state_to_eba() 132 BUG_ON(!cma_obj); in drm_plane_state_to_eba() 142 struct drm_gem_cma_object *cma_obj; in drm_plane_state_to_ubo() local 147 cma_obj = drm_fb_cma_get_gem_obj(fb, 1); in drm_plane_state_to_ubo() 148 BUG_ON(!cma_obj); in drm_plane_state_to_ubo() 153 return cma_obj->paddr + fb->offsets[1] + fb->pitches[1] * y + in drm_plane_state_to_ubo() 161 struct drm_gem_cma_object *cma_obj; in drm_plane_state_to_vbo() local 166 cma_obj = drm_fb_cma_get_gem_obj(fb, 2); in drm_plane_state_to_vbo() 167 BUG_ON(!cma_obj); in drm_plane_state_to_vbo() [all …]
|
/linux/drivers/gpu/drm/imx/dcss/ |
A D | dcss-plane.c | 148 struct drm_gem_cma_object *cma_obj; in dcss_plane_atomic_check() local 157 cma_obj = drm_fb_cma_get_gem_obj(fb, 0); in dcss_plane_atomic_check() 158 WARN_ON(!cma_obj); in dcss_plane_atomic_check() 219 struct drm_gem_cma_object *cma_obj = drm_fb_cma_get_gem_obj(fb, 0); in dcss_plane_atomic_set_base() local 225 p1_ba = cma_obj->paddr + fb->offsets[0] + in dcss_plane_atomic_set_base() 232 p1_ba = cma_obj->paddr + fb->offsets[0] + in dcss_plane_atomic_set_base() 238 p2_ba = cma_obj->paddr + fb->offsets[1] + in dcss_plane_atomic_set_base()
|
/linux/drivers/gpu/drm/rcar-du/ |
A D | rcar_du_kms.c | 343 struct drm_gem_cma_object *cma_obj; in rcar_du_gem_prime_import_sg_table() local 351 cma_obj = kzalloc(sizeof(*cma_obj), GFP_KERNEL); in rcar_du_gem_prime_import_sg_table() 352 if (!cma_obj) in rcar_du_gem_prime_import_sg_table() 355 gem_obj = &cma_obj->base; in rcar_du_gem_prime_import_sg_table() 359 cma_obj->map_noncoherent = false; in rcar_du_gem_prime_import_sg_table() 364 kfree(cma_obj); in rcar_du_gem_prime_import_sg_table() 368 cma_obj->paddr = 0; in rcar_du_gem_prime_import_sg_table() 369 cma_obj->sgt = sgt; in rcar_du_gem_prime_import_sg_table()
|
/linux/drivers/gpu/drm/vc4/ |
A D | vc4_bo.c | 415 struct drm_gem_cma_object *cma_obj; in vc4_bo_create() local 429 cma_obj = drm_gem_cma_create(dev, size); in vc4_bo_create() 430 if (IS_ERR(cma_obj)) { in vc4_bo_create() 436 cma_obj = drm_gem_cma_create(dev, size); in vc4_bo_create() 439 if (IS_ERR(cma_obj)) { in vc4_bo_create() 452 cma_obj = drm_gem_cma_create(dev, size); in vc4_bo_create() 455 if (IS_ERR(cma_obj)) { in vc4_bo_create() 461 bo = to_vc4_bo(&cma_obj->base); in vc4_bo_create() 470 vc4_bo_set_label(&cma_obj->base, type); in vc4_bo_create()
|
/linux/drivers/gpu/drm/arm/ |
A D | malidp_planes.c | 339 struct drm_gem_cma_object *cma_obj; in malidp_check_pages_threshold() local 344 cma_obj = to_drm_gem_cma_obj(obj); in malidp_check_pages_threshold() 346 if (cma_obj->sgt) in malidp_check_pages_threshold() 347 sgt = cma_obj->sgt; in malidp_check_pages_threshold() 358 if (!cma_obj->sgt) in malidp_check_pages_threshold() 365 if (!cma_obj->sgt) in malidp_check_pages_threshold()
|
/linux/drivers/gpu/drm/sti/ |
A D | sti_cursor.c | 269 struct drm_gem_cma_object *cma_obj; in sti_cursor_atomic_update() local 280 cma_obj = drm_fb_cma_get_gem_obj(fb, 0); in sti_cursor_atomic_update() 283 sti_cursor_argb8888_to_clut8(cursor, (u32 *)cma_obj->vaddr); in sti_cursor_atomic_update()
|
A D | sti_gdp.c | 715 struct drm_gem_cma_object *cma_obj; in sti_gdp_atomic_update() local 779 cma_obj = drm_fb_cma_get_gem_obj(fb, 0); in sti_gdp_atomic_update() 783 (unsigned long)cma_obj->paddr); in sti_gdp_atomic_update() 787 top_field->gam_gdp_pml = (u32)cma_obj->paddr + fb->offsets[0]; in sti_gdp_atomic_update() 832 (unsigned long)cma_obj->paddr, in sti_gdp_atomic_update()
|
A D | sti_hqvdp.c | 1125 struct drm_gem_cma_object *cma_obj; in sti_hqvdp_atomic_update() local 1179 cma_obj = drm_fb_cma_get_gem_obj(fb, 0); in sti_hqvdp_atomic_update() 1183 (unsigned long)cma_obj->paddr); in sti_hqvdp_atomic_update() 1186 cmd->top.current_luma = (u32)cma_obj->paddr + fb->offsets[0]; in sti_hqvdp_atomic_update() 1187 cmd->top.current_chroma = (u32)cma_obj->paddr + fb->offsets[1]; in sti_hqvdp_atomic_update()
|
/linux/drivers/gpu/drm/tiny/ |
A D | st7586.c | 94 struct drm_gem_cma_object *cma_obj = drm_fb_cma_get_gem_obj(fb, 0); in st7586_buf_copy() local 95 void *src = cma_obj->vaddr; in st7586_buf_copy()
|
A D | ili9225.c | 80 struct drm_gem_cma_object *cma_obj = drm_fb_cma_get_gem_obj(fb, 0); in ili9225_fb_dirty() local 106 tr = cma_obj->vaddr; in ili9225_fb_dirty()
|
A D | repaper.c | 533 struct drm_gem_cma_object *cma_obj = drm_fb_cma_get_gem_obj(fb, 0); in repaper_fb_dirty() local 563 drm_fb_xrgb8888_to_gray8(buf, cma_obj->vaddr, fb, &clip); in repaper_fb_dirty()
|