/linux/drivers/gpu/drm/nouveau/ |
A D | nouveau_bo.c | 155 kfree(nvbo); in nouveau_bo_del_ttm() 215 if (!nvbo) in nouveau_bo_alloc() 237 kfree(nvbo); in nouveau_bo_alloc() 241 nvbo->comp = mmu->kind[nvbo->kind] != nvbo->kind; in nouveau_bo_alloc() 247 kfree(nvbo); in nouveau_bo_alloc() 290 nvbo->kind = mmu->kind[nvbo->kind]; in nouveau_bo_alloc() 297 return nvbo; in nouveau_bo_alloc() 310 ret = ttm_bo_init(nvbo->bo.bdev, &nvbo->bo, size, type, in nouveau_bo_init() 533 ret = ttm_bo_kmap(&nvbo->bo, 0, nvbo->bo.resource->num_pages, &nvbo->kmap); in nouveau_bo_map() 542 if (!nvbo) in nouveau_bo_unmap() [all …]
|
A D | nouveau_gem.c | 93 ttm_bo_put(&nvbo->bo); in nouveau_gem_object_del() 235 if (IS_ERR(nvbo)) in nouveau_gem_new() 236 return PTR_ERR(nvbo); in nouveau_gem_new() 245 kfree(nvbo); in nouveau_gem_new() 262 *pnvbo = nvbo; in nouveau_gem_new() 298 rep->tile_flags |= nvbo->kind << 8 | nvbo->comp << 16; in nouveau_gem_info() 454 if (nvbo->reserved_by && nvbo->reserved_by == file_priv) { in validate_init() 472 res_bo = nvbo; in validate_init() 513 if (nvbo == res_bo) in validate_init() 682 ret = ttm_bo_kmap(&nvbo->bo, 0, nvbo->bo.resource->num_pages, in nouveau_gem_pushbuf_reloc_apply() [all …]
|
A D | nouveau_prime.c | 35 nvbo->bo.ttm->num_pages); in nouveau_gem_prime_get_sg_table() 44 struct nouveau_bo *nvbo; in nouveau_gem_prime_import_sg_table() local 53 if (IS_ERR(nvbo)) { in nouveau_gem_prime_import_sg_table() 54 obj = ERR_CAST(nvbo); in nouveau_gem_prime_import_sg_table() 66 nouveau_bo_ref(NULL, &nvbo); in nouveau_gem_prime_import_sg_table() 74 nouveau_bo_ref(NULL, &nvbo); in nouveau_gem_prime_import_sg_table() 79 obj = &nvbo->bo.base; in nouveau_gem_prime_import_sg_table() 100 if (nvbo->bo.moving) in nouveau_gem_prime_pin() 103 ttm_bo_unreserve(&nvbo->bo); in nouveau_gem_prime_pin() 110 nouveau_bo_unpin(nvbo); in nouveau_gem_prime_pin() [all …]
|
A D | nouveau_vmm.c | 49 nouveau_vma_find(struct nouveau_bo *nvbo, struct nouveau_vmm *vmm) in nouveau_vma_find() argument 53 list_for_each_entry(vma, &nvbo->vma_list, head) { in nouveau_vma_find() 77 nouveau_vma_new(struct nouveau_bo *nvbo, struct nouveau_vmm *vmm, in nouveau_vma_new() argument 80 struct nouveau_mem *mem = nouveau_mem(nvbo->bo.resource); in nouveau_vma_new() 85 if ((vma = *pvma = nouveau_vma_find(nvbo, vmm))) { in nouveau_vma_new() 97 list_add_tail(&vma->head, &nvbo->vma_list); in nouveau_vma_new() 99 if (nvbo->bo.resource->mem_type != TTM_PL_SYSTEM && in nouveau_vma_new() 100 mem->mem.page == nvbo->page) { in nouveau_vma_new()
|
A D | nouveau_fbcon.c | 319 struct nouveau_bo *nvbo; in nouveau_fbcon_create() local 334 0, 0x0000, &nvbo); in nouveau_fbcon_create() 350 ret = nouveau_bo_map(nvbo); in nouveau_fbcon_create() 385 info->screen_size = nvbo->bo.base.size; in nouveau_fbcon_create() 397 fb->width, fb->height, nvbo->offset, nvbo); in nouveau_fbcon_create() 407 nouveau_bo_unmap(nvbo); in nouveau_fbcon_create() 409 nouveau_bo_unpin(nvbo); in nouveau_fbcon_create() 411 nouveau_bo_ref(NULL, &nvbo); in nouveau_fbcon_create() 420 struct nouveau_bo *nvbo; in nouveau_fbcon_destroy() local 428 nouveau_bo_unmap(nvbo); in nouveau_fbcon_destroy() [all …]
|
A D | nouveau_ttm.c | 50 struct nouveau_bo *nvbo = nouveau_bo(bo); in nouveau_vram_manager_new() local 57 ret = nouveau_mem_new(&drm->master, nvbo->kind, nvbo->comp, res); in nouveau_vram_manager_new() 63 ret = nouveau_mem_vram(*res, nvbo->contig, nvbo->page); in nouveau_vram_manager_new() 83 struct nouveau_bo *nvbo = nouveau_bo(bo); in nouveau_gart_manager_new() local 87 ret = nouveau_mem_new(&drm->master, nvbo->kind, nvbo->comp, res); in nouveau_gart_manager_new() 107 struct nouveau_bo *nvbo = nouveau_bo(bo); in nv04_gart_manager_new() local 112 ret = nouveau_mem_new(&drm->master, nvbo->kind, nvbo->comp, res); in nv04_gart_manager_new()
|
A D | nouveau_display.c | 178 const struct nouveau_bo *nvbo = nouveau_gem_object(fb->obj[0]); in nouveau_framebuffer_get_layout() local 180 *tile_mode = nvbo->mode; in nouveau_framebuffer_get_layout() 181 *kind = nvbo->kind; in nouveau_framebuffer_get_layout() 261 nouveau_check_bl_size(struct nouveau_drm *drm, struct nouveau_bo *nvbo, in nouveau_check_bl_size() argument 291 nvbo->bo.base.size); in nouveau_check_bl_size() 293 if (bl_size + offset > nvbo->bo.base.size) in nouveau_check_bl_size() 306 struct nouveau_bo *nvbo = nouveau_gem_object(gem); in nouveau_framebuffer_new() local 339 tile_mode = nvbo->mode; in nouveau_framebuffer_new() 340 kind = nvbo->kind; in nouveau_framebuffer_new() 354 ret = nouveau_check_bl_size(drm, nvbo, in nouveau_framebuffer_new() [all …]
|
A D | nouveau_bo.h | 94 void nouveau_bo_sync_for_device(struct nouveau_bo *nvbo); 95 void nouveau_bo_sync_for_cpu(struct nouveau_bo *nvbo); 101 nvbo_kmap_obj_iovirtual(struct nouveau_bo *nvbo) in nvbo_kmap_obj_iovirtual() argument 105 &nvbo->kmap, &is_iomem); in nvbo_kmap_obj_iovirtual()
|
A D | nouveau_sgdma.c | 70 struct nouveau_bo *nvbo = nouveau_bo(bo); in nouveau_sgdma_create_ttm() local 74 if (nvbo->force_coherent) in nouveau_sgdma_create_ttm()
|
A D | nouveau_crtc.h | 56 struct nouveau_bo *nvbo; member
|
A D | nouveau_fence.c | 342 nouveau_fence_sync(struct nouveau_bo *nvbo, struct nouveau_channel *chan, bool exclusive, bool intr) in nouveau_fence_sync() argument 346 struct dma_resv *resv = nvbo->bo.base.resv; in nouveau_fence_sync()
|
/linux/drivers/gpu/drm/nouveau/dispnv04/ |
A D | disp.c | 80 struct nouveau_bo *nvbo; in nv04_display_fini() local 84 nvbo = nouveau_gem_object(fb->obj[0]); in nv04_display_fini() 85 nouveau_bo_unpin(nvbo); in nv04_display_fini() 90 if (nv_crtc->cursor.nvbo) { in nv04_display_fini() 92 nouveau_bo_unmap(nv_crtc->cursor.nvbo); in nv04_display_fini() 93 nouveau_bo_unpin(nv_crtc->cursor.nvbo); in nv04_display_fini() 132 struct nouveau_bo *nvbo; in nv04_display_init() local 136 nvbo = nouveau_gem_object(fb->obj[0]); in nv04_display_init() 144 if (!nv_crtc->cursor.nvbo) in nv04_display_init() 175 if (!nv_crtc->cursor.nvbo) in nv04_display_init() [all …]
|
A D | overlay.c | 126 struct nouveau_bo *nvbo; in nv10_update_plane() local 144 nvbo = nouveau_gem_object(fb->obj[0]); in nv10_update_plane() 145 ret = nouveau_bo_pin(nvbo, NOUVEAU_GEM_DOMAIN_VRAM, false); in nv10_update_plane() 149 nv_plane->cur = nvbo; in nv10_update_plane() 155 nvif_wr32(dev, NV_PVIDEO_OFFSET_BUFF(flip), nvbo->offset); in nv10_update_plane() 177 nvbo->offset + fb->offsets[1]); in nv10_update_plane() 374 struct nouveau_bo *nvbo; in nv04_update_plane() local 389 nvbo = nouveau_gem_object(fb->obj[0]); in nv04_update_plane() 390 ret = nouveau_bo_pin(nvbo, NOUVEAU_GEM_DOMAIN_VRAM, false); in nv04_update_plane() 394 nv_plane->cur = nvbo; in nv04_update_plane() [all …]
|
A D | crtc.c | 761 nouveau_bo_unmap(nv_crtc->cursor.nvbo); in nv_crtc_destroy() 762 nouveau_bo_unpin(nv_crtc->cursor.nvbo); in nv_crtc_destroy() 763 nouveau_bo_ref(NULL, &nv_crtc->cursor.nvbo); in nv_crtc_destroy() 832 struct nouveau_bo *nvbo; in nv04_crtc_do_mode_set_base() local 853 nvbo = nouveau_gem_object(drm_fb->obj[0]); in nv04_crtc_do_mode_set_base() 854 nv_crtc->fb.offset = nvbo->offset; in nv04_crtc_do_mode_set_base() 1340 &nv_crtc->cursor.nvbo); in nv04_crtc_create() 1342 ret = nouveau_bo_pin(nv_crtc->cursor.nvbo, in nv04_crtc_create() 1345 ret = nouveau_bo_map(nv_crtc->cursor.nvbo); in nv04_crtc_create() 1347 nouveau_bo_unpin(nv_crtc->cursor.nvbo); in nv04_crtc_create() [all …]
|
/linux/drivers/gpu/drm/nouveau/dispnv50/ |
A D | wndw.c | 522 struct nouveau_bo *nvbo; in nv50_wndw_cleanup_fb() local 528 nvbo = nouveau_gem_object(old_state->fb->obj[0]); in nv50_wndw_cleanup_fb() 529 nouveau_bo_unpin(nvbo); in nv50_wndw_cleanup_fb() 539 struct nouveau_bo *nvbo; in nv50_wndw_prepare_fb() local 548 nvbo = nouveau_gem_object(fb->obj[0]); in nv50_wndw_prepare_fb() 549 ret = nouveau_bo_pin(nvbo, NOUVEAU_GEM_DOMAIN_VRAM, true); in nv50_wndw_prepare_fb() 556 nouveau_bo_unpin(nvbo); in nv50_wndw_prepare_fb() 564 asyw->state.fence = dma_resv_get_excl_unlocked(nvbo->bo.base.resv); in nv50_wndw_prepare_fb() 565 asyw->image.offset[0] = nvbo->offset; in nv50_wndw_prepare_fb()
|