Lines Matching refs:gfp
88 gfp_t gfp) in __dma_direct_alloc_pages() argument
96 gfp |= dma_direct_optimal_gfp_mask(dev, dev->coherent_dma_mask, in __dma_direct_alloc_pages()
108 page = dma_alloc_contiguous(dev, size, gfp); in __dma_direct_alloc_pages()
115 page = alloc_pages_node(node, gfp, get_order(size)); in __dma_direct_alloc_pages()
122 !(gfp & (GFP_DMA32 | GFP_DMA))) { in __dma_direct_alloc_pages()
123 gfp |= GFP_DMA32; in __dma_direct_alloc_pages()
127 if (IS_ENABLED(CONFIG_ZONE_DMA) && !(gfp & GFP_DMA)) { in __dma_direct_alloc_pages()
128 gfp = (gfp & ~GFP_DMA32) | GFP_DMA; in __dma_direct_alloc_pages()
137 dma_addr_t *dma_handle, gfp_t gfp) in dma_direct_alloc_from_pool() argument
143 gfp |= dma_direct_optimal_gfp_mask(dev, dev->coherent_dma_mask, in dma_direct_alloc_from_pool()
145 page = dma_alloc_from_pool(dev, size, &ret, gfp, dma_coherent_ok); in dma_direct_alloc_from_pool()
153 dma_addr_t *dma_handle, gfp_t gfp, unsigned long attrs) in dma_direct_alloc() argument
161 gfp |= __GFP_NOWARN; in dma_direct_alloc()
165 page = __dma_direct_alloc_pages(dev, size, gfp & ~__GFP_ZERO); in dma_direct_alloc()
181 return arch_dma_alloc(dev, size, dma_handle, gfp, attrs); in dma_direct_alloc()
195 !gfpflags_allow_blocking(gfp) && in dma_direct_alloc()
200 return dma_direct_alloc_from_pool(dev, size, dma_handle, gfp); in dma_direct_alloc()
203 page = __dma_direct_alloc_pages(dev, size, gfp & ~__GFP_ZERO); in dma_direct_alloc()
319 dma_addr_t *dma_handle, enum dma_data_direction dir, gfp_t gfp) in dma_direct_alloc_pages() argument
325 force_dma_unencrypted(dev) && !gfpflags_allow_blocking(gfp) && in dma_direct_alloc_pages()
327 return dma_direct_alloc_from_pool(dev, size, dma_handle, gfp); in dma_direct_alloc_pages()
329 page = __dma_direct_alloc_pages(dev, size, gfp); in dma_direct_alloc_pages()