Use kzalloc() with GFP_ATOMIC instead of vzalloc(). At freeing, __in_atomic_pool() checks if it comes from atomic_pool or not.
Signed-off-by: Hiroshi Doyu hdoyu@nvidia.com --- arch/arm/mm/dma-mapping.c | 7 ++++--- 1 files changed, 4 insertions(+), 3 deletions(-)
diff --git a/arch/arm/mm/dma-mapping.c b/arch/arm/mm/dma-mapping.c index aca2fd0..b64475a 100644 --- a/arch/arm/mm/dma-mapping.c +++ b/arch/arm/mm/dma-mapping.c @@ -1029,7 +1029,7 @@ static struct page **__iommu_alloc_buffer(struct device *dev, size_t size, gfp_t int array_size = count * sizeof(struct page *); int i = 0;
- if (array_size <= PAGE_SIZE) + if ((array_size <= PAGE_SIZE) || (gfp & GFP_ATOMIC)) pages = kzalloc(array_size, gfp); else pages = vzalloc(array_size); @@ -1061,7 +1061,7 @@ error: while (i--) if (pages[i]) __free_pages(pages[i], 0); - if (array_size <= PAGE_SIZE) + if ((array_size <= PAGE_SIZE) || (gfp & GFP_ATOMIC)) kfree(pages); else vfree(pages); @@ -1076,7 +1076,8 @@ static int __iommu_free_buffer(struct device *dev, struct page **pages, size_t s for (i = 0; i < count; i++) if (pages[i]) __free_pages(pages[i], 0); - if (array_size <= PAGE_SIZE) + if ((array_size <= PAGE_SIZE) || + __in_atomic_pool(page_address(pages[0]), size)) kfree(pages); else vfree(pages);