On Friday 19 February 2016 09:22:44 Marek Szyprowski wrote:
This patch replaces ARM-specific IOMMU-based DMA-mapping implementation with generic IOMMU DMA-mapping code shared with ARM64 architecture. The side-effect of this change is a switch from bitmap-based IO address space management to tree-based code. There should be no functional changes for drivers, which rely on initialization from generic arch_setup_dna_ops() interface. Code, which used old arm_iommu_* functions must be updated to new interface.
Signed-off-by: Marek Szyprowski m.szyprowski@samsung.com
I like the overall idea. However, this interface from the iommu subsystem into architecture specific code:
+/*
- The DMA API is built upon the notion of "buffer ownership". A buffer
- is either exclusively owned by the CPU (and therefore may be accessed
- by it) or exclusively owned by the DMA device. These helper functions
- represent the transitions between these two ownership states.
- Note, however, that on later ARMs, this notion does not work due to
- speculative prefetches. We model our approach on the assumption that
- the CPU does do speculative prefetches, which means we clean caches
- before transfers and delay cache invalidation until transfer completion.
- */
+extern void __dma_page_cpu_to_dev(struct page *, unsigned long, size_t,
enum dma_data_direction);
+extern void __dma_page_dev_to_cpu(struct page *, unsigned long, size_t,
enum dma_data_direction);
+static inline void arch_flush_page(struct device *dev, const void *virt,
phys_addr_t phys)
+{
- dmac_flush_range(virt, virt + PAGE_SIZE);
- outer_flush_range(phys, phys + PAGE_SIZE);
+}
+static inline void arch_dma_map_area(phys_addr_t phys, size_t size,
enum dma_data_direction dir)
+{
- unsigned int offset = phys & ~PAGE_MASK;
- __dma_page_cpu_to_dev(phys_to_page(phys & PAGE_MASK), offset, size, dir);
+}
+static inline void arch_dma_unmap_area(phys_addr_t phys, size_t size,
enum dma_data_direction dir)
+{
- unsigned int offset = phys & ~PAGE_MASK;
- __dma_page_dev_to_cpu(phys_to_page(phys & PAGE_MASK), offset, size, dir);
+}
+static inline pgprot_t arch_get_dma_pgprot(struct dma_attrs *attrs,
pgprot_t prot, bool coherent)
+{
- if (coherent)
return prot;
- prot = dma_get_attr(DMA_ATTR_WRITE_COMBINE, attrs) ?
pgprot_writecombine(prot) :
pgprot_dmacoherent(prot);
- return prot;
+}
+extern void *arch_alloc_from_atomic_pool(size_t size, struct page **ret_page,
gfp_t flags);
+extern bool arch_in_atomic_pool(void *start, size_t size); +extern int arch_free_from_atomic_pool(void *start, size_t size);
doesn't feel completely right yet. In particular the arch_flush_page() interface is probably still too specific to ARM/ARM64 and won't work that way on other architectures.
I think it would be better to do this either more generic, or less generic:
a) leave the iommu_dma_map_ops definition in the architecture specific code, but make it call helper functions in the drivers/iommu to do all of the really generic parts.
b) clarify that this is only applicable to arch/arm and arch/arm64, and unify things further between these two, as they have very similar requirements in the CPU architecture.
Arnd