diff options
| author | Christoph Hellwig <hch@lst.de> | 2018-05-18 15:24:13 +0200 | 
|---|---|---|
| committer | Christoph Hellwig <hch@lst.de> | 2018-05-19 08:46:12 +0200 | 
| commit | a8eb92d02dd7ffc7f04c48da3f2f80dbb6c74e5e (patch) | |
| tree | e00686eca5c0cd67d67fbf91caded50d7e678c5c | |
| parent | b591741072271f3a7cbf5b976e64ffbfa9a7ce8c (diff) | |
arc: fix arc_dma_{map,unmap}_page
These functions should perform the same cache synchronoization as calling
arc_dma_sync_single_for_{cpu,device} in addition to doing any required
address translation or mapping [1].  Ensure they actually do that by calling
arc_dma_sync_single_for_{cpu,device} instead of passing the dir argument
along to _dma_cache_sync.
The now unused _dma_cache_sync function is removed as well.
[1] in fact various drivers rely on that by passing DMA_ATTR_SKIP_CPU_SYNC
to the map/unmap routines and doing the cache synchronization manually.
Signed-off-by: Christoph Hellwig <hch@lst.de>
Tested-by: Alexey Brodkin <abrodkin@synopsys.com>
Acked-by: Vineet Gupta <vgupta@synopsys.com>
| -rw-r--r-- | arch/arc/mm/dma.c | 27 | 
1 files changed, 2 insertions, 25 deletions
| diff --git a/arch/arc/mm/dma.c b/arch/arc/mm/dma.c index eafdbd2ad20a..08d91c13ac52 100644 --- a/arch/arc/mm/dma.c +++ b/arch/arc/mm/dma.c @@ -130,29 +130,6 @@ static int arc_dma_mmap(struct device *dev, struct vm_area_struct *vma,  	return ret;  } -/* - * streaming DMA Mapping API... - * CPU accesses page via normal paddr, thus needs to explicitly made - * consistent before each use - */ -static void _dma_cache_sync(phys_addr_t paddr, size_t size, -		enum dma_data_direction dir) -{ -	switch (dir) { -	case DMA_FROM_DEVICE: -		dma_cache_inv(paddr, size); -		break; -	case DMA_TO_DEVICE: -		dma_cache_wback(paddr, size); -		break; -	case DMA_BIDIRECTIONAL: -		dma_cache_wback_inv(paddr, size); -		break; -	default: -		pr_err("Invalid DMA dir [%d] for OP @ %pa[p]\n", dir, &paddr); -	} -} -  static void arc_dma_sync_single_for_device(struct device *dev,  		dma_addr_t dma_handle, size_t size, enum dma_data_direction dir)  { @@ -185,7 +162,7 @@ static dma_addr_t arc_dma_map_page(struct device *dev, struct page *page,  	phys_addr_t paddr = page_to_phys(page) + offset;  	if (!(attrs & DMA_ATTR_SKIP_CPU_SYNC)) -		_dma_cache_sync(paddr, size, dir); +		arc_dma_sync_single_for_device(dev, paddr, size, dir);  	return paddr;  } @@ -205,7 +182,7 @@ static void arc_dma_unmap_page(struct device *dev, dma_addr_t handle,  	phys_addr_t paddr = handle;  	if (!(attrs & DMA_ATTR_SKIP_CPU_SYNC)) -		_dma_cache_sync(paddr, size, dir); +		arc_dma_sync_single_for_cpu(dev, paddr, size, dir);  }  static int arc_dma_map_sg(struct device *dev, struct scatterlist *sg, | 
