summary refs log tree commit diff
path: root/arch/hexagon
diff options
context:
space:
mode:
authorChristoph Hellwig <hch@lst.de>2018-05-19 08:55:39 +0200
committerChristoph Hellwig <hch@lst.de>2018-09-08 11:17:30 +0200
commit37fca4bd018f9959cf28ea4a6da9f74ccc4d4025 (patch)
tree7643afa9a04190127d62db8c53df389751f050d2 /arch/hexagon
parentd7b686ebf704e3d91925a535a0905ba6be23757c (diff)
downloadlinux-37fca4bd018f9959cf28ea4a6da9f74ccc4d4025.tar.gz
hexagon: remove the sync_single_for_cpu DMA operation
hexagon does all the required cache maintainance at dma map time, and none
at unmap time.  It thus has to implement sync_single_for_device to match
the map cace for buffer reuse, but there is no point in doing another
invalidation in the sync_single_cpu_case, which in terms of cache
maintainance is equivalent to the unmap case.

Signed-off-by: Christoph Hellwig <hch@lst.de>
Diffstat (limited to 'arch/hexagon')
-rw-r--r--arch/hexagon/kernel/dma.c8
1 files changed, 0 insertions, 8 deletions
diff --git a/arch/hexagon/kernel/dma.c b/arch/hexagon/kernel/dma.c
index 77459df34e2e..d2b717f352f4 100644
--- a/arch/hexagon/kernel/dma.c
+++ b/arch/hexagon/kernel/dma.c
@@ -181,13 +181,6 @@ static dma_addr_t hexagon_map_page(struct device *dev, struct page *page,
 	return bus;
 }
 
-static void hexagon_sync_single_for_cpu(struct device *dev,
-					dma_addr_t dma_handle, size_t size,
-					enum dma_data_direction dir)
-{
-	dma_sync(dma_addr_to_virt(dma_handle), size, dir);
-}
-
 static void hexagon_sync_single_for_device(struct device *dev,
 					dma_addr_t dma_handle, size_t size,
 					enum dma_data_direction dir)
@@ -205,7 +198,6 @@ const struct dma_map_ops hexagon_dma_ops = {
 	.free		= hexagon_free_coherent,
 	.map_sg		= hexagon_map_sg,
 	.map_page	= hexagon_map_page,
-	.sync_single_for_cpu = hexagon_sync_single_for_cpu,
 	.sync_single_for_device = hexagon_sync_single_for_device,
 	.mapping_error	= hexagon_mapping_error,
 };