diff options
author | Catalin Marinas <catalin.marinas@arm.com> | 2014-04-01 18:32:55 +0100 |
---|---|---|
committer | Catalin Marinas <catalin.marinas@arm.com> | 2014-04-08 11:45:08 +0100 |
commit | ebf81a938dade3b450eb11c57fa744cfac4b523f (patch) | |
tree | fb7c10fd49f9ee3a71a7d5c0620d3a6745a082ea /arch/arm64 | |
parent | d253b4406df69fa7a74231769d6f6ad80dc33063 (diff) | |
download | linux-ebf81a938dade3b450eb11c57fa744cfac4b523f.tar.bz2 |
arm64: Fix DMA range invalidation for cache line unaligned buffers
If the buffer needing cache invalidation for inbound DMA does start or
end on a cache line aligned address, we need to use the non-destructive
clean&invalidate operation. This issue was introduced by commit
7363590d2c46 (arm64: Implement coherent DMA API based on swiotlb).
Signed-off-by: Catalin Marinas <catalin.marinas@arm.com>
Reported-by: Jon Medhurst (Tixy) <tixy@linaro.org>
Diffstat (limited to 'arch/arm64')
-rw-r--r-- | arch/arm64/mm/cache.S | 15 |
1 files changed, 11 insertions, 4 deletions
diff --git a/arch/arm64/mm/cache.S b/arch/arm64/mm/cache.S index e803a62e0e45..fda756875fa6 100644 --- a/arch/arm64/mm/cache.S +++ b/arch/arm64/mm/cache.S @@ -183,12 +183,19 @@ ENTRY(__inval_cache_range) __dma_inv_range: dcache_line_size x2, x3 sub x3, x2, #1 - bic x0, x0, x3 + tst x1, x3 // end cache line aligned? bic x1, x1, x3 -1: dc ivac, x0 // invalidate D / U line - add x0, x0, x2 + b.eq 1f + dc civac, x1 // clean & invalidate D / U line +1: tst x0, x3 // start cache line aligned? + bic x0, x0, x3 + b.eq 2f + dc civac, x0 // clean & invalidate D / U line + b 3f +2: dc ivac, x0 // invalidate D / U line +3: add x0, x0, x2 cmp x0, x1 - b.lo 1b + b.lo 2b dsb sy ret ENDPROC(__inval_cache_range) |