James Hogan | f507758 | 2012-10-05 16:27:03 +0100 | [diff] [blame] | 1 | #ifndef _ASM_METAG_DMA_MAPPING_H |
| 2 | #define _ASM_METAG_DMA_MAPPING_H |
| 3 | |
Christoph Hellwig | 5348c1e | 2016-01-20 15:01:56 -0800 | [diff] [blame] | 4 | extern struct dma_map_ops metag_dma_ops; |
James Hogan | f507758 | 2012-10-05 16:27:03 +0100 | [diff] [blame] | 5 | |
Christoph Hellwig | 5348c1e | 2016-01-20 15:01:56 -0800 | [diff] [blame] | 6 | static inline struct dma_map_ops *get_dma_ops(struct device *dev) |
James Hogan | f507758 | 2012-10-05 16:27:03 +0100 | [diff] [blame] | 7 | { |
Christoph Hellwig | 5348c1e | 2016-01-20 15:01:56 -0800 | [diff] [blame] | 8 | return &metag_dma_ops; |
James Hogan | f507758 | 2012-10-05 16:27:03 +0100 | [diff] [blame] | 9 | } |
| 10 | |
James Hogan | f507758 | 2012-10-05 16:27:03 +0100 | [diff] [blame] | 11 | /* |
| 12 | * dma_alloc_noncoherent() returns non-cacheable memory, so there's no need to |
| 13 | * do any flushing here. |
| 14 | */ |
| 15 | static inline void |
| 16 | dma_cache_sync(struct device *dev, void *vaddr, size_t size, |
| 17 | enum dma_data_direction direction) |
| 18 | { |
| 19 | } |
| 20 | |
James Hogan | f507758 | 2012-10-05 16:27:03 +0100 | [diff] [blame] | 21 | #endif |