1#ifndef _ASM_IA64_DMA_MAPPING_H 2#define _ASM_IA64_DMA_MAPPING_H 3 4/* 5 * Copyright (C) 2003-2004 Hewlett-Packard Co 6 * David Mosberger-Tang <davidm@hpl.hp.com> 7 */ 8#include <asm/machvec.h> 9 10#define dma_alloc_coherent platform_dma_alloc_coherent 11/* coherent mem. is cheap */ 12static inline void * 13dma_alloc_noncoherent(struct device *dev, size_t size, dma_addr_t *dma_handle, 14 gfp_t flag) 15{ 16 return dma_alloc_coherent(dev, size, dma_handle, flag); 17} 18#define dma_free_coherent platform_dma_free_coherent 19static inline void 20dma_free_noncoherent(struct device *dev, size_t size, void *cpu_addr, 21 dma_addr_t dma_handle) 22{ 23 dma_free_coherent(dev, size, cpu_addr, dma_handle); 24} 25#define dma_map_single platform_dma_map_single 26#define dma_map_sg platform_dma_map_sg 27#define dma_unmap_single platform_dma_unmap_single 28#define dma_unmap_sg platform_dma_unmap_sg 29#define dma_sync_single_for_cpu platform_dma_sync_single_for_cpu 30#define dma_sync_sg_for_cpu platform_dma_sync_sg_for_cpu 31#define dma_sync_single_for_device platform_dma_sync_single_for_device 32#define dma_sync_sg_for_device platform_dma_sync_sg_for_device 33#define dma_mapping_error platform_dma_mapping_error 34 35#define dma_map_page(dev, pg, off, size, dir) \ 36 dma_map_single(dev, page_address(pg) + (off), (size), (dir)) 37#define dma_unmap_page(dev, dma_addr, size, dir) \ 38 dma_unmap_single(dev, dma_addr, size, dir) 39 40/* 41 * Rest of this file is part of the "Advanced DMA API". Use at your own risk. 42 * See Documentation/DMA-API.txt for details. 43 */ 44 45#define dma_sync_single_range_for_cpu(dev, dma_handle, offset, size, dir) \ 46 dma_sync_single_for_cpu(dev, dma_handle, size, dir) 47#define dma_sync_single_range_for_device(dev, dma_handle, offset, size, dir) \ 48 dma_sync_single_for_device(dev, dma_handle, size, dir) 49 50#define dma_supported platform_dma_supported 51 52static inline int 53dma_set_mask (struct device *dev, u64 mask) 54{ 55 if (!dev->dma_mask || !dma_supported(dev, mask)) 56 return -EIO; 57 *dev->dma_mask = mask; 58 return 0; 59} 60 61extern int dma_get_cache_alignment(void); 62 63static inline void 64dma_cache_sync (struct device *dev, void *vaddr, size_t size, 65 enum dma_data_direction dir) 66{ 67 /* 68 * IA-64 is cache-coherent, so this is mostly a no-op. However, we do need to 69 * ensure that dma_cache_sync() enforces order, hence the mb(). 70 */ 71 mb(); 72} 73 74#define dma_is_consistent(d, h) (1) /* all we do is coherent memory... */ 75 76#endif /* _ASM_IA64_DMA_MAPPING_H */ 77