1 /*
2  * Copyright 2010 Tilera Corporation. All Rights Reserved.
3  *
4  *   This program is free software; you can redistribute it and/or
5  *   modify it under the terms of the GNU General Public License
6  *   as published by the Free Software Foundation, version 2.
7  *
8  *   This program is distributed in the hope that it will be useful, but
9  *   WITHOUT ANY WARRANTY; without even the implied warranty of
10  *   MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE, GOOD TITLE or
11  *   NON INFRINGEMENT.  See the GNU General Public License for
12  *   more details.
13  */
14 
15 #ifndef _ASM_TILE_DMA_MAPPING_H
16 #define _ASM_TILE_DMA_MAPPING_H
17 
18 #include <linux/mm.h>
19 #include <linux/scatterlist.h>
20 #include <linux/cache.h>
21 #include <linux/io.h>
22 
23 /*
24  * Note that on x86 and powerpc, there is a "struct dma_mapping_ops"
25  * that is used for all the DMA operations.  For now, we don't have an
26  * equivalent on tile, because we only have a single way of doing DMA.
27  * (Tilera bug 7994 to use dma_mapping_ops.)
28  */
29 
30 #define dma_alloc_noncoherent(d, s, h, f) dma_alloc_coherent(d, s, h, f)
31 #define dma_free_noncoherent(d, s, v, h) dma_free_coherent(d, s, v, h)
32 
33 extern dma_addr_t dma_map_single(struct device *dev, void *ptr, size_t size,
34 			  enum dma_data_direction);
35 extern void dma_unmap_single(struct device *dev, dma_addr_t dma_addr,
36 			     size_t size, enum dma_data_direction);
37 extern int dma_map_sg(struct device *dev, struct scatterlist *sg, int nents,
38 	       enum dma_data_direction);
39 extern void dma_unmap_sg(struct device *dev, struct scatterlist *sg,
40 			 int nhwentries, enum dma_data_direction);
41 extern dma_addr_t dma_map_page(struct device *dev, struct page *page,
42 			       unsigned long offset, size_t size,
43 			       enum dma_data_direction);
44 extern void dma_unmap_page(struct device *dev, dma_addr_t dma_address,
45 			   size_t size, enum dma_data_direction);
46 extern void dma_sync_sg_for_cpu(struct device *dev, struct scatterlist *sg,
47 				int nelems, enum dma_data_direction);
48 extern void dma_sync_sg_for_device(struct device *dev, struct scatterlist *sg,
49 				   int nelems, enum dma_data_direction);
50 
51 
52 void *dma_alloc_coherent(struct device *dev, size_t size,
53 			   dma_addr_t *dma_handle, gfp_t flag);
54 
55 void dma_free_coherent(struct device *dev, size_t size,
56 			 void *vaddr, dma_addr_t dma_handle);
57 
58 extern void dma_sync_single_for_cpu(struct device *, dma_addr_t, size_t,
59 				    enum dma_data_direction);
60 extern void dma_sync_single_for_device(struct device *, dma_addr_t,
61 				       size_t, enum dma_data_direction);
62 extern void dma_sync_single_range_for_cpu(struct device *, dma_addr_t,
63 					  unsigned long offset, size_t,
64 					  enum dma_data_direction);
65 extern void dma_sync_single_range_for_device(struct device *, dma_addr_t,
66 					     unsigned long offset, size_t,
67 					     enum dma_data_direction);
68 extern void dma_cache_sync(void *vaddr, size_t, enum dma_data_direction);
69 
70 static inline int
dma_mapping_error(struct device * dev,dma_addr_t dma_addr)71 dma_mapping_error(struct device *dev, dma_addr_t dma_addr)
72 {
73 	return 0;
74 }
75 
76 static inline int
dma_supported(struct device * dev,u64 mask)77 dma_supported(struct device *dev, u64 mask)
78 {
79 	return 1;
80 }
81 
82 static inline int
dma_set_mask(struct device * dev,u64 mask)83 dma_set_mask(struct device *dev, u64 mask)
84 {
85 	if (!dev->dma_mask || !dma_supported(dev, mask))
86 		return -EIO;
87 
88 	*dev->dma_mask = mask;
89 
90 	return 0;
91 }
92 
93 #endif /* _ASM_TILE_DMA_MAPPING_H */
94