2 * This file is subject to the terms and conditions of the GNU General Public
3 * License. See the file "COPYING" in the main directory of this archive
6 * Copyright (C) 2000 Ani Joshi <ajoshi@unixbox.com>
7 * Copyright (C) 2000, 2001 Ralf Baechle <ralf@gnu.org>
8 * swiped from i386, and cloned for MIPS by Geert, polished by Ralf.
10 #include <linux/types.h>
12 #include <linux/module.h>
13 #include <linux/string.h>
14 #include <linux/pci.h>
16 #include <asm/cache.h>
17 #include <asm/pci/bridge.h>
19 #define pdev_to_baddr(pdev, addr) \
20 (BRIDGE_CONTROLLER(pdev->bus)->baddr + (addr))
21 #define dev_to_baddr(dev, addr) \
22 pdev_to_baddr(to_pci_dev(dev), (addr))
24 void *dma_alloc_noncoherent(struct device *dev, size_t size,
25 dma_addr_t * dma_handle, gfp_t gfp)
29 /* ignore region specifiers */
30 gfp &= ~(__GFP_DMA | __GFP_HIGHMEM);
32 if (dev == NULL || (dev->coherent_dma_mask < 0xffffffff))
34 ret = (void *) __get_free_pages(gfp, get_order(size));
38 *dma_handle = dev_to_baddr(dev, virt_to_phys(ret));
44 EXPORT_SYMBOL(dma_alloc_noncoherent);
46 void *dma_alloc_coherent(struct device *dev, size_t size,
47 dma_addr_t * dma_handle, gfp_t gfp)
48 __attribute__((alias("dma_alloc_noncoherent")));
50 EXPORT_SYMBOL(dma_alloc_coherent);
52 void dma_free_noncoherent(struct device *dev, size_t size, void *vaddr,
53 dma_addr_t dma_handle)
55 unsigned long addr = (unsigned long) vaddr;
57 free_pages(addr, get_order(size));
60 EXPORT_SYMBOL(dma_free_noncoherent);
62 void dma_free_coherent(struct device *dev, size_t size, void *vaddr,
63 dma_addr_t dma_handle) __attribute__((alias("dma_free_noncoherent")));
65 EXPORT_SYMBOL(dma_free_coherent);
67 dma_addr_t dma_map_single(struct device *dev, void *ptr, size_t size,
68 enum dma_data_direction direction)
70 BUG_ON(direction == DMA_NONE);
72 return dev_to_baddr(dev, __pa(ptr));
75 EXPORT_SYMBOL(dma_map_single);
77 void dma_unmap_single(struct device *dev, dma_addr_t dma_addr, size_t size,
78 enum dma_data_direction direction)
80 BUG_ON(direction == DMA_NONE);
83 EXPORT_SYMBOL(dma_unmap_single);
85 int dma_map_sg(struct device *dev, struct scatterlist *sg, int nents,
86 enum dma_data_direction direction)
90 BUG_ON(direction == DMA_NONE);
92 for (i = 0; i < nents; i++, sg++) {
93 sg->dma_address = (dma_addr_t) dev_to_baddr(dev,
94 page_to_phys(sg->page) + sg->offset);
100 EXPORT_SYMBOL(dma_map_sg);
102 dma_addr_t dma_map_page(struct device *dev, struct page *page,
103 unsigned long offset, size_t size, enum dma_data_direction direction)
105 BUG_ON(direction == DMA_NONE);
107 return dev_to_baddr(dev, page_to_phys(page) + offset);
110 EXPORT_SYMBOL(dma_map_page);
112 void dma_unmap_page(struct device *dev, dma_addr_t dma_address, size_t size,
113 enum dma_data_direction direction)
115 BUG_ON(direction == DMA_NONE);
118 EXPORT_SYMBOL(dma_unmap_page);
120 void dma_unmap_sg(struct device *dev, struct scatterlist *sg, int nhwentries,
121 enum dma_data_direction direction)
123 BUG_ON(direction == DMA_NONE);
126 EXPORT_SYMBOL(dma_unmap_sg);
128 void dma_sync_single_for_cpu(struct device *dev, dma_addr_t dma_handle, size_t size,
129 enum dma_data_direction direction)
131 BUG_ON(direction == DMA_NONE);
134 EXPORT_SYMBOL(dma_sync_single_for_cpu);
136 void dma_sync_single_for_device(struct device *dev, dma_addr_t dma_handle, size_t size,
137 enum dma_data_direction direction)
139 BUG_ON(direction == DMA_NONE);
142 EXPORT_SYMBOL(dma_sync_single_for_device);
144 void dma_sync_single_range_for_cpu(struct device *dev, dma_addr_t dma_handle,
145 unsigned long offset, size_t size,
146 enum dma_data_direction direction)
148 BUG_ON(direction == DMA_NONE);
151 EXPORT_SYMBOL(dma_sync_single_range_for_cpu);
153 void dma_sync_single_range_for_device(struct device *dev, dma_addr_t dma_handle,
154 unsigned long offset, size_t size,
155 enum dma_data_direction direction)
157 BUG_ON(direction == DMA_NONE);
160 EXPORT_SYMBOL(dma_sync_single_range_for_device);
162 void dma_sync_sg_for_cpu(struct device *dev, struct scatterlist *sg, int nelems,
163 enum dma_data_direction direction)
165 BUG_ON(direction == DMA_NONE);
168 EXPORT_SYMBOL(dma_sync_sg_for_cpu);
170 void dma_sync_sg_for_device(struct device *dev, struct scatterlist *sg, int nelems,
171 enum dma_data_direction direction)
173 BUG_ON(direction == DMA_NONE);
176 EXPORT_SYMBOL(dma_sync_sg_for_device);
178 int dma_mapping_error(dma_addr_t dma_addr)
183 EXPORT_SYMBOL(dma_mapping_error);
185 int dma_supported(struct device *dev, u64 mask)
188 * we fall back to GFP_DMA when the mask isn't all 1s,
189 * so we can't guarantee allocations that must be
190 * within a tighter range than GFP_DMA..
192 if (mask < 0x00ffffff)
198 EXPORT_SYMBOL(dma_supported);
200 int dma_is_consistent(struct device *dev, dma_addr_t dma_addr)
205 EXPORT_SYMBOL(dma_is_consistent);
207 void dma_cache_sync(struct device *dev, void *vaddr, size_t size,
208 enum dma_data_direction direction)
210 BUG_ON(direction == DMA_NONE);
213 EXPORT_SYMBOL(dma_cache_sync);
215 dma64_addr_t pci_dac_page_to_dma(struct pci_dev *pdev,
216 struct page *page, unsigned long offset, int direction)
218 dma64_addr_t addr = page_to_phys(page) + offset;
220 return (dma64_addr_t) pdev_to_baddr(pdev, addr);
223 EXPORT_SYMBOL(pci_dac_page_to_dma);
225 struct page *pci_dac_dma_to_page(struct pci_dev *pdev,
226 dma64_addr_t dma_addr)
228 struct bridge_controller *bc = BRIDGE_CONTROLLER(pdev->bus);
230 return pfn_to_page((dma_addr - bc->baddr) >> PAGE_SHIFT);
233 EXPORT_SYMBOL(pci_dac_dma_to_page);
235 unsigned long pci_dac_dma_to_offset(struct pci_dev *pdev,
236 dma64_addr_t dma_addr)
238 return dma_addr & ~PAGE_MASK;
241 EXPORT_SYMBOL(pci_dac_dma_to_offset);
243 void pci_dac_dma_sync_single_for_cpu(struct pci_dev *pdev,
244 dma64_addr_t dma_addr, size_t len, int direction)
246 BUG_ON(direction == PCI_DMA_NONE);
249 EXPORT_SYMBOL(pci_dac_dma_sync_single_for_cpu);
251 void pci_dac_dma_sync_single_for_device(struct pci_dev *pdev,
252 dma64_addr_t dma_addr, size_t len, int direction)
254 BUG_ON(direction == PCI_DMA_NONE);
257 EXPORT_SYMBOL(pci_dac_dma_sync_single_for_device);