Loading...
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 | #ifndef __LINUX_CMA_H #define __LINUX_CMA_H /* * Contiguous Memory Allocator for DMA mapping framework * Copyright (c) 2010-2011 by Samsung Electronics. * Written by: * Marek Szyprowski <m.szyprowski@samsung.com> * Michal Nazarewicz <mina86@mina86.com> * * This program is free software; you can redistribute it and/or * modify it under the terms of the GNU General Public License as * published by the Free Software Foundation; either version 2 of the * License or (at your optional) any later version of the license. */ /* * Contiguous Memory Allocator * * The Contiguous Memory Allocator (CMA) makes it possible to * allocate big contiguous chunks of memory after the system has * booted. * * Why is it needed? * * Various devices on embedded systems have no scatter-getter and/or * IO map support and require contiguous blocks of memory to * operate. They include devices such as cameras, hardware video * coders, etc. * * Such devices often require big memory buffers (a full HD frame * is, for instance, more then 2 mega pixels large, i.e. more than 6 * MB of memory), which makes mechanisms such as kmalloc() or * alloc_page() ineffective. * * At the same time, a solution where a big memory region is * reserved for a device is suboptimal since often more memory is * reserved then strictly required and, moreover, the memory is * inaccessible to page system even if device drivers don't use it. * * CMA tries to solve this issue by operating on memory regions * where only movable pages can be allocated from. This way, kernel * can use the memory for pagecache and when device driver requests * it, allocated pages can be migrated. * * Driver usage * * CMA should not be used by the device drivers directly. It is * only a helper framework for dma-mapping subsystem. * * For more information, see kernel-docs in drivers/base/dma-contiguous.c */ #ifdef __KERNEL__ #include <linux/device.h> struct cma; struct page; #ifdef CONFIG_DMA_CMA extern struct cma *dma_contiguous_default_area; static inline struct cma *dev_get_cma_area(struct device *dev) { if (dev && dev->cma_area) return dev->cma_area; return dma_contiguous_default_area; } static inline void dev_set_cma_area(struct device *dev, struct cma *cma) { if (dev) dev->cma_area = cma; } static inline void dma_contiguous_set_default(struct cma *cma) { dma_contiguous_default_area = cma; } void dma_contiguous_reserve(phys_addr_t addr_limit); int __init dma_contiguous_reserve_area(phys_addr_t size, phys_addr_t base, phys_addr_t limit, struct cma **res_cma, bool fixed); /** * dma_declare_contiguous() - reserve area for contiguous memory handling * for particular device * @dev: Pointer to device structure. * @size: Size of the reserved memory. * @base: Start address of the reserved memory (optional, 0 for any). * @limit: End address of the reserved memory (optional, 0 for any). * * This function reserves memory for specified device. It should be * called by board specific code when early allocator (memblock or bootmem) * is still activate. */ static inline int dma_declare_contiguous(struct device *dev, phys_addr_t size, phys_addr_t base, phys_addr_t limit) { struct cma *cma; int ret; ret = dma_contiguous_reserve_area(size, base, limit, &cma, true); if (ret == 0) dev_set_cma_area(dev, cma); return ret; } struct page *dma_alloc_from_contiguous(struct device *dev, int count, unsigned int order); bool dma_release_from_contiguous(struct device *dev, struct page *pages, int count); #else static inline struct cma *dev_get_cma_area(struct device *dev) { return NULL; } static inline void dev_set_cma_area(struct device *dev, struct cma *cma) { } static inline void dma_contiguous_set_default(struct cma *cma) { } static inline void dma_contiguous_reserve(phys_addr_t limit) { } static inline int dma_contiguous_reserve_area(phys_addr_t size, phys_addr_t base, phys_addr_t limit, struct cma **res_cma, bool fixed) { return -ENOSYS; } static inline int dma_declare_contiguous(struct device *dev, phys_addr_t size, phys_addr_t base, phys_addr_t limit) { return -ENOSYS; } static inline struct page *dma_alloc_from_contiguous(struct device *dev, int count, unsigned int order) { return NULL; } static inline bool dma_release_from_contiguous(struct device *dev, struct page *pages, int count) { return false; } #endif #endif #endif |