#ifndef _LINUX_MM_H #define _LINUX_MM_H #include #include #include #include #include #include typedef unsigned long dma_addr_t; #define unlikely #define BUG_ON(x) assert(!(x)) #define WARN_ON(condition) ({ \ int __ret_warn_on = !!(condition); \ unlikely(__ret_warn_on); \ }) #define WARN_ON_ONCE(condition) ({ \ int __ret_warn_on = !!(condition); \ if (unlikely(__ret_warn_on)) \ assert(0); \ unlikely(__ret_warn_on); \ }) #define PAGE_SIZE (4096) #define PAGE_SHIFT (12) #define PAGE_MASK (~(PAGE_SIZE-1)) #define __ALIGN_KERNEL(x, a) __ALIGN_KERNEL_MASK(x, (typeof(x))(a) - 1) #define __ALIGN_KERNEL_MASK(x, mask) (((x) + (mask)) & ~(mask)) #define ALIGN(x, a) __ALIGN_KERNEL((x), (a)) #define ALIGN_DOWN(x, a) __ALIGN_KERNEL((x) - ((a) - 1), (a)) #define PAGE_ALIGN(addr) ALIGN(addr, PAGE_SIZE) #define offset_in_page(p) ((unsigned long)(p) & ~PAGE_MASK) #define virt_to_page(x) ((void *)x) #define page_address(x) ((void *)x) static inline unsigned long page_to_phys(struct page *page) { assert(0); return 0; } #define page_to_pfn(page) ((unsigned long)(page) / PAGE_SIZE) #define pfn_to_page(pfn) (void *)((pfn) * PAGE_SIZE) #define nth_page(page,n) pfn_to_page(page_to_pfn((page)) + (n)) #define __min(t1, t2, min1, min2, x, y) ({ \ t1 min1 = (x); \ t2 min2 = (y); \ (void) (&min1 == &min2); \ min1 < min2 ? min1 : min2; }) #define ___PASTE(a,b) a##b #define __PASTE(a,b) ___PASTE(a,b) #define __UNIQUE_ID(prefix) __PASTE(__PASTE(__UNIQUE_ID_, prefix), __COUNTER__) #define min(x, y) \ __min(typeof(x), typeof(y), \ __UNIQUE_ID(min1_), __UNIQUE_ID(min2_), \ x, y) #define min_t(type, x, y) \ __min(type, type, \ __UNIQUE_ID(min1_), __UNIQUE_ID(min2_), \ x, y) #define pagefault_disabled() (0) static inline void *kmap(struct page *page) { assert(0); return NULL; } static inline void *kmap_atomic(struct page *page) { assert(0); return NULL; } static inline void kunmap(void *addr) { assert(0); } static inline void kunmap_atomic(void *addr) { assert(0); } static inline unsigned long __get_free_page(unsigned int flags) { return (unsigned long)malloc(PAGE_SIZE); } static inline void free_page(unsigned long page) { free((void *)page); } static inline void *kmalloc(unsigned int size, unsigned int flags) { return malloc(size); } static inline void * kmalloc_array(unsigned int n, unsigned int size, unsigned int flags) { return malloc(n * size); } #define kfree(x) free(x) #define kmemleak_alloc(a, b, c, d) #define kmemleak_free(a) #define PageSlab(p) (0) #define flush_dcache_page(p) #define MAX_ERRNO 4095 #define IS_ERR_VALUE(x) unlikely((unsigned long)(void *)(x) >= (unsigned long)-MAX_ERRNO) static inline void * __must_check ERR_PTR(long error) { return (void *) error; } static inline long __must_check PTR_ERR(__force const void *ptr) { return (long) ptr; } static inline bool __must_check IS_ERR(__force const void *ptr) { return IS_ERR_VALUE((unsigned long)ptr); } static inline int __must_check PTR_ERR_OR_ZERO(__force const void *ptr) { if (IS_ERR(ptr)) return PTR_ERR(ptr); else return 0; } #define IS_ENABLED(x) (0) #endif