1#ifndef _H8300_PGTABLE_H 2#define _H8300_PGTABLE_H 3 4#include <asm-generic/4level-fixup.h> 5 6#include <linux/slab.h> 7#include <asm/processor.h> 8#include <asm/page.h> 9#include <asm/io.h> 10 11#define pgd_present(pgd) (1) /* pages are always present on NO_MM */ 12#define pgd_none(pgd) (0) 13#define pgd_bad(pgd) (0) 14#define pgd_clear(pgdp) 15#define kern_addr_valid(addr) (1) 16#define pmd_offset(a, b) ((void *)0) 17#define pmd_none(pmd) (1) 18#define pgd_offset_k(adrdress) ((pgd_t *)0) 19#define pte_offset_kernel(dir, address) ((pte_t *)0) 20 21#define PAGE_NONE __pgprot(0) /* these mean nothing to NO_MM */ 22#define PAGE_SHARED __pgprot(0) /* these mean nothing to NO_MM */ 23#define PAGE_COPY __pgprot(0) /* these mean nothing to NO_MM */ 24#define PAGE_READONLY __pgprot(0) /* these mean nothing to NO_MM */ 25#define PAGE_KERNEL __pgprot(0) /* these mean nothing to NO_MM */ 26 27extern void paging_init(void); 28#define swapper_pg_dir ((pgd_t *) 0) 29 30#define __swp_type(x) (0) 31#define __swp_offset(x) (0) 32#define __swp_entry(typ,off) ((swp_entry_t) { ((typ) | ((off) << 7)) }) 33#define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val(pte) }) 34#define __swp_entry_to_pte(x) ((pte_t) { (x).val }) 35 36static inline int pte_file(pte_t pte) { return 0; } 37 38/* 39 * ZERO_PAGE is a global shared page that is always zero: used 40 * for zero-mapped memory areas etc.. 41 */ 42#define ZERO_PAGE(vaddr) (virt_to_page(0)) 43 44/* 45 * These would be in other places but having them here reduces the diffs. 46 */ 47extern unsigned int kobjsize(const void *objp); 48extern int is_in_rom(unsigned long); 49 50/* 51 * No page table caches to initialise 52 */ 53#define pgtable_cache_init() do { } while (0) 54 55#define io_remap_pfn_range(vma, vaddr, pfn, size, prot) \ 56 remap_pfn_range(vma, vaddr, pfn, size, prot) 57 58/* 59 * All 32bit addresses are effectively valid for vmalloc... 60 * Sort of meaningless for non-VM targets. 61 */ 62#define VMALLOC_START 0 63#define VMALLOC_END 0xffffffff 64 65/* 66 * All 32bit addresses are effectively valid for vmalloc... 67 * Sort of meaningless for non-VM targets. 68 */ 69#define VMALLOC_START 0 70#define VMALLOC_END 0xffffffff 71 72#define arch_enter_lazy_cpu_mode() do {} while (0) 73#endif /* _H8300_PGTABLE_H */ 74