forked from Qortal/Brooklyn
You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
181 lines
4.8 KiB
181 lines
4.8 KiB
/* SPDX-License-Identifier: GPL-2.0 */ |
|
#ifndef _M68K_PGTABLE_H |
|
#define _M68K_PGTABLE_H |
|
|
|
|
|
#if defined(CONFIG_SUN3) || defined(CONFIG_COLDFIRE) |
|
#include <asm-generic/pgtable-nopmd.h> |
|
#else |
|
#include <asm-generic/pgtable-nopud.h> |
|
#endif |
|
|
|
#include <asm/setup.h> |
|
|
|
#ifndef __ASSEMBLY__ |
|
#include <asm/processor.h> |
|
#include <linux/sched.h> |
|
#include <linux/threads.h> |
|
|
|
/* |
|
* This file contains the functions and defines necessary to modify and use |
|
* the m68k page table tree. |
|
*/ |
|
|
|
#include <asm/virtconvert.h> |
|
|
|
/* Certain architectures need to do special things when pte's |
|
* within a page table are directly modified. Thus, the following |
|
* hook is made available. |
|
*/ |
|
#define set_pte(pteptr, pteval) \ |
|
do{ \ |
|
*(pteptr) = (pteval); \ |
|
} while(0) |
|
#define set_pte_at(mm,addr,ptep,pteval) set_pte(ptep,pteval) |
|
|
|
|
|
/* PMD_SHIFT determines the size of the area a second-level page table can map */ |
|
#if CONFIG_PGTABLE_LEVELS == 3 |
|
#define PMD_SHIFT 18 |
|
#endif |
|
#define PMD_SIZE (1UL << PMD_SHIFT) |
|
#define PMD_MASK (~(PMD_SIZE-1)) |
|
|
|
/* PGDIR_SHIFT determines what a third-level page table entry can map */ |
|
#ifdef CONFIG_SUN3 |
|
#define PGDIR_SHIFT 17 |
|
#elif defined(CONFIG_COLDFIRE) |
|
#define PGDIR_SHIFT 22 |
|
#else |
|
#define PGDIR_SHIFT 25 |
|
#endif |
|
#define PGDIR_SIZE (1UL << PGDIR_SHIFT) |
|
#define PGDIR_MASK (~(PGDIR_SIZE-1)) |
|
|
|
/* |
|
* entries per page directory level: the m68k is configured as three-level, |
|
* so we do have PMD level physically. |
|
*/ |
|
#ifdef CONFIG_SUN3 |
|
#define PTRS_PER_PTE 16 |
|
#define __PAGETABLE_PMD_FOLDED 1 |
|
#define PTRS_PER_PMD 1 |
|
#define PTRS_PER_PGD 2048 |
|
#elif defined(CONFIG_COLDFIRE) |
|
#define PTRS_PER_PTE 512 |
|
#define __PAGETABLE_PMD_FOLDED 1 |
|
#define PTRS_PER_PMD 1 |
|
#define PTRS_PER_PGD 1024 |
|
#else |
|
#define PTRS_PER_PTE 64 |
|
#define PTRS_PER_PMD 128 |
|
#define PTRS_PER_PGD 128 |
|
#endif |
|
#define USER_PTRS_PER_PGD (TASK_SIZE/PGDIR_SIZE) |
|
#define FIRST_USER_ADDRESS 0UL |
|
|
|
/* Virtual address region for use by kernel_map() */ |
|
#ifdef CONFIG_SUN3 |
|
#define KMAP_START 0x0dc00000 |
|
#define KMAP_END 0x0e000000 |
|
#elif defined(CONFIG_COLDFIRE) |
|
#define KMAP_START 0xe0000000 |
|
#define KMAP_END 0xf0000000 |
|
#else |
|
#define KMAP_START 0xd0000000 |
|
#define KMAP_END 0xf0000000 |
|
#endif |
|
|
|
#ifdef CONFIG_SUN3 |
|
extern unsigned long m68k_vmalloc_end; |
|
#define VMALLOC_START 0x0f800000 |
|
#define VMALLOC_END m68k_vmalloc_end |
|
#elif defined(CONFIG_COLDFIRE) |
|
#define VMALLOC_START 0xd0000000 |
|
#define VMALLOC_END 0xe0000000 |
|
#else |
|
/* Just any arbitrary offset to the start of the vmalloc VM area: the |
|
* current 8MB value just means that there will be a 8MB "hole" after the |
|
* physical memory until the kernel virtual memory starts. That means that |
|
* any out-of-bounds memory accesses will hopefully be caught. |
|
* The vmalloc() routines leaves a hole of 4kB between each vmalloced |
|
* area for the same reason. ;) |
|
*/ |
|
#define VMALLOC_OFFSET (8*1024*1024) |
|
#define VMALLOC_START (((unsigned long) high_memory + VMALLOC_OFFSET) & ~(VMALLOC_OFFSET-1)) |
|
#define VMALLOC_END KMAP_START |
|
#endif |
|
|
|
/* zero page used for uninitialized stuff */ |
|
extern void *empty_zero_page; |
|
|
|
/* |
|
* ZERO_PAGE is a global shared page that is always zero: used |
|
* for zero-mapped memory areas etc.. |
|
*/ |
|
#define ZERO_PAGE(vaddr) (virt_to_page(empty_zero_page)) |
|
|
|
/* number of bits that fit into a memory pointer */ |
|
#define BITS_PER_PTR (8*sizeof(unsigned long)) |
|
|
|
/* to align the pointer to a pointer address */ |
|
#define PTR_MASK (~(sizeof(void*)-1)) |
|
|
|
/* sizeof(void*)==1<<SIZEOF_PTR_LOG2 */ |
|
/* 64-bit machines, beware! SRB. */ |
|
#define SIZEOF_PTR_LOG2 2 |
|
|
|
extern void kernel_set_cachemode(void *addr, unsigned long size, int cmode); |
|
|
|
/* |
|
* The m68k doesn't have any external MMU info: the kernel page |
|
* tables contain all the necessary information. The Sun3 does, but |
|
* they are updated on demand. |
|
*/ |
|
static inline void update_mmu_cache(struct vm_area_struct *vma, |
|
unsigned long address, pte_t *ptep) |
|
{ |
|
} |
|
|
|
#endif /* !__ASSEMBLY__ */ |
|
|
|
#define kern_addr_valid(addr) (1) |
|
|
|
/* MMU-specific headers */ |
|
|
|
#ifdef CONFIG_SUN3 |
|
#include <asm/sun3_pgtable.h> |
|
#elif defined(CONFIG_COLDFIRE) |
|
#include <asm/mcf_pgtable.h> |
|
#else |
|
#include <asm/motorola_pgtable.h> |
|
#endif |
|
|
|
#ifndef __ASSEMBLY__ |
|
/* |
|
* Macro to mark a page protection value as "uncacheable". |
|
*/ |
|
#ifdef CONFIG_COLDFIRE |
|
# define pgprot_noncached(prot) (__pgprot(pgprot_val(prot) | CF_PAGE_NOCACHE)) |
|
#else |
|
#ifdef SUN3_PAGE_NOCACHE |
|
# define __SUN3_PAGE_NOCACHE SUN3_PAGE_NOCACHE |
|
#else |
|
# define __SUN3_PAGE_NOCACHE 0 |
|
#endif |
|
#define pgprot_noncached(prot) \ |
|
(MMU_IS_SUN3 \ |
|
? (__pgprot(pgprot_val(prot) | __SUN3_PAGE_NOCACHE)) \ |
|
: ((MMU_IS_851 || MMU_IS_030) \ |
|
? (__pgprot(pgprot_val(prot) | _PAGE_NOCACHE030)) \ |
|
: (MMU_IS_040 || MMU_IS_060) \ |
|
? (__pgprot((pgprot_val(prot) & _CACHEMASK040) | _PAGE_NOCACHE_S)) \ |
|
: (prot))) |
|
|
|
pgprot_t pgprot_dmacoherent(pgprot_t prot); |
|
#define pgprot_dmacoherent(prot) pgprot_dmacoherent(prot) |
|
|
|
#endif /* CONFIG_COLDFIRE */ |
|
#endif /* !__ASSEMBLY__ */ |
|
|
|
#endif /* _M68K_PGTABLE_H */
|
|
|