|
@@ -0,0 +1,91 @@
|
|
|
+/*
|
|
|
+ * Copyright (C) 2004-2006 Atmel Corporation
|
|
|
+ *
|
|
|
+ * This program is free software; you can redistribute it and/or modify
|
|
|
+ * it under the terms of the GNU General Public License version 2 as
|
|
|
+ * published by the Free Software Foundation.
|
|
|
+ */
|
|
|
+#ifndef __ASM_AVR32_PGTABLE_H
|
|
|
+#define __ASM_AVR32_PGTABLE_H
|
|
|
+
|
|
|
+#include <asm/addrspace.h>
|
|
|
+
|
|
|
+#ifndef __ASSEMBLY__
|
|
|
+#include <linux/sched.h>
|
|
|
+
|
|
|
+#endif /* !__ASSEMBLY__ */
|
|
|
+
|
|
|
+/*
|
|
|
+ * Use two-level page tables just as the i386 (without PAE)
|
|
|
+ */
|
|
|
+#include <asm/pgtable-2level.h>
|
|
|
+
|
|
|
+/*
|
|
|
+ * The following code might need some cleanup when the values are
|
|
|
+ * final...
|
|
|
+ */
|
|
|
+#define PMD_SIZE (1UL << PMD_SHIFT)
|
|
|
+#define PMD_MASK (~(PMD_SIZE-1))
|
|
|
+#define PGDIR_SIZE (1UL << PGDIR_SHIFT)
|
|
|
+#define PGDIR_MASK (~(PGDIR_SIZE-1))
|
|
|
+
|
|
|
+#define USER_PTRS_PER_PGD (TASK_SIZE / PGDIR_SIZE)
|
|
|
+#define FIRST_USER_ADDRESS 0
|
|
|
+
|
|
|
+#ifndef __ASSEMBLY__
|
|
|
+extern pgd_t swapper_pg_dir[PTRS_PER_PGD];
|
|
|
+extern void paging_init(void);
|
|
|
+
|
|
|
+/*
|
|
|
+ * ZERO_PAGE is a global shared page that is always zero: used for
|
|
|
+ * zero-mapped memory areas etc.
|
|
|
+ */
|
|
|
+extern struct page *empty_zero_page;
|
|
|
+#define ZERO_PAGE(vaddr) (empty_zero_page)
|
|
|
+
|
|
|
+/*
|
|
|
+ * Just any arbitrary offset to the start of the vmalloc VM area: the
|
|
|
+ * current 8 MiB value just means that there will be a 8 MiB "hole"
|
|
|
+ * after the uncached physical memory (P2 segment) until the vmalloc
|
|
|
+ * area starts. That means that any out-of-bounds memory accesses will
|
|
|
+ * hopefully be caught; we don't know if the end of the P1/P2 segments
|
|
|
+ * are actually used for anything, but it is anyway safer to let the
|
|
|
+ * MMU catch these kinds of errors than to rely on the memory bus.
|
|
|
+ *
|
|
|
+ * A "hole" of the same size is added to the end of the P3 segment as
|
|
|
+ * well. It might seem wasteful to use 16 MiB of virtual address space
|
|
|
+ * on this, but we do have 512 MiB of it...
|
|
|
+ *
|
|
|
+ * The vmalloc() routines leave a hole of 4 KiB between each vmalloced
|
|
|
+ * area for the same reason.
|
|
|
+ */
|
|
|
+#define VMALLOC_OFFSET (8 * 1024 * 1024)
|
|
|
+#define VMALLOC_START (P3SEG + VMALLOC_OFFSET)
|
|
|
+#define VMALLOC_END (P4SEG - VMALLOC_OFFSET)
|
|
|
+#endif /* !__ASSEMBLY__ */
|
|
|
+
|
|
|
+/*
|
|
|
+ * Page flags. Some of these flags are not directly supported by
|
|
|
+ * hardware, so we have to emulate them.
|
|
|
+ */
|
|
|
+#define _TLBEHI_BIT_VALID 9
|
|
|
+#define _TLBEHI_VALID (1 << _TLBEHI_BIT_VALID)
|
|
|
+
|
|
|
+#define _PAGE_BIT_WT 0 /* W-bit : write-through */
|
|
|
+#define _PAGE_BIT_DIRTY 1 /* D-bit : page changed */
|
|
|
+#define _PAGE_BIT_SZ0 2 /* SZ0-bit : Size of page */
|
|
|
+#define _PAGE_BIT_SZ1 3 /* SZ1-bit : Size of page */
|
|
|
+#define _PAGE_BIT_EXECUTE 4 /* X-bit : execute access allowed */
|
|
|
+#define _PAGE_BIT_RW 5 /* AP0-bit : write access allowed */
|
|
|
+#define _PAGE_BIT_USER 6 /* AP1-bit : user space access allowed */
|
|
|
+#define _PAGE_BIT_BUFFER 7 /* B-bit : bufferable */
|
|
|
+#define _PAGE_BIT_GLOBAL 8 /* G-bit : global (ignore ASID) */
|
|
|
+#define _PAGE_BIT_CACHABLE 9 /* C-bit : cachable */
|
|
|
+
|
|
|
+/* If we drop support for 1K pages, we get two extra bits */
|
|
|
+#define _PAGE_BIT_PRESENT 10
|
|
|
+#define _PAGE_BIT_ACCESSED 11 /* software: page was accessed */
|
|
|
+
|
|
|
+/* The following flags are only valid when !PRESENT */
|
|
|
+#define _PAGE_BIT_FILE 0 /* software: pagecache or swap? */
|
|
|
+
|