summaryrefslogtreecommitdiff
diff options
context:
space:
mode:
authorAnthony G. Basile <blueness@gentoo.org>2011-03-27 01:04:52 -0400
committerAnthony G. Basile <blueness@gentoo.org>2011-03-27 01:04:52 -0400
commitde7e71c902329b0dfde8a5ebea1f0eb2b29ea2aa (patch)
tree854bf0745f0d506a008d40893cbb86121d83c165
parentUpdate Grsec/PaX (diff)
downloadhardened-patchset-de7e71c902329b0dfde8a5ebea1f0eb2b29ea2aa.tar.gz
hardened-patchset-de7e71c902329b0dfde8a5ebea1f0eb2b29ea2aa.tar.bz2
hardened-patchset-de7e71c902329b0dfde8a5ebea1f0eb2b29ea2aa.zip
EOL 2.6.37
-rw-r--r--2.6.37/0000_README48
-rw-r--r--2.6.37/4420_grsecurity-2.2.1-2.6.37.4-201103211735.patch60162
-rw-r--r--2.6.37/4421_grsec-remove-localversion-grsec.patch9
-rw-r--r--2.6.37/4422_grsec-mute-warnings.patch35
-rw-r--r--2.6.37/4423_grsec-remove-protected-paths.patch20
-rw-r--r--2.6.37/4425_grsec-pax-without-grsec.patch92
-rw-r--r--2.6.37/4430_grsec-kconfig-default-gids.patch76
-rw-r--r--2.6.37/4435_grsec-kconfig-gentoo.patch316
-rw-r--r--2.6.37/4440_selinux-avc_audit-log-curr_ip.patch71
-rw-r--r--2.6.37/4445_disable-compat_vdso.patch46
10 files changed, 0 insertions, 60875 deletions
diff --git a/2.6.37/0000_README b/2.6.37/0000_README
deleted file mode 100644
index 556e1fe..0000000
--- a/2.6.37/0000_README
+++ /dev/null
@@ -1,48 +0,0 @@
-README
------------------------------------------------------------------------------
-
-Individual Patch Descriptions:
------------------------------------------------------------------------------
-Patch: 4420_grsecurity-2.2.1-2.6.37.4-201103211735.patch
-From: http://www.grsecurity.net
-Desc: hardened-sources base patch from upstream grsecurity
-
-Patch: 4421_grsec-remove-localversion-grsec.patch
-From: Kerin Millar <kerframil@gmail.com>
-Desc: Removes grsecurity's localversion-grsec file
-
-Patch: 4422_grsec-mute-warnings.patch
-From: Alexander Gabert <gaberta@fh-trier.de>
- Gordon Malm <gengor@gentoo.org>
-Desc: Removes verbose compile warning settings from grsecurity, restores
- mainline Linux kernel behavior
-
-Patch: 4423_grsec-remove-protected-paths.patch
-From: Anthony G. Basile <blueness@gentoo.org>
-Desc: Removes chmod statements from grsecurity/Makefile
-
-Patch: 4425_grsec-pax-without-grsec.patch
-From: Gordon Malm <gengor@gentoo.org>
-Desc: Allows PaX features to be selected without enabling GRKERNSEC
-
-Patch: 4430_grsec-kconfig-default-gids.patch
-From: Kerin Millar <kerframil@gmail.com>
-Desc: Sets sane(r) default GIDs on various grsecurity group-dependent
- features
-
-Patch: 4435_grsec-kconfig-gentoo.patch
-From: Gordon Malm <gengor@gentoo.org>
- Kerin Millar <kerframil@gmail.com>
- Anthony G. Basile <blueness@gentoo.org>
-Desc: Adds Hardened Gentoo [server/workstation/virtualization] security levels,
- sets Hardened Gentoo [workstation] as default
-
-Patch: 4440_selinux-avc_audit-log-curr_ip.patch
-From: Gordon Malm <gengor@gentoo.org>
- Anthony G. Basile <blueness@gentoo.org>
-Desc: Configurable option to add src IP address to SELinux log messages
-
-Patch: 4445_disable-compat_vdso.patch
-From: Gordon Malm <gengor@gentoo.org>
- Kerin Millar <kerframil@gmail.com>
-Desc: Disables VDSO_COMPAT operation completely
diff --git a/2.6.37/4420_grsecurity-2.2.1-2.6.37.4-201103211735.patch b/2.6.37/4420_grsecurity-2.2.1-2.6.37.4-201103211735.patch
deleted file mode 100644
index d402c08..0000000
--- a/2.6.37/4420_grsecurity-2.2.1-2.6.37.4-201103211735.patch
+++ /dev/null
@@ -1,60162 +0,0 @@
-diff -urNp linux-2.6.37.4/arch/alpha/include/asm/dma-mapping.h linux-2.6.37.4/arch/alpha/include/asm/dma-mapping.h
---- linux-2.6.37.4/arch/alpha/include/asm/dma-mapping.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/alpha/include/asm/dma-mapping.h 2011-01-17 02:41:00.000000000 -0500
-@@ -3,9 +3,9 @@
-
- #include <linux/dma-attrs.h>
-
--extern struct dma_map_ops *dma_ops;
-+extern const struct dma_map_ops *dma_ops;
-
--static inline struct dma_map_ops *get_dma_ops(struct device *dev)
-+static inline const struct dma_map_ops *get_dma_ops(struct device *dev)
- {
- return dma_ops;
- }
-diff -urNp linux-2.6.37.4/arch/alpha/include/asm/elf.h linux-2.6.37.4/arch/alpha/include/asm/elf.h
---- linux-2.6.37.4/arch/alpha/include/asm/elf.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/alpha/include/asm/elf.h 2011-01-17 02:41:00.000000000 -0500
-@@ -90,6 +90,13 @@ typedef elf_fpreg_t elf_fpregset_t[ELF_N
-
- #define ELF_ET_DYN_BASE (TASK_UNMAPPED_BASE + 0x1000000)
-
-+#ifdef CONFIG_PAX_ASLR
-+#define PAX_ELF_ET_DYN_BASE (current->personality & ADDR_LIMIT_32BIT ? 0x10000 : 0x120000000UL)
-+
-+#define PAX_DELTA_MMAP_LEN (current->personality & ADDR_LIMIT_32BIT ? 14 : 28)
-+#define PAX_DELTA_STACK_LEN (current->personality & ADDR_LIMIT_32BIT ? 14 : 19)
-+#endif
-+
- /* $0 is set by ld.so to a pointer to a function which might be
- registered using atexit. This provides a mean for the dynamic
- linker to call DT_FINI functions for shared libraries that have
-diff -urNp linux-2.6.37.4/arch/alpha/include/asm/pgtable.h linux-2.6.37.4/arch/alpha/include/asm/pgtable.h
---- linux-2.6.37.4/arch/alpha/include/asm/pgtable.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/alpha/include/asm/pgtable.h 2011-01-17 02:41:00.000000000 -0500
-@@ -101,6 +101,17 @@ struct vm_area_struct;
- #define PAGE_SHARED __pgprot(_PAGE_VALID | __ACCESS_BITS)
- #define PAGE_COPY __pgprot(_PAGE_VALID | __ACCESS_BITS | _PAGE_FOW)
- #define PAGE_READONLY __pgprot(_PAGE_VALID | __ACCESS_BITS | _PAGE_FOW)
-+
-+#ifdef CONFIG_PAX_PAGEEXEC
-+# define PAGE_SHARED_NOEXEC __pgprot(_PAGE_VALID | __ACCESS_BITS | _PAGE_FOE)
-+# define PAGE_COPY_NOEXEC __pgprot(_PAGE_VALID | __ACCESS_BITS | _PAGE_FOW | _PAGE_FOE)
-+# define PAGE_READONLY_NOEXEC __pgprot(_PAGE_VALID | __ACCESS_BITS | _PAGE_FOW | _PAGE_FOE)
-+#else
-+# define PAGE_SHARED_NOEXEC PAGE_SHARED
-+# define PAGE_COPY_NOEXEC PAGE_COPY
-+# define PAGE_READONLY_NOEXEC PAGE_READONLY
-+#endif
-+
- #define PAGE_KERNEL __pgprot(_PAGE_VALID | _PAGE_ASM | _PAGE_KRE | _PAGE_KWE)
-
- #define _PAGE_NORMAL(x) __pgprot(_PAGE_VALID | __ACCESS_BITS | (x))
-diff -urNp linux-2.6.37.4/arch/alpha/kernel/module.c linux-2.6.37.4/arch/alpha/kernel/module.c
---- linux-2.6.37.4/arch/alpha/kernel/module.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/alpha/kernel/module.c 2011-01-17 02:41:00.000000000 -0500
-@@ -182,7 +182,7 @@ apply_relocate_add(Elf64_Shdr *sechdrs,
-
- /* The small sections were sorted to the end of the segment.
- The following should definitely cover them. */
-- gp = (u64)me->module_core + me->core_size - 0x8000;
-+ gp = (u64)me->module_core_rw + me->core_size_rw - 0x8000;
- got = sechdrs[me->arch.gotsecindex].sh_addr;
-
- for (i = 0; i < n; i++) {
-diff -urNp linux-2.6.37.4/arch/alpha/kernel/osf_sys.c linux-2.6.37.4/arch/alpha/kernel/osf_sys.c
---- linux-2.6.37.4/arch/alpha/kernel/osf_sys.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/alpha/kernel/osf_sys.c 2011-01-17 02:41:00.000000000 -0500
-@@ -1165,7 +1165,7 @@ arch_get_unmapped_area_1(unsigned long a
- /* At this point: (!vma || addr < vma->vm_end). */
- if (limit - len < addr)
- return -ENOMEM;
-- if (!vma || addr + len <= vma->vm_start)
-+ if (check_heap_stack_gap(vma, addr, len))
- return addr;
- addr = vma->vm_end;
- vma = vma->vm_next;
-@@ -1201,6 +1201,10 @@ arch_get_unmapped_area(struct file *filp
- merely specific addresses, but regions of memory -- perhaps
- this feature should be incorporated into all ports? */
-
-+#ifdef CONFIG_PAX_RANDMMAP
-+ if (!(current->mm->pax_flags & MF_PAX_RANDMMAP))
-+#endif
-+
- if (addr) {
- addr = arch_get_unmapped_area_1 (PAGE_ALIGN(addr), len, limit);
- if (addr != (unsigned long) -ENOMEM)
-@@ -1208,8 +1212,8 @@ arch_get_unmapped_area(struct file *filp
- }
-
- /* Next, try allocating at TASK_UNMAPPED_BASE. */
-- addr = arch_get_unmapped_area_1 (PAGE_ALIGN(TASK_UNMAPPED_BASE),
-- len, limit);
-+ addr = arch_get_unmapped_area_1 (PAGE_ALIGN(current->mm->mmap_base), len, limit);
-+
- if (addr != (unsigned long) -ENOMEM)
- return addr;
-
-diff -urNp linux-2.6.37.4/arch/alpha/kernel/pci_iommu.c linux-2.6.37.4/arch/alpha/kernel/pci_iommu.c
---- linux-2.6.37.4/arch/alpha/kernel/pci_iommu.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/alpha/kernel/pci_iommu.c 2011-01-17 02:41:00.000000000 -0500
-@@ -950,7 +950,7 @@ static int alpha_pci_set_mask(struct dev
- return 0;
- }
-
--struct dma_map_ops alpha_pci_ops = {
-+const struct dma_map_ops alpha_pci_ops = {
- .alloc_coherent = alpha_pci_alloc_coherent,
- .free_coherent = alpha_pci_free_coherent,
- .map_page = alpha_pci_map_page,
-@@ -962,5 +962,5 @@ struct dma_map_ops alpha_pci_ops = {
- .set_dma_mask = alpha_pci_set_mask,
- };
-
--struct dma_map_ops *dma_ops = &alpha_pci_ops;
-+const struct dma_map_ops *dma_ops = &alpha_pci_ops;
- EXPORT_SYMBOL(dma_ops);
-diff -urNp linux-2.6.37.4/arch/alpha/kernel/pci-noop.c linux-2.6.37.4/arch/alpha/kernel/pci-noop.c
---- linux-2.6.37.4/arch/alpha/kernel/pci-noop.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/alpha/kernel/pci-noop.c 2011-01-17 02:41:00.000000000 -0500
-@@ -173,7 +173,7 @@ static int alpha_noop_set_mask(struct de
- return 0;
- }
-
--struct dma_map_ops alpha_noop_ops = {
-+const struct dma_map_ops alpha_noop_ops = {
- .alloc_coherent = alpha_noop_alloc_coherent,
- .free_coherent = alpha_noop_free_coherent,
- .map_page = alpha_noop_map_page,
-@@ -183,7 +183,7 @@ struct dma_map_ops alpha_noop_ops = {
- .set_dma_mask = alpha_noop_set_mask,
- };
-
--struct dma_map_ops *dma_ops = &alpha_noop_ops;
-+const struct dma_map_ops *dma_ops = &alpha_noop_ops;
- EXPORT_SYMBOL(dma_ops);
-
- void __iomem *pci_iomap(struct pci_dev *dev, int bar, unsigned long maxlen)
-diff -urNp linux-2.6.37.4/arch/alpha/mm/fault.c linux-2.6.37.4/arch/alpha/mm/fault.c
---- linux-2.6.37.4/arch/alpha/mm/fault.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/alpha/mm/fault.c 2011-01-17 02:41:00.000000000 -0500
-@@ -54,6 +54,124 @@ __load_new_mm_context(struct mm_struct *
- __reload_thread(pcb);
- }
-
-+#ifdef CONFIG_PAX_PAGEEXEC
-+/*
-+ * PaX: decide what to do with offenders (regs->pc = fault address)
-+ *
-+ * returns 1 when task should be killed
-+ * 2 when patched PLT trampoline was detected
-+ * 3 when unpatched PLT trampoline was detected
-+ */
-+static int pax_handle_fetch_fault(struct pt_regs *regs)
-+{
-+
-+#ifdef CONFIG_PAX_EMUPLT
-+ int err;
-+
-+ do { /* PaX: patched PLT emulation #1 */
-+ unsigned int ldah, ldq, jmp;
-+
-+ err = get_user(ldah, (unsigned int *)regs->pc);
-+ err |= get_user(ldq, (unsigned int *)(regs->pc+4));
-+ err |= get_user(jmp, (unsigned int *)(regs->pc+8));
-+
-+ if (err)
-+ break;
-+
-+ if ((ldah & 0xFFFF0000U) == 0x277B0000U &&
-+ (ldq & 0xFFFF0000U) == 0xA77B0000U &&
-+ jmp == 0x6BFB0000U)
-+ {
-+ unsigned long r27, addr;
-+ unsigned long addrh = (ldah | 0xFFFFFFFFFFFF0000UL) << 16;
-+ unsigned long addrl = ldq | 0xFFFFFFFFFFFF0000UL;
-+
-+ addr = regs->r27 + ((addrh ^ 0x80000000UL) + 0x80000000UL) + ((addrl ^ 0x8000UL) + 0x8000UL);
-+ err = get_user(r27, (unsigned long *)addr);
-+ if (err)
-+ break;
-+
-+ regs->r27 = r27;
-+ regs->pc = r27;
-+ return 2;
-+ }
-+ } while (0);
-+
-+ do { /* PaX: patched PLT emulation #2 */
-+ unsigned int ldah, lda, br;
-+
-+ err = get_user(ldah, (unsigned int *)regs->pc);
-+ err |= get_user(lda, (unsigned int *)(regs->pc+4));
-+ err |= get_user(br, (unsigned int *)(regs->pc+8));
-+
-+ if (err)
-+ break;
-+
-+ if ((ldah & 0xFFFF0000U) == 0x277B0000U &&
-+ (lda & 0xFFFF0000U) == 0xA77B0000U &&
-+ (br & 0xFFE00000U) == 0xC3E00000U)
-+ {
-+ unsigned long addr = br | 0xFFFFFFFFFFE00000UL;
-+ unsigned long addrh = (ldah | 0xFFFFFFFFFFFF0000UL) << 16;
-+ unsigned long addrl = lda | 0xFFFFFFFFFFFF0000UL;
-+
-+ regs->r27 += ((addrh ^ 0x80000000UL) + 0x80000000UL) + ((addrl ^ 0x8000UL) + 0x8000UL);
-+ regs->pc += 12 + (((addr ^ 0x00100000UL) + 0x00100000UL) << 2);
-+ return 2;
-+ }
-+ } while (0);
-+
-+ do { /* PaX: unpatched PLT emulation */
-+ unsigned int br;
-+
-+ err = get_user(br, (unsigned int *)regs->pc);
-+
-+ if (!err && (br & 0xFFE00000U) == 0xC3800000U) {
-+ unsigned int br2, ldq, nop, jmp;
-+ unsigned long addr = br | 0xFFFFFFFFFFE00000UL, resolver;
-+
-+ addr = regs->pc + 4 + (((addr ^ 0x00100000UL) + 0x00100000UL) << 2);
-+ err = get_user(br2, (unsigned int *)addr);
-+ err |= get_user(ldq, (unsigned int *)(addr+4));
-+ err |= get_user(nop, (unsigned int *)(addr+8));
-+ err |= get_user(jmp, (unsigned int *)(addr+12));
-+ err |= get_user(resolver, (unsigned long *)(addr+16));
-+
-+ if (err)
-+ break;
-+
-+ if (br2 == 0xC3600000U &&
-+ ldq == 0xA77B000CU &&
-+ nop == 0x47FF041FU &&
-+ jmp == 0x6B7B0000U)
-+ {
-+ regs->r28 = regs->pc+4;
-+ regs->r27 = addr+16;
-+ regs->pc = resolver;
-+ return 3;
-+ }
-+ }
-+ } while (0);
-+#endif
-+
-+ return 1;
-+}
-+
-+void pax_report_insns(void *pc, void *sp)
-+{
-+ unsigned long i;
-+
-+ printk(KERN_ERR "PAX: bytes at PC: ");
-+ for (i = 0; i < 5; i++) {
-+ unsigned int c;
-+ if (get_user(c, (unsigned int *)pc+i))
-+ printk(KERN_CONT "???????? ");
-+ else
-+ printk(KERN_CONT "%08x ", c);
-+ }
-+ printk("\n");
-+}
-+#endif
-
- /*
- * This routine handles page faults. It determines the address,
-@@ -131,8 +249,29 @@ do_page_fault(unsigned long address, uns
- good_area:
- si_code = SEGV_ACCERR;
- if (cause < 0) {
-- if (!(vma->vm_flags & VM_EXEC))
-+ if (!(vma->vm_flags & VM_EXEC)) {
-+
-+#ifdef CONFIG_PAX_PAGEEXEC
-+ if (!(mm->pax_flags & MF_PAX_PAGEEXEC) || address != regs->pc)
-+ goto bad_area;
-+
-+ up_read(&mm->mmap_sem);
-+ switch (pax_handle_fetch_fault(regs)) {
-+
-+#ifdef CONFIG_PAX_EMUPLT
-+ case 2:
-+ case 3:
-+ return;
-+#endif
-+
-+ }
-+ pax_report_fault(regs, (void *)regs->pc, (void *)rdusp());
-+ do_group_exit(SIGKILL);
-+#else
- goto bad_area;
-+#endif
-+
-+ }
- } else if (!cause) {
- /* Allow reads even for write-only mappings */
- if (!(vma->vm_flags & (VM_READ | VM_WRITE)))
-diff -urNp linux-2.6.37.4/arch/arm/include/asm/elf.h linux-2.6.37.4/arch/arm/include/asm/elf.h
---- linux-2.6.37.4/arch/arm/include/asm/elf.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/arm/include/asm/elf.h 2011-01-17 02:41:00.000000000 -0500
-@@ -113,7 +113,14 @@ int dump_task_regs(struct task_struct *t
- the loader. We need to make sure that it is out of the way of the program
- that it will "exec", and that there is sufficient room for the brk. */
-
--#define ELF_ET_DYN_BASE (2 * TASK_SIZE / 3)
-+#define ELF_ET_DYN_BASE (TASK_SIZE / 3 * 2)
-+
-+#ifdef CONFIG_PAX_ASLR
-+#define PAX_ELF_ET_DYN_BASE 0x00008000UL
-+
-+#define PAX_DELTA_MMAP_LEN ((current->personality == PER_LINUX_32BIT) ? 16 : 10)
-+#define PAX_DELTA_STACK_LEN ((current->personality == PER_LINUX_32BIT) ? 16 : 10)
-+#endif
-
- /* When the program starts, a1 contains a pointer to a function to be
- registered with atexit, as per the SVR4 ABI. A value of 0 means we
-@@ -123,10 +130,6 @@ int dump_task_regs(struct task_struct *t
- extern void elf_set_personality(const struct elf32_hdr *);
- #define SET_PERSONALITY(ex) elf_set_personality(&(ex))
-
--struct mm_struct;
--extern unsigned long arch_randomize_brk(struct mm_struct *mm);
--#define arch_randomize_brk arch_randomize_brk
--
- extern int vectors_user_mapping(void);
- #define arch_setup_additional_pages(bprm, uses_interp) vectors_user_mapping()
- #define ARCH_HAS_SETUP_ADDITIONAL_PAGES
-diff -urNp linux-2.6.37.4/arch/arm/include/asm/kmap_types.h linux-2.6.37.4/arch/arm/include/asm/kmap_types.h
---- linux-2.6.37.4/arch/arm/include/asm/kmap_types.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/arm/include/asm/kmap_types.h 2011-01-17 02:41:00.000000000 -0500
-@@ -21,6 +21,7 @@ enum km_type {
- KM_L1_CACHE,
- KM_L2_CACHE,
- KM_KDB,
-+ KM_CLEARPAGE,
- KM_TYPE_NR
- };
-
-diff -urNp linux-2.6.37.4/arch/arm/include/asm/uaccess.h linux-2.6.37.4/arch/arm/include/asm/uaccess.h
---- linux-2.6.37.4/arch/arm/include/asm/uaccess.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/arm/include/asm/uaccess.h 2011-01-17 02:41:00.000000000 -0500
-@@ -403,6 +403,9 @@ extern unsigned long __must_check __strn
-
- static inline unsigned long __must_check copy_from_user(void *to, const void __user *from, unsigned long n)
- {
-+ if ((long)n < 0)
-+ return n;
-+
- if (access_ok(VERIFY_READ, from, n))
- n = __copy_from_user(to, from, n);
- else /* security hole - plug it */
-@@ -412,6 +415,9 @@ static inline unsigned long __must_check
-
- static inline unsigned long __must_check copy_to_user(void __user *to, const void *from, unsigned long n)
- {
-+ if ((long)n < 0)
-+ return n;
-+
- if (access_ok(VERIFY_WRITE, to, n))
- n = __copy_to_user(to, from, n);
- return n;
-diff -urNp linux-2.6.37.4/arch/arm/kernel/kgdb.c linux-2.6.37.4/arch/arm/kernel/kgdb.c
---- linux-2.6.37.4/arch/arm/kernel/kgdb.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/arm/kernel/kgdb.c 2011-01-17 02:41:00.000000000 -0500
-@@ -246,7 +246,7 @@ void kgdb_arch_exit(void)
- * and we handle the normal undef case within the do_undefinstr
- * handler.
- */
--struct kgdb_arch arch_kgdb_ops = {
-+const struct kgdb_arch arch_kgdb_ops = {
- #ifndef __ARMEB__
- .gdb_bpt_instr = {0xfe, 0xde, 0xff, 0xe7}
- #else /* ! __ARMEB__ */
-diff -urNp linux-2.6.37.4/arch/arm/kernel/process.c linux-2.6.37.4/arch/arm/kernel/process.c
---- linux-2.6.37.4/arch/arm/kernel/process.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/arm/kernel/process.c 2011-01-17 02:41:00.000000000 -0500
-@@ -28,7 +28,6 @@
- #include <linux/tick.h>
- #include <linux/utsname.h>
- #include <linux/uaccess.h>
--#include <linux/random.h>
- #include <linux/hw_breakpoint.h>
-
- #include <asm/cacheflush.h>
-@@ -477,12 +476,6 @@ unsigned long get_wchan(struct task_stru
- return 0;
- }
-
--unsigned long arch_randomize_brk(struct mm_struct *mm)
--{
-- unsigned long range_end = mm->brk + 0x02000000;
-- return randomize_range(mm->brk, range_end, 0) ? : mm->brk;
--}
--
- /*
- * The vectors page is always readable from user space for the
- * atomic helpers and the signal restart code. Let's declare a mapping
-diff -urNp linux-2.6.37.4/arch/arm/mach-at91/pm.c linux-2.6.37.4/arch/arm/mach-at91/pm.c
---- linux-2.6.37.4/arch/arm/mach-at91/pm.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/arm/mach-at91/pm.c 2011-01-17 02:41:00.000000000 -0500
-@@ -301,7 +301,7 @@ static void at91_pm_end(void)
- }
-
-
--static struct platform_suspend_ops at91_pm_ops ={
-+static const struct platform_suspend_ops at91_pm_ops ={
- .valid = at91_pm_valid_state,
- .begin = at91_pm_begin,
- .enter = at91_pm_enter,
-diff -urNp linux-2.6.37.4/arch/arm/mach-davinci/pm.c linux-2.6.37.4/arch/arm/mach-davinci/pm.c
---- linux-2.6.37.4/arch/arm/mach-davinci/pm.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/arm/mach-davinci/pm.c 2011-01-17 02:41:00.000000000 -0500
-@@ -110,7 +110,7 @@ static int davinci_pm_enter(suspend_stat
- return ret;
- }
-
--static struct platform_suspend_ops davinci_pm_ops = {
-+static const struct platform_suspend_ops davinci_pm_ops = {
- .enter = davinci_pm_enter,
- .valid = suspend_valid_only_mem,
- };
-diff -urNp linux-2.6.37.4/arch/arm/mach-imx/pm-imx27.c linux-2.6.37.4/arch/arm/mach-imx/pm-imx27.c
---- linux-2.6.37.4/arch/arm/mach-imx/pm-imx27.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/arm/mach-imx/pm-imx27.c 2011-01-17 02:41:00.000000000 -0500
-@@ -32,7 +32,7 @@ static int mx27_suspend_enter(suspend_st
- return 0;
- }
-
--static struct platform_suspend_ops mx27_suspend_ops = {
-+static const struct platform_suspend_ops mx27_suspend_ops = {
- .enter = mx27_suspend_enter,
- .valid = suspend_valid_only_mem,
- };
-diff -urNp linux-2.6.37.4/arch/arm/mach-lpc32xx/pm.c linux-2.6.37.4/arch/arm/mach-lpc32xx/pm.c
---- linux-2.6.37.4/arch/arm/mach-lpc32xx/pm.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/arm/mach-lpc32xx/pm.c 2011-01-17 02:41:00.000000000 -0500
-@@ -123,7 +123,7 @@ static int lpc32xx_pm_enter(suspend_stat
- return 0;
- }
-
--static struct platform_suspend_ops lpc32xx_pm_ops = {
-+static const struct platform_suspend_ops lpc32xx_pm_ops = {
- .valid = suspend_valid_only_mem,
- .enter = lpc32xx_pm_enter,
- };
-diff -urNp linux-2.6.37.4/arch/arm/mach-msm/last_radio_log.c linux-2.6.37.4/arch/arm/mach-msm/last_radio_log.c
---- linux-2.6.37.4/arch/arm/mach-msm/last_radio_log.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/arm/mach-msm/last_radio_log.c 2011-01-17 02:41:00.000000000 -0500
-@@ -47,6 +47,7 @@ static ssize_t last_radio_log_read(struc
- return count;
- }
-
-+/* cannot be const, see msm_init_last_radio_log */
- static struct file_operations last_radio_log_fops = {
- .read = last_radio_log_read,
- .llseek = default_llseek,
-diff -urNp linux-2.6.37.4/arch/arm/mach-omap1/pm.c linux-2.6.37.4/arch/arm/mach-omap1/pm.c
---- linux-2.6.37.4/arch/arm/mach-omap1/pm.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/arm/mach-omap1/pm.c 2011-01-17 02:41:00.000000000 -0500
-@@ -647,7 +647,7 @@ static struct irqaction omap_wakeup_irq
-
-
-
--static struct platform_suspend_ops omap_pm_ops ={
-+static const struct platform_suspend_ops omap_pm_ops ={
- .prepare = omap_pm_prepare,
- .enter = omap_pm_enter,
- .finish = omap_pm_finish,
-diff -urNp linux-2.6.37.4/arch/arm/mach-omap2/mux.c linux-2.6.37.4/arch/arm/mach-omap2/mux.c
---- linux-2.6.37.4/arch/arm/mach-omap2/mux.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/arm/mach-omap2/mux.c 2011-02-27 15:27:38.000000000 -0500
-@@ -393,7 +393,7 @@ static void __init omap_mux_dbg_init(voi
- list_for_each_entry(e, &muxmodes, node) {
- struct omap_mux *m = &e->mux;
-
-- (void)debugfs_create_file(m->muxnames[0], S_IWUGO, mux_dbg_dir,
-+ (void)debugfs_create_file(m->muxnames[0], S_IWUSR, mux_dbg_dir,
- m, &omap_mux_dbg_signal_fops);
- }
- }
-diff -urNp linux-2.6.37.4/arch/arm/mach-omap2/pm24xx.c linux-2.6.37.4/arch/arm/mach-omap2/pm24xx.c
---- linux-2.6.37.4/arch/arm/mach-omap2/pm24xx.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/arm/mach-omap2/pm24xx.c 2011-01-17 02:41:00.000000000 -0500
-@@ -359,7 +359,7 @@ static void omap2_pm_end(void)
- suspend_state = PM_SUSPEND_ON;
- }
-
--static struct platform_suspend_ops omap_pm_ops = {
-+static const struct platform_suspend_ops omap_pm_ops = {
- .begin = omap2_pm_begin,
- .prepare = omap2_pm_prepare,
- .enter = omap2_pm_enter,
-diff -urNp linux-2.6.37.4/arch/arm/mach-omap2/pm34xx.c linux-2.6.37.4/arch/arm/mach-omap2/pm34xx.c
---- linux-2.6.37.4/arch/arm/mach-omap2/pm34xx.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/arm/mach-omap2/pm34xx.c 2011-01-17 02:41:00.000000000 -0500
-@@ -617,7 +617,7 @@ static void omap3_pm_end(void)
- return;
- }
-
--static struct platform_suspend_ops omap_pm_ops = {
-+static const struct platform_suspend_ops omap_pm_ops = {
- .begin = omap3_pm_begin,
- .end = omap3_pm_end,
- .prepare = omap3_pm_prepare,
-diff -urNp linux-2.6.37.4/arch/arm/mach-omap2/pm44xx.c linux-2.6.37.4/arch/arm/mach-omap2/pm44xx.c
---- linux-2.6.37.4/arch/arm/mach-omap2/pm44xx.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/arm/mach-omap2/pm44xx.c 2011-01-17 02:41:00.000000000 -0500
-@@ -75,7 +75,7 @@ static void omap4_pm_end(void)
- return;
- }
-
--static struct platform_suspend_ops omap_pm_ops = {
-+static const struct platform_suspend_ops omap_pm_ops = {
- .begin = omap4_pm_begin,
- .end = omap4_pm_end,
- .prepare = omap4_pm_prepare,
-diff -urNp linux-2.6.37.4/arch/arm/mach-omap2/pm-debug.c linux-2.6.37.4/arch/arm/mach-omap2/pm-debug.c
---- linux-2.6.37.4/arch/arm/mach-omap2/pm-debug.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/arm/mach-omap2/pm-debug.c 2011-02-27 15:28:01.000000000 -0500
-@@ -632,14 +632,14 @@ static int __init pm_dbg_init(void)
-
- }
-
-- (void) debugfs_create_file("enable_off_mode", S_IRUGO | S_IWUGO, d,
-+ (void) debugfs_create_file("enable_off_mode", S_IRUGO | S_IWUSR, d,
- &enable_off_mode, &pm_dbg_option_fops);
-- (void) debugfs_create_file("sleep_while_idle", S_IRUGO | S_IWUGO, d,
-+ (void) debugfs_create_file("sleep_while_idle", S_IRUGO | S_IWUSR, d,
- &sleep_while_idle, &pm_dbg_option_fops);
-- (void) debugfs_create_file("wakeup_timer_seconds", S_IRUGO | S_IWUGO, d,
-+ (void) debugfs_create_file("wakeup_timer_seconds", S_IRUGO | S_IWUSR, d,
- &wakeup_timer_seconds, &pm_dbg_option_fops);
- (void) debugfs_create_file("wakeup_timer_milliseconds",
-- S_IRUGO | S_IWUGO, d, &wakeup_timer_milliseconds,
-+ S_IRUGO | S_IWUSR, d, &wakeup_timer_milliseconds,
- &pm_dbg_option_fops);
- pm_dbg_init_done = 1;
-
-diff -urNp linux-2.6.37.4/arch/arm/mach-pnx4008/pm.c linux-2.6.37.4/arch/arm/mach-pnx4008/pm.c
---- linux-2.6.37.4/arch/arm/mach-pnx4008/pm.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/arm/mach-pnx4008/pm.c 2011-01-17 02:41:00.000000000 -0500
-@@ -119,7 +119,7 @@ static int pnx4008_pm_valid(suspend_stat
- (state == PM_SUSPEND_MEM);
- }
-
--static struct platform_suspend_ops pnx4008_pm_ops = {
-+static const struct platform_suspend_ops pnx4008_pm_ops = {
- .enter = pnx4008_pm_enter,
- .valid = pnx4008_pm_valid,
- };
-diff -urNp linux-2.6.37.4/arch/arm/mach-pxa/pm.c linux-2.6.37.4/arch/arm/mach-pxa/pm.c
---- linux-2.6.37.4/arch/arm/mach-pxa/pm.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/arm/mach-pxa/pm.c 2011-01-17 02:41:00.000000000 -0500
-@@ -96,7 +96,7 @@ void pxa_pm_finish(void)
- pxa_cpu_pm_fns->finish();
- }
-
--static struct platform_suspend_ops pxa_pm_ops = {
-+static const struct platform_suspend_ops pxa_pm_ops = {
- .valid = pxa_pm_valid,
- .enter = pxa_pm_enter,
- .prepare = pxa_pm_prepare,
-diff -urNp linux-2.6.37.4/arch/arm/mach-pxa/sharpsl_pm.c linux-2.6.37.4/arch/arm/mach-pxa/sharpsl_pm.c
---- linux-2.6.37.4/arch/arm/mach-pxa/sharpsl_pm.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/arm/mach-pxa/sharpsl_pm.c 2011-01-17 02:41:00.000000000 -0500
-@@ -868,7 +868,7 @@ static void sharpsl_apm_get_power_status
- }
-
- #ifdef CONFIG_PM
--static struct platform_suspend_ops sharpsl_pm_ops = {
-+static const struct platform_suspend_ops sharpsl_pm_ops = {
- .prepare = pxa_pm_prepare,
- .finish = pxa_pm_finish,
- .enter = corgi_pxa_pm_enter,
-diff -urNp linux-2.6.37.4/arch/arm/mach-sa1100/pm.c linux-2.6.37.4/arch/arm/mach-sa1100/pm.c
---- linux-2.6.37.4/arch/arm/mach-sa1100/pm.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/arm/mach-sa1100/pm.c 2011-01-17 02:41:00.000000000 -0500
-@@ -120,7 +120,7 @@ unsigned long sleep_phys_sp(void *sp)
- return virt_to_phys(sp);
- }
-
--static struct platform_suspend_ops sa11x0_pm_ops = {
-+static const struct platform_suspend_ops sa11x0_pm_ops = {
- .enter = sa11x0_pm_enter,
- .valid = suspend_valid_only_mem,
- };
-diff -urNp linux-2.6.37.4/arch/arm/mach-ux500/mbox.c linux-2.6.37.4/arch/arm/mach-ux500/mbox.c
---- linux-2.6.37.4/arch/arm/mach-ux500/mbox.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/arm/mach-ux500/mbox.c 2011-02-27 15:26:39.000000000 -0500
-@@ -168,7 +168,7 @@ static ssize_t mbox_read_fifo(struct dev
- return sprintf(buf, "0x%X\n", mbox_value);
- }
-
--static DEVICE_ATTR(fifo, S_IWUGO | S_IRUGO, mbox_read_fifo, mbox_write_fifo);
-+static DEVICE_ATTR(fifo, S_IWUSR | S_IRUGO, mbox_read_fifo, mbox_write_fifo);
-
- static int mbox_show(struct seq_file *s, void *data)
- {
-diff -urNp linux-2.6.37.4/arch/arm/mm/fault.c linux-2.6.37.4/arch/arm/mm/fault.c
---- linux-2.6.37.4/arch/arm/mm/fault.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/arm/mm/fault.c 2011-01-17 02:41:00.000000000 -0500
-@@ -167,6 +167,13 @@ __do_user_fault(struct task_struct *tsk,
- }
- #endif
-
-+#ifdef CONFIG_PAX_PAGEEXEC
-+ if (fsr & FSR_LNX_PF) {
-+ pax_report_fault(regs, (void *)regs->ARM_pc, (void *)regs->ARM_sp);
-+ do_group_exit(SIGKILL);
-+ }
-+#endif
-+
- tsk->thread.address = addr;
- tsk->thread.error_code = fsr;
- tsk->thread.trap_no = 14;
-@@ -364,6 +371,33 @@ do_page_fault(unsigned long addr, unsign
- }
- #endif /* CONFIG_MMU */
-
-+#ifdef CONFIG_PAX_PAGEEXEC
-+void pax_report_insns(void *pc, void *sp)
-+{
-+ long i;
-+
-+ printk(KERN_ERR "PAX: bytes at PC: ");
-+ for (i = 0; i < 20; i++) {
-+ unsigned char c;
-+ if (get_user(c, (__force unsigned char __user *)pc+i))
-+ printk(KERN_CONT "?? ");
-+ else
-+ printk(KERN_CONT "%02x ", c);
-+ }
-+ printk("\n");
-+
-+ printk(KERN_ERR "PAX: bytes at SP-4: ");
-+ for (i = -1; i < 20; i++) {
-+ unsigned long c;
-+ if (get_user(c, (__force unsigned long __user *)sp+i))
-+ printk(KERN_CONT "???????? ");
-+ else
-+ printk(KERN_CONT "%08lx ", c);
-+ }
-+ printk("\n");
-+}
-+#endif
-+
- /*
- * First Level Translation Fault Handler
- *
-diff -urNp linux-2.6.37.4/arch/arm/mm/mmap.c linux-2.6.37.4/arch/arm/mm/mmap.c
---- linux-2.6.37.4/arch/arm/mm/mmap.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/arm/mm/mmap.c 2011-01-17 02:41:00.000000000 -0500
-@@ -64,6 +64,10 @@ arch_get_unmapped_area(struct file *filp
- if (len > TASK_SIZE)
- return -ENOMEM;
-
-+#ifdef CONFIG_PAX_RANDMMAP
-+ if (!(mm->pax_flags & MF_PAX_RANDMMAP))
-+#endif
-+
- if (addr) {
- if (do_align)
- addr = COLOUR_ALIGN(addr, pgoff);
-@@ -71,15 +75,14 @@ arch_get_unmapped_area(struct file *filp
- addr = PAGE_ALIGN(addr);
-
- vma = find_vma(mm, addr);
-- if (TASK_SIZE - len >= addr &&
-- (!vma || addr + len <= vma->vm_start))
-+ if (TASK_SIZE - len >= addr && check_heap_stack_gap(vma, addr, len))
- return addr;
- }
- if (len > mm->cached_hole_size) {
-- start_addr = addr = mm->free_area_cache;
-+ start_addr = addr = mm->free_area_cache;
- } else {
-- start_addr = addr = TASK_UNMAPPED_BASE;
-- mm->cached_hole_size = 0;
-+ start_addr = addr = mm->mmap_base;
-+ mm->cached_hole_size = 0;
- }
- /* 8 bits of randomness in 20 address space bits */
- if (current->flags & PF_RANDOMIZE)
-@@ -98,14 +101,14 @@ full_search:
- * Start a new search - just in case we missed
- * some holes.
- */
-- if (start_addr != TASK_UNMAPPED_BASE) {
-- start_addr = addr = TASK_UNMAPPED_BASE;
-+ if (start_addr != mm->mmap_base) {
-+ start_addr = addr = mm->mmap_base;
- mm->cached_hole_size = 0;
- goto full_search;
- }
- return -ENOMEM;
- }
-- if (!vma || addr + len <= vma->vm_start) {
-+ if (check_heap_stack_gap(vma, addr, len)) {
- /*
- * Remember the place where we stopped the search:
- */
-diff -urNp linux-2.6.37.4/arch/arm/plat-samsung/pm.c linux-2.6.37.4/arch/arm/plat-samsung/pm.c
---- linux-2.6.37.4/arch/arm/plat-samsung/pm.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/arm/plat-samsung/pm.c 2011-01-17 02:41:00.000000000 -0500
-@@ -355,7 +355,7 @@ static void s3c_pm_finish(void)
- s3c_pm_check_cleanup();
- }
-
--static struct platform_suspend_ops s3c_pm_ops = {
-+static const struct platform_suspend_ops s3c_pm_ops = {
- .enter = s3c_pm_enter,
- .prepare = s3c_pm_prepare,
- .finish = s3c_pm_finish,
-diff -urNp linux-2.6.37.4/arch/avr32/include/asm/elf.h linux-2.6.37.4/arch/avr32/include/asm/elf.h
---- linux-2.6.37.4/arch/avr32/include/asm/elf.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/avr32/include/asm/elf.h 2011-01-17 02:41:00.000000000 -0500
-@@ -84,8 +84,14 @@ typedef struct user_fpu_struct elf_fpreg
- the loader. We need to make sure that it is out of the way of the program
- that it will "exec", and that there is sufficient room for the brk. */
-
--#define ELF_ET_DYN_BASE (2 * TASK_SIZE / 3)
-+#define ELF_ET_DYN_BASE (TASK_SIZE / 3 * 2)
-
-+#ifdef CONFIG_PAX_ASLR
-+#define PAX_ELF_ET_DYN_BASE 0x00001000UL
-+
-+#define PAX_DELTA_MMAP_LEN 15
-+#define PAX_DELTA_STACK_LEN 15
-+#endif
-
- /* This yields a mask that user programs can use to figure out what
- instruction set this CPU supports. This could be done in user space,
-diff -urNp linux-2.6.37.4/arch/avr32/include/asm/kmap_types.h linux-2.6.37.4/arch/avr32/include/asm/kmap_types.h
---- linux-2.6.37.4/arch/avr32/include/asm/kmap_types.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/avr32/include/asm/kmap_types.h 2011-01-17 02:41:00.000000000 -0500
-@@ -22,7 +22,8 @@ D(10) KM_IRQ0,
- D(11) KM_IRQ1,
- D(12) KM_SOFTIRQ0,
- D(13) KM_SOFTIRQ1,
--D(14) KM_TYPE_NR
-+D(14) KM_CLEARPAGE,
-+D(15) KM_TYPE_NR
- };
-
- #undef D
-diff -urNp linux-2.6.37.4/arch/avr32/mach-at32ap/pm.c linux-2.6.37.4/arch/avr32/mach-at32ap/pm.c
---- linux-2.6.37.4/arch/avr32/mach-at32ap/pm.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/avr32/mach-at32ap/pm.c 2011-01-17 02:41:00.000000000 -0500
-@@ -176,7 +176,7 @@ out:
- return 0;
- }
-
--static struct platform_suspend_ops avr32_pm_ops = {
-+static const struct platform_suspend_ops avr32_pm_ops = {
- .valid = avr32_pm_valid_state,
- .enter = avr32_pm_enter,
- };
-diff -urNp linux-2.6.37.4/arch/avr32/mm/fault.c linux-2.6.37.4/arch/avr32/mm/fault.c
---- linux-2.6.37.4/arch/avr32/mm/fault.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/avr32/mm/fault.c 2011-01-17 02:41:00.000000000 -0500
-@@ -41,6 +41,23 @@ static inline int notify_page_fault(stru
-
- int exception_trace = 1;
-
-+#ifdef CONFIG_PAX_PAGEEXEC
-+void pax_report_insns(void *pc, void *sp)
-+{
-+ unsigned long i;
-+
-+ printk(KERN_ERR "PAX: bytes at PC: ");
-+ for (i = 0; i < 20; i++) {
-+ unsigned char c;
-+ if (get_user(c, (unsigned char *)pc+i))
-+ printk(KERN_CONT "???????? ");
-+ else
-+ printk(KERN_CONT "%02x ", c);
-+ }
-+ printk("\n");
-+}
-+#endif
-+
- /*
- * This routine handles page faults. It determines the address and the
- * problem, and then passes it off to one of the appropriate routines.
-@@ -156,6 +173,16 @@ bad_area:
- up_read(&mm->mmap_sem);
-
- if (user_mode(regs)) {
-+
-+#ifdef CONFIG_PAX_PAGEEXEC
-+ if (mm->pax_flags & MF_PAX_PAGEEXEC) {
-+ if (ecr == ECR_PROTECTION_X || ecr == ECR_TLB_MISS_X) {
-+ pax_report_fault(regs, (void *)regs->pc, (void *)regs->sp);
-+ do_group_exit(SIGKILL);
-+ }
-+ }
-+#endif
-+
- if (exception_trace && printk_ratelimit())
- printk("%s%s[%d]: segfault at %08lx pc %08lx "
- "sp %08lx ecr %lu\n",
-diff -urNp linux-2.6.37.4/arch/blackfin/kernel/kgdb.c linux-2.6.37.4/arch/blackfin/kernel/kgdb.c
---- linux-2.6.37.4/arch/blackfin/kernel/kgdb.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/blackfin/kernel/kgdb.c 2011-01-17 02:41:00.000000000 -0500
-@@ -397,7 +397,7 @@ int kgdb_arch_handle_exception(int vecto
- return -1; /* this means that we do not want to exit from the handler */
- }
-
--struct kgdb_arch arch_kgdb_ops = {
-+const struct kgdb_arch arch_kgdb_ops = {
- .gdb_bpt_instr = {0xa1},
- #ifdef CONFIG_SMP
- .flags = KGDB_HW_BREAKPOINT|KGDB_THR_PROC_SWAP,
-diff -urNp linux-2.6.37.4/arch/blackfin/mach-common/pm.c linux-2.6.37.4/arch/blackfin/mach-common/pm.c
---- linux-2.6.37.4/arch/blackfin/mach-common/pm.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/blackfin/mach-common/pm.c 2011-01-17 02:41:00.000000000 -0500
-@@ -233,7 +233,7 @@ static int bfin_pm_enter(suspend_state_t
- return 0;
- }
-
--struct platform_suspend_ops bfin_pm_ops = {
-+const struct platform_suspend_ops bfin_pm_ops = {
- .enter = bfin_pm_enter,
- .valid = bfin_pm_valid,
- };
-diff -urNp linux-2.6.37.4/arch/blackfin/mm/maccess.c linux-2.6.37.4/arch/blackfin/mm/maccess.c
---- linux-2.6.37.4/arch/blackfin/mm/maccess.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/blackfin/mm/maccess.c 2011-01-17 02:41:00.000000000 -0500
-@@ -16,7 +16,7 @@ static int validate_memory_access_addres
- return bfin_mem_access_type(addr, size);
- }
-
--long probe_kernel_read(void *dst, void *src, size_t size)
-+long probe_kernel_read(void *dst, const void *src, size_t size)
- {
- unsigned long lsrc = (unsigned long)src;
- int mem_type;
-@@ -55,7 +55,7 @@ long probe_kernel_read(void *dst, void *
- return -EFAULT;
- }
-
--long probe_kernel_write(void *dst, void *src, size_t size)
-+long probe_kernel_write(void *dst, const void *src, size_t size)
- {
- unsigned long ldst = (unsigned long)dst;
- int mem_type;
-diff -urNp linux-2.6.37.4/arch/frv/include/asm/kmap_types.h linux-2.6.37.4/arch/frv/include/asm/kmap_types.h
---- linux-2.6.37.4/arch/frv/include/asm/kmap_types.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/frv/include/asm/kmap_types.h 2011-01-17 02:41:00.000000000 -0500
-@@ -23,6 +23,7 @@ enum km_type {
- KM_IRQ1,
- KM_SOFTIRQ0,
- KM_SOFTIRQ1,
-+ KM_CLEARPAGE,
- KM_TYPE_NR
- };
-
-diff -urNp linux-2.6.37.4/arch/frv/mm/elf-fdpic.c linux-2.6.37.4/arch/frv/mm/elf-fdpic.c
---- linux-2.6.37.4/arch/frv/mm/elf-fdpic.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/frv/mm/elf-fdpic.c 2011-01-17 02:41:00.000000000 -0500
-@@ -73,8 +73,7 @@ unsigned long arch_get_unmapped_area(str
- if (addr) {
- addr = PAGE_ALIGN(addr);
- vma = find_vma(current->mm, addr);
-- if (TASK_SIZE - len >= addr &&
-- (!vma || addr + len <= vma->vm_start))
-+ if (TASK_SIZE - len >= addr && check_heap_stack_gap(vma, addr, len))
- goto success;
- }
-
-@@ -89,7 +88,7 @@ unsigned long arch_get_unmapped_area(str
- for (; vma; vma = vma->vm_next) {
- if (addr > limit)
- break;
-- if (addr + len <= vma->vm_start)
-+ if (check_heap_stack_gap(vma, addr, len))
- goto success;
- addr = vma->vm_end;
- }
-@@ -104,7 +103,7 @@ unsigned long arch_get_unmapped_area(str
- for (; vma; vma = vma->vm_next) {
- if (addr > limit)
- break;
-- if (addr + len <= vma->vm_start)
-+ if (check_heap_stack_gap(vma, addr, len))
- goto success;
- addr = vma->vm_end;
- }
-diff -urNp linux-2.6.37.4/arch/ia64/hp/common/hwsw_iommu.c linux-2.6.37.4/arch/ia64/hp/common/hwsw_iommu.c
---- linux-2.6.37.4/arch/ia64/hp/common/hwsw_iommu.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/ia64/hp/common/hwsw_iommu.c 2011-01-17 02:41:00.000000000 -0500
-@@ -17,7 +17,7 @@
- #include <linux/swiotlb.h>
- #include <asm/machvec.h>
-
--extern struct dma_map_ops sba_dma_ops, swiotlb_dma_ops;
-+extern const struct dma_map_ops sba_dma_ops, swiotlb_dma_ops;
-
- /* swiotlb declarations & definitions: */
- extern int swiotlb_late_init_with_default_size (size_t size);
-@@ -33,7 +33,7 @@ static inline int use_swiotlb(struct dev
- !sba_dma_ops.dma_supported(dev, *dev->dma_mask);
- }
-
--struct dma_map_ops *hwsw_dma_get_ops(struct device *dev)
-+const struct dma_map_ops *hwsw_dma_get_ops(struct device *dev)
- {
- if (use_swiotlb(dev))
- return &swiotlb_dma_ops;
-diff -urNp linux-2.6.37.4/arch/ia64/hp/common/sba_iommu.c linux-2.6.37.4/arch/ia64/hp/common/sba_iommu.c
---- linux-2.6.37.4/arch/ia64/hp/common/sba_iommu.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/ia64/hp/common/sba_iommu.c 2011-01-17 02:41:00.000000000 -0500
-@@ -2097,7 +2097,7 @@ static struct acpi_driver acpi_sba_ioc_d
- },
- };
-
--extern struct dma_map_ops swiotlb_dma_ops;
-+extern const struct dma_map_ops swiotlb_dma_ops;
-
- static int __init
- sba_init(void)
-@@ -2211,7 +2211,7 @@ sba_page_override(char *str)
-
- __setup("sbapagesize=",sba_page_override);
-
--struct dma_map_ops sba_dma_ops = {
-+const struct dma_map_ops sba_dma_ops = {
- .alloc_coherent = sba_alloc_coherent,
- .free_coherent = sba_free_coherent,
- .map_page = sba_map_page,
-diff -urNp linux-2.6.37.4/arch/ia64/include/asm/dma-mapping.h linux-2.6.37.4/arch/ia64/include/asm/dma-mapping.h
---- linux-2.6.37.4/arch/ia64/include/asm/dma-mapping.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/ia64/include/asm/dma-mapping.h 2011-01-17 02:41:00.000000000 -0500
-@@ -12,7 +12,7 @@
-
- #define ARCH_HAS_DMA_GET_REQUIRED_MASK
-
--extern struct dma_map_ops *dma_ops;
-+extern const struct dma_map_ops *dma_ops;
- extern struct ia64_machine_vector ia64_mv;
- extern void set_iommu_machvec(void);
-
-@@ -24,7 +24,7 @@ extern void machvec_dma_sync_sg(struct d
- static inline void *dma_alloc_coherent(struct device *dev, size_t size,
- dma_addr_t *daddr, gfp_t gfp)
- {
-- struct dma_map_ops *ops = platform_dma_get_ops(dev);
-+ const struct dma_map_ops *ops = platform_dma_get_ops(dev);
- void *caddr;
-
- caddr = ops->alloc_coherent(dev, size, daddr, gfp);
-@@ -35,7 +35,7 @@ static inline void *dma_alloc_coherent(s
- static inline void dma_free_coherent(struct device *dev, size_t size,
- void *caddr, dma_addr_t daddr)
- {
-- struct dma_map_ops *ops = platform_dma_get_ops(dev);
-+ const struct dma_map_ops *ops = platform_dma_get_ops(dev);
- debug_dma_free_coherent(dev, size, caddr, daddr);
- ops->free_coherent(dev, size, caddr, daddr);
- }
-@@ -49,13 +49,13 @@ static inline void dma_free_coherent(str
-
- static inline int dma_mapping_error(struct device *dev, dma_addr_t daddr)
- {
-- struct dma_map_ops *ops = platform_dma_get_ops(dev);
-+ const struct dma_map_ops *ops = platform_dma_get_ops(dev);
- return ops->mapping_error(dev, daddr);
- }
-
- static inline int dma_supported(struct device *dev, u64 mask)
- {
-- struct dma_map_ops *ops = platform_dma_get_ops(dev);
-+ const struct dma_map_ops *ops = platform_dma_get_ops(dev);
- return ops->dma_supported(dev, mask);
- }
-
-diff -urNp linux-2.6.37.4/arch/ia64/include/asm/elf.h linux-2.6.37.4/arch/ia64/include/asm/elf.h
---- linux-2.6.37.4/arch/ia64/include/asm/elf.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/ia64/include/asm/elf.h 2011-01-17 02:41:00.000000000 -0500
-@@ -42,6 +42,13 @@
- */
- #define ELF_ET_DYN_BASE (TASK_UNMAPPED_BASE + 0x800000000UL)
-
-+#ifdef CONFIG_PAX_ASLR
-+#define PAX_ELF_ET_DYN_BASE (current->personality == PER_LINUX32 ? 0x08048000UL : 0x4000000000000000UL)
-+
-+#define PAX_DELTA_MMAP_LEN (current->personality == PER_LINUX32 ? 16 : 3*PAGE_SHIFT - 13)
-+#define PAX_DELTA_STACK_LEN (current->personality == PER_LINUX32 ? 16 : 3*PAGE_SHIFT - 13)
-+#endif
-+
- #define PT_IA_64_UNWIND 0x70000001
-
- /* IA-64 relocations: */
-diff -urNp linux-2.6.37.4/arch/ia64/include/asm/machvec.h linux-2.6.37.4/arch/ia64/include/asm/machvec.h
---- linux-2.6.37.4/arch/ia64/include/asm/machvec.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/ia64/include/asm/machvec.h 2011-01-17 02:41:00.000000000 -0500
-@@ -45,7 +45,7 @@ typedef void ia64_mv_kernel_launch_event
- /* DMA-mapping interface: */
- typedef void ia64_mv_dma_init (void);
- typedef u64 ia64_mv_dma_get_required_mask (struct device *);
--typedef struct dma_map_ops *ia64_mv_dma_get_ops(struct device *);
-+typedef const struct dma_map_ops *ia64_mv_dma_get_ops(struct device *);
-
- /*
- * WARNING: The legacy I/O space is _architected_. Platforms are
-@@ -251,7 +251,7 @@ extern void machvec_init_from_cmdline(co
- # endif /* CONFIG_IA64_GENERIC */
-
- extern void swiotlb_dma_init(void);
--extern struct dma_map_ops *dma_get_ops(struct device *);
-+extern const struct dma_map_ops *dma_get_ops(struct device *);
-
- /*
- * Define default versions so we can extend machvec for new platforms without having
-diff -urNp linux-2.6.37.4/arch/ia64/include/asm/pgtable.h linux-2.6.37.4/arch/ia64/include/asm/pgtable.h
---- linux-2.6.37.4/arch/ia64/include/asm/pgtable.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/ia64/include/asm/pgtable.h 2011-01-17 02:41:00.000000000 -0500
-@@ -12,7 +12,7 @@
- * David Mosberger-Tang <davidm@hpl.hp.com>
- */
-
--
-+#include <linux/const.h>
- #include <asm/mman.h>
- #include <asm/page.h>
- #include <asm/processor.h>
-@@ -143,6 +143,17 @@
- #define PAGE_READONLY __pgprot(__ACCESS_BITS | _PAGE_PL_3 | _PAGE_AR_R)
- #define PAGE_COPY __pgprot(__ACCESS_BITS | _PAGE_PL_3 | _PAGE_AR_R)
- #define PAGE_COPY_EXEC __pgprot(__ACCESS_BITS | _PAGE_PL_3 | _PAGE_AR_RX)
-+
-+#ifdef CONFIG_PAX_PAGEEXEC
-+# define PAGE_SHARED_NOEXEC __pgprot(__ACCESS_BITS | _PAGE_PL_3 | _PAGE_AR_RW)
-+# define PAGE_READONLY_NOEXEC __pgprot(__ACCESS_BITS | _PAGE_PL_3 | _PAGE_AR_R)
-+# define PAGE_COPY_NOEXEC __pgprot(__ACCESS_BITS | _PAGE_PL_3 | _PAGE_AR_R)
-+#else
-+# define PAGE_SHARED_NOEXEC PAGE_SHARED
-+# define PAGE_READONLY_NOEXEC PAGE_READONLY
-+# define PAGE_COPY_NOEXEC PAGE_COPY
-+#endif
-+
- #define PAGE_GATE __pgprot(__ACCESS_BITS | _PAGE_PL_0 | _PAGE_AR_X_RX)
- #define PAGE_KERNEL __pgprot(__DIRTY_BITS | _PAGE_PL_0 | _PAGE_AR_RWX)
- #define PAGE_KERNELRX __pgprot(__ACCESS_BITS | _PAGE_PL_0 | _PAGE_AR_RX)
-diff -urNp linux-2.6.37.4/arch/ia64/include/asm/spinlock.h linux-2.6.37.4/arch/ia64/include/asm/spinlock.h
---- linux-2.6.37.4/arch/ia64/include/asm/spinlock.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/ia64/include/asm/spinlock.h 2011-02-27 14:48:10.000000000 -0500
-@@ -72,7 +72,7 @@ static __always_inline void __ticket_spi
- unsigned short *p = (unsigned short *)&lock->lock + 1, tmp;
-
- asm volatile ("ld2.bias %0=[%1]" : "=r"(tmp) : "r"(p));
-- ACCESS_ONCE(*p) = (tmp + 2) & ~1;
-+ ACCESS_ONCE_RW(*p) = (tmp + 2) & ~1;
- }
-
- static __always_inline void __ticket_spin_unlock_wait(arch_spinlock_t *lock)
-diff -urNp linux-2.6.37.4/arch/ia64/include/asm/uaccess.h linux-2.6.37.4/arch/ia64/include/asm/uaccess.h
---- linux-2.6.37.4/arch/ia64/include/asm/uaccess.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/ia64/include/asm/uaccess.h 2011-01-17 02:41:00.000000000 -0500
-@@ -257,7 +257,7 @@ __copy_from_user (void *to, const void _
- const void *__cu_from = (from); \
- long __cu_len = (n); \
- \
-- if (__access_ok(__cu_to, __cu_len, get_fs())) \
-+ if (__cu_len > 0 && __cu_len <= INT_MAX && __access_ok(__cu_to, __cu_len, get_fs())) \
- __cu_len = __copy_user(__cu_to, (__force void __user *) __cu_from, __cu_len); \
- __cu_len; \
- })
-@@ -269,7 +269,7 @@ __copy_from_user (void *to, const void _
- long __cu_len = (n); \
- \
- __chk_user_ptr(__cu_from); \
-- if (__access_ok(__cu_from, __cu_len, get_fs())) \
-+ if (__cu_len > 0 && __cu_len <= INT_MAX && __access_ok(__cu_from, __cu_len, get_fs())) \
- __cu_len = __copy_user((__force void __user *) __cu_to, __cu_from, __cu_len); \
- __cu_len; \
- })
-diff -urNp linux-2.6.37.4/arch/ia64/kernel/dma-mapping.c linux-2.6.37.4/arch/ia64/kernel/dma-mapping.c
---- linux-2.6.37.4/arch/ia64/kernel/dma-mapping.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/ia64/kernel/dma-mapping.c 2011-01-17 02:41:00.000000000 -0500
-@@ -3,7 +3,7 @@
- /* Set this to 1 if there is a HW IOMMU in the system */
- int iommu_detected __read_mostly;
-
--struct dma_map_ops *dma_ops;
-+const struct dma_map_ops *dma_ops;
- EXPORT_SYMBOL(dma_ops);
-
- #define PREALLOC_DMA_DEBUG_ENTRIES (1 << 16)
-@@ -16,7 +16,7 @@ static int __init dma_init(void)
- }
- fs_initcall(dma_init);
-
--struct dma_map_ops *dma_get_ops(struct device *dev)
-+const struct dma_map_ops *dma_get_ops(struct device *dev)
- {
- return dma_ops;
- }
-diff -urNp linux-2.6.37.4/arch/ia64/kernel/module.c linux-2.6.37.4/arch/ia64/kernel/module.c
---- linux-2.6.37.4/arch/ia64/kernel/module.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/ia64/kernel/module.c 2011-01-17 02:41:00.000000000 -0500
-@@ -315,8 +315,7 @@ module_alloc (unsigned long size)
- void
- module_free (struct module *mod, void *module_region)
- {
-- if (mod && mod->arch.init_unw_table &&
-- module_region == mod->module_init) {
-+ if (mod && mod->arch.init_unw_table && module_region == mod->module_init_rx) {
- unw_remove_unwind_table(mod->arch.init_unw_table);
- mod->arch.init_unw_table = NULL;
- }
-@@ -502,15 +501,39 @@ module_frob_arch_sections (Elf_Ehdr *ehd
- }
-
- static inline int
-+in_init_rx (const struct module *mod, uint64_t addr)
-+{
-+ return addr - (uint64_t) mod->module_init_rx < mod->init_size_rx;
-+}
-+
-+static inline int
-+in_init_rw (const struct module *mod, uint64_t addr)
-+{
-+ return addr - (uint64_t) mod->module_init_rw < mod->init_size_rw;
-+}
-+
-+static inline int
- in_init (const struct module *mod, uint64_t addr)
- {
-- return addr - (uint64_t) mod->module_init < mod->init_size;
-+ return in_init_rx(mod, addr) || in_init_rw(mod, addr);
-+}
-+
-+static inline int
-+in_core_rx (const struct module *mod, uint64_t addr)
-+{
-+ return addr - (uint64_t) mod->module_core_rx < mod->core_size_rx;
-+}
-+
-+static inline int
-+in_core_rw (const struct module *mod, uint64_t addr)
-+{
-+ return addr - (uint64_t) mod->module_core_rw < mod->core_size_rw;
- }
-
- static inline int
- in_core (const struct module *mod, uint64_t addr)
- {
-- return addr - (uint64_t) mod->module_core < mod->core_size;
-+ return in_core_rx(mod, addr) || in_core_rw(mod, addr);
- }
-
- static inline int
-@@ -693,7 +716,14 @@ do_reloc (struct module *mod, uint8_t r_
- break;
-
- case RV_BDREL:
-- val -= (uint64_t) (in_init(mod, val) ? mod->module_init : mod->module_core);
-+ if (in_init_rx(mod, val))
-+ val -= (uint64_t) mod->module_init_rx;
-+ else if (in_init_rw(mod, val))
-+ val -= (uint64_t) mod->module_init_rw;
-+ else if (in_core_rx(mod, val))
-+ val -= (uint64_t) mod->module_core_rx;
-+ else if (in_core_rw(mod, val))
-+ val -= (uint64_t) mod->module_core_rw;
- break;
-
- case RV_LTV:
-@@ -828,15 +858,15 @@ apply_relocate_add (Elf64_Shdr *sechdrs,
- * addresses have been selected...
- */
- uint64_t gp;
-- if (mod->core_size > MAX_LTOFF)
-+ if (mod->core_size_rx + mod->core_size_rw > MAX_LTOFF)
- /*
- * This takes advantage of fact that SHF_ARCH_SMALL gets allocated
- * at the end of the module.
- */
-- gp = mod->core_size - MAX_LTOFF / 2;
-+ gp = mod->core_size_rx + mod->core_size_rw - MAX_LTOFF / 2;
- else
-- gp = mod->core_size / 2;
-- gp = (uint64_t) mod->module_core + ((gp + 7) & -8);
-+ gp = (mod->core_size_rx + mod->core_size_rw) / 2;
-+ gp = (uint64_t) mod->module_core_rx + ((gp + 7) & -8);
- mod->arch.gp = gp;
- DEBUGP("%s: placing gp at 0x%lx\n", __func__, gp);
- }
-diff -urNp linux-2.6.37.4/arch/ia64/kernel/pci-dma.c linux-2.6.37.4/arch/ia64/kernel/pci-dma.c
---- linux-2.6.37.4/arch/ia64/kernel/pci-dma.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/ia64/kernel/pci-dma.c 2011-01-17 02:41:00.000000000 -0500
-@@ -43,7 +43,7 @@ struct device fallback_dev = {
- .dma_mask = &fallback_dev.coherent_dma_mask,
- };
-
--extern struct dma_map_ops intel_dma_ops;
-+extern const struct dma_map_ops intel_dma_ops;
-
- static int __init pci_iommu_init(void)
- {
-diff -urNp linux-2.6.37.4/arch/ia64/kernel/pci-swiotlb.c linux-2.6.37.4/arch/ia64/kernel/pci-swiotlb.c
---- linux-2.6.37.4/arch/ia64/kernel/pci-swiotlb.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/ia64/kernel/pci-swiotlb.c 2011-01-17 02:41:00.000000000 -0500
-@@ -22,7 +22,7 @@ static void *ia64_swiotlb_alloc_coherent
- return swiotlb_alloc_coherent(dev, size, dma_handle, gfp);
- }
-
--struct dma_map_ops swiotlb_dma_ops = {
-+const struct dma_map_ops swiotlb_dma_ops = {
- .alloc_coherent = ia64_swiotlb_alloc_coherent,
- .free_coherent = swiotlb_free_coherent,
- .map_page = swiotlb_map_page,
-diff -urNp linux-2.6.37.4/arch/ia64/kernel/sys_ia64.c linux-2.6.37.4/arch/ia64/kernel/sys_ia64.c
---- linux-2.6.37.4/arch/ia64/kernel/sys_ia64.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/ia64/kernel/sys_ia64.c 2011-01-17 02:41:00.000000000 -0500
-@@ -43,6 +43,13 @@ arch_get_unmapped_area (struct file *fil
- if (REGION_NUMBER(addr) == RGN_HPAGE)
- addr = 0;
- #endif
-+
-+#ifdef CONFIG_PAX_RANDMMAP
-+ if (mm->pax_flags & MF_PAX_RANDMMAP)
-+ addr = mm->free_area_cache;
-+ else
-+#endif
-+
- if (!addr)
- addr = mm->free_area_cache;
-
-@@ -61,14 +68,14 @@ arch_get_unmapped_area (struct file *fil
- for (vma = find_vma(mm, addr); ; vma = vma->vm_next) {
- /* At this point: (!vma || addr < vma->vm_end). */
- if (TASK_SIZE - len < addr || RGN_MAP_LIMIT - len < REGION_OFFSET(addr)) {
-- if (start_addr != TASK_UNMAPPED_BASE) {
-+ if (start_addr != mm->mmap_base) {
- /* Start a new search --- just in case we missed some holes. */
-- addr = TASK_UNMAPPED_BASE;
-+ addr = mm->mmap_base;
- goto full_search;
- }
- return -ENOMEM;
- }
-- if (!vma || addr + len <= vma->vm_start) {
-+ if (check_heap_stack_gap(vma, addr, len)) {
- /* Remember the address where we stopped this search: */
- mm->free_area_cache = addr + len;
- return addr;
-diff -urNp linux-2.6.37.4/arch/ia64/kernel/vmlinux.lds.S linux-2.6.37.4/arch/ia64/kernel/vmlinux.lds.S
---- linux-2.6.37.4/arch/ia64/kernel/vmlinux.lds.S 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/ia64/kernel/vmlinux.lds.S 2011-01-17 02:41:00.000000000 -0500
-@@ -199,7 +199,7 @@ SECTIONS {
- /* Per-cpu data: */
- . = ALIGN(PERCPU_PAGE_SIZE);
- PERCPU_VADDR(PERCPU_ADDR, :percpu)
-- __phys_per_cpu_start = __per_cpu_load;
-+ __phys_per_cpu_start = per_cpu_load;
- /*
- * ensure percpu data fits
- * into percpu page size
-diff -urNp linux-2.6.37.4/arch/ia64/mm/fault.c linux-2.6.37.4/arch/ia64/mm/fault.c
---- linux-2.6.37.4/arch/ia64/mm/fault.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/ia64/mm/fault.c 2011-01-17 02:41:00.000000000 -0500
-@@ -72,6 +72,23 @@ mapped_kernel_page_is_present (unsigned
- return pte_present(pte);
- }
-
-+#ifdef CONFIG_PAX_PAGEEXEC
-+void pax_report_insns(void *pc, void *sp)
-+{
-+ unsigned long i;
-+
-+ printk(KERN_ERR "PAX: bytes at PC: ");
-+ for (i = 0; i < 8; i++) {
-+ unsigned int c;
-+ if (get_user(c, (unsigned int *)pc+i))
-+ printk(KERN_CONT "???????? ");
-+ else
-+ printk(KERN_CONT "%08x ", c);
-+ }
-+ printk("\n");
-+}
-+#endif
-+
- void __kprobes
- ia64_do_page_fault (unsigned long address, unsigned long isr, struct pt_regs *regs)
- {
-@@ -145,9 +162,23 @@ ia64_do_page_fault (unsigned long addres
- mask = ( (((isr >> IA64_ISR_X_BIT) & 1UL) << VM_EXEC_BIT)
- | (((isr >> IA64_ISR_W_BIT) & 1UL) << VM_WRITE_BIT));
-
-- if ((vma->vm_flags & mask) != mask)
-+ if ((vma->vm_flags & mask) != mask) {
-+
-+#ifdef CONFIG_PAX_PAGEEXEC
-+ if (!(vma->vm_flags & VM_EXEC) && (mask & VM_EXEC)) {
-+ if (!(mm->pax_flags & MF_PAX_PAGEEXEC) || address != regs->cr_iip)
-+ goto bad_area;
-+
-+ up_read(&mm->mmap_sem);
-+ pax_report_fault(regs, (void *)regs->cr_iip, (void *)regs->r12);
-+ do_group_exit(SIGKILL);
-+ }
-+#endif
-+
- goto bad_area;
-
-+ }
-+
- /*
- * If for any reason at all we couldn't handle the fault, make
- * sure we exit gracefully rather than endlessly redo the
-diff -urNp linux-2.6.37.4/arch/ia64/mm/hugetlbpage.c linux-2.6.37.4/arch/ia64/mm/hugetlbpage.c
---- linux-2.6.37.4/arch/ia64/mm/hugetlbpage.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/ia64/mm/hugetlbpage.c 2011-01-17 02:41:00.000000000 -0500
-@@ -171,7 +171,7 @@ unsigned long hugetlb_get_unmapped_area(
- /* At this point: (!vmm || addr < vmm->vm_end). */
- if (REGION_OFFSET(addr) + len > RGN_MAP_LIMIT)
- return -ENOMEM;
-- if (!vmm || (addr + len) <= vmm->vm_start)
-+ if (check_heap_stack_gap(vmm, addr, len))
- return addr;
- addr = ALIGN(vmm->vm_end, HPAGE_SIZE);
- }
-diff -urNp linux-2.6.37.4/arch/ia64/mm/init.c linux-2.6.37.4/arch/ia64/mm/init.c
---- linux-2.6.37.4/arch/ia64/mm/init.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/ia64/mm/init.c 2011-01-17 02:41:00.000000000 -0500
-@@ -122,6 +122,19 @@ ia64_init_addr_space (void)
- vma->vm_start = current->thread.rbs_bot & PAGE_MASK;
- vma->vm_end = vma->vm_start + PAGE_SIZE;
- vma->vm_flags = VM_DATA_DEFAULT_FLAGS|VM_GROWSUP|VM_ACCOUNT;
-+
-+#ifdef CONFIG_PAX_PAGEEXEC
-+ if (current->mm->pax_flags & MF_PAX_PAGEEXEC) {
-+ vma->vm_flags &= ~VM_EXEC;
-+
-+#ifdef CONFIG_PAX_MPROTECT
-+ if (current->mm->pax_flags & MF_PAX_MPROTECT)
-+ vma->vm_flags &= ~VM_MAYEXEC;
-+#endif
-+
-+ }
-+#endif
-+
- vma->vm_page_prot = vm_get_page_prot(vma->vm_flags);
- down_write(&current->mm->mmap_sem);
- if (insert_vm_struct(current->mm, vma)) {
-diff -urNp linux-2.6.37.4/arch/ia64/sn/pci/pci_dma.c linux-2.6.37.4/arch/ia64/sn/pci/pci_dma.c
---- linux-2.6.37.4/arch/ia64/sn/pci/pci_dma.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/ia64/sn/pci/pci_dma.c 2011-01-17 02:41:00.000000000 -0500
-@@ -465,7 +465,7 @@ int sn_pci_legacy_write(struct pci_bus *
- return ret;
- }
-
--static struct dma_map_ops sn_dma_ops = {
-+static const struct dma_map_ops sn_dma_ops = {
- .alloc_coherent = sn_dma_alloc_coherent,
- .free_coherent = sn_dma_free_coherent,
- .map_page = sn_dma_map_page,
-diff -urNp linux-2.6.37.4/arch/m32r/lib/usercopy.c linux-2.6.37.4/arch/m32r/lib/usercopy.c
---- linux-2.6.37.4/arch/m32r/lib/usercopy.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/m32r/lib/usercopy.c 2011-01-17 02:41:00.000000000 -0500
-@@ -14,6 +14,9 @@
- unsigned long
- __generic_copy_to_user(void __user *to, const void *from, unsigned long n)
- {
-+ if ((long)n < 0)
-+ return n;
-+
- prefetch(from);
- if (access_ok(VERIFY_WRITE, to, n))
- __copy_user(to,from,n);
-@@ -23,6 +26,9 @@ __generic_copy_to_user(void __user *to,
- unsigned long
- __generic_copy_from_user(void *to, const void __user *from, unsigned long n)
- {
-+ if ((long)n < 0)
-+ return n;
-+
- prefetchw(to);
- if (access_ok(VERIFY_READ, from, n))
- __copy_user_zeroing(to,from,n);
-diff -urNp linux-2.6.37.4/arch/microblaze/include/asm/device.h linux-2.6.37.4/arch/microblaze/include/asm/device.h
---- linux-2.6.37.4/arch/microblaze/include/asm/device.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/microblaze/include/asm/device.h 2011-01-17 02:41:00.000000000 -0500
-@@ -13,7 +13,7 @@ struct device_node;
-
- struct dev_archdata {
- /* DMA operations on that device */
-- struct dma_map_ops *dma_ops;
-+ const struct dma_map_ops *dma_ops;
- void *dma_data;
- };
-
-diff -urNp linux-2.6.37.4/arch/microblaze/include/asm/dma-mapping.h linux-2.6.37.4/arch/microblaze/include/asm/dma-mapping.h
---- linux-2.6.37.4/arch/microblaze/include/asm/dma-mapping.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/microblaze/include/asm/dma-mapping.h 2011-01-17 02:41:00.000000000 -0500
-@@ -43,14 +43,14 @@ static inline unsigned long device_to_ma
- return 0xfffffffful;
- }
-
--extern struct dma_map_ops *dma_ops;
-+extern const struct dma_map_ops *dma_ops;
-
- /*
- * Available generic sets of operations
- */
--extern struct dma_map_ops dma_direct_ops;
-+extern const struct dma_map_ops dma_direct_ops;
-
--static inline struct dma_map_ops *get_dma_ops(struct device *dev)
-+static inline const struct dma_map_ops *get_dma_ops(struct device *dev)
- {
- /* We don't handle the NULL dev case for ISA for now. We could
- * do it via an out of line call but it is not needed for now. The
-@@ -63,14 +63,14 @@ static inline struct dma_map_ops *get_dm
- return dev->archdata.dma_ops;
- }
-
--static inline void set_dma_ops(struct device *dev, struct dma_map_ops *ops)
-+static inline void set_dma_ops(struct device *dev, const struct dma_map_ops *ops)
- {
- dev->archdata.dma_ops = ops;
- }
-
- static inline int dma_supported(struct device *dev, u64 mask)
- {
-- struct dma_map_ops *ops = get_dma_ops(dev);
-+ const struct dma_map_ops *ops = get_dma_ops(dev);
-
- if (unlikely(!ops))
- return 0;
-@@ -81,7 +81,7 @@ static inline int dma_supported(struct d
-
- static inline int dma_set_mask(struct device *dev, u64 dma_mask)
- {
-- struct dma_map_ops *ops = get_dma_ops(dev);
-+ const struct dma_map_ops *ops = get_dma_ops(dev);
-
- if (unlikely(ops == NULL))
- return -EIO;
-@@ -97,7 +97,7 @@ static inline int dma_set_mask(struct de
-
- static inline int dma_mapping_error(struct device *dev, dma_addr_t dma_addr)
- {
-- struct dma_map_ops *ops = get_dma_ops(dev);
-+ const struct dma_map_ops *ops = get_dma_ops(dev);
- if (ops->mapping_error)
- return ops->mapping_error(dev, dma_addr);
-
-@@ -110,7 +110,7 @@ static inline int dma_mapping_error(stru
- static inline void *dma_alloc_coherent(struct device *dev, size_t size,
- dma_addr_t *dma_handle, gfp_t flag)
- {
-- struct dma_map_ops *ops = get_dma_ops(dev);
-+ const struct dma_map_ops *ops = get_dma_ops(dev);
- void *memory;
-
- BUG_ON(!ops);
-@@ -124,7 +124,7 @@ static inline void *dma_alloc_coherent(s
- static inline void dma_free_coherent(struct device *dev, size_t size,
- void *cpu_addr, dma_addr_t dma_handle)
- {
-- struct dma_map_ops *ops = get_dma_ops(dev);
-+ const struct dma_map_ops *ops = get_dma_ops(dev);
-
- BUG_ON(!ops);
- debug_dma_free_coherent(dev, size, cpu_addr, dma_handle);
-diff -urNp linux-2.6.37.4/arch/microblaze/include/asm/pci.h linux-2.6.37.4/arch/microblaze/include/asm/pci.h
---- linux-2.6.37.4/arch/microblaze/include/asm/pci.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/microblaze/include/asm/pci.h 2011-01-17 02:41:00.000000000 -0500
-@@ -54,8 +54,8 @@ static inline void pcibios_penalize_isa_
- }
-
- #ifdef CONFIG_PCI
--extern void set_pci_dma_ops(struct dma_map_ops *dma_ops);
--extern struct dma_map_ops *get_pci_dma_ops(void);
-+extern void set_pci_dma_ops(const struct dma_map_ops *dma_ops);
-+extern const struct dma_map_ops *get_pci_dma_ops(void);
- #else /* CONFIG_PCI */
- #define set_pci_dma_ops(d)
- #define get_pci_dma_ops() NULL
-diff -urNp linux-2.6.37.4/arch/microblaze/kernel/dma.c linux-2.6.37.4/arch/microblaze/kernel/dma.c
---- linux-2.6.37.4/arch/microblaze/kernel/dma.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/microblaze/kernel/dma.c 2011-01-17 02:41:00.000000000 -0500
-@@ -133,7 +133,7 @@ static inline void dma_direct_unmap_page
- __dma_sync_page(dma_address, 0 , size, direction);
- }
-
--struct dma_map_ops dma_direct_ops = {
-+const struct dma_map_ops dma_direct_ops = {
- .alloc_coherent = dma_direct_alloc_coherent,
- .free_coherent = dma_direct_free_coherent,
- .map_sg = dma_direct_map_sg,
-diff -urNp linux-2.6.37.4/arch/microblaze/kernel/kgdb.c linux-2.6.37.4/arch/microblaze/kernel/kgdb.c
---- linux-2.6.37.4/arch/microblaze/kernel/kgdb.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/microblaze/kernel/kgdb.c 2011-01-17 02:41:00.000000000 -0500
-@@ -141,10 +141,11 @@ void kgdb_arch_exit(void)
- /*
- * Global data
- */
--struct kgdb_arch arch_kgdb_ops = {
-+const struct kgdb_arch arch_kgdb_ops = {
- #ifdef __MICROBLAZEEL__
- .gdb_bpt_instr = {0x18, 0x00, 0x0c, 0xba}, /* brki r16, 0x18 */
- #else
-+>>>>>>> master
- .gdb_bpt_instr = {0xba, 0x0c, 0x00, 0x18}, /* brki r16, 0x18 */
- #endif
- };
-diff -urNp linux-2.6.37.4/arch/microblaze/pci/pci-common.c linux-2.6.37.4/arch/microblaze/pci/pci-common.c
---- linux-2.6.37.4/arch/microblaze/pci/pci-common.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/microblaze/pci/pci-common.c 2011-01-17 02:41:00.000000000 -0500
-@@ -47,14 +47,14 @@ resource_size_t isa_mem_base;
- /* Default PCI flags is 0 on ppc32, modified at boot on ppc64 */
- unsigned int pci_flags;
-
--static struct dma_map_ops *pci_dma_ops = &dma_direct_ops;
-+static const struct dma_map_ops *pci_dma_ops = &dma_direct_ops;
-
--void set_pci_dma_ops(struct dma_map_ops *dma_ops)
-+void set_pci_dma_ops(const struct dma_map_ops *dma_ops)
- {
- pci_dma_ops = dma_ops;
- }
-
--struct dma_map_ops *get_pci_dma_ops(void)
-+const struct dma_map_ops *get_pci_dma_ops(void)
- {
- return pci_dma_ops;
- }
-diff -urNp linux-2.6.37.4/arch/mips/alchemy/devboards/pm.c linux-2.6.37.4/arch/mips/alchemy/devboards/pm.c
---- linux-2.6.37.4/arch/mips/alchemy/devboards/pm.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/mips/alchemy/devboards/pm.c 2011-01-17 02:41:00.000000000 -0500
-@@ -110,7 +110,7 @@ static void db1x_pm_end(void)
-
- }
-
--static struct platform_suspend_ops db1x_pm_ops = {
-+static const struct platform_suspend_ops db1x_pm_ops = {
- .valid = suspend_valid_only_mem,
- .begin = db1x_pm_begin,
- .enter = db1x_pm_enter,
-diff -urNp linux-2.6.37.4/arch/mips/cavium-octeon/dma-octeon.c linux-2.6.37.4/arch/mips/cavium-octeon/dma-octeon.c
---- linux-2.6.37.4/arch/mips/cavium-octeon/dma-octeon.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/mips/cavium-octeon/dma-octeon.c 2011-01-17 02:41:00.000000000 -0500
-@@ -202,7 +202,7 @@ static phys_addr_t octeon_unity_dma_to_p
- }
-
- struct octeon_dma_map_ops {
-- struct dma_map_ops dma_map_ops;
-+ const struct dma_map_ops dma_map_ops;
- dma_addr_t (*phys_to_dma)(struct device *dev, phys_addr_t paddr);
- phys_addr_t (*dma_to_phys)(struct device *dev, dma_addr_t daddr);
- };
-@@ -324,7 +324,7 @@ static struct octeon_dma_map_ops _octeon
- },
- };
-
--struct dma_map_ops *octeon_pci_dma_map_ops;
-+const struct dma_map_ops *octeon_pci_dma_map_ops;
-
- void __init octeon_pci_dma_init(void)
- {
-diff -urNp linux-2.6.37.4/arch/mips/include/asm/device.h linux-2.6.37.4/arch/mips/include/asm/device.h
---- linux-2.6.37.4/arch/mips/include/asm/device.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/mips/include/asm/device.h 2011-01-17 02:41:00.000000000 -0500
-@@ -10,7 +10,7 @@ struct dma_map_ops;
-
- struct dev_archdata {
- /* DMA operations on that device */
-- struct dma_map_ops *dma_ops;
-+ const struct dma_map_ops *dma_ops;
- };
-
- struct pdev_archdata {
-diff -urNp linux-2.6.37.4/arch/mips/include/asm/dma-mapping.h linux-2.6.37.4/arch/mips/include/asm/dma-mapping.h
---- linux-2.6.37.4/arch/mips/include/asm/dma-mapping.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/mips/include/asm/dma-mapping.h 2011-01-17 02:41:00.000000000 -0500
-@@ -7,9 +7,9 @@
-
- #include <dma-coherence.h>
-
--extern struct dma_map_ops *mips_dma_map_ops;
-+extern const struct dma_map_ops *mips_dma_map_ops;
-
--static inline struct dma_map_ops *get_dma_ops(struct device *dev)
-+static inline const struct dma_map_ops *get_dma_ops(struct device *dev)
- {
- if (dev && dev->archdata.dma_ops)
- return dev->archdata.dma_ops;
-@@ -31,13 +31,13 @@ static inline void dma_mark_clean(void *
-
- static inline int dma_supported(struct device *dev, u64 mask)
- {
-- struct dma_map_ops *ops = get_dma_ops(dev);
-+ const struct dma_map_ops *ops = get_dma_ops(dev);
- return ops->dma_supported(dev, mask);
- }
-
- static inline int dma_mapping_error(struct device *dev, u64 mask)
- {
-- struct dma_map_ops *ops = get_dma_ops(dev);
-+ const struct dma_map_ops *ops = get_dma_ops(dev);
- return ops->mapping_error(dev, mask);
- }
-
-@@ -59,7 +59,7 @@ static inline void *dma_alloc_coherent(s
- dma_addr_t *dma_handle, gfp_t gfp)
- {
- void *ret;
-- struct dma_map_ops *ops = get_dma_ops(dev);
-+ const struct dma_map_ops *ops = get_dma_ops(dev);
-
- ret = ops->alloc_coherent(dev, size, dma_handle, gfp);
-
-@@ -71,7 +71,7 @@ static inline void *dma_alloc_coherent(s
- static inline void dma_free_coherent(struct device *dev, size_t size,
- void *vaddr, dma_addr_t dma_handle)
- {
-- struct dma_map_ops *ops = get_dma_ops(dev);
-+ const struct dma_map_ops *ops = get_dma_ops(dev);
-
- ops->free_coherent(dev, size, vaddr, dma_handle);
-
-diff -urNp linux-2.6.37.4/arch/mips/include/asm/elf.h linux-2.6.37.4/arch/mips/include/asm/elf.h
---- linux-2.6.37.4/arch/mips/include/asm/elf.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/mips/include/asm/elf.h 2011-01-17 02:41:00.000000000 -0500
-@@ -372,13 +372,16 @@ extern const char *__elf_platform;
- #define ELF_ET_DYN_BASE (TASK_SIZE / 3 * 2)
- #endif
-
-+#ifdef CONFIG_PAX_ASLR
-+#define PAX_ELF_ET_DYN_BASE (test_thread_flag(TIF_32BIT_ADDR) ? 0x00400000UL : 0x00400000UL)
-+
-+#define PAX_DELTA_MMAP_LEN (test_thread_flag(TIF_32BIT_ADDR) ? 27-PAGE_SHIFT : 36-PAGE_SHIFT)
-+#define PAX_DELTA_STACK_LEN (test_thread_flag(TIF_32BIT_ADDR) ? 27-PAGE_SHIFT : 36-PAGE_SHIFT)
-+#endif
-+
- #define ARCH_HAS_SETUP_ADDITIONAL_PAGES 1
- struct linux_binprm;
- extern int arch_setup_additional_pages(struct linux_binprm *bprm,
- int uses_interp);
-
--struct mm_struct;
--extern unsigned long arch_randomize_brk(struct mm_struct *mm);
--#define arch_randomize_brk arch_randomize_brk
--
- #endif /* _ASM_ELF_H */
-diff -urNp linux-2.6.37.4/arch/mips/include/asm/mach-cavium-octeon/dma-coherence.h linux-2.6.37.4/arch/mips/include/asm/mach-cavium-octeon/dma-coherence.h
---- linux-2.6.37.4/arch/mips/include/asm/mach-cavium-octeon/dma-coherence.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/mips/include/asm/mach-cavium-octeon/dma-coherence.h 2011-01-17 02:41:00.000000000 -0500
-@@ -66,7 +66,7 @@ dma_addr_t phys_to_dma(struct device *de
- phys_addr_t dma_to_phys(struct device *dev, dma_addr_t daddr);
-
- struct dma_map_ops;
--extern struct dma_map_ops *octeon_pci_dma_map_ops;
-+extern const struct dma_map_ops *octeon_pci_dma_map_ops;
- extern char *octeon_swiotlb;
-
- #endif /* __ASM_MACH_CAVIUM_OCTEON_DMA_COHERENCE_H */
-diff -urNp linux-2.6.37.4/arch/mips/include/asm/page.h linux-2.6.37.4/arch/mips/include/asm/page.h
---- linux-2.6.37.4/arch/mips/include/asm/page.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/mips/include/asm/page.h 2011-01-17 02:41:00.000000000 -0500
-@@ -93,7 +93,7 @@ extern void copy_user_highpage(struct pa
- #ifdef CONFIG_CPU_MIPS32
- typedef struct { unsigned long pte_low, pte_high; } pte_t;
- #define pte_val(x) ((x).pte_low | ((unsigned long long)(x).pte_high << 32))
-- #define __pte(x) ({ pte_t __pte = {(x), ((unsigned long long)(x)) >> 32}; __pte; })
-+ #define __pte(x) ({ pte_t __pte = {(x), (x) >> 32}; __pte; })
- #else
- typedef struct { unsigned long long pte; } pte_t;
- #define pte_val(x) ((x).pte)
-diff -urNp linux-2.6.37.4/arch/mips/include/asm/system.h linux-2.6.37.4/arch/mips/include/asm/system.h
---- linux-2.6.37.4/arch/mips/include/asm/system.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/mips/include/asm/system.h 2011-01-17 02:41:00.000000000 -0500
-@@ -23,6 +23,7 @@
- #include <asm/dsp.h>
- #include <asm/watch.h>
- #include <asm/war.h>
-+#include <asm/asm.h>
-
-
- /*
-@@ -230,6 +231,6 @@ extern void per_cpu_trap_init(void);
- */
- #define __ARCH_WANT_UNLOCKED_CTXSW
-
--extern unsigned long arch_align_stack(unsigned long sp);
-+#define arch_align_stack(x) ((x) & ALMASK)
-
- #endif /* _ASM_SYSTEM_H */
-diff -urNp linux-2.6.37.4/arch/mips/jz4740/pm.c linux-2.6.37.4/arch/mips/jz4740/pm.c
---- linux-2.6.37.4/arch/mips/jz4740/pm.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/mips/jz4740/pm.c 2011-01-17 02:41:00.000000000 -0500
-@@ -42,7 +42,7 @@ static int jz4740_pm_enter(suspend_state
- return 0;
- }
-
--static struct platform_suspend_ops jz4740_pm_ops = {
-+static const struct platform_suspend_ops jz4740_pm_ops = {
- .valid = suspend_valid_only_mem,
- .enter = jz4740_pm_enter,
- };
-diff -urNp linux-2.6.37.4/arch/mips/kernel/binfmt_elfn32.c linux-2.6.37.4/arch/mips/kernel/binfmt_elfn32.c
---- linux-2.6.37.4/arch/mips/kernel/binfmt_elfn32.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/mips/kernel/binfmt_elfn32.c 2011-01-17 02:41:00.000000000 -0500
-@@ -50,6 +50,13 @@ typedef elf_fpreg_t elf_fpregset_t[ELF_N
- #undef ELF_ET_DYN_BASE
- #define ELF_ET_DYN_BASE (TASK32_SIZE / 3 * 2)
-
-+#ifdef CONFIG_PAX_ASLR
-+#define PAX_ELF_ET_DYN_BASE (test_thread_flag(TIF_32BIT_ADDR) ? 0x00400000UL : 0x00400000UL)
-+
-+#define PAX_DELTA_MMAP_LEN (test_thread_flag(TIF_32BIT_ADDR) ? 27-PAGE_SHIFT : 36-PAGE_SHIFT)
-+#define PAX_DELTA_STACK_LEN (test_thread_flag(TIF_32BIT_ADDR) ? 27-PAGE_SHIFT : 36-PAGE_SHIFT)
-+#endif
-+
- #include <asm/processor.h>
- #include <linux/module.h>
- #include <linux/elfcore.h>
-diff -urNp linux-2.6.37.4/arch/mips/kernel/binfmt_elfo32.c linux-2.6.37.4/arch/mips/kernel/binfmt_elfo32.c
---- linux-2.6.37.4/arch/mips/kernel/binfmt_elfo32.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/mips/kernel/binfmt_elfo32.c 2011-01-17 02:41:00.000000000 -0500
-@@ -52,6 +52,13 @@ typedef elf_fpreg_t elf_fpregset_t[ELF_N
- #undef ELF_ET_DYN_BASE
- #define ELF_ET_DYN_BASE (TASK32_SIZE / 3 * 2)
-
-+#ifdef CONFIG_PAX_ASLR
-+#define PAX_ELF_ET_DYN_BASE (test_thread_flag(TIF_32BIT_ADDR) ? 0x00400000UL : 0x00400000UL)
-+
-+#define PAX_DELTA_MMAP_LEN (test_thread_flag(TIF_32BIT_ADDR) ? 27-PAGE_SHIFT : 36-PAGE_SHIFT)
-+#define PAX_DELTA_STACK_LEN (test_thread_flag(TIF_32BIT_ADDR) ? 27-PAGE_SHIFT : 36-PAGE_SHIFT)
-+#endif
-+
- #include <asm/processor.h>
-
- /*
-diff -urNp linux-2.6.37.4/arch/mips/kernel/kgdb.c linux-2.6.37.4/arch/mips/kernel/kgdb.c
---- linux-2.6.37.4/arch/mips/kernel/kgdb.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/mips/kernel/kgdb.c 2011-01-17 02:41:00.000000000 -0500
-@@ -351,6 +351,7 @@ int kgdb_arch_handle_exception(int vecto
- return -1;
- }
-
-+/* cannot be const, see kgdb_arch_init */
- struct kgdb_arch arch_kgdb_ops;
-
- /*
-diff -urNp linux-2.6.37.4/arch/mips/kernel/process.c linux-2.6.37.4/arch/mips/kernel/process.c
---- linux-2.6.37.4/arch/mips/kernel/process.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/mips/kernel/process.c 2011-01-17 02:41:00.000000000 -0500
-@@ -473,15 +473,3 @@ unsigned long get_wchan(struct task_stru
- out:
- return pc;
- }
--
--/*
-- * Don't forget that the stack pointer must be aligned on a 8 bytes
-- * boundary for 32-bits ABI and 16 bytes for 64-bits ABI.
-- */
--unsigned long arch_align_stack(unsigned long sp)
--{
-- if (!(current->personality & ADDR_NO_RANDOMIZE) && randomize_va_space)
-- sp -= get_random_int() & ~PAGE_MASK;
--
-- return sp & ALMASK;
--}
-diff -urNp linux-2.6.37.4/arch/mips/kernel/syscall.c linux-2.6.37.4/arch/mips/kernel/syscall.c
---- linux-2.6.37.4/arch/mips/kernel/syscall.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/mips/kernel/syscall.c 2011-01-17 02:41:00.000000000 -0500
-@@ -108,14 +108,18 @@ unsigned long arch_get_unmapped_area(str
- do_color_align = 0;
- if (filp || (flags & MAP_SHARED))
- do_color_align = 1;
-+
-+#ifdef CONFIG_PAX_RANDMMAP
-+ if (!(current->mm->pax_flags & MF_PAX_RANDMMAP))
-+#endif
-+
- if (addr) {
- if (do_color_align)
- addr = COLOUR_ALIGN(addr, pgoff);
- else
- addr = PAGE_ALIGN(addr);
- vmm = find_vma(current->mm, addr);
-- if (task_size - len >= addr &&
-- (!vmm || addr + len <= vmm->vm_start))
-+ if (task_size - len >= addr && check_heap_stack_gap(vmm, addr, len))
- return addr;
- }
- addr = current->mm->mmap_base;
-@@ -128,7 +132,7 @@ unsigned long arch_get_unmapped_area(str
- /* At this point: (!vmm || addr < vmm->vm_end). */
- if (task_size - len < addr)
- return -ENOMEM;
-- if (!vmm || addr + len <= vmm->vm_start)
-+ if (check_heap_stack_gap(vmm, addr, len))
- return addr;
- addr = vmm->vm_end;
- if (do_color_align)
-@@ -168,19 +172,6 @@ static inline unsigned long brk_rnd(void
- return rnd;
- }
-
--unsigned long arch_randomize_brk(struct mm_struct *mm)
--{
-- unsigned long base = mm->brk;
-- unsigned long ret;
--
-- ret = PAGE_ALIGN(base + brk_rnd());
--
-- if (ret < mm->brk)
-- return mm->brk;
--
-- return ret;
--}
--
- SYSCALL_DEFINE6(mips_mmap, unsigned long, addr, unsigned long, len,
- unsigned long, prot, unsigned long, flags, unsigned long,
- fd, off_t, offset)
-diff -urNp linux-2.6.37.4/arch/mips/loongson/common/pm.c linux-2.6.37.4/arch/mips/loongson/common/pm.c
---- linux-2.6.37.4/arch/mips/loongson/common/pm.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/mips/loongson/common/pm.c 2011-01-17 02:41:00.000000000 -0500
-@@ -147,7 +147,7 @@ static int loongson_pm_valid_state(suspe
- }
- }
-
--static struct platform_suspend_ops loongson_pm_ops = {
-+static const struct platform_suspend_ops loongson_pm_ops = {
- .valid = loongson_pm_valid_state,
- .enter = loongson_pm_enter,
- };
-diff -urNp linux-2.6.37.4/arch/mips/mm/dma-default.c linux-2.6.37.4/arch/mips/mm/dma-default.c
---- linux-2.6.37.4/arch/mips/mm/dma-default.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/mips/mm/dma-default.c 2011-01-17 02:41:00.000000000 -0500
-@@ -300,7 +300,7 @@ void dma_cache_sync(struct device *dev,
-
- EXPORT_SYMBOL(dma_cache_sync);
-
--static struct dma_map_ops mips_default_dma_map_ops = {
-+static const struct dma_map_ops mips_default_dma_map_ops = {
- .alloc_coherent = mips_dma_alloc_coherent,
- .free_coherent = mips_dma_free_coherent,
- .map_page = mips_dma_map_page,
-@@ -315,7 +315,7 @@ static struct dma_map_ops mips_default_d
- .dma_supported = mips_dma_supported
- };
-
--struct dma_map_ops *mips_dma_map_ops = &mips_default_dma_map_ops;
-+const struct dma_map_ops *mips_dma_map_ops = &mips_default_dma_map_ops;
- EXPORT_SYMBOL(mips_dma_map_ops);
-
- #define PREALLOC_DMA_DEBUG_ENTRIES (1 << 16)
-diff -urNp linux-2.6.37.4/arch/mips/mm/fault.c linux-2.6.37.4/arch/mips/mm/fault.c
---- linux-2.6.37.4/arch/mips/mm/fault.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/mips/mm/fault.c 2011-01-17 02:41:00.000000000 -0500
-@@ -28,6 +28,23 @@
- #include <asm/highmem.h> /* For VMALLOC_END */
- #include <linux/kdebug.h>
-
-+#ifdef CONFIG_PAX_PAGEEXEC
-+void pax_report_insns(void *pc, void *sp)
-+{
-+ unsigned long i;
-+
-+ printk(KERN_ERR "PAX: bytes at PC: ");
-+ for (i = 0; i < 5; i++) {
-+ unsigned int c;
-+ if (get_user(c, (unsigned int *)pc+i))
-+ printk(KERN_CONT "???????? ");
-+ else
-+ printk(KERN_CONT "%08x ", c);
-+ }
-+ printk("\n");
-+}
-+#endif
-+
- /*
- * This routine handles page faults. It determines the address,
- * and the problem, and then passes it off to one of the appropriate
-diff -urNp linux-2.6.37.4/arch/parisc/include/asm/elf.h linux-2.6.37.4/arch/parisc/include/asm/elf.h
---- linux-2.6.37.4/arch/parisc/include/asm/elf.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/parisc/include/asm/elf.h 2011-01-17 02:41:00.000000000 -0500
-@@ -342,6 +342,13 @@ struct pt_regs; /* forward declaration..
-
- #define ELF_ET_DYN_BASE (TASK_UNMAPPED_BASE + 0x01000000)
-
-+#ifdef CONFIG_PAX_ASLR
-+#define PAX_ELF_ET_DYN_BASE 0x10000UL
-+
-+#define PAX_DELTA_MMAP_LEN 16
-+#define PAX_DELTA_STACK_LEN 16
-+#endif
-+
- /* This yields a mask that user programs can use to figure out what
- instruction set this CPU supports. This could be done in user space,
- but it's not easy, and we've already done it here. */
-diff -urNp linux-2.6.37.4/arch/parisc/include/asm/pgtable.h linux-2.6.37.4/arch/parisc/include/asm/pgtable.h
---- linux-2.6.37.4/arch/parisc/include/asm/pgtable.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/parisc/include/asm/pgtable.h 2011-01-17 02:41:00.000000000 -0500
-@@ -207,6 +207,17 @@
- #define PAGE_EXECREAD __pgprot(_PAGE_PRESENT | _PAGE_USER | _PAGE_READ | _PAGE_EXEC |_PAGE_ACCESSED)
- #define PAGE_COPY PAGE_EXECREAD
- #define PAGE_RWX __pgprot(_PAGE_PRESENT | _PAGE_USER | _PAGE_READ | _PAGE_WRITE | _PAGE_EXEC |_PAGE_ACCESSED)
-+
-+#ifdef CONFIG_PAX_PAGEEXEC
-+# define PAGE_SHARED_NOEXEC __pgprot(_PAGE_PRESENT | _PAGE_USER | _PAGE_READ | _PAGE_WRITE | _PAGE_ACCESSED)
-+# define PAGE_COPY_NOEXEC __pgprot(_PAGE_PRESENT | _PAGE_USER | _PAGE_READ | _PAGE_ACCESSED)
-+# define PAGE_READONLY_NOEXEC __pgprot(_PAGE_PRESENT | _PAGE_USER | _PAGE_READ | _PAGE_ACCESSED)
-+#else
-+# define PAGE_SHARED_NOEXEC PAGE_SHARED
-+# define PAGE_COPY_NOEXEC PAGE_COPY
-+# define PAGE_READONLY_NOEXEC PAGE_READONLY
-+#endif
-+
- #define PAGE_KERNEL __pgprot(_PAGE_KERNEL)
- #define PAGE_KERNEL_RO __pgprot(_PAGE_KERNEL & ~_PAGE_WRITE)
- #define PAGE_KERNEL_UNC __pgprot(_PAGE_KERNEL | _PAGE_NO_CACHE)
-diff -urNp linux-2.6.37.4/arch/parisc/kernel/module.c linux-2.6.37.4/arch/parisc/kernel/module.c
---- linux-2.6.37.4/arch/parisc/kernel/module.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/parisc/kernel/module.c 2011-01-17 02:41:00.000000000 -0500
-@@ -96,16 +96,38 @@
-
- /* three functions to determine where in the module core
- * or init pieces the location is */
-+static inline int in_init_rx(struct module *me, void *loc)
-+{
-+ return (loc >= me->module_init_rx &&
-+ loc < (me->module_init_rx + me->init_size_rx));
-+}
-+
-+static inline int in_init_rw(struct module *me, void *loc)
-+{
-+ return (loc >= me->module_init_rw &&
-+ loc < (me->module_init_rw + me->init_size_rw));
-+}
-+
- static inline int in_init(struct module *me, void *loc)
- {
-- return (loc >= me->module_init &&
-- loc <= (me->module_init + me->init_size));
-+ return in_init_rx(me, loc) || in_init_rw(me, loc);
-+}
-+
-+static inline int in_core_rx(struct module *me, void *loc)
-+{
-+ return (loc >= me->module_core_rx &&
-+ loc < (me->module_core_rx + me->core_size_rx));
-+}
-+
-+static inline int in_core_rw(struct module *me, void *loc)
-+{
-+ return (loc >= me->module_core_rw &&
-+ loc < (me->module_core_rw + me->core_size_rw));
- }
-
- static inline int in_core(struct module *me, void *loc)
- {
-- return (loc >= me->module_core &&
-- loc <= (me->module_core + me->core_size));
-+ return in_core_rx(me, loc) || in_core_rw(me, loc);
- }
-
- static inline int in_local(struct module *me, void *loc)
-@@ -365,13 +387,13 @@ int module_frob_arch_sections(CONST Elf_
- }
-
- /* align things a bit */
-- me->core_size = ALIGN(me->core_size, 16);
-- me->arch.got_offset = me->core_size;
-- me->core_size += gots * sizeof(struct got_entry);
--
-- me->core_size = ALIGN(me->core_size, 16);
-- me->arch.fdesc_offset = me->core_size;
-- me->core_size += fdescs * sizeof(Elf_Fdesc);
-+ me->core_size_rw = ALIGN(me->core_size_rw, 16);
-+ me->arch.got_offset = me->core_size_rw;
-+ me->core_size_rw += gots * sizeof(struct got_entry);
-+
-+ me->core_size_rw = ALIGN(me->core_size_rw, 16);
-+ me->arch.fdesc_offset = me->core_size_rw;
-+ me->core_size_rw += fdescs * sizeof(Elf_Fdesc);
-
- me->arch.got_max = gots;
- me->arch.fdesc_max = fdescs;
-@@ -389,7 +411,7 @@ static Elf64_Word get_got(struct module
-
- BUG_ON(value == 0);
-
-- got = me->module_core + me->arch.got_offset;
-+ got = me->module_core_rw + me->arch.got_offset;
- for (i = 0; got[i].addr; i++)
- if (got[i].addr == value)
- goto out;
-@@ -407,7 +429,7 @@ static Elf64_Word get_got(struct module
- #ifdef CONFIG_64BIT
- static Elf_Addr get_fdesc(struct module *me, unsigned long value)
- {
-- Elf_Fdesc *fdesc = me->module_core + me->arch.fdesc_offset;
-+ Elf_Fdesc *fdesc = me->module_core_rw + me->arch.fdesc_offset;
-
- if (!value) {
- printk(KERN_ERR "%s: zero OPD requested!\n", me->name);
-@@ -425,7 +447,7 @@ static Elf_Addr get_fdesc(struct module
-
- /* Create new one */
- fdesc->addr = value;
-- fdesc->gp = (Elf_Addr)me->module_core + me->arch.got_offset;
-+ fdesc->gp = (Elf_Addr)me->module_core_rw + me->arch.got_offset;
- return (Elf_Addr)fdesc;
- }
- #endif /* CONFIG_64BIT */
-@@ -849,7 +871,7 @@ register_unwind_table(struct module *me,
-
- table = (unsigned char *)sechdrs[me->arch.unwind_section].sh_addr;
- end = table + sechdrs[me->arch.unwind_section].sh_size;
-- gp = (Elf_Addr)me->module_core + me->arch.got_offset;
-+ gp = (Elf_Addr)me->module_core_rw + me->arch.got_offset;
-
- DEBUGP("register_unwind_table(), sect = %d at 0x%p - 0x%p (gp=0x%lx)\n",
- me->arch.unwind_section, table, end, gp);
-diff -urNp linux-2.6.37.4/arch/parisc/kernel/sys_parisc.c linux-2.6.37.4/arch/parisc/kernel/sys_parisc.c
---- linux-2.6.37.4/arch/parisc/kernel/sys_parisc.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/parisc/kernel/sys_parisc.c 2011-01-17 02:41:00.000000000 -0500
-@@ -43,7 +43,7 @@ static unsigned long get_unshared_area(u
- /* At this point: (!vma || addr < vma->vm_end). */
- if (TASK_SIZE - len < addr)
- return -ENOMEM;
-- if (!vma || addr + len <= vma->vm_start)
-+ if (check_heap_stack_gap(vma, addr, len))
- return addr;
- addr = vma->vm_end;
- }
-@@ -79,7 +79,7 @@ static unsigned long get_shared_area(str
- /* At this point: (!vma || addr < vma->vm_end). */
- if (TASK_SIZE - len < addr)
- return -ENOMEM;
-- if (!vma || addr + len <= vma->vm_start)
-+ if (check_heap_stack_gap(vma, addr, len))
- return addr;
- addr = DCACHE_ALIGN(vma->vm_end - offset) + offset;
- if (addr < vma->vm_end) /* handle wraparound */
-@@ -98,7 +98,7 @@ unsigned long arch_get_unmapped_area(str
- if (flags & MAP_FIXED)
- return addr;
- if (!addr)
-- addr = TASK_UNMAPPED_BASE;
-+ addr = current->mm->mmap_base;
-
- if (filp) {
- addr = get_shared_area(filp->f_mapping, addr, len, pgoff);
-diff -urNp linux-2.6.37.4/arch/parisc/kernel/traps.c linux-2.6.37.4/arch/parisc/kernel/traps.c
---- linux-2.6.37.4/arch/parisc/kernel/traps.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/parisc/kernel/traps.c 2011-01-17 02:41:00.000000000 -0500
-@@ -733,9 +733,7 @@ void notrace handle_interruption(int cod
-
- down_read(&current->mm->mmap_sem);
- vma = find_vma(current->mm,regs->iaoq[0]);
-- if (vma && (regs->iaoq[0] >= vma->vm_start)
-- && (vma->vm_flags & VM_EXEC)) {
--
-+ if (vma && (regs->iaoq[0] >= vma->vm_start)) {
- fault_address = regs->iaoq[0];
- fault_space = regs->iasq[0];
-
-diff -urNp linux-2.6.37.4/arch/parisc/mm/fault.c linux-2.6.37.4/arch/parisc/mm/fault.c
---- linux-2.6.37.4/arch/parisc/mm/fault.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/parisc/mm/fault.c 2011-01-17 02:41:00.000000000 -0500
-@@ -15,6 +15,7 @@
- #include <linux/sched.h>
- #include <linux/interrupt.h>
- #include <linux/module.h>
-+#include <linux/unistd.h>
-
- #include <asm/uaccess.h>
- #include <asm/traps.h>
-@@ -52,7 +53,7 @@ DEFINE_PER_CPU(struct exception_data, ex
- static unsigned long
- parisc_acctyp(unsigned long code, unsigned int inst)
- {
-- if (code == 6 || code == 16)
-+ if (code == 6 || code == 7 || code == 16)
- return VM_EXEC;
-
- switch (inst & 0xf0000000) {
-@@ -138,6 +139,116 @@ parisc_acctyp(unsigned long code, unsign
- }
- #endif
-
-+#ifdef CONFIG_PAX_PAGEEXEC
-+/*
-+ * PaX: decide what to do with offenders (instruction_pointer(regs) = fault address)
-+ *
-+ * returns 1 when task should be killed
-+ * 2 when rt_sigreturn trampoline was detected
-+ * 3 when unpatched PLT trampoline was detected
-+ */
-+static int pax_handle_fetch_fault(struct pt_regs *regs)
-+{
-+
-+#ifdef CONFIG_PAX_EMUPLT
-+ int err;
-+
-+ do { /* PaX: unpatched PLT emulation */
-+ unsigned int bl, depwi;
-+
-+ err = get_user(bl, (unsigned int *)instruction_pointer(regs));
-+ err |= get_user(depwi, (unsigned int *)(instruction_pointer(regs)+4));
-+
-+ if (err)
-+ break;
-+
-+ if (bl == 0xEA9F1FDDU && depwi == 0xD6801C1EU) {
-+ unsigned int ldw, bv, ldw2, addr = instruction_pointer(regs)-12;
-+
-+ err = get_user(ldw, (unsigned int *)addr);
-+ err |= get_user(bv, (unsigned int *)(addr+4));
-+ err |= get_user(ldw2, (unsigned int *)(addr+8));
-+
-+ if (err)
-+ break;
-+
-+ if (ldw == 0x0E801096U &&
-+ bv == 0xEAC0C000U &&
-+ ldw2 == 0x0E881095U)
-+ {
-+ unsigned int resolver, map;
-+
-+ err = get_user(resolver, (unsigned int *)(instruction_pointer(regs)+8));
-+ err |= get_user(map, (unsigned int *)(instruction_pointer(regs)+12));
-+ if (err)
-+ break;
-+
-+ regs->gr[20] = instruction_pointer(regs)+8;
-+ regs->gr[21] = map;
-+ regs->gr[22] = resolver;
-+ regs->iaoq[0] = resolver | 3UL;
-+ regs->iaoq[1] = regs->iaoq[0] + 4;
-+ return 3;
-+ }
-+ }
-+ } while (0);
-+#endif
-+
-+#ifdef CONFIG_PAX_EMUTRAMP
-+
-+#ifndef CONFIG_PAX_EMUSIGRT
-+ if (!(current->mm->pax_flags & MF_PAX_EMUTRAMP))
-+ return 1;
-+#endif
-+
-+ do { /* PaX: rt_sigreturn emulation */
-+ unsigned int ldi1, ldi2, bel, nop;
-+
-+ err = get_user(ldi1, (unsigned int *)instruction_pointer(regs));
-+ err |= get_user(ldi2, (unsigned int *)(instruction_pointer(regs)+4));
-+ err |= get_user(bel, (unsigned int *)(instruction_pointer(regs)+8));
-+ err |= get_user(nop, (unsigned int *)(instruction_pointer(regs)+12));
-+
-+ if (err)
-+ break;
-+
-+ if ((ldi1 == 0x34190000U || ldi1 == 0x34190002U) &&
-+ ldi2 == 0x3414015AU &&
-+ bel == 0xE4008200U &&
-+ nop == 0x08000240U)
-+ {
-+ regs->gr[25] = (ldi1 & 2) >> 1;
-+ regs->gr[20] = __NR_rt_sigreturn;
-+ regs->gr[31] = regs->iaoq[1] + 16;
-+ regs->sr[0] = regs->iasq[1];
-+ regs->iaoq[0] = 0x100UL;
-+ regs->iaoq[1] = regs->iaoq[0] + 4;
-+ regs->iasq[0] = regs->sr[2];
-+ regs->iasq[1] = regs->sr[2];
-+ return 2;
-+ }
-+ } while (0);
-+#endif
-+
-+ return 1;
-+}
-+
-+void pax_report_insns(void *pc, void *sp)
-+{
-+ unsigned long i;
-+
-+ printk(KERN_ERR "PAX: bytes at PC: ");
-+ for (i = 0; i < 5; i++) {
-+ unsigned int c;
-+ if (get_user(c, (unsigned int *)pc+i))
-+ printk(KERN_CONT "???????? ");
-+ else
-+ printk(KERN_CONT "%08x ", c);
-+ }
-+ printk("\n");
-+}
-+#endif
-+
- int fixup_exception(struct pt_regs *regs)
- {
- const struct exception_table_entry *fix;
-@@ -192,8 +303,33 @@ good_area:
-
- acc_type = parisc_acctyp(code,regs->iir);
-
-- if ((vma->vm_flags & acc_type) != acc_type)
-+ if ((vma->vm_flags & acc_type) != acc_type) {
-+
-+#ifdef CONFIG_PAX_PAGEEXEC
-+ if ((mm->pax_flags & MF_PAX_PAGEEXEC) && (acc_type & VM_EXEC) &&
-+ (address & ~3UL) == instruction_pointer(regs))
-+ {
-+ up_read(&mm->mmap_sem);
-+ switch (pax_handle_fetch_fault(regs)) {
-+
-+#ifdef CONFIG_PAX_EMUPLT
-+ case 3:
-+ return;
-+#endif
-+
-+#ifdef CONFIG_PAX_EMUTRAMP
-+ case 2:
-+ return;
-+#endif
-+
-+ }
-+ pax_report_fault(regs, (void *)instruction_pointer(regs), (void *)regs->gr[30]);
-+ do_group_exit(SIGKILL);
-+ }
-+#endif
-+
- goto bad_area;
-+ }
-
- /*
- * If for any reason at all we couldn't handle the fault, make
-diff -urNp linux-2.6.37.4/arch/powerpc/include/asm/device.h linux-2.6.37.4/arch/powerpc/include/asm/device.h
---- linux-2.6.37.4/arch/powerpc/include/asm/device.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/powerpc/include/asm/device.h 2011-01-17 02:41:00.000000000 -0500
-@@ -11,7 +11,7 @@ struct device_node;
-
- struct dev_archdata {
- /* DMA operations on that device */
-- struct dma_map_ops *dma_ops;
-+ const struct dma_map_ops *dma_ops;
-
- /*
- * When an iommu is in use, dma_data is used as a ptr to the base of the
-diff -urNp linux-2.6.37.4/arch/powerpc/include/asm/dma-mapping.h linux-2.6.37.4/arch/powerpc/include/asm/dma-mapping.h
---- linux-2.6.37.4/arch/powerpc/include/asm/dma-mapping.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/powerpc/include/asm/dma-mapping.h 2011-01-17 02:41:00.000000000 -0500
-@@ -66,12 +66,13 @@ static inline unsigned long device_to_ma
- /*
- * Available generic sets of operations
- */
-+/* cannot be const */
- #ifdef CONFIG_PPC64
--extern struct dma_map_ops dma_iommu_ops;
-+extern const struct dma_map_ops dma_iommu_ops;
- #endif
--extern struct dma_map_ops dma_direct_ops;
-+extern const struct dma_map_ops dma_direct_ops;
-
--static inline struct dma_map_ops *get_dma_ops(struct device *dev)
-+static inline const struct dma_map_ops *get_dma_ops(struct device *dev)
- {
- /* We don't handle the NULL dev case for ISA for now. We could
- * do it via an out of line call but it is not needed for now. The
-@@ -84,7 +85,7 @@ static inline struct dma_map_ops *get_dm
- return dev->archdata.dma_ops;
- }
-
--static inline void set_dma_ops(struct device *dev, struct dma_map_ops *ops)
-+static inline void set_dma_ops(struct device *dev, const struct dma_map_ops *ops)
- {
- dev->archdata.dma_ops = ops;
- }
-@@ -118,7 +119,7 @@ static inline void set_dma_offset(struct
-
- static inline int dma_supported(struct device *dev, u64 mask)
- {
-- struct dma_map_ops *dma_ops = get_dma_ops(dev);
-+ const struct dma_map_ops *dma_ops = get_dma_ops(dev);
-
- if (unlikely(dma_ops == NULL))
- return 0;
-@@ -132,7 +133,7 @@ extern int dma_set_mask(struct device *d
- static inline void *dma_alloc_coherent(struct device *dev, size_t size,
- dma_addr_t *dma_handle, gfp_t flag)
- {
-- struct dma_map_ops *dma_ops = get_dma_ops(dev);
-+ const struct dma_map_ops *dma_ops = get_dma_ops(dev);
- void *cpu_addr;
-
- BUG_ON(!dma_ops);
-@@ -147,7 +148,7 @@ static inline void *dma_alloc_coherent(s
- static inline void dma_free_coherent(struct device *dev, size_t size,
- void *cpu_addr, dma_addr_t dma_handle)
- {
-- struct dma_map_ops *dma_ops = get_dma_ops(dev);
-+ const struct dma_map_ops *dma_ops = get_dma_ops(dev);
-
- BUG_ON(!dma_ops);
-
-@@ -158,7 +159,7 @@ static inline void dma_free_coherent(str
-
- static inline int dma_mapping_error(struct device *dev, dma_addr_t dma_addr)
- {
-- struct dma_map_ops *dma_ops = get_dma_ops(dev);
-+ const struct dma_map_ops *dma_ops = get_dma_ops(dev);
-
- if (dma_ops->mapping_error)
- return dma_ops->mapping_error(dev, dma_addr);
-diff -urNp linux-2.6.37.4/arch/powerpc/include/asm/elf.h linux-2.6.37.4/arch/powerpc/include/asm/elf.h
---- linux-2.6.37.4/arch/powerpc/include/asm/elf.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/powerpc/include/asm/elf.h 2011-01-17 02:41:00.000000000 -0500
-@@ -178,8 +178,19 @@ typedef elf_fpreg_t elf_vsrreghalf_t32[E
- the loader. We need to make sure that it is out of the way of the program
- that it will "exec", and that there is sufficient room for the brk. */
-
--extern unsigned long randomize_et_dyn(unsigned long base);
--#define ELF_ET_DYN_BASE (randomize_et_dyn(0x20000000))
-+#define ELF_ET_DYN_BASE (0x20000000)
-+
-+#ifdef CONFIG_PAX_ASLR
-+#define PAX_ELF_ET_DYN_BASE (0x10000000UL)
-+
-+#ifdef __powerpc64__
-+#define PAX_DELTA_MMAP_LEN (is_32bit_task() ? 16 : 28)
-+#define PAX_DELTA_STACK_LEN (is_32bit_task() ? 16 : 28)
-+#else
-+#define PAX_DELTA_MMAP_LEN 15
-+#define PAX_DELTA_STACK_LEN 15
-+#endif
-+#endif
-
- /*
- * Our registers are always unsigned longs, whether we're a 32 bit
-@@ -274,9 +285,6 @@ extern int arch_setup_additional_pages(s
- (0x7ff >> (PAGE_SHIFT - 12)) : \
- (0x3ffff >> (PAGE_SHIFT - 12)))
-
--extern unsigned long arch_randomize_brk(struct mm_struct *mm);
--#define arch_randomize_brk arch_randomize_brk
--
- #endif /* __KERNEL__ */
-
- /*
-diff -urNp linux-2.6.37.4/arch/powerpc/include/asm/iommu.h linux-2.6.37.4/arch/powerpc/include/asm/iommu.h
---- linux-2.6.37.4/arch/powerpc/include/asm/iommu.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/powerpc/include/asm/iommu.h 2011-01-17 02:41:00.000000000 -0500
-@@ -116,6 +116,9 @@ extern void iommu_init_early_iSeries(voi
- extern void iommu_init_early_dart(void);
- extern void iommu_init_early_pasemi(void);
-
-+/* dma-iommu.c */
-+extern int dma_iommu_dma_supported(struct device *dev, u64 mask);
-+
- #ifdef CONFIG_PCI
- extern void pci_iommu_init(void);
- extern void pci_direct_iommu_init(void);
-diff -urNp linux-2.6.37.4/arch/powerpc/include/asm/kmap_types.h linux-2.6.37.4/arch/powerpc/include/asm/kmap_types.h
---- linux-2.6.37.4/arch/powerpc/include/asm/kmap_types.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/powerpc/include/asm/kmap_types.h 2011-01-17 02:41:00.000000000 -0500
-@@ -27,6 +27,7 @@ enum km_type {
- KM_PPC_SYNC_PAGE,
- KM_PPC_SYNC_ICACHE,
- KM_KDB,
-+ KM_CLEARPAGE,
- KM_TYPE_NR
- };
-
-diff -urNp linux-2.6.37.4/arch/powerpc/include/asm/page_64.h linux-2.6.37.4/arch/powerpc/include/asm/page_64.h
---- linux-2.6.37.4/arch/powerpc/include/asm/page_64.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/powerpc/include/asm/page_64.h 2011-01-17 02:41:00.000000000 -0500
-@@ -172,15 +172,18 @@ do { \
- * stack by default, so in the absense of a PT_GNU_STACK program header
- * we turn execute permission off.
- */
--#define VM_STACK_DEFAULT_FLAGS32 (VM_READ | VM_WRITE | VM_EXEC | \
-- VM_MAYREAD | VM_MAYWRITE | VM_MAYEXEC)
-+#define VM_STACK_DEFAULT_FLAGS32 \
-+ (((current->personality & READ_IMPLIES_EXEC) ? VM_EXEC : 0) | \
-+ VM_READ | VM_WRITE | VM_MAYREAD | VM_MAYWRITE | VM_MAYEXEC)
-
- #define VM_STACK_DEFAULT_FLAGS64 (VM_READ | VM_WRITE | \
- VM_MAYREAD | VM_MAYWRITE | VM_MAYEXEC)
-
-+#ifndef CONFIG_PAX_PAGEEXEC
- #define VM_STACK_DEFAULT_FLAGS \
- (is_32bit_task() ? \
- VM_STACK_DEFAULT_FLAGS32 : VM_STACK_DEFAULT_FLAGS64)
-+#endif
-
- #include <asm-generic/getorder.h>
-
-diff -urNp linux-2.6.37.4/arch/powerpc/include/asm/page.h linux-2.6.37.4/arch/powerpc/include/asm/page.h
---- linux-2.6.37.4/arch/powerpc/include/asm/page.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/powerpc/include/asm/page.h 2011-01-17 02:41:00.000000000 -0500
-@@ -129,8 +129,9 @@ extern phys_addr_t kernstart_addr;
- * and needs to be executable. This means the whole heap ends
- * up being executable.
- */
--#define VM_DATA_DEFAULT_FLAGS32 (VM_READ | VM_WRITE | VM_EXEC | \
-- VM_MAYREAD | VM_MAYWRITE | VM_MAYEXEC)
-+#define VM_DATA_DEFAULT_FLAGS32 \
-+ (((current->personality & READ_IMPLIES_EXEC) ? VM_EXEC : 0) | \
-+ VM_READ | VM_WRITE | VM_MAYREAD | VM_MAYWRITE | VM_MAYEXEC)
-
- #define VM_DATA_DEFAULT_FLAGS64 (VM_READ | VM_WRITE | \
- VM_MAYREAD | VM_MAYWRITE | VM_MAYEXEC)
-@@ -158,6 +159,9 @@ extern phys_addr_t kernstart_addr;
- #define is_kernel_addr(x) ((x) >= PAGE_OFFSET)
- #endif
-
-+#define ktla_ktva(addr) (addr)
-+#define ktva_ktla(addr) (addr)
-+
- #ifndef __ASSEMBLY__
-
- #undef STRICT_MM_TYPECHECKS
-diff -urNp linux-2.6.37.4/arch/powerpc/include/asm/pci.h linux-2.6.37.4/arch/powerpc/include/asm/pci.h
---- linux-2.6.37.4/arch/powerpc/include/asm/pci.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/powerpc/include/asm/pci.h 2011-01-17 02:41:00.000000000 -0500
-@@ -65,8 +65,8 @@ static inline int pci_get_legacy_ide_irq
- }
-
- #ifdef CONFIG_PCI
--extern void set_pci_dma_ops(struct dma_map_ops *dma_ops);
--extern struct dma_map_ops *get_pci_dma_ops(void);
-+extern void set_pci_dma_ops(const struct dma_map_ops *dma_ops);
-+extern const struct dma_map_ops *get_pci_dma_ops(void);
- #else /* CONFIG_PCI */
- #define set_pci_dma_ops(d)
- #define get_pci_dma_ops() NULL
-diff -urNp linux-2.6.37.4/arch/powerpc/include/asm/pgtable.h linux-2.6.37.4/arch/powerpc/include/asm/pgtable.h
---- linux-2.6.37.4/arch/powerpc/include/asm/pgtable.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/powerpc/include/asm/pgtable.h 2011-02-27 14:48:10.000000000 -0500
-@@ -2,6 +2,7 @@
- #define _ASM_POWERPC_PGTABLE_H
- #ifdef __KERNEL__
-
-+#include <linux/const.h>
- #ifndef __ASSEMBLY__
- #include <asm/processor.h> /* For TASK_SIZE */
- #include <asm/mmu.h>
-diff -urNp linux-2.6.37.4/arch/powerpc/include/asm/pte-hash32.h linux-2.6.37.4/arch/powerpc/include/asm/pte-hash32.h
---- linux-2.6.37.4/arch/powerpc/include/asm/pte-hash32.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/powerpc/include/asm/pte-hash32.h 2011-01-17 02:41:00.000000000 -0500
-@@ -21,6 +21,7 @@
- #define _PAGE_FILE 0x004 /* when !present: nonlinear file mapping */
- #define _PAGE_USER 0x004 /* usermode access allowed */
- #define _PAGE_GUARDED 0x008 /* G: prohibit speculative access */
-+#define _PAGE_EXEC _PAGE_GUARDED
- #define _PAGE_COHERENT 0x010 /* M: enforce memory coherence (SMP systems) */
- #define _PAGE_NO_CACHE 0x020 /* I: cache inhibit */
- #define _PAGE_WRITETHRU 0x040 /* W: cache write-through */
-diff -urNp linux-2.6.37.4/arch/powerpc/include/asm/reg.h linux-2.6.37.4/arch/powerpc/include/asm/reg.h
---- linux-2.6.37.4/arch/powerpc/include/asm/reg.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/powerpc/include/asm/reg.h 2011-01-17 02:41:00.000000000 -0500
-@@ -191,6 +191,7 @@
- #define SPRN_DBCR 0x136 /* e300 Data Breakpoint Control Reg */
- #define SPRN_DSISR 0x012 /* Data Storage Interrupt Status Register */
- #define DSISR_NOHPTE 0x40000000 /* no translation found */
-+#define DSISR_GUARDED 0x10000000 /* fetch from guarded storage */
- #define DSISR_PROTFAULT 0x08000000 /* protection fault */
- #define DSISR_ISSTORE 0x02000000 /* access was a store */
- #define DSISR_DABRMATCH 0x00400000 /* hit data breakpoint */
-diff -urNp linux-2.6.37.4/arch/powerpc/include/asm/swiotlb.h linux-2.6.37.4/arch/powerpc/include/asm/swiotlb.h
---- linux-2.6.37.4/arch/powerpc/include/asm/swiotlb.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/powerpc/include/asm/swiotlb.h 2011-01-17 02:41:00.000000000 -0500
-@@ -13,7 +13,7 @@
-
- #include <linux/swiotlb.h>
-
--extern struct dma_map_ops swiotlb_dma_ops;
-+extern const struct dma_map_ops swiotlb_dma_ops;
-
- static inline void dma_mark_clean(void *addr, size_t size) {}
-
-diff -urNp linux-2.6.37.4/arch/powerpc/include/asm/system.h linux-2.6.37.4/arch/powerpc/include/asm/system.h
---- linux-2.6.37.4/arch/powerpc/include/asm/system.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/powerpc/include/asm/system.h 2011-01-17 02:41:00.000000000 -0500
-@@ -533,7 +533,7 @@ __cmpxchg_local(volatile void *ptr, unsi
- #define cmpxchg64_local(ptr, o, n) __cmpxchg64_local_generic((ptr), (o), (n))
- #endif
-
--extern unsigned long arch_align_stack(unsigned long sp);
-+#define arch_align_stack(x) ((x) & ~0xfUL)
-
- /* Used in very early kernel initialization. */
- extern unsigned long reloc_offset(void);
-diff -urNp linux-2.6.37.4/arch/powerpc/include/asm/uaccess.h linux-2.6.37.4/arch/powerpc/include/asm/uaccess.h
---- linux-2.6.37.4/arch/powerpc/include/asm/uaccess.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/powerpc/include/asm/uaccess.h 2011-01-17 02:41:00.000000000 -0500
-@@ -13,6 +13,8 @@
- #define VERIFY_READ 0
- #define VERIFY_WRITE 1
-
-+extern void check_object_size(const void *ptr, unsigned long n, bool to);
-+
- /*
- * The fs value determines whether argument validity checking should be
- * performed or not. If get_fs() == USER_DS, checking is performed, with
-@@ -327,52 +329,6 @@ do { \
- extern unsigned long __copy_tofrom_user(void __user *to,
- const void __user *from, unsigned long size);
-
--#ifndef __powerpc64__
--
--static inline unsigned long copy_from_user(void *to,
-- const void __user *from, unsigned long n)
--{
-- unsigned long over;
--
-- if (access_ok(VERIFY_READ, from, n))
-- return __copy_tofrom_user((__force void __user *)to, from, n);
-- if ((unsigned long)from < TASK_SIZE) {
-- over = (unsigned long)from + n - TASK_SIZE;
-- return __copy_tofrom_user((__force void __user *)to, from,
-- n - over) + over;
-- }
-- return n;
--}
--
--static inline unsigned long copy_to_user(void __user *to,
-- const void *from, unsigned long n)
--{
-- unsigned long over;
--
-- if (access_ok(VERIFY_WRITE, to, n))
-- return __copy_tofrom_user(to, (__force void __user *)from, n);
-- if ((unsigned long)to < TASK_SIZE) {
-- over = (unsigned long)to + n - TASK_SIZE;
-- return __copy_tofrom_user(to, (__force void __user *)from,
-- n - over) + over;
-- }
-- return n;
--}
--
--#else /* __powerpc64__ */
--
--#define __copy_in_user(to, from, size) \
-- __copy_tofrom_user((to), (from), (size))
--
--extern unsigned long copy_from_user(void *to, const void __user *from,
-- unsigned long n);
--extern unsigned long copy_to_user(void __user *to, const void *from,
-- unsigned long n);
--extern unsigned long copy_in_user(void __user *to, const void __user *from,
-- unsigned long n);
--
--#endif /* __powerpc64__ */
--
- static inline unsigned long __copy_from_user_inatomic(void *to,
- const void __user *from, unsigned long n)
- {
-@@ -396,6 +352,10 @@ static inline unsigned long __copy_from_
- if (ret == 0)
- return 0;
- }
-+
-+ if (!__builtin_constant_p(n))
-+ check_object_size(to, n, false);
-+
- return __copy_tofrom_user((__force void __user *)to, from, n);
- }
-
-@@ -422,6 +382,10 @@ static inline unsigned long __copy_to_us
- if (ret == 0)
- return 0;
- }
-+
-+ if (!__builtin_constant_p(n))
-+ check_object_size(from, n, true);
-+
- return __copy_tofrom_user(to, (__force const void __user *)from, n);
- }
-
-@@ -439,6 +403,92 @@ static inline unsigned long __copy_to_us
- return __copy_to_user_inatomic(to, from, size);
- }
-
-+#ifndef __powerpc64__
-+
-+static inline unsigned long __must_check copy_from_user(void *to,
-+ const void __user *from, unsigned long n)
-+{
-+ unsigned long over;
-+
-+ if ((long)n < 0)
-+ return n;
-+
-+ if (access_ok(VERIFY_READ, from, n)) {
-+ if (!__builtin_constant_p(n))
-+ check_object_size(to, n, false);
-+ return __copy_tofrom_user((__force void __user *)to, from, n);
-+ }
-+ if ((unsigned long)from < TASK_SIZE) {
-+ over = (unsigned long)from + n - TASK_SIZE;
-+ if (!__builtin_constant_p(n - over))
-+ check_object_size(to, n - over, false);
-+ return __copy_tofrom_user((__force void __user *)to, from,
-+ n - over) + over;
-+ }
-+ return n;
-+}
-+
-+static inline unsigned long __must_check copy_to_user(void __user *to,
-+ const void *from, unsigned long n)
-+{
-+ unsigned long over;
-+
-+ if ((long)n < 0)
-+ return n;
-+
-+ if (access_ok(VERIFY_WRITE, to, n)) {
-+ if (!__builtin_constant_p(n))
-+ check_object_size(from, n, true);
-+ return __copy_tofrom_user(to, (__force void __user *)from, n);
-+ }
-+ if ((unsigned long)to < TASK_SIZE) {
-+ over = (unsigned long)to + n - TASK_SIZE;
-+ if (!__builtin_constant_p(n))
-+ check_object_size(from, n - over, true);
-+ return __copy_tofrom_user(to, (__force void __user *)from,
-+ n - over) + over;
-+ }
-+ return n;
-+}
-+
-+#else /* __powerpc64__ */
-+
-+#define __copy_in_user(to, from, size) \
-+ __copy_tofrom_user((to), (from), (size))
-+
-+static inline unsigned long __must_check copy_from_user(void *to, const void __user *from, unsigned long n)
-+{
-+ if ((long)n < 0 || n > INT_MAX)
-+ return n;
-+
-+ if (!__builtin_constant_p(n))
-+ check_object_size(to, n, false);
-+
-+ if (likely(access_ok(VERIFY_READ, from, n)))
-+ n = __copy_from_user(to, from, n);
-+ else
-+ memset(to, 0, n);
-+ return n;
-+}
-+
-+static inline unsigned long __must_check copy_to_user(void __user *to, const void *from, unsigned long n)
-+{
-+ if ((long)n < 0 || n > INT_MAX)
-+ return n;
-+
-+ if (likely(access_ok(VERIFY_WRITE, to, n))) {
-+ if (!__builtin_constant_p(n))
-+ check_object_size(from, n, true);
-+ n = __copy_to_user(to, from, n);
-+ }
-+ return n;
-+}
-+
-+extern unsigned long copy_in_user(void __user *to, const void __user *from,
-+ unsigned long n);
-+
-+#endif /* __powerpc64__ */
-+
- extern unsigned long __clear_user(void __user *addr, unsigned long size);
-
- static inline unsigned long clear_user(void __user *addr, unsigned long size)
-diff -urNp linux-2.6.37.4/arch/powerpc/kernel/dma.c linux-2.6.37.4/arch/powerpc/kernel/dma.c
---- linux-2.6.37.4/arch/powerpc/kernel/dma.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/powerpc/kernel/dma.c 2011-01-17 02:41:00.000000000 -0500
-@@ -136,7 +136,7 @@ static inline void dma_direct_sync_singl
- }
- #endif
-
--struct dma_map_ops dma_direct_ops = {
-+const struct dma_map_ops dma_direct_ops = {
- .alloc_coherent = dma_direct_alloc_coherent,
- .free_coherent = dma_direct_free_coherent,
- .map_sg = dma_direct_map_sg,
-@@ -157,7 +157,7 @@ EXPORT_SYMBOL(dma_direct_ops);
-
- int dma_set_mask(struct device *dev, u64 dma_mask)
- {
-- struct dma_map_ops *dma_ops = get_dma_ops(dev);
-+ const struct dma_map_ops *dma_ops = get_dma_ops(dev);
-
- if (ppc_md.dma_set_mask)
- return ppc_md.dma_set_mask(dev, dma_mask);
-diff -urNp linux-2.6.37.4/arch/powerpc/kernel/dma-iommu.c linux-2.6.37.4/arch/powerpc/kernel/dma-iommu.c
---- linux-2.6.37.4/arch/powerpc/kernel/dma-iommu.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/powerpc/kernel/dma-iommu.c 2011-01-17 02:41:00.000000000 -0500
-@@ -70,7 +70,7 @@ static void dma_iommu_unmap_sg(struct de
- }
-
- /* We support DMA to/from any memory page via the iommu */
--static int dma_iommu_dma_supported(struct device *dev, u64 mask)
-+int dma_iommu_dma_supported(struct device *dev, u64 mask)
- {
- struct iommu_table *tbl = get_iommu_table_base(dev);
-
-@@ -90,6 +90,7 @@ static int dma_iommu_dma_supported(struc
- return 1;
- }
-
-+/* cannot be const, see arch/powerpc/platforms/cell/iommu.c */
- struct dma_map_ops dma_iommu_ops = {
- .alloc_coherent = dma_iommu_alloc_coherent,
- .free_coherent = dma_iommu_free_coherent,
-diff -urNp linux-2.6.37.4/arch/powerpc/kernel/dma-swiotlb.c linux-2.6.37.4/arch/powerpc/kernel/dma-swiotlb.c
---- linux-2.6.37.4/arch/powerpc/kernel/dma-swiotlb.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/powerpc/kernel/dma-swiotlb.c 2011-01-17 02:41:00.000000000 -0500
-@@ -31,7 +31,7 @@ unsigned int ppc_swiotlb_enable;
- * map_page, and unmap_page on highmem, use normal dma_ops
- * for everything else.
- */
--struct dma_map_ops swiotlb_dma_ops = {
-+const struct dma_map_ops swiotlb_dma_ops = {
- .alloc_coherent = dma_direct_alloc_coherent,
- .free_coherent = dma_direct_free_coherent,
- .map_sg = swiotlb_map_sg_attrs,
-diff -urNp linux-2.6.37.4/arch/powerpc/kernel/exceptions-64e.S linux-2.6.37.4/arch/powerpc/kernel/exceptions-64e.S
---- linux-2.6.37.4/arch/powerpc/kernel/exceptions-64e.S 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/powerpc/kernel/exceptions-64e.S 2011-01-17 02:41:00.000000000 -0500
-@@ -495,6 +495,7 @@ storage_fault_common:
- std r14,_DAR(r1)
- std r15,_DSISR(r1)
- addi r3,r1,STACK_FRAME_OVERHEAD
-+ bl .save_nvgprs
- mr r4,r14
- mr r5,r15
- ld r14,PACA_EXGEN+EX_R14(r13)
-@@ -504,8 +505,7 @@ storage_fault_common:
- cmpdi r3,0
- bne- 1f
- b .ret_from_except_lite
--1: bl .save_nvgprs
-- mr r5,r3
-+1: mr r5,r3
- addi r3,r1,STACK_FRAME_OVERHEAD
- ld r4,_DAR(r1)
- bl .bad_page_fault
-diff -urNp linux-2.6.37.4/arch/powerpc/kernel/exceptions-64s.S linux-2.6.37.4/arch/powerpc/kernel/exceptions-64s.S
---- linux-2.6.37.4/arch/powerpc/kernel/exceptions-64s.S 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/powerpc/kernel/exceptions-64s.S 2011-01-17 02:41:00.000000000 -0500
-@@ -847,10 +847,10 @@ handle_page_fault:
- 11: ld r4,_DAR(r1)
- ld r5,_DSISR(r1)
- addi r3,r1,STACK_FRAME_OVERHEAD
-+ bl .save_nvgprs
- bl .do_page_fault
- cmpdi r3,0
- beq+ 13f
-- bl .save_nvgprs
- mr r5,r3
- addi r3,r1,STACK_FRAME_OVERHEAD
- lwz r4,_DAR(r1)
-diff -urNp linux-2.6.37.4/arch/powerpc/kernel/ibmebus.c linux-2.6.37.4/arch/powerpc/kernel/ibmebus.c
---- linux-2.6.37.4/arch/powerpc/kernel/ibmebus.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/powerpc/kernel/ibmebus.c 2011-01-17 02:41:00.000000000 -0500
-@@ -128,7 +128,7 @@ static int ibmebus_dma_supported(struct
- return 1;
- }
-
--static struct dma_map_ops ibmebus_dma_ops = {
-+static const struct dma_map_ops ibmebus_dma_ops = {
- .alloc_coherent = ibmebus_alloc_coherent,
- .free_coherent = ibmebus_free_coherent,
- .map_sg = ibmebus_map_sg,
-diff -urNp linux-2.6.37.4/arch/powerpc/kernel/kgdb.c linux-2.6.37.4/arch/powerpc/kernel/kgdb.c
---- linux-2.6.37.4/arch/powerpc/kernel/kgdb.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/powerpc/kernel/kgdb.c 2011-01-17 02:41:00.000000000 -0500
-@@ -422,7 +422,7 @@ int kgdb_arch_handle_exception(int vecto
- /*
- * Global data
- */
--struct kgdb_arch arch_kgdb_ops = {
-+const struct kgdb_arch arch_kgdb_ops = {
- .gdb_bpt_instr = {0x7d, 0x82, 0x10, 0x08},
- };
-
-diff -urNp linux-2.6.37.4/arch/powerpc/kernel/module_32.c linux-2.6.37.4/arch/powerpc/kernel/module_32.c
---- linux-2.6.37.4/arch/powerpc/kernel/module_32.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/powerpc/kernel/module_32.c 2011-01-17 02:41:00.000000000 -0500
-@@ -162,7 +162,7 @@ int module_frob_arch_sections(Elf32_Ehdr
- me->arch.core_plt_section = i;
- }
- if (!me->arch.core_plt_section || !me->arch.init_plt_section) {
-- printk("Module doesn't contain .plt or .init.plt sections.\n");
-+ printk("Module %s doesn't contain .plt or .init.plt sections.\n", me->name);
- return -ENOEXEC;
- }
-
-@@ -203,11 +203,16 @@ static uint32_t do_plt_call(void *locati
-
- DEBUGP("Doing plt for call to 0x%x at 0x%x\n", val, (unsigned int)location);
- /* Init, or core PLT? */
-- if (location >= mod->module_core
-- && location < mod->module_core + mod->core_size)
-+ if ((location >= mod->module_core_rx && location < mod->module_core_rx + mod->core_size_rx) ||
-+ (location >= mod->module_core_rw && location < mod->module_core_rw + mod->core_size_rw))
- entry = (void *)sechdrs[mod->arch.core_plt_section].sh_addr;
-- else
-+ else if ((location >= mod->module_init_rx && location < mod->module_init_rx + mod->init_size_rx) ||
-+ (location >= mod->module_init_rw && location < mod->module_init_rw + mod->init_size_rw))
- entry = (void *)sechdrs[mod->arch.init_plt_section].sh_addr;
-+ else {
-+ printk(KERN_ERR "%s: invalid R_PPC_REL24 entry found\n", mod->name);
-+ return ~0UL;
-+ }
-
- /* Find this entry, or if that fails, the next avail. entry */
- while (entry->jump[0]) {
-diff -urNp linux-2.6.37.4/arch/powerpc/kernel/module.c linux-2.6.37.4/arch/powerpc/kernel/module.c
---- linux-2.6.37.4/arch/powerpc/kernel/module.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/powerpc/kernel/module.c 2011-01-17 02:41:00.000000000 -0500
-@@ -31,11 +31,24 @@
-
- LIST_HEAD(module_bug_list);
-
-+#ifdef CONFIG_PAX_KERNEXEC
- void *module_alloc(unsigned long size)
- {
- if (size == 0)
- return NULL;
-
-+ return vmalloc(size);
-+}
-+
-+void *module_alloc_exec(unsigned long size)
-+#else
-+void *module_alloc(unsigned long size)
-+#endif
-+
-+{
-+ if (size == 0)
-+ return NULL;
-+
- return vmalloc_exec(size);
- }
-
-@@ -45,6 +58,13 @@ void module_free(struct module *mod, voi
- vfree(module_region);
- }
-
-+#ifdef CONFIG_PAX_KERNEXEC
-+void module_free_exec(struct module *mod, void *module_region)
-+{
-+ module_free(mod, module_region);
-+}
-+#endif
-+
- static const Elf_Shdr *find_section(const Elf_Ehdr *hdr,
- const Elf_Shdr *sechdrs,
- const char *name)
-diff -urNp linux-2.6.37.4/arch/powerpc/kernel/pci-common.c linux-2.6.37.4/arch/powerpc/kernel/pci-common.c
---- linux-2.6.37.4/arch/powerpc/kernel/pci-common.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/powerpc/kernel/pci-common.c 2011-01-17 02:41:00.000000000 -0500
-@@ -52,14 +52,14 @@ resource_size_t isa_mem_base;
- unsigned int ppc_pci_flags = 0;
-
-
--static struct dma_map_ops *pci_dma_ops = &dma_direct_ops;
-+static const struct dma_map_ops *pci_dma_ops = &dma_direct_ops;
-
--void set_pci_dma_ops(struct dma_map_ops *dma_ops)
-+void set_pci_dma_ops(const struct dma_map_ops *dma_ops)
- {
- pci_dma_ops = dma_ops;
- }
-
--struct dma_map_ops *get_pci_dma_ops(void)
-+const struct dma_map_ops *get_pci_dma_ops(void)
- {
- return pci_dma_ops;
- }
-diff -urNp linux-2.6.37.4/arch/powerpc/kernel/process.c linux-2.6.37.4/arch/powerpc/kernel/process.c
---- linux-2.6.37.4/arch/powerpc/kernel/process.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/powerpc/kernel/process.c 2011-01-17 02:41:00.000000000 -0500
-@@ -653,8 +653,8 @@ void show_regs(struct pt_regs * regs)
- * Lookup NIP late so we have the best change of getting the
- * above info out without failing
- */
-- printk("NIP ["REG"] %pS\n", regs->nip, (void *)regs->nip);
-- printk("LR ["REG"] %pS\n", regs->link, (void *)regs->link);
-+ printk("NIP ["REG"] %pA\n", regs->nip, (void *)regs->nip);
-+ printk("LR ["REG"] %pA\n", regs->link, (void *)regs->link);
- #endif
- show_stack(current, (unsigned long *) regs->gpr[1]);
- if (!user_mode(regs))
-@@ -1144,10 +1144,10 @@ void show_stack(struct task_struct *tsk,
- newsp = stack[0];
- ip = stack[STACK_FRAME_LR_SAVE];
- if (!firstframe || ip != lr) {
-- printk("["REG"] ["REG"] %pS", sp, ip, (void *)ip);
-+ printk("["REG"] ["REG"] %pA", sp, ip, (void *)ip);
- #ifdef CONFIG_FUNCTION_GRAPH_TRACER
- if ((ip == rth || ip == mrth) && curr_frame >= 0) {
-- printk(" (%pS)",
-+ printk(" (%pA)",
- (void *)current->ret_stack[curr_frame].ret);
- curr_frame--;
- }
-@@ -1167,7 +1167,7 @@ void show_stack(struct task_struct *tsk,
- struct pt_regs *regs = (struct pt_regs *)
- (sp + STACK_FRAME_OVERHEAD);
- lr = regs->link;
-- printk("--- Exception: %lx at %pS\n LR = %pS\n",
-+ printk("--- Exception: %lx at %pA\n LR = %pA\n",
- regs->trap, (void *)regs->nip, (void *)lr);
- firstframe = 1;
- }
-@@ -1242,58 +1242,3 @@ void thread_info_cache_init(void)
- }
-
- #endif /* THREAD_SHIFT < PAGE_SHIFT */
--
--unsigned long arch_align_stack(unsigned long sp)
--{
-- if (!(current->personality & ADDR_NO_RANDOMIZE) && randomize_va_space)
-- sp -= get_random_int() & ~PAGE_MASK;
-- return sp & ~0xf;
--}
--
--static inline unsigned long brk_rnd(void)
--{
-- unsigned long rnd = 0;
--
-- /* 8MB for 32bit, 1GB for 64bit */
-- if (is_32bit_task())
-- rnd = (long)(get_random_int() % (1<<(23-PAGE_SHIFT)));
-- else
-- rnd = (long)(get_random_int() % (1<<(30-PAGE_SHIFT)));
--
-- return rnd << PAGE_SHIFT;
--}
--
--unsigned long arch_randomize_brk(struct mm_struct *mm)
--{
-- unsigned long base = mm->brk;
-- unsigned long ret;
--
--#ifdef CONFIG_PPC_STD_MMU_64
-- /*
-- * If we are using 1TB segments and we are allowed to randomise
-- * the heap, we can put it above 1TB so it is backed by a 1TB
-- * segment. Otherwise the heap will be in the bottom 1TB
-- * which always uses 256MB segments and this may result in a
-- * performance penalty.
-- */
-- if (!is_32bit_task() && (mmu_highuser_ssize == MMU_SEGSIZE_1T))
-- base = max_t(unsigned long, mm->brk, 1UL << SID_SHIFT_1T);
--#endif
--
-- ret = PAGE_ALIGN(base + brk_rnd());
--
-- if (ret < mm->brk)
-- return mm->brk;
--
-- return ret;
--}
--
--unsigned long randomize_et_dyn(unsigned long base)
--{
-- unsigned long ret = PAGE_ALIGN(base + brk_rnd());
--
-- if (ret < base)
-- return base;
--
-- return ret;
--}
-diff -urNp linux-2.6.37.4/arch/powerpc/kernel/signal_32.c linux-2.6.37.4/arch/powerpc/kernel/signal_32.c
---- linux-2.6.37.4/arch/powerpc/kernel/signal_32.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/powerpc/kernel/signal_32.c 2011-01-17 02:41:00.000000000 -0500
-@@ -858,7 +858,7 @@ int handle_rt_signal32(unsigned long sig
- /* Save user registers on the stack */
- frame = &rt_sf->uc.uc_mcontext;
- addr = frame;
-- if (vdso32_rt_sigtramp && current->mm->context.vdso_base) {
-+ if (vdso32_rt_sigtramp && current->mm->context.vdso_base != ~0UL) {
- if (save_user_regs(regs, frame, 0, 1))
- goto badframe;
- regs->link = current->mm->context.vdso_base + vdso32_rt_sigtramp;
-diff -urNp linux-2.6.37.4/arch/powerpc/kernel/signal_64.c linux-2.6.37.4/arch/powerpc/kernel/signal_64.c
---- linux-2.6.37.4/arch/powerpc/kernel/signal_64.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/powerpc/kernel/signal_64.c 2011-01-17 02:41:00.000000000 -0500
-@@ -429,7 +429,7 @@ int handle_rt_signal64(int signr, struct
- current->thread.fpscr.val = 0;
-
- /* Set up to return from userspace. */
-- if (vdso64_rt_sigtramp && current->mm->context.vdso_base) {
-+ if (vdso64_rt_sigtramp && current->mm->context.vdso_base != ~0UL) {
- regs->link = current->mm->context.vdso_base + vdso64_rt_sigtramp;
- } else {
- err |= setup_trampoline(__NR_rt_sigreturn, &frame->tramp[0]);
-diff -urNp linux-2.6.37.4/arch/powerpc/kernel/vdso.c linux-2.6.37.4/arch/powerpc/kernel/vdso.c
---- linux-2.6.37.4/arch/powerpc/kernel/vdso.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/powerpc/kernel/vdso.c 2011-01-17 02:41:00.000000000 -0500
-@@ -36,6 +36,7 @@
- #include <asm/firmware.h>
- #include <asm/vdso.h>
- #include <asm/vdso_datapage.h>
-+#include <asm/mman.h>
-
- #include "setup.h"
-
-@@ -220,7 +221,7 @@ int arch_setup_additional_pages(struct l
- vdso_base = VDSO32_MBASE;
- #endif
-
-- current->mm->context.vdso_base = 0;
-+ current->mm->context.vdso_base = ~0UL;
-
- /* vDSO has a problem and was disabled, just don't "enable" it for the
- * process
-@@ -240,7 +241,7 @@ int arch_setup_additional_pages(struct l
- vdso_base = get_unmapped_area(NULL, vdso_base,
- (vdso_pages << PAGE_SHIFT) +
- ((VDSO_ALIGNMENT - 1) & PAGE_MASK),
-- 0, 0);
-+ 0, MAP_PRIVATE | MAP_EXECUTABLE);
- if (IS_ERR_VALUE(vdso_base)) {
- rc = vdso_base;
- goto fail_mmapsem;
-diff -urNp linux-2.6.37.4/arch/powerpc/kernel/vio.c linux-2.6.37.4/arch/powerpc/kernel/vio.c
---- linux-2.6.37.4/arch/powerpc/kernel/vio.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/powerpc/kernel/vio.c 2011-01-17 02:41:00.000000000 -0500
-@@ -600,11 +600,12 @@ static void vio_dma_iommu_unmap_sg(struc
- vio_cmo_dealloc(viodev, alloc_size);
- }
-
--struct dma_map_ops vio_dma_mapping_ops = {
-+static const struct dma_map_ops vio_dma_mapping_ops = {
- .alloc_coherent = vio_dma_iommu_alloc_coherent,
- .free_coherent = vio_dma_iommu_free_coherent,
- .map_sg = vio_dma_iommu_map_sg,
- .unmap_sg = vio_dma_iommu_unmap_sg,
-+ .dma_supported = dma_iommu_dma_supported,
- .map_page = vio_dma_iommu_map_page,
- .unmap_page = vio_dma_iommu_unmap_page,
-
-@@ -858,7 +859,6 @@ static void vio_cmo_bus_remove(struct vi
-
- static void vio_cmo_set_dma_ops(struct vio_dev *viodev)
- {
-- vio_dma_mapping_ops.dma_supported = dma_iommu_ops.dma_supported;
- viodev->dev.archdata.dma_ops = &vio_dma_mapping_ops;
- }
-
-diff -urNp linux-2.6.37.4/arch/powerpc/lib/usercopy_64.c linux-2.6.37.4/arch/powerpc/lib/usercopy_64.c
---- linux-2.6.37.4/arch/powerpc/lib/usercopy_64.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/powerpc/lib/usercopy_64.c 2011-01-17 02:41:00.000000000 -0500
-@@ -9,22 +9,6 @@
- #include <linux/module.h>
- #include <asm/uaccess.h>
-
--unsigned long copy_from_user(void *to, const void __user *from, unsigned long n)
--{
-- if (likely(access_ok(VERIFY_READ, from, n)))
-- n = __copy_from_user(to, from, n);
-- else
-- memset(to, 0, n);
-- return n;
--}
--
--unsigned long copy_to_user(void __user *to, const void *from, unsigned long n)
--{
-- if (likely(access_ok(VERIFY_WRITE, to, n)))
-- n = __copy_to_user(to, from, n);
-- return n;
--}
--
- unsigned long copy_in_user(void __user *to, const void __user *from,
- unsigned long n)
- {
-@@ -35,7 +19,5 @@ unsigned long copy_in_user(void __user *
- return n;
- }
-
--EXPORT_SYMBOL(copy_from_user);
--EXPORT_SYMBOL(copy_to_user);
- EXPORT_SYMBOL(copy_in_user);
-
-diff -urNp linux-2.6.37.4/arch/powerpc/mm/fault.c linux-2.6.37.4/arch/powerpc/mm/fault.c
---- linux-2.6.37.4/arch/powerpc/mm/fault.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/powerpc/mm/fault.c 2011-01-17 02:41:00.000000000 -0500
-@@ -31,6 +31,10 @@
- #include <linux/kdebug.h>
- #include <linux/perf_event.h>
- #include <linux/magic.h>
-+#include <linux/slab.h>
-+#include <linux/pagemap.h>
-+#include <linux/compiler.h>
-+#include <linux/unistd.h>
-
- #include <asm/firmware.h>
- #include <asm/page.h>
-@@ -42,6 +46,7 @@
- #include <asm/tlbflush.h>
- #include <asm/siginfo.h>
- #include <mm/mmu_decl.h>
-+#include <asm/ptrace.h>
-
- #ifdef CONFIG_KPROBES
- static inline int notify_page_fault(struct pt_regs *regs)
-@@ -65,6 +70,33 @@ static inline int notify_page_fault(stru
- }
- #endif
-
-+#ifdef CONFIG_PAX_PAGEEXEC
-+/*
-+ * PaX: decide what to do with offenders (regs->nip = fault address)
-+ *
-+ * returns 1 when task should be killed
-+ */
-+static int pax_handle_fetch_fault(struct pt_regs *regs)
-+{
-+ return 1;
-+}
-+
-+void pax_report_insns(void *pc, void *sp)
-+{
-+ unsigned long i;
-+
-+ printk(KERN_ERR "PAX: bytes at PC: ");
-+ for (i = 0; i < 5; i++) {
-+ unsigned int c;
-+ if (get_user(c, (unsigned int __user *)pc+i))
-+ printk(KERN_CONT "???????? ");
-+ else
-+ printk(KERN_CONT "%08x ", c);
-+ }
-+ printk("\n");
-+}
-+#endif
-+
- /*
- * Check whether the instruction at regs->nip is a store using
- * an update addressing form which will update r1.
-@@ -135,7 +167,7 @@ int __kprobes do_page_fault(struct pt_re
- * indicate errors in DSISR but can validly be set in SRR1.
- */
- if (trap == 0x400)
-- error_code &= 0x48200000;
-+ error_code &= 0x58200000;
- else
- is_write = error_code & DSISR_ISSTORE;
- #else
-@@ -258,7 +290,7 @@ good_area:
- * "undefined". Of those that can be set, this is the only
- * one which seems bad.
- */
-- if (error_code & 0x10000000)
-+ if (error_code & DSISR_GUARDED)
- /* Guarded storage error. */
- goto bad_area;
- #endif /* CONFIG_8xx */
-@@ -273,7 +305,7 @@ good_area:
- * processors use the same I/D cache coherency mechanism
- * as embedded.
- */
-- if (error_code & DSISR_PROTFAULT)
-+ if (error_code & (DSISR_PROTFAULT | DSISR_GUARDED))
- goto bad_area;
- #endif /* CONFIG_PPC_STD_MMU */
-
-@@ -342,6 +374,23 @@ bad_area:
- bad_area_nosemaphore:
- /* User mode accesses cause a SIGSEGV */
- if (user_mode(regs)) {
-+
-+#ifdef CONFIG_PAX_PAGEEXEC
-+ if (mm->pax_flags & MF_PAX_PAGEEXEC) {
-+#ifdef CONFIG_PPC_STD_MMU
-+ if (is_exec && (error_code & (DSISR_PROTFAULT | DSISR_GUARDED))) {
-+#else
-+ if (is_exec && regs->nip == address) {
-+#endif
-+ switch (pax_handle_fetch_fault(regs)) {
-+ }
-+
-+ pax_report_fault(regs, (void *)regs->nip, (void *)regs->gpr[PT_R1]);
-+ do_group_exit(SIGKILL);
-+ }
-+ }
-+#endif
-+
- _exception(SIGSEGV, regs, code, address);
- return 0;
- }
-diff -urNp linux-2.6.37.4/arch/powerpc/mm/mmap_64.c linux-2.6.37.4/arch/powerpc/mm/mmap_64.c
---- linux-2.6.37.4/arch/powerpc/mm/mmap_64.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/powerpc/mm/mmap_64.c 2011-01-17 02:41:00.000000000 -0500
-@@ -99,10 +99,22 @@ void arch_pick_mmap_layout(struct mm_str
- */
- if (mmap_is_legacy()) {
- mm->mmap_base = TASK_UNMAPPED_BASE;
-+
-+#ifdef CONFIG_PAX_RANDMMAP
-+ if (mm->pax_flags & MF_PAX_RANDMMAP)
-+ mm->mmap_base += mm->delta_mmap;
-+#endif
-+
- mm->get_unmapped_area = arch_get_unmapped_area;
- mm->unmap_area = arch_unmap_area;
- } else {
- mm->mmap_base = mmap_base();
-+
-+#ifdef CONFIG_PAX_RANDMMAP
-+ if (mm->pax_flags & MF_PAX_RANDMMAP)
-+ mm->mmap_base -= mm->delta_mmap + mm->delta_stack;
-+#endif
-+
- mm->get_unmapped_area = arch_get_unmapped_area_topdown;
- mm->unmap_area = arch_unmap_area_topdown;
- }
-diff -urNp linux-2.6.37.4/arch/powerpc/mm/slice.c linux-2.6.37.4/arch/powerpc/mm/slice.c
---- linux-2.6.37.4/arch/powerpc/mm/slice.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/powerpc/mm/slice.c 2011-03-21 17:35:08.000000000 -0400
-@@ -98,7 +98,7 @@ static int slice_area_is_free(struct mm_
- if ((mm->task_size - len) < addr)
- return 0;
- vma = find_vma(mm, addr);
-- return (!vma || (addr + len) <= vma->vm_start);
-+ return check_heap_stack_gap(vma, addr, len);
- }
-
- static int slice_low_has_vma(struct mm_struct *mm, unsigned long slice)
-@@ -256,7 +256,7 @@ full_search:
- addr = _ALIGN_UP(addr + 1, 1ul << SLICE_HIGH_SHIFT);
- continue;
- }
-- if (!vma || addr + len <= vma->vm_start) {
-+ if (check_heap_stack_gap(vma, addr, len)) {
- /*
- * Remember the place where we stopped the search:
- */
-@@ -313,10 +313,14 @@ static unsigned long slice_find_area_top
- }
- }
-
-- addr = mm->mmap_base;
-- while (addr > len) {
-+ if (mm->mmap_base < len)
-+ addr = -ENOMEM;
-+ else
-+ addr = mm->mmap_base - len;
-+
-+ while (!IS_ERR_VALUE(addr)) {
- /* Go down by chunk size */
-- addr = _ALIGN_DOWN(addr - len, 1ul << pshift);
-+ addr = _ALIGN_DOWN(addr, 1ul << pshift);
-
- /* Check for hit with different page size */
- mask = slice_range_to_mask(addr, len);
-@@ -336,7 +340,7 @@ static unsigned long slice_find_area_top
- * return with success:
- */
- vma = find_vma(mm, addr);
-- if (!vma || (addr + len) <= vma->vm_start) {
-+ if (check_heap_stack_gap(vma, addr, len)) {
- /* remember the address as a hint for next time */
- if (use_cache)
- mm->free_area_cache = addr;
-@@ -348,7 +352,7 @@ static unsigned long slice_find_area_top
- mm->cached_hole_size = vma->vm_start - addr;
-
- /* try just below the current vma->vm_start */
-- addr = vma->vm_start;
-+ addr = skip_heap_stack_gap(vma, len);
- }
-
- /*
-@@ -426,6 +430,11 @@ unsigned long slice_get_unmapped_area(un
- if (fixed && addr > (mm->task_size - len))
- return -EINVAL;
-
-+#ifdef CONFIG_PAX_RANDMMAP
-+ if (!fixed && (mm->pax_flags & MF_PAX_RANDMMAP))
-+ addr = 0;
-+#endif
-+
- /* If hint, make sure it matches our alignment restrictions */
- if (!fixed && addr) {
- addr = _ALIGN_UP(addr, 1ul << pshift);
-diff -urNp linux-2.6.37.4/arch/powerpc/platforms/52xx/lite5200_pm.c linux-2.6.37.4/arch/powerpc/platforms/52xx/lite5200_pm.c
---- linux-2.6.37.4/arch/powerpc/platforms/52xx/lite5200_pm.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/powerpc/platforms/52xx/lite5200_pm.c 2011-01-17 02:41:00.000000000 -0500
-@@ -232,7 +232,7 @@ static void lite5200_pm_end(void)
- lite5200_pm_target_state = PM_SUSPEND_ON;
- }
-
--static struct platform_suspend_ops lite5200_pm_ops = {
-+static const struct platform_suspend_ops lite5200_pm_ops = {
- .valid = lite5200_pm_valid,
- .begin = lite5200_pm_begin,
- .prepare = lite5200_pm_prepare,
-diff -urNp linux-2.6.37.4/arch/powerpc/platforms/52xx/mpc52xx_pm.c linux-2.6.37.4/arch/powerpc/platforms/52xx/mpc52xx_pm.c
---- linux-2.6.37.4/arch/powerpc/platforms/52xx/mpc52xx_pm.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/powerpc/platforms/52xx/mpc52xx_pm.c 2011-01-17 02:41:00.000000000 -0500
-@@ -186,7 +186,7 @@ void mpc52xx_pm_finish(void)
- iounmap(mbar);
- }
-
--static struct platform_suspend_ops mpc52xx_pm_ops = {
-+static const struct platform_suspend_ops mpc52xx_pm_ops = {
- .valid = mpc52xx_pm_valid,
- .prepare = mpc52xx_pm_prepare,
- .enter = mpc52xx_pm_enter,
-diff -urNp linux-2.6.37.4/arch/powerpc/platforms/83xx/suspend.c linux-2.6.37.4/arch/powerpc/platforms/83xx/suspend.c
---- linux-2.6.37.4/arch/powerpc/platforms/83xx/suspend.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/powerpc/platforms/83xx/suspend.c 2011-01-17 02:41:00.000000000 -0500
-@@ -311,7 +311,7 @@ static int mpc83xx_is_pci_agent(void)
- return ret;
- }
-
--static struct platform_suspend_ops mpc83xx_suspend_ops = {
-+static const struct platform_suspend_ops mpc83xx_suspend_ops = {
- .valid = mpc83xx_suspend_valid,
- .begin = mpc83xx_suspend_begin,
- .enter = mpc83xx_suspend_enter,
-diff -urNp linux-2.6.37.4/arch/powerpc/platforms/cell/iommu.c linux-2.6.37.4/arch/powerpc/platforms/cell/iommu.c
---- linux-2.6.37.4/arch/powerpc/platforms/cell/iommu.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/powerpc/platforms/cell/iommu.c 2011-01-17 02:41:00.000000000 -0500
-@@ -642,7 +642,7 @@ static int dma_fixed_dma_supported(struc
-
- static int dma_set_mask_and_switch(struct device *dev, u64 dma_mask);
-
--struct dma_map_ops dma_iommu_fixed_ops = {
-+const struct dma_map_ops dma_iommu_fixed_ops = {
- .alloc_coherent = dma_fixed_alloc_coherent,
- .free_coherent = dma_fixed_free_coherent,
- .map_sg = dma_fixed_map_sg,
-diff -urNp linux-2.6.37.4/arch/powerpc/platforms/ps3/system-bus.c linux-2.6.37.4/arch/powerpc/platforms/ps3/system-bus.c
---- linux-2.6.37.4/arch/powerpc/platforms/ps3/system-bus.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/powerpc/platforms/ps3/system-bus.c 2011-01-17 02:41:00.000000000 -0500
-@@ -695,7 +695,7 @@ static int ps3_dma_supported(struct devi
- return mask >= DMA_BIT_MASK(32);
- }
-
--static struct dma_map_ops ps3_sb_dma_ops = {
-+static const struct dma_map_ops ps3_sb_dma_ops = {
- .alloc_coherent = ps3_alloc_coherent,
- .free_coherent = ps3_free_coherent,
- .map_sg = ps3_sb_map_sg,
-@@ -705,7 +705,7 @@ static struct dma_map_ops ps3_sb_dma_ops
- .unmap_page = ps3_unmap_page,
- };
-
--static struct dma_map_ops ps3_ioc0_dma_ops = {
-+static const struct dma_map_ops ps3_ioc0_dma_ops = {
- .alloc_coherent = ps3_alloc_coherent,
- .free_coherent = ps3_free_coherent,
- .map_sg = ps3_ioc0_map_sg,
-diff -urNp linux-2.6.37.4/arch/powerpc/platforms/pseries/suspend.c linux-2.6.37.4/arch/powerpc/platforms/pseries/suspend.c
---- linux-2.6.37.4/arch/powerpc/platforms/pseries/suspend.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/powerpc/platforms/pseries/suspend.c 2011-01-17 02:41:00.000000000 -0500
-@@ -153,7 +153,7 @@ static struct sysdev_class suspend_sysde
- .name = "power",
- };
-
--static struct platform_suspend_ops pseries_suspend_ops = {
-+static const struct platform_suspend_ops pseries_suspend_ops = {
- .valid = suspend_valid_only_mem,
- .begin = pseries_suspend_begin,
- .prepare_late = pseries_prepare_late,
-diff -urNp linux-2.6.37.4/arch/powerpc/sysdev/fsl_pmc.c linux-2.6.37.4/arch/powerpc/sysdev/fsl_pmc.c
---- linux-2.6.37.4/arch/powerpc/sysdev/fsl_pmc.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/powerpc/sysdev/fsl_pmc.c 2011-01-17 02:41:00.000000000 -0500
-@@ -53,7 +53,7 @@ static int pmc_suspend_valid(suspend_sta
- return 1;
- }
-
--static struct platform_suspend_ops pmc_suspend_ops = {
-+static const struct platform_suspend_ops pmc_suspend_ops = {
- .valid = pmc_suspend_valid,
- .enter = pmc_suspend_enter,
- };
-diff -urNp linux-2.6.37.4/arch/s390/include/asm/elf.h linux-2.6.37.4/arch/s390/include/asm/elf.h
---- linux-2.6.37.4/arch/s390/include/asm/elf.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/s390/include/asm/elf.h 2011-01-17 02:41:00.000000000 -0500
-@@ -163,6 +163,13 @@ extern unsigned int vdso_enabled;
- that it will "exec", and that there is sufficient room for the brk. */
- #define ELF_ET_DYN_BASE (STACK_TOP / 3 * 2)
-
-+#ifdef CONFIG_PAX_ASLR
-+#define PAX_ELF_ET_DYN_BASE (test_thread_flag(TIF_31BIT) ? 0x10000UL : 0x80000000UL)
-+
-+#define PAX_DELTA_MMAP_LEN (test_thread_flag(TIF_31BIT) ? 15 : 26 )
-+#define PAX_DELTA_STACK_LEN (test_thread_flag(TIF_31BIT) ? 15 : 26 )
-+#endif
-+
- /* This yields a mask that user programs can use to figure out what
- instruction set this CPU supports. */
-
-diff -urNp linux-2.6.37.4/arch/s390/include/asm/uaccess.h linux-2.6.37.4/arch/s390/include/asm/uaccess.h
---- linux-2.6.37.4/arch/s390/include/asm/uaccess.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/s390/include/asm/uaccess.h 2011-01-17 02:41:00.000000000 -0500
-@@ -234,6 +234,10 @@ static inline unsigned long __must_check
- copy_to_user(void __user *to, const void *from, unsigned long n)
- {
- might_fault();
-+
-+ if ((long)n < 0)
-+ return n;
-+
- if (access_ok(VERIFY_WRITE, to, n))
- n = __copy_to_user(to, from, n);
- return n;
-@@ -259,6 +263,9 @@ copy_to_user(void __user *to, const void
- static inline unsigned long __must_check
- __copy_from_user(void *to, const void __user *from, unsigned long n)
- {
-+ if ((long)n < 0)
-+ return n;
-+
- if (__builtin_constant_p(n) && (n <= 256))
- return uaccess.copy_from_user_small(n, from, to);
- else
-@@ -293,6 +300,10 @@ copy_from_user(void *to, const void __us
- unsigned int sz = __compiletime_object_size(to);
-
- might_fault();
-+
-+ if ((long)n < 0)
-+ return n;
-+
- if (unlikely(sz != -1 && sz < n)) {
- copy_from_user_overflow();
- return n;
-diff -urNp linux-2.6.37.4/arch/s390/Kconfig linux-2.6.37.4/arch/s390/Kconfig
---- linux-2.6.37.4/arch/s390/Kconfig 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/s390/Kconfig 2011-01-17 02:41:00.000000000 -0500
-@@ -242,13 +242,12 @@ config AUDIT_ARCH
-
- config S390_EXEC_PROTECT
- bool "Data execute protection"
-+ default y
- help
- This option allows to enable a buffer overflow protection for user
-- space programs and it also selects the addressing mode option above.
-- The kernel parameter noexec=on will enable this feature and also
-- switch the addressing modes, default is disabled. Enabling this (via
-- kernel parameter) on machines earlier than IBM System z9 this will
-- reduce system performance.
-+ space programs.
-+ Enabling this (via kernel parameter) on machines earlier than IBM
-+ System z9 will reduce system performance.
-
- comment "Code generation options"
-
-diff -urNp linux-2.6.37.4/arch/s390/kernel/module.c linux-2.6.37.4/arch/s390/kernel/module.c
---- linux-2.6.37.4/arch/s390/kernel/module.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/s390/kernel/module.c 2011-01-17 02:41:00.000000000 -0500
-@@ -168,11 +168,11 @@ module_frob_arch_sections(Elf_Ehdr *hdr,
-
- /* Increase core size by size of got & plt and set start
- offsets for got and plt. */
-- me->core_size = ALIGN(me->core_size, 4);
-- me->arch.got_offset = me->core_size;
-- me->core_size += me->arch.got_size;
-- me->arch.plt_offset = me->core_size;
-- me->core_size += me->arch.plt_size;
-+ me->core_size_rw = ALIGN(me->core_size_rw, 4);
-+ me->arch.got_offset = me->core_size_rw;
-+ me->core_size_rw += me->arch.got_size;
-+ me->arch.plt_offset = me->core_size_rx;
-+ me->core_size_rx += me->arch.plt_size;
- return 0;
- }
-
-@@ -258,7 +258,7 @@ apply_rela(Elf_Rela *rela, Elf_Addr base
- if (info->got_initialized == 0) {
- Elf_Addr *gotent;
-
-- gotent = me->module_core + me->arch.got_offset +
-+ gotent = me->module_core_rw + me->arch.got_offset +
- info->got_offset;
- *gotent = val;
- info->got_initialized = 1;
-@@ -282,7 +282,7 @@ apply_rela(Elf_Rela *rela, Elf_Addr base
- else if (r_type == R_390_GOTENT ||
- r_type == R_390_GOTPLTENT)
- *(unsigned int *) loc =
-- (val + (Elf_Addr) me->module_core - loc) >> 1;
-+ (val + (Elf_Addr) me->module_core_rw - loc) >> 1;
- else if (r_type == R_390_GOT64 ||
- r_type == R_390_GOTPLT64)
- *(unsigned long *) loc = val;
-@@ -296,7 +296,7 @@ apply_rela(Elf_Rela *rela, Elf_Addr base
- case R_390_PLTOFF64: /* 16 bit offset from GOT to PLT. */
- if (info->plt_initialized == 0) {
- unsigned int *ip;
-- ip = me->module_core + me->arch.plt_offset +
-+ ip = me->module_core_rx + me->arch.plt_offset +
- info->plt_offset;
- #ifndef CONFIG_64BIT
- ip[0] = 0x0d105810; /* basr 1,0; l 1,6(1); br 1 */
-@@ -321,7 +321,7 @@ apply_rela(Elf_Rela *rela, Elf_Addr base
- val - loc + 0xffffUL < 0x1ffffeUL) ||
- (r_type == R_390_PLT32DBL &&
- val - loc + 0xffffffffULL < 0x1fffffffeULL)))
-- val = (Elf_Addr) me->module_core +
-+ val = (Elf_Addr) me->module_core_rx +
- me->arch.plt_offset +
- info->plt_offset;
- val += rela->r_addend - loc;
-@@ -343,7 +343,7 @@ apply_rela(Elf_Rela *rela, Elf_Addr base
- case R_390_GOTOFF32: /* 32 bit offset to GOT. */
- case R_390_GOTOFF64: /* 64 bit offset to GOT. */
- val = val + rela->r_addend -
-- ((Elf_Addr) me->module_core + me->arch.got_offset);
-+ ((Elf_Addr) me->module_core_rw + me->arch.got_offset);
- if (r_type == R_390_GOTOFF16)
- *(unsigned short *) loc = val;
- else if (r_type == R_390_GOTOFF32)
-@@ -353,7 +353,7 @@ apply_rela(Elf_Rela *rela, Elf_Addr base
- break;
- case R_390_GOTPC: /* 32 bit PC relative offset to GOT. */
- case R_390_GOTPCDBL: /* 32 bit PC rel. off. to GOT shifted by 1. */
-- val = (Elf_Addr) me->module_core + me->arch.got_offset +
-+ val = (Elf_Addr) me->module_core_rw + me->arch.got_offset +
- rela->r_addend - loc;
- if (r_type == R_390_GOTPC)
- *(unsigned int *) loc = val;
-diff -urNp linux-2.6.37.4/arch/s390/kernel/setup.c linux-2.6.37.4/arch/s390/kernel/setup.c
---- linux-2.6.37.4/arch/s390/kernel/setup.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/s390/kernel/setup.c 2011-01-17 02:41:00.000000000 -0500
-@@ -281,7 +281,7 @@ static int __init early_parse_mem(char *
- }
- early_param("mem", early_parse_mem);
-
--unsigned int user_mode = HOME_SPACE_MODE;
-+unsigned int user_mode = SECONDARY_SPACE_MODE;
- EXPORT_SYMBOL_GPL(user_mode);
-
- static int set_amode_and_uaccess(unsigned long user_amode,
-@@ -310,17 +310,6 @@ static int set_amode_and_uaccess(unsigne
- }
- }
-
--/*
-- * Switch kernel/user addressing modes?
-- */
--static int __init early_parse_switch_amode(char *p)
--{
-- if (user_mode != SECONDARY_SPACE_MODE)
-- user_mode = PRIMARY_SPACE_MODE;
-- return 0;
--}
--early_param("switch_amode", early_parse_switch_amode);
--
- static int __init early_parse_user_mode(char *p)
- {
- if (p && strcmp(p, "primary") == 0)
-@@ -337,20 +326,6 @@ static int __init early_parse_user_mode(
- }
- early_param("user_mode", early_parse_user_mode);
-
--#ifdef CONFIG_S390_EXEC_PROTECT
--/*
-- * Enable execute protection?
-- */
--static int __init early_parse_noexec(char *p)
--{
-- if (!strncmp(p, "off", 3))
-- return 0;
-- user_mode = SECONDARY_SPACE_MODE;
-- return 0;
--}
--early_param("noexec", early_parse_noexec);
--#endif /* CONFIG_S390_EXEC_PROTECT */
--
- static void setup_addressing_mode(void)
- {
- if (user_mode == SECONDARY_SPACE_MODE) {
-diff -urNp linux-2.6.37.4/arch/s390/mm/maccess.c linux-2.6.37.4/arch/s390/mm/maccess.c
---- linux-2.6.37.4/arch/s390/mm/maccess.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/s390/mm/maccess.c 2011-01-17 02:41:00.000000000 -0500
-@@ -45,7 +45,7 @@ static long probe_kernel_write_odd(void
- return rc ? rc : count;
- }
-
--long probe_kernel_write(void *dst, void *src, size_t size)
-+long probe_kernel_write(void *dst, const void *src, size_t size)
- {
- long copied = 0;
-
-diff -urNp linux-2.6.37.4/arch/s390/mm/mmap.c linux-2.6.37.4/arch/s390/mm/mmap.c
---- linux-2.6.37.4/arch/s390/mm/mmap.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/s390/mm/mmap.c 2011-01-17 02:41:00.000000000 -0500
-@@ -78,10 +78,22 @@ void arch_pick_mmap_layout(struct mm_str
- */
- if (mmap_is_legacy()) {
- mm->mmap_base = TASK_UNMAPPED_BASE;
-+
-+#ifdef CONFIG_PAX_RANDMMAP
-+ if (mm->pax_flags & MF_PAX_RANDMMAP)
-+ mm->mmap_base += mm->delta_mmap;
-+#endif
-+
- mm->get_unmapped_area = arch_get_unmapped_area;
- mm->unmap_area = arch_unmap_area;
- } else {
- mm->mmap_base = mmap_base();
-+
-+#ifdef CONFIG_PAX_RANDMMAP
-+ if (mm->pax_flags & MF_PAX_RANDMMAP)
-+ mm->mmap_base -= mm->delta_mmap + mm->delta_stack;
-+#endif
-+
- mm->get_unmapped_area = arch_get_unmapped_area_topdown;
- mm->unmap_area = arch_unmap_area_topdown;
- }
-@@ -153,10 +165,22 @@ void arch_pick_mmap_layout(struct mm_str
- */
- if (mmap_is_legacy()) {
- mm->mmap_base = TASK_UNMAPPED_BASE;
-+
-+#ifdef CONFIG_PAX_RANDMMAP
-+ if (mm->pax_flags & MF_PAX_RANDMMAP)
-+ mm->mmap_base += mm->delta_mmap;
-+#endif
-+
- mm->get_unmapped_area = s390_get_unmapped_area;
- mm->unmap_area = arch_unmap_area;
- } else {
- mm->mmap_base = mmap_base();
-+
-+#ifdef CONFIG_PAX_RANDMMAP
-+ if (mm->pax_flags & MF_PAX_RANDMMAP)
-+ mm->mmap_base -= mm->delta_mmap + mm->delta_stack;
-+#endif
-+
- mm->get_unmapped_area = s390_get_unmapped_area_topdown;
- mm->unmap_area = arch_unmap_area_topdown;
- }
-diff -urNp linux-2.6.37.4/arch/score/include/asm/system.h linux-2.6.37.4/arch/score/include/asm/system.h
---- linux-2.6.37.4/arch/score/include/asm/system.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/score/include/asm/system.h 2011-01-17 02:41:00.000000000 -0500
-@@ -17,7 +17,7 @@ do { \
- #define finish_arch_switch(prev) do {} while (0)
-
- typedef void (*vi_handler_t)(void);
--extern unsigned long arch_align_stack(unsigned long sp);
-+#define arch_align_stack(x) (x)
-
- #define mb() barrier()
- #define rmb() barrier()
-diff -urNp linux-2.6.37.4/arch/score/kernel/process.c linux-2.6.37.4/arch/score/kernel/process.c
---- linux-2.6.37.4/arch/score/kernel/process.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/score/kernel/process.c 2011-01-17 02:41:00.000000000 -0500
-@@ -161,8 +161,3 @@ unsigned long get_wchan(struct task_stru
-
- return task_pt_regs(task)->cp0_epc;
- }
--
--unsigned long arch_align_stack(unsigned long sp)
--{
-- return sp;
--}
-diff -urNp linux-2.6.37.4/arch/sh/boards/mach-hp6xx/pm.c linux-2.6.37.4/arch/sh/boards/mach-hp6xx/pm.c
---- linux-2.6.37.4/arch/sh/boards/mach-hp6xx/pm.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/sh/boards/mach-hp6xx/pm.c 2011-01-17 02:41:00.000000000 -0500
-@@ -143,7 +143,7 @@ static int hp6x0_pm_enter(suspend_state_
- return 0;
- }
-
--static struct platform_suspend_ops hp6x0_pm_ops = {
-+static const struct platform_suspend_ops hp6x0_pm_ops = {
- .enter = hp6x0_pm_enter,
- .valid = suspend_valid_only_mem,
- };
-diff -urNp linux-2.6.37.4/arch/sh/include/asm/dma-mapping.h linux-2.6.37.4/arch/sh/include/asm/dma-mapping.h
---- linux-2.6.37.4/arch/sh/include/asm/dma-mapping.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/sh/include/asm/dma-mapping.h 2011-01-17 02:41:00.000000000 -0500
-@@ -1,10 +1,10 @@
- #ifndef __ASM_SH_DMA_MAPPING_H
- #define __ASM_SH_DMA_MAPPING_H
-
--extern struct dma_map_ops *dma_ops;
-+extern const struct dma_map_ops *dma_ops;
- extern void no_iommu_init(void);
-
--static inline struct dma_map_ops *get_dma_ops(struct device *dev)
-+static inline const struct dma_map_ops *get_dma_ops(struct device *dev)
- {
- return dma_ops;
- }
-@@ -14,7 +14,7 @@ static inline struct dma_map_ops *get_dm
-
- static inline int dma_supported(struct device *dev, u64 mask)
- {
-- struct dma_map_ops *ops = get_dma_ops(dev);
-+ const struct dma_map_ops *ops = get_dma_ops(dev);
-
- if (ops->dma_supported)
- return ops->dma_supported(dev, mask);
-@@ -24,7 +24,7 @@ static inline int dma_supported(struct d
-
- static inline int dma_set_mask(struct device *dev, u64 mask)
- {
-- struct dma_map_ops *ops = get_dma_ops(dev);
-+ const struct dma_map_ops *ops = get_dma_ops(dev);
-
- if (!dev->dma_mask || !dma_supported(dev, mask))
- return -EIO;
-@@ -44,7 +44,7 @@ void dma_cache_sync(struct device *dev,
-
- static inline int dma_mapping_error(struct device *dev, dma_addr_t dma_addr)
- {
-- struct dma_map_ops *ops = get_dma_ops(dev);
-+ const struct dma_map_ops *ops = get_dma_ops(dev);
-
- if (ops->mapping_error)
- return ops->mapping_error(dev, dma_addr);
-@@ -55,7 +55,7 @@ static inline int dma_mapping_error(stru
- static inline void *dma_alloc_coherent(struct device *dev, size_t size,
- dma_addr_t *dma_handle, gfp_t gfp)
- {
-- struct dma_map_ops *ops = get_dma_ops(dev);
-+ const struct dma_map_ops *ops = get_dma_ops(dev);
- void *memory;
-
- if (dma_alloc_from_coherent(dev, size, dma_handle, &memory))
-@@ -72,7 +72,7 @@ static inline void *dma_alloc_coherent(s
- static inline void dma_free_coherent(struct device *dev, size_t size,
- void *vaddr, dma_addr_t dma_handle)
- {
-- struct dma_map_ops *ops = get_dma_ops(dev);
-+ const struct dma_map_ops *ops = get_dma_ops(dev);
-
- if (dma_release_from_coherent(dev, get_order(size), vaddr))
- return;
-diff -urNp linux-2.6.37.4/arch/sh/kernel/cpu/shmobile/pm.c linux-2.6.37.4/arch/sh/kernel/cpu/shmobile/pm.c
---- linux-2.6.37.4/arch/sh/kernel/cpu/shmobile/pm.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/sh/kernel/cpu/shmobile/pm.c 2011-01-17 02:41:00.000000000 -0500
-@@ -141,7 +141,7 @@ static int sh_pm_enter(suspend_state_t s
- return 0;
- }
-
--static struct platform_suspend_ops sh_pm_ops = {
-+static const struct platform_suspend_ops sh_pm_ops = {
- .enter = sh_pm_enter,
- .valid = suspend_valid_only_mem,
- };
-diff -urNp linux-2.6.37.4/arch/sh/kernel/dma-nommu.c linux-2.6.37.4/arch/sh/kernel/dma-nommu.c
---- linux-2.6.37.4/arch/sh/kernel/dma-nommu.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/sh/kernel/dma-nommu.c 2011-01-17 02:41:00.000000000 -0500
-@@ -62,7 +62,7 @@ static void nommu_sync_sg(struct device
- }
- #endif
-
--struct dma_map_ops nommu_dma_ops = {
-+const struct dma_map_ops nommu_dma_ops = {
- .alloc_coherent = dma_generic_alloc_coherent,
- .free_coherent = dma_generic_free_coherent,
- .map_page = nommu_map_page,
-diff -urNp linux-2.6.37.4/arch/sh/kernel/kgdb.c linux-2.6.37.4/arch/sh/kernel/kgdb.c
---- linux-2.6.37.4/arch/sh/kernel/kgdb.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/sh/kernel/kgdb.c 2011-01-17 02:41:00.000000000 -0500
-@@ -319,7 +319,7 @@ void kgdb_arch_exit(void)
- unregister_die_notifier(&kgdb_notifier);
- }
-
--struct kgdb_arch arch_kgdb_ops = {
-+const struct kgdb_arch arch_kgdb_ops = {
- /* Breakpoint instruction: trapa #0x3c */
- #ifdef CONFIG_CPU_LITTLE_ENDIAN
- .gdb_bpt_instr = { 0x3c, 0xc3 },
-diff -urNp linux-2.6.37.4/arch/sh/mm/consistent.c linux-2.6.37.4/arch/sh/mm/consistent.c
---- linux-2.6.37.4/arch/sh/mm/consistent.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/sh/mm/consistent.c 2011-01-17 02:41:00.000000000 -0500
-@@ -22,7 +22,7 @@
-
- #define PREALLOC_DMA_DEBUG_ENTRIES 4096
-
--struct dma_map_ops *dma_ops;
-+const struct dma_map_ops *dma_ops;
- EXPORT_SYMBOL(dma_ops);
-
- static int __init dma_init(void)
-diff -urNp linux-2.6.37.4/arch/sh/mm/mmap.c linux-2.6.37.4/arch/sh/mm/mmap.c
---- linux-2.6.37.4/arch/sh/mm/mmap.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/sh/mm/mmap.c 2011-03-21 17:35:08.000000000 -0400
-@@ -74,8 +74,7 @@ unsigned long arch_get_unmapped_area(str
- addr = PAGE_ALIGN(addr);
-
- vma = find_vma(mm, addr);
-- if (TASK_SIZE - len >= addr &&
-- (!vma || addr + len <= vma->vm_start))
-+ if (TASK_SIZE - len >= addr && check_heap_stack_gap(vma, addr, len))
- return addr;
- }
-
-@@ -106,7 +105,7 @@ full_search:
- }
- return -ENOMEM;
- }
-- if (likely(!vma || addr + len <= vma->vm_start)) {
-+ if (likely(check_heap_stack_gap(vma, addr, len))) {
- /*
- * Remember the place where we stopped the search:
- */
-@@ -157,8 +156,7 @@ arch_get_unmapped_area_topdown(struct fi
- addr = PAGE_ALIGN(addr);
-
- vma = find_vma(mm, addr);
-- if (TASK_SIZE - len >= addr &&
-- (!vma || addr + len <= vma->vm_start))
-+ if (TASK_SIZE - len >= addr && check_heap_stack_gap(vma, addr, len))
- return addr;
- }
-
-@@ -179,7 +177,7 @@ arch_get_unmapped_area_topdown(struct fi
- /* make sure it can fit in the remaining address space */
- if (likely(addr > len)) {
- vma = find_vma(mm, addr-len);
-- if (!vma || addr <= vma->vm_start) {
-+ if (check_heap_stack_gap(vma, addr - len, len)) {
- /* remember the address as a hint for next time */
- return (mm->free_area_cache = addr-len);
- }
-@@ -188,18 +186,18 @@ arch_get_unmapped_area_topdown(struct fi
- if (unlikely(mm->mmap_base < len))
- goto bottomup;
-
-- addr = mm->mmap_base-len;
-- if (do_colour_align)
-- addr = COLOUR_ALIGN_DOWN(addr, pgoff);
-+ addr = mm->mmap_base - len;
-
- do {
-+ if (do_colour_align)
-+ addr = COLOUR_ALIGN_DOWN(addr, pgoff);
- /*
- * Lookup failure means no vma is above this address,
- * else if new region fits below vma->vm_start,
- * return with success:
- */
- vma = find_vma(mm, addr);
-- if (likely(!vma || addr+len <= vma->vm_start)) {
-+ if (likely(check_heap_stack_gap(vma, addr, len))) {
- /* remember the address as a hint for next time */
- return (mm->free_area_cache = addr);
- }
-@@ -209,10 +207,8 @@ arch_get_unmapped_area_topdown(struct fi
- mm->cached_hole_size = vma->vm_start - addr;
-
- /* try just below the current vma->vm_start */
-- addr = vma->vm_start-len;
-- if (do_colour_align)
-- addr = COLOUR_ALIGN_DOWN(addr, pgoff);
-- } while (likely(len < vma->vm_start));
-+ addr = skip_heap_stack_gap(vma, len);
-+ } while (!IS_ERR_VALUE(addr));
-
- bottomup:
- /*
-diff -urNp linux-2.6.37.4/arch/sparc/include/asm/atomic_64.h linux-2.6.37.4/arch/sparc/include/asm/atomic_64.h
---- linux-2.6.37.4/arch/sparc/include/asm/atomic_64.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/sparc/include/asm/atomic_64.h 2011-01-17 02:41:00.000000000 -0500
-@@ -14,18 +14,40 @@
- #define ATOMIC64_INIT(i) { (i) }
-
- #define atomic_read(v) (*(volatile int *)&(v)->counter)
-+static inline int atomic_read_unchecked(const atomic_unchecked_t *v)
-+{
-+ return v->counter;
-+}
- #define atomic64_read(v) (*(volatile long *)&(v)->counter)
-+static inline long atomic64_read_unchecked(const atomic64_unchecked_t *v)
-+{
-+ return v->counter;
-+}
-
- #define atomic_set(v, i) (((v)->counter) = i)
-+static inline void atomic_set_unchecked(atomic_unchecked_t *v, int i)
-+{
-+ v->counter = i;
-+}
- #define atomic64_set(v, i) (((v)->counter) = i)
-+static inline void atomic64_set_unchecked(atomic64_unchecked_t *v, long i)
-+{
-+ v->counter = i;
-+}
-
- extern void atomic_add(int, atomic_t *);
-+extern void atomic_add_unchecked(int, atomic_unchecked_t *);
- extern void atomic64_add(long, atomic64_t *);
-+extern void atomic64_add_unchecked(long, atomic64_unchecked_t *);
- extern void atomic_sub(int, atomic_t *);
-+extern void atomic_sub_unchecked(int, atomic_unchecked_t *);
- extern void atomic64_sub(long, atomic64_t *);
-+extern void atomic64_sub_unchecked(long, atomic64_unchecked_t *);
-
- extern int atomic_add_ret(int, atomic_t *);
-+extern int atomic_add_ret_unchecked(int, atomic_unchecked_t *);
- extern long atomic64_add_ret(long, atomic64_t *);
-+extern long atomic64_add_ret_unchecked(long, atomic64_unchecked_t *);
- extern int atomic_sub_ret(int, atomic_t *);
- extern long atomic64_sub_ret(long, atomic64_t *);
-
-@@ -33,12 +55,24 @@ extern long atomic64_sub_ret(long, atomi
- #define atomic64_dec_return(v) atomic64_sub_ret(1, v)
-
- #define atomic_inc_return(v) atomic_add_ret(1, v)
-+static inline int atomic_inc_return_unchecked(atomic_unchecked_t *v)
-+{
-+ return atomic_add_ret_unchecked(1, v);
-+}
- #define atomic64_inc_return(v) atomic64_add_ret(1, v)
-+static inline long atomic64_inc_return_unchecked(atomic64_unchecked_t *v)
-+{
-+ return atomic64_add_ret_unchecked(1, v);
-+}
-
- #define atomic_sub_return(i, v) atomic_sub_ret(i, v)
- #define atomic64_sub_return(i, v) atomic64_sub_ret(i, v)
-
- #define atomic_add_return(i, v) atomic_add_ret(i, v)
-+static inline int atomic_add_return_unchecked(int i, atomic_unchecked_t *v)
-+{
-+ return atomic_add_ret_unchecked(i, v);
-+}
- #define atomic64_add_return(i, v) atomic64_add_ret(i, v)
-
- /*
-@@ -59,10 +93,26 @@ extern long atomic64_sub_ret(long, atomi
- #define atomic64_dec_and_test(v) (atomic64_sub_ret(1, v) == 0)
-
- #define atomic_inc(v) atomic_add(1, v)
-+static inline void atomic_inc_unchecked(atomic_unchecked_t *v)
-+{
-+ atomic_add_unchecked(1, v);
-+}
- #define atomic64_inc(v) atomic64_add(1, v)
-+static inline void atomic64_inc_unchecked(atomic64_unchecked_t *v)
-+{
-+ atomic64_add_unchecked(1, v);
-+}
-
- #define atomic_dec(v) atomic_sub(1, v)
-+static inline void atomic_dec_unchecked(atomic_unchecked_t *v)
-+{
-+ atomic_sub_unchecked(1, v);
-+}
- #define atomic64_dec(v) atomic64_sub(1, v)
-+static inline void atomic64_dec_unchecked(atomic64_unchecked_t *v)
-+{
-+ atomic64_sub_unchecked(1, v);
-+}
-
- #define atomic_add_negative(i, v) (atomic_add_ret(i, v) < 0)
- #define atomic64_add_negative(i, v) (atomic64_add_ret(i, v) < 0)
-@@ -72,17 +122,28 @@ extern long atomic64_sub_ret(long, atomi
-
- static inline int atomic_add_unless(atomic_t *v, int a, int u)
- {
-- int c, old;
-+ int c, old, new;
- c = atomic_read(v);
- for (;;) {
-- if (unlikely(c == (u)))
-+ if (unlikely(c == u))
- break;
-- old = atomic_cmpxchg((v), c, c + (a));
-+
-+ asm volatile("addcc %2, %0, %0\n"
-+
-+#ifdef CONFIG_PAX_REFCOUNT
-+ "tvs %%icc, 6\n"
-+#endif
-+
-+ : "=r" (new)
-+ : "0" (c), "ir" (a)
-+ : "cc");
-+
-+ old = atomic_cmpxchg(v, c, new);
- if (likely(old == c))
- break;
- c = old;
- }
-- return c != (u);
-+ return c != u;
- }
-
- #define atomic_inc_not_zero(v) atomic_add_unless((v), 1, 0)
-@@ -93,17 +154,28 @@ static inline int atomic_add_unless(atom
-
- static inline long atomic64_add_unless(atomic64_t *v, long a, long u)
- {
-- long c, old;
-+ long c, old, new;
- c = atomic64_read(v);
- for (;;) {
-- if (unlikely(c == (u)))
-+ if (unlikely(c == u))
- break;
-- old = atomic64_cmpxchg((v), c, c + (a));
-+
-+ asm volatile("addcc %2, %0, %0\n"
-+
-+#ifdef CONFIG_PAX_REFCOUNT
-+ "tvs %%xcc, 6\n"
-+#endif
-+
-+ : "=r" (new)
-+ : "0" (c), "ir" (a)
-+ : "cc");
-+
-+ old = atomic64_cmpxchg(v, c, new);
- if (likely(old == c))
- break;
- c = old;
- }
-- return c != (u);
-+ return c != u;
- }
-
- #define atomic64_inc_not_zero(v) atomic64_add_unless((v), 1, 0)
-diff -urNp linux-2.6.37.4/arch/sparc/include/asm/dma-mapping.h linux-2.6.37.4/arch/sparc/include/asm/dma-mapping.h
---- linux-2.6.37.4/arch/sparc/include/asm/dma-mapping.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/sparc/include/asm/dma-mapping.h 2011-01-17 02:41:00.000000000 -0500
-@@ -12,10 +12,10 @@ extern int dma_supported(struct device *
- #define dma_alloc_noncoherent(d, s, h, f) dma_alloc_coherent(d, s, h, f)
- #define dma_free_noncoherent(d, s, v, h) dma_free_coherent(d, s, v, h)
-
--extern struct dma_map_ops *dma_ops, pci32_dma_ops;
-+extern const struct dma_map_ops *dma_ops, pci32_dma_ops;
- extern struct bus_type pci_bus_type;
-
--static inline struct dma_map_ops *get_dma_ops(struct device *dev)
-+static inline const struct dma_map_ops *get_dma_ops(struct device *dev)
- {
- #if defined(CONFIG_SPARC32) && defined(CONFIG_PCI)
- if (dev->bus == &pci_bus_type)
-@@ -29,7 +29,7 @@ static inline struct dma_map_ops *get_dm
- static inline void *dma_alloc_coherent(struct device *dev, size_t size,
- dma_addr_t *dma_handle, gfp_t flag)
- {
-- struct dma_map_ops *ops = get_dma_ops(dev);
-+ const struct dma_map_ops *ops = get_dma_ops(dev);
- void *cpu_addr;
-
- cpu_addr = ops->alloc_coherent(dev, size, dma_handle, flag);
-@@ -40,7 +40,7 @@ static inline void *dma_alloc_coherent(s
- static inline void dma_free_coherent(struct device *dev, size_t size,
- void *cpu_addr, dma_addr_t dma_handle)
- {
-- struct dma_map_ops *ops = get_dma_ops(dev);
-+ const struct dma_map_ops *ops = get_dma_ops(dev);
-
- debug_dma_free_coherent(dev, size, cpu_addr, dma_handle);
- ops->free_coherent(dev, size, cpu_addr, dma_handle);
-diff -urNp linux-2.6.37.4/arch/sparc/include/asm/elf_32.h linux-2.6.37.4/arch/sparc/include/asm/elf_32.h
---- linux-2.6.37.4/arch/sparc/include/asm/elf_32.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/sparc/include/asm/elf_32.h 2011-01-17 02:41:00.000000000 -0500
-@@ -114,6 +114,13 @@ typedef struct {
-
- #define ELF_ET_DYN_BASE (TASK_UNMAPPED_BASE)
-
-+#ifdef CONFIG_PAX_ASLR
-+#define PAX_ELF_ET_DYN_BASE 0x10000UL
-+
-+#define PAX_DELTA_MMAP_LEN 16
-+#define PAX_DELTA_STACK_LEN 16
-+#endif
-+
- /* This yields a mask that user programs can use to figure out what
- instruction set this cpu supports. This can NOT be done in userspace
- on Sparc. */
-diff -urNp linux-2.6.37.4/arch/sparc/include/asm/elf_64.h linux-2.6.37.4/arch/sparc/include/asm/elf_64.h
---- linux-2.6.37.4/arch/sparc/include/asm/elf_64.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/sparc/include/asm/elf_64.h 2011-01-17 02:41:00.000000000 -0500
-@@ -162,6 +162,12 @@ typedef struct {
- #define ELF_ET_DYN_BASE 0x0000010000000000UL
- #define COMPAT_ELF_ET_DYN_BASE 0x0000000070000000UL
-
-+#ifdef CONFIG_PAX_ASLR
-+#define PAX_ELF_ET_DYN_BASE (test_thread_flag(TIF_32BIT) ? 0x10000UL : 0x100000UL)
-+
-+#define PAX_DELTA_MMAP_LEN (test_thread_flag(TIF_32BIT) ? 14 : 28)
-+#define PAX_DELTA_STACK_LEN (test_thread_flag(TIF_32BIT) ? 15 : 29)
-+#endif
-
- /* This yields a mask that user programs can use to figure out what
- instruction set this cpu supports. */
-diff -urNp linux-2.6.37.4/arch/sparc/include/asm/pgtable_32.h linux-2.6.37.4/arch/sparc/include/asm/pgtable_32.h
---- linux-2.6.37.4/arch/sparc/include/asm/pgtable_32.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/sparc/include/asm/pgtable_32.h 2011-01-17 02:41:00.000000000 -0500
-@@ -43,6 +43,13 @@ BTFIXUPDEF_SIMM13(user_ptrs_per_pgd)
- BTFIXUPDEF_INT(page_none)
- BTFIXUPDEF_INT(page_copy)
- BTFIXUPDEF_INT(page_readonly)
-+
-+#ifdef CONFIG_PAX_PAGEEXEC
-+BTFIXUPDEF_INT(page_shared_noexec)
-+BTFIXUPDEF_INT(page_copy_noexec)
-+BTFIXUPDEF_INT(page_readonly_noexec)
-+#endif
-+
- BTFIXUPDEF_INT(page_kernel)
-
- #define PMD_SHIFT SUN4C_PMD_SHIFT
-@@ -64,6 +71,16 @@ extern pgprot_t PAGE_SHARED;
- #define PAGE_COPY __pgprot(BTFIXUP_INT(page_copy))
- #define PAGE_READONLY __pgprot(BTFIXUP_INT(page_readonly))
-
-+#ifdef CONFIG_PAX_PAGEEXEC
-+extern pgprot_t PAGE_SHARED_NOEXEC;
-+# define PAGE_COPY_NOEXEC __pgprot(BTFIXUP_INT(page_copy_noexec))
-+# define PAGE_READONLY_NOEXEC __pgprot(BTFIXUP_INT(page_readonly_noexec))
-+#else
-+# define PAGE_SHARED_NOEXEC PAGE_SHARED
-+# define PAGE_COPY_NOEXEC PAGE_COPY
-+# define PAGE_READONLY_NOEXEC PAGE_READONLY
-+#endif
-+
- extern unsigned long page_kernel;
-
- #ifdef MODULE
-diff -urNp linux-2.6.37.4/arch/sparc/include/asm/pgtsrmmu.h linux-2.6.37.4/arch/sparc/include/asm/pgtsrmmu.h
---- linux-2.6.37.4/arch/sparc/include/asm/pgtsrmmu.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/sparc/include/asm/pgtsrmmu.h 2011-01-17 02:41:00.000000000 -0500
-@@ -115,6 +115,13 @@
- SRMMU_EXEC | SRMMU_REF)
- #define SRMMU_PAGE_RDONLY __pgprot(SRMMU_VALID | SRMMU_CACHE | \
- SRMMU_EXEC | SRMMU_REF)
-+
-+#ifdef CONFIG_PAX_PAGEEXEC
-+#define SRMMU_PAGE_SHARED_NOEXEC __pgprot(SRMMU_VALID | SRMMU_CACHE | SRMMU_WRITE | SRMMU_REF)
-+#define SRMMU_PAGE_COPY_NOEXEC __pgprot(SRMMU_VALID | SRMMU_CACHE | SRMMU_REF)
-+#define SRMMU_PAGE_RDONLY_NOEXEC __pgprot(SRMMU_VALID | SRMMU_CACHE | SRMMU_REF)
-+#endif
-+
- #define SRMMU_PAGE_KERNEL __pgprot(SRMMU_VALID | SRMMU_CACHE | SRMMU_PRIV | \
- SRMMU_DIRTY | SRMMU_REF)
-
-diff -urNp linux-2.6.37.4/arch/sparc/include/asm/spinlock_64.h linux-2.6.37.4/arch/sparc/include/asm/spinlock_64.h
---- linux-2.6.37.4/arch/sparc/include/asm/spinlock_64.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/sparc/include/asm/spinlock_64.h 2011-01-17 02:41:00.000000000 -0500
-@@ -99,7 +99,12 @@ static void inline arch_read_lock(arch_r
- __asm__ __volatile__ (
- "1: ldsw [%2], %0\n"
- " brlz,pn %0, 2f\n"
--"4: add %0, 1, %1\n"
-+"4: addcc %0, 1, %1\n"
-+
-+#ifdef CONFIG_PAX_REFCOUNT
-+" tvs %%icc, 6\n"
-+#endif
-+
- " cas [%2], %0, %1\n"
- " cmp %0, %1\n"
- " bne,pn %%icc, 1b\n"
-@@ -112,7 +117,7 @@ static void inline arch_read_lock(arch_r
- " .previous"
- : "=&r" (tmp1), "=&r" (tmp2)
- : "r" (lock)
-- : "memory");
-+ : "memory", "cc");
- }
-
- static int inline arch_read_trylock(arch_rwlock_t *lock)
-@@ -123,7 +128,12 @@ static int inline arch_read_trylock(arch
- "1: ldsw [%2], %0\n"
- " brlz,a,pn %0, 2f\n"
- " mov 0, %0\n"
--" add %0, 1, %1\n"
-+" addcc %0, 1, %1\n"
-+
-+#ifdef CONFIG_PAX_REFCOUNT
-+" tvs %%icc, 6\n"
-+#endif
-+
- " cas [%2], %0, %1\n"
- " cmp %0, %1\n"
- " bne,pn %%icc, 1b\n"
-@@ -142,7 +152,12 @@ static void inline arch_read_unlock(arch
-
- __asm__ __volatile__(
- "1: lduw [%2], %0\n"
--" sub %0, 1, %1\n"
-+" subcc %0, 1, %1\n"
-+
-+#ifdef CONFIG_PAX_REFCOUNT
-+" tvs %%icc, 6\n"
-+#endif
-+
- " cas [%2], %0, %1\n"
- " cmp %0, %1\n"
- " bne,pn %%xcc, 1b\n"
-diff -urNp linux-2.6.37.4/arch/sparc/include/asm/uaccess_32.h linux-2.6.37.4/arch/sparc/include/asm/uaccess_32.h
---- linux-2.6.37.4/arch/sparc/include/asm/uaccess_32.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/sparc/include/asm/uaccess_32.h 2011-01-17 02:41:00.000000000 -0500
-@@ -249,27 +249,46 @@ extern unsigned long __copy_user(void __
-
- static inline unsigned long copy_to_user(void __user *to, const void *from, unsigned long n)
- {
-- if (n && __access_ok((unsigned long) to, n))
-+ if ((long)n < 0)
-+ return n;
-+
-+ if (n && __access_ok((unsigned long) to, n)) {
-+ if (!__builtin_constant_p(n))
-+ check_object_size(from, n, true);
- return __copy_user(to, (__force void __user *) from, n);
-- else
-+ } else
- return n;
- }
-
- static inline unsigned long __copy_to_user(void __user *to, const void *from, unsigned long n)
- {
-+ if ((long)n < 0)
-+ return n;
-+
-+ if (!__builtin_constant_p(n))
-+ check_object_size(from, n, true);
-+
- return __copy_user(to, (__force void __user *) from, n);
- }
-
- static inline unsigned long copy_from_user(void *to, const void __user *from, unsigned long n)
- {
-- if (n && __access_ok((unsigned long) from, n))
-+ if ((long)n < 0)
-+ return n;
-+
-+ if (n && __access_ok((unsigned long) from, n)) {
-+ if (!__builtin_constant_p(n))
-+ check_object_size(to, n, false);
- return __copy_user((__force void __user *) to, from, n);
-- else
-+ } else
- return n;
- }
-
- static inline unsigned long __copy_from_user(void *to, const void __user *from, unsigned long n)
- {
-+ if ((long)n < 0)
-+ return n;
-+
- return __copy_user((__force void __user *) to, from, n);
- }
-
-diff -urNp linux-2.6.37.4/arch/sparc/include/asm/uaccess_64.h linux-2.6.37.4/arch/sparc/include/asm/uaccess_64.h
---- linux-2.6.37.4/arch/sparc/include/asm/uaccess_64.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/sparc/include/asm/uaccess_64.h 2011-01-17 02:41:00.000000000 -0500
-@@ -10,6 +10,7 @@
- #include <linux/compiler.h>
- #include <linux/string.h>
- #include <linux/thread_info.h>
-+#include <linux/kernel.h>
- #include <asm/asi.h>
- #include <asm/system.h>
- #include <asm/spitfire.h>
-@@ -213,8 +214,15 @@ extern unsigned long copy_from_user_fixu
- static inline unsigned long __must_check
- copy_from_user(void *to, const void __user *from, unsigned long size)
- {
-- unsigned long ret = ___copy_from_user(to, from, size);
-+ unsigned long ret;
-
-+ if ((long)size < 0 || size > INT_MAX)
-+ return size;
-+
-+ if (!__builtin_constant_p(size))
-+ check_object_size(to, size, false);
-+
-+ ret = ___copy_from_user(to, from, size);
- if (unlikely(ret))
- ret = copy_from_user_fixup(to, from, size);
-
-@@ -230,8 +238,15 @@ extern unsigned long copy_to_user_fixup(
- static inline unsigned long __must_check
- copy_to_user(void __user *to, const void *from, unsigned long size)
- {
-- unsigned long ret = ___copy_to_user(to, from, size);
-+ unsigned long ret;
-+
-+ if ((long)size < 0 || size > INT_MAX)
-+ return size;
-+
-+ if (!__builtin_constant_p(size))
-+ check_object_size(from, size, true);
-
-+ ret = ___copy_to_user(to, from, size);
- if (unlikely(ret))
- ret = copy_to_user_fixup(to, from, size);
- return ret;
-diff -urNp linux-2.6.37.4/arch/sparc/include/asm/uaccess.h linux-2.6.37.4/arch/sparc/include/asm/uaccess.h
---- linux-2.6.37.4/arch/sparc/include/asm/uaccess.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/sparc/include/asm/uaccess.h 2011-01-17 02:41:00.000000000 -0500
-@@ -1,5 +1,13 @@
- #ifndef ___ASM_SPARC_UACCESS_H
- #define ___ASM_SPARC_UACCESS_H
-+
-+#ifdef __KERNEL__
-+#ifndef __ASSEMBLY__
-+#include <linux/types.h>
-+extern void check_object_size(const void *ptr, unsigned long n, bool to);
-+#endif
-+#endif
-+
- #if defined(__sparc__) && defined(__arch64__)
- #include <asm/uaccess_64.h>
- #else
-diff -urNp linux-2.6.37.4/arch/sparc/kernel/iommu.c linux-2.6.37.4/arch/sparc/kernel/iommu.c
---- linux-2.6.37.4/arch/sparc/kernel/iommu.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/sparc/kernel/iommu.c 2011-01-17 02:41:00.000000000 -0500
-@@ -828,7 +828,7 @@ static void dma_4u_sync_sg_for_cpu(struc
- spin_unlock_irqrestore(&iommu->lock, flags);
- }
-
--static struct dma_map_ops sun4u_dma_ops = {
-+static const struct dma_map_ops sun4u_dma_ops = {
- .alloc_coherent = dma_4u_alloc_coherent,
- .free_coherent = dma_4u_free_coherent,
- .map_page = dma_4u_map_page,
-@@ -839,7 +839,7 @@ static struct dma_map_ops sun4u_dma_ops
- .sync_sg_for_cpu = dma_4u_sync_sg_for_cpu,
- };
-
--struct dma_map_ops *dma_ops = &sun4u_dma_ops;
-+const struct dma_map_ops *dma_ops = &sun4u_dma_ops;
- EXPORT_SYMBOL(dma_ops);
-
- extern int pci64_dma_supported(struct pci_dev *pdev, u64 device_mask);
-diff -urNp linux-2.6.37.4/arch/sparc/kernel/ioport.c linux-2.6.37.4/arch/sparc/kernel/ioport.c
---- linux-2.6.37.4/arch/sparc/kernel/ioport.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/sparc/kernel/ioport.c 2011-01-17 02:41:00.000000000 -0500
-@@ -397,7 +397,7 @@ static void sbus_sync_sg_for_device(stru
- BUG();
- }
-
--struct dma_map_ops sbus_dma_ops = {
-+const struct dma_map_ops sbus_dma_ops = {
- .alloc_coherent = sbus_alloc_coherent,
- .free_coherent = sbus_free_coherent,
- .map_page = sbus_map_page,
-@@ -408,7 +408,7 @@ struct dma_map_ops sbus_dma_ops = {
- .sync_sg_for_device = sbus_sync_sg_for_device,
- };
-
--struct dma_map_ops *dma_ops = &sbus_dma_ops;
-+const struct dma_map_ops *dma_ops = &sbus_dma_ops;
- EXPORT_SYMBOL(dma_ops);
-
- static int __init sparc_register_ioport(void)
-@@ -645,7 +645,7 @@ static void pci32_sync_sg_for_device(str
- }
- }
-
--struct dma_map_ops pci32_dma_ops = {
-+const struct dma_map_ops pci32_dma_ops = {
- .alloc_coherent = pci32_alloc_coherent,
- .free_coherent = pci32_free_coherent,
- .map_page = pci32_map_page,
-diff -urNp linux-2.6.37.4/arch/sparc/kernel/kgdb_32.c linux-2.6.37.4/arch/sparc/kernel/kgdb_32.c
---- linux-2.6.37.4/arch/sparc/kernel/kgdb_32.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/sparc/kernel/kgdb_32.c 2011-01-17 02:41:00.000000000 -0500
-@@ -164,7 +164,7 @@ void kgdb_arch_set_pc(struct pt_regs *re
- regs->npc = regs->pc + 4;
- }
-
--struct kgdb_arch arch_kgdb_ops = {
-+const struct kgdb_arch arch_kgdb_ops = {
- /* Breakpoint instruction: ta 0x7d */
- .gdb_bpt_instr = { 0x91, 0xd0, 0x20, 0x7d },
- };
-diff -urNp linux-2.6.37.4/arch/sparc/kernel/kgdb_64.c linux-2.6.37.4/arch/sparc/kernel/kgdb_64.c
---- linux-2.6.37.4/arch/sparc/kernel/kgdb_64.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/sparc/kernel/kgdb_64.c 2011-01-17 02:41:00.000000000 -0500
-@@ -187,7 +187,7 @@ void kgdb_arch_set_pc(struct pt_regs *re
- regs->tnpc = regs->tpc + 4;
- }
-
--struct kgdb_arch arch_kgdb_ops = {
-+const struct kgdb_arch arch_kgdb_ops = {
- /* Breakpoint instruction: ta 0x72 */
- .gdb_bpt_instr = { 0x91, 0xd0, 0x20, 0x72 },
- };
-diff -urNp linux-2.6.37.4/arch/sparc/kernel/Makefile linux-2.6.37.4/arch/sparc/kernel/Makefile
---- linux-2.6.37.4/arch/sparc/kernel/Makefile 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/sparc/kernel/Makefile 2011-01-17 02:41:00.000000000 -0500
-@@ -3,7 +3,7 @@
- #
-
- asflags-y := -ansi
--ccflags-y := -Werror
-+#ccflags-y := -Werror
-
- extra-y := head_$(BITS).o
- extra-y += init_task.o
-diff -urNp linux-2.6.37.4/arch/sparc/kernel/pci_sun4v.c linux-2.6.37.4/arch/sparc/kernel/pci_sun4v.c
---- linux-2.6.37.4/arch/sparc/kernel/pci_sun4v.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/sparc/kernel/pci_sun4v.c 2011-01-17 02:41:00.000000000 -0500
-@@ -525,7 +525,7 @@ static void dma_4v_unmap_sg(struct devic
- spin_unlock_irqrestore(&iommu->lock, flags);
- }
-
--static struct dma_map_ops sun4v_dma_ops = {
-+static const struct dma_map_ops sun4v_dma_ops = {
- .alloc_coherent = dma_4v_alloc_coherent,
- .free_coherent = dma_4v_free_coherent,
- .map_page = dma_4v_map_page,
-diff -urNp linux-2.6.37.4/arch/sparc/kernel/process_32.c linux-2.6.37.4/arch/sparc/kernel/process_32.c
---- linux-2.6.37.4/arch/sparc/kernel/process_32.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/sparc/kernel/process_32.c 2011-01-17 02:41:00.000000000 -0500
-@@ -196,7 +196,7 @@ void __show_backtrace(unsigned long fp)
- rw->ins[4], rw->ins[5],
- rw->ins[6],
- rw->ins[7]);
-- printk("%pS\n", (void *) rw->ins[7]);
-+ printk("%pA\n", (void *) rw->ins[7]);
- rw = (struct reg_window32 *) rw->ins[6];
- }
- spin_unlock_irqrestore(&sparc_backtrace_lock, flags);
-@@ -263,14 +263,14 @@ void show_regs(struct pt_regs *r)
-
- printk("PSR: %08lx PC: %08lx NPC: %08lx Y: %08lx %s\n",
- r->psr, r->pc, r->npc, r->y, print_tainted());
-- printk("PC: <%pS>\n", (void *) r->pc);
-+ printk("PC: <%pA>\n", (void *) r->pc);
- printk("%%G: %08lx %08lx %08lx %08lx %08lx %08lx %08lx %08lx\n",
- r->u_regs[0], r->u_regs[1], r->u_regs[2], r->u_regs[3],
- r->u_regs[4], r->u_regs[5], r->u_regs[6], r->u_regs[7]);
- printk("%%O: %08lx %08lx %08lx %08lx %08lx %08lx %08lx %08lx\n",
- r->u_regs[8], r->u_regs[9], r->u_regs[10], r->u_regs[11],
- r->u_regs[12], r->u_regs[13], r->u_regs[14], r->u_regs[15]);
-- printk("RPC: <%pS>\n", (void *) r->u_regs[15]);
-+ printk("RPC: <%pA>\n", (void *) r->u_regs[15]);
-
- printk("%%L: %08lx %08lx %08lx %08lx %08lx %08lx %08lx %08lx\n",
- rw->locals[0], rw->locals[1], rw->locals[2], rw->locals[3],
-@@ -305,7 +305,7 @@ void show_stack(struct task_struct *tsk,
- rw = (struct reg_window32 *) fp;
- pc = rw->ins[7];
- printk("[%08lx : ", pc);
-- printk("%pS ] ", (void *) pc);
-+ printk("%pA ] ", (void *) pc);
- fp = rw->ins[6];
- } while (++count < 16);
- printk("\n");
-diff -urNp linux-2.6.37.4/arch/sparc/kernel/process_64.c linux-2.6.37.4/arch/sparc/kernel/process_64.c
---- linux-2.6.37.4/arch/sparc/kernel/process_64.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/sparc/kernel/process_64.c 2011-01-17 02:41:00.000000000 -0500
-@@ -180,14 +180,14 @@ static void show_regwindow(struct pt_reg
- printk("i4: %016lx i5: %016lx i6: %016lx i7: %016lx\n",
- rwk->ins[4], rwk->ins[5], rwk->ins[6], rwk->ins[7]);
- if (regs->tstate & TSTATE_PRIV)
-- printk("I7: <%pS>\n", (void *) rwk->ins[7]);
-+ printk("I7: <%pA>\n", (void *) rwk->ins[7]);
- }
-
- void show_regs(struct pt_regs *regs)
- {
- printk("TSTATE: %016lx TPC: %016lx TNPC: %016lx Y: %08x %s\n", regs->tstate,
- regs->tpc, regs->tnpc, regs->y, print_tainted());
-- printk("TPC: <%pS>\n", (void *) regs->tpc);
-+ printk("TPC: <%pA>\n", (void *) regs->tpc);
- printk("g0: %016lx g1: %016lx g2: %016lx g3: %016lx\n",
- regs->u_regs[0], regs->u_regs[1], regs->u_regs[2],
- regs->u_regs[3]);
-@@ -200,7 +200,7 @@ void show_regs(struct pt_regs *regs)
- printk("o4: %016lx o5: %016lx sp: %016lx ret_pc: %016lx\n",
- regs->u_regs[12], regs->u_regs[13], regs->u_regs[14],
- regs->u_regs[15]);
-- printk("RPC: <%pS>\n", (void *) regs->u_regs[15]);
-+ printk("RPC: <%pA>\n", (void *) regs->u_regs[15]);
- show_regwindow(regs);
- show_stack(current, (unsigned long *) regs->u_regs[UREG_FP]);
- }
-@@ -285,7 +285,7 @@ void arch_trigger_all_cpu_backtrace(void
- ((tp && tp->task) ? tp->task->pid : -1));
-
- if (gp->tstate & TSTATE_PRIV) {
-- printk(" TPC[%pS] O7[%pS] I7[%pS] RPC[%pS]\n",
-+ printk(" TPC[%pA] O7[%pA] I7[%pA] RPC[%pA]\n",
- (void *) gp->tpc,
- (void *) gp->o7,
- (void *) gp->i7,
-diff -urNp linux-2.6.37.4/arch/sparc/kernel/sys_sparc_32.c linux-2.6.37.4/arch/sparc/kernel/sys_sparc_32.c
---- linux-2.6.37.4/arch/sparc/kernel/sys_sparc_32.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/sparc/kernel/sys_sparc_32.c 2011-01-17 02:41:00.000000000 -0500
-@@ -56,7 +56,7 @@ unsigned long arch_get_unmapped_area(str
- if (ARCH_SUN4C && len > 0x20000000)
- return -ENOMEM;
- if (!addr)
-- addr = TASK_UNMAPPED_BASE;
-+ addr = current->mm->mmap_base;
-
- if (flags & MAP_SHARED)
- addr = COLOUR_ALIGN(addr);
-@@ -71,7 +71,7 @@ unsigned long arch_get_unmapped_area(str
- }
- if (TASK_SIZE - PAGE_SIZE - len < addr)
- return -ENOMEM;
-- if (!vmm || addr + len <= vmm->vm_start)
-+ if (check_heap_stack_gap(vmm, addr, len))
- return addr;
- addr = vmm->vm_end;
- if (flags & MAP_SHARED)
-diff -urNp linux-2.6.37.4/arch/sparc/kernel/sys_sparc_64.c linux-2.6.37.4/arch/sparc/kernel/sys_sparc_64.c
---- linux-2.6.37.4/arch/sparc/kernel/sys_sparc_64.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/sparc/kernel/sys_sparc_64.c 2011-03-21 17:35:08.000000000 -0400
-@@ -124,7 +124,7 @@ unsigned long arch_get_unmapped_area(str
- /* We do not accept a shared mapping if it would violate
- * cache aliasing constraints.
- */
-- if ((flags & MAP_SHARED) &&
-+ if ((filp || (flags & MAP_SHARED)) &&
- ((addr - (pgoff << PAGE_SHIFT)) & (SHMLBA - 1)))
- return -EINVAL;
- return addr;
-@@ -139,6 +139,10 @@ unsigned long arch_get_unmapped_area(str
- if (filp || (flags & MAP_SHARED))
- do_color_align = 1;
-
-+#ifdef CONFIG_PAX_RANDMMAP
-+ if (!(mm->pax_flags & MF_PAX_RANDMMAP))
-+#endif
-+
- if (addr) {
- if (do_color_align)
- addr = COLOUR_ALIGN(addr, pgoff);
-@@ -146,15 +150,14 @@ unsigned long arch_get_unmapped_area(str
- addr = PAGE_ALIGN(addr);
-
- vma = find_vma(mm, addr);
-- if (task_size - len >= addr &&
-- (!vma || addr + len <= vma->vm_start))
-+ if (task_size - len >= addr && check_heap_stack_gap(vma, addr, len))
- return addr;
- }
-
- if (len > mm->cached_hole_size) {
-- start_addr = addr = mm->free_area_cache;
-+ start_addr = addr = mm->free_area_cache;
- } else {
-- start_addr = addr = TASK_UNMAPPED_BASE;
-+ start_addr = addr = mm->mmap_base;
- mm->cached_hole_size = 0;
- }
-
-@@ -174,14 +177,14 @@ full_search:
- vma = find_vma(mm, VA_EXCLUDE_END);
- }
- if (unlikely(task_size < addr)) {
-- if (start_addr != TASK_UNMAPPED_BASE) {
-- start_addr = addr = TASK_UNMAPPED_BASE;
-+ if (start_addr != mm->mmap_base) {
-+ start_addr = addr = mm->mmap_base;
- mm->cached_hole_size = 0;
- goto full_search;
- }
- return -ENOMEM;
- }
-- if (likely(!vma || addr + len <= vma->vm_start)) {
-+ if (likely(check_heap_stack_gap(vma, addr, len))) {
- /*
- * Remember the place where we stopped the search:
- */
-@@ -215,7 +218,7 @@ arch_get_unmapped_area_topdown(struct fi
- /* We do not accept a shared mapping if it would violate
- * cache aliasing constraints.
- */
-- if ((flags & MAP_SHARED) &&
-+ if ((filp || (flags & MAP_SHARED)) &&
- ((addr - (pgoff << PAGE_SHIFT)) & (SHMLBA - 1)))
- return -EINVAL;
- return addr;
-@@ -236,8 +239,7 @@ arch_get_unmapped_area_topdown(struct fi
- addr = PAGE_ALIGN(addr);
-
- vma = find_vma(mm, addr);
-- if (task_size - len >= addr &&
-- (!vma || addr + len <= vma->vm_start))
-+ if (task_size - len >= addr && check_heap_stack_gap(vma, addr, len))
- return addr;
- }
-
-@@ -258,7 +260,7 @@ arch_get_unmapped_area_topdown(struct fi
- /* make sure it can fit in the remaining address space */
- if (likely(addr > len)) {
- vma = find_vma(mm, addr-len);
-- if (!vma || addr <= vma->vm_start) {
-+ if (check_heap_stack_gap(vma, addr - len, len)) {
- /* remember the address as a hint for next time */
- return (mm->free_area_cache = addr-len);
- }
-@@ -267,18 +269,18 @@ arch_get_unmapped_area_topdown(struct fi
- if (unlikely(mm->mmap_base < len))
- goto bottomup;
-
-- addr = mm->mmap_base-len;
-- if (do_color_align)
-- addr = COLOUR_ALIGN_DOWN(addr, pgoff);
-+ addr = mm->mmap_base - len;
-
- do {
-+ if (do_color_align)
-+ addr = COLOUR_ALIGN_DOWN(addr, pgoff);
- /*
- * Lookup failure means no vma is above this address,
- * else if new region fits below vma->vm_start,
- * return with success:
- */
- vma = find_vma(mm, addr);
-- if (likely(!vma || addr+len <= vma->vm_start)) {
-+ if (likely(check_heap_stack_gap(vma, addr, len))) {
- /* remember the address as a hint for next time */
- return (mm->free_area_cache = addr);
- }
-@@ -288,10 +290,8 @@ arch_get_unmapped_area_topdown(struct fi
- mm->cached_hole_size = vma->vm_start - addr;
-
- /* try just below the current vma->vm_start */
-- addr = vma->vm_start-len;
-- if (do_color_align)
-- addr = COLOUR_ALIGN_DOWN(addr, pgoff);
-- } while (likely(len < vma->vm_start));
-+ addr = skip_heap_stack_gap(vma, len);
-+ } while (!IS_ERR_VALUE(addr));
-
- bottomup:
- /*
-@@ -385,6 +385,12 @@ void arch_pick_mmap_layout(struct mm_str
- gap == RLIM_INFINITY ||
- sysctl_legacy_va_layout) {
- mm->mmap_base = TASK_UNMAPPED_BASE + random_factor;
-+
-+#ifdef CONFIG_PAX_RANDMMAP
-+ if (mm->pax_flags & MF_PAX_RANDMMAP)
-+ mm->mmap_base += mm->delta_mmap;
-+#endif
-+
- mm->get_unmapped_area = arch_get_unmapped_area;
- mm->unmap_area = arch_unmap_area;
- } else {
-@@ -397,6 +403,12 @@ void arch_pick_mmap_layout(struct mm_str
- gap = (task_size / 6 * 5);
-
- mm->mmap_base = PAGE_ALIGN(task_size - gap - random_factor);
-+
-+#ifdef CONFIG_PAX_RANDMMAP
-+ if (mm->pax_flags & MF_PAX_RANDMMAP)
-+ mm->mmap_base -= mm->delta_mmap + mm->delta_stack;
-+#endif
-+
- mm->get_unmapped_area = arch_get_unmapped_area_topdown;
- mm->unmap_area = arch_unmap_area_topdown;
- }
-diff -urNp linux-2.6.37.4/arch/sparc/kernel/traps_32.c linux-2.6.37.4/arch/sparc/kernel/traps_32.c
---- linux-2.6.37.4/arch/sparc/kernel/traps_32.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/sparc/kernel/traps_32.c 2011-01-17 02:41:00.000000000 -0500
-@@ -76,7 +76,7 @@ void die_if_kernel(char *str, struct pt_
- count++ < 30 &&
- (((unsigned long) rw) >= PAGE_OFFSET) &&
- !(((unsigned long) rw) & 0x7)) {
-- printk("Caller[%08lx]: %pS\n", rw->ins[7],
-+ printk("Caller[%08lx]: %pA\n", rw->ins[7],
- (void *) rw->ins[7]);
- rw = (struct reg_window32 *)rw->ins[6];
- }
-diff -urNp linux-2.6.37.4/arch/sparc/kernel/traps_64.c linux-2.6.37.4/arch/sparc/kernel/traps_64.c
---- linux-2.6.37.4/arch/sparc/kernel/traps_64.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/sparc/kernel/traps_64.c 2011-01-17 02:41:00.000000000 -0500
-@@ -75,7 +75,7 @@ static void dump_tl1_traplog(struct tl1_
- i + 1,
- p->trapstack[i].tstate, p->trapstack[i].tpc,
- p->trapstack[i].tnpc, p->trapstack[i].tt);
-- printk("TRAPLOG: TPC<%pS>\n", (void *) p->trapstack[i].tpc);
-+ printk("TRAPLOG: TPC<%pA>\n", (void *) p->trapstack[i].tpc);
- }
- }
-
-@@ -95,6 +95,12 @@ void bad_trap(struct pt_regs *regs, long
-
- lvl -= 0x100;
- if (regs->tstate & TSTATE_PRIV) {
-+
-+#ifdef CONFIG_PAX_REFCOUNT
-+ if (lvl == 6)
-+ pax_report_refcount_overflow(regs);
-+#endif
-+
- sprintf(buffer, "Kernel bad sw trap %lx", lvl);
- die_if_kernel(buffer, regs);
- }
-@@ -113,11 +119,16 @@ void bad_trap(struct pt_regs *regs, long
- void bad_trap_tl1(struct pt_regs *regs, long lvl)
- {
- char buffer[32];
--
-+
- if (notify_die(DIE_TRAP_TL1, "bad trap tl1", regs,
- 0, lvl, SIGTRAP) == NOTIFY_STOP)
- return;
-
-+#ifdef CONFIG_PAX_REFCOUNT
-+ if (lvl == 6)
-+ pax_report_refcount_overflow(regs);
-+#endif
-+
- dump_tl1_traplog((struct tl1_traplog *)(regs + 1));
-
- sprintf (buffer, "Bad trap %lx at tl>0", lvl);
-@@ -1141,7 +1152,7 @@ static void cheetah_log_errors(struct pt
- regs->tpc, regs->tnpc, regs->u_regs[UREG_I7], regs->tstate);
- printk("%s" "ERROR(%d): ",
- (recoverable ? KERN_WARNING : KERN_CRIT), smp_processor_id());
-- printk("TPC<%pS>\n", (void *) regs->tpc);
-+ printk("TPC<%pA>\n", (void *) regs->tpc);
- printk("%s" "ERROR(%d): M_SYND(%lx), E_SYND(%lx)%s%s\n",
- (recoverable ? KERN_WARNING : KERN_CRIT), smp_processor_id(),
- (afsr & CHAFSR_M_SYNDROME) >> CHAFSR_M_SYNDROME_SHIFT,
-@@ -1748,7 +1759,7 @@ void cheetah_plus_parity_error(int type,
- smp_processor_id(),
- (type & 0x1) ? 'I' : 'D',
- regs->tpc);
-- printk(KERN_EMERG "TPC<%pS>\n", (void *) regs->tpc);
-+ printk(KERN_EMERG "TPC<%pA>\n", (void *) regs->tpc);
- panic("Irrecoverable Cheetah+ parity error.");
- }
-
-@@ -1756,7 +1767,7 @@ void cheetah_plus_parity_error(int type,
- smp_processor_id(),
- (type & 0x1) ? 'I' : 'D',
- regs->tpc);
-- printk(KERN_WARNING "TPC<%pS>\n", (void *) regs->tpc);
-+ printk(KERN_WARNING "TPC<%pA>\n", (void *) regs->tpc);
- }
-
- struct sun4v_error_entry {
-@@ -1963,9 +1974,9 @@ void sun4v_itlb_error_report(struct pt_r
-
- printk(KERN_EMERG "SUN4V-ITLB: Error at TPC[%lx], tl %d\n",
- regs->tpc, tl);
-- printk(KERN_EMERG "SUN4V-ITLB: TPC<%pS>\n", (void *) regs->tpc);
-+ printk(KERN_EMERG "SUN4V-ITLB: TPC<%pA>\n", (void *) regs->tpc);
- printk(KERN_EMERG "SUN4V-ITLB: O7[%lx]\n", regs->u_regs[UREG_I7]);
-- printk(KERN_EMERG "SUN4V-ITLB: O7<%pS>\n",
-+ printk(KERN_EMERG "SUN4V-ITLB: O7<%pA>\n",
- (void *) regs->u_regs[UREG_I7]);
- printk(KERN_EMERG "SUN4V-ITLB: vaddr[%lx] ctx[%lx] "
- "pte[%lx] error[%lx]\n",
-@@ -1987,9 +1998,9 @@ void sun4v_dtlb_error_report(struct pt_r
-
- printk(KERN_EMERG "SUN4V-DTLB: Error at TPC[%lx], tl %d\n",
- regs->tpc, tl);
-- printk(KERN_EMERG "SUN4V-DTLB: TPC<%pS>\n", (void *) regs->tpc);
-+ printk(KERN_EMERG "SUN4V-DTLB: TPC<%pA>\n", (void *) regs->tpc);
- printk(KERN_EMERG "SUN4V-DTLB: O7[%lx]\n", regs->u_regs[UREG_I7]);
-- printk(KERN_EMERG "SUN4V-DTLB: O7<%pS>\n",
-+ printk(KERN_EMERG "SUN4V-DTLB: O7<%pA>\n",
- (void *) regs->u_regs[UREG_I7]);
- printk(KERN_EMERG "SUN4V-DTLB: vaddr[%lx] ctx[%lx] "
- "pte[%lx] error[%lx]\n",
-@@ -2196,13 +2207,13 @@ void show_stack(struct task_struct *tsk,
- fp = (unsigned long)sf->fp + STACK_BIAS;
- }
-
-- printk(" [%016lx] %pS\n", pc, (void *) pc);
-+ printk(" [%016lx] %pA\n", pc, (void *) pc);
- #ifdef CONFIG_FUNCTION_GRAPH_TRACER
- if ((pc + 8UL) == (unsigned long) &return_to_handler) {
- int index = tsk->curr_ret_stack;
- if (tsk->ret_stack && index >= graph) {
- pc = tsk->ret_stack[index - graph].ret;
-- printk(" [%016lx] %pS\n", pc, (void *) pc);
-+ printk(" [%016lx] %pA\n", pc, (void *) pc);
- graph++;
- }
- }
-@@ -2255,7 +2266,7 @@ void die_if_kernel(char *str, struct pt_
- while (rw &&
- count++ < 30 &&
- kstack_valid(tp, (unsigned long) rw)) {
-- printk("Caller[%016lx]: %pS\n", rw->ins[7],
-+ printk("Caller[%016lx]: %pA\n", rw->ins[7],
- (void *) rw->ins[7]);
-
- rw = kernel_stack_up(rw);
-diff -urNp linux-2.6.37.4/arch/sparc/kernel/unaligned_64.c linux-2.6.37.4/arch/sparc/kernel/unaligned_64.c
---- linux-2.6.37.4/arch/sparc/kernel/unaligned_64.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/sparc/kernel/unaligned_64.c 2011-01-17 02:41:00.000000000 -0500
-@@ -278,7 +278,7 @@ static void log_unaligned(struct pt_regs
- static DEFINE_RATELIMIT_STATE(ratelimit, 5 * HZ, 5);
-
- if (__ratelimit(&ratelimit)) {
-- printk("Kernel unaligned access at TPC[%lx] %pS\n",
-+ printk("Kernel unaligned access at TPC[%lx] %pA\n",
- regs->tpc, (void *) regs->tpc);
- }
- }
-diff -urNp linux-2.6.37.4/arch/sparc/lib/atomic_64.S linux-2.6.37.4/arch/sparc/lib/atomic_64.S
---- linux-2.6.37.4/arch/sparc/lib/atomic_64.S 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/sparc/lib/atomic_64.S 2011-01-17 02:41:00.000000000 -0500
-@@ -18,7 +18,12 @@
- atomic_add: /* %o0 = increment, %o1 = atomic_ptr */
- BACKOFF_SETUP(%o2)
- 1: lduw [%o1], %g1
-- add %g1, %o0, %g7
-+ addcc %g1, %o0, %g7
-+
-+#ifdef CONFIG_PAX_REFCOUNT
-+ tvs %icc, 6
-+#endif
-+
- cas [%o1], %g1, %g7
- cmp %g1, %g7
- bne,pn %icc, BACKOFF_LABEL(2f, 1b)
-@@ -28,12 +33,32 @@ atomic_add: /* %o0 = increment, %o1 = at
- 2: BACKOFF_SPIN(%o2, %o3, 1b)
- .size atomic_add, .-atomic_add
-
-+ .globl atomic_add_unchecked
-+ .type atomic_add_unchecked,#function
-+atomic_add_unchecked: /* %o0 = increment, %o1 = atomic_ptr */
-+ BACKOFF_SETUP(%o2)
-+1: lduw [%o1], %g1
-+ add %g1, %o0, %g7
-+ cas [%o1], %g1, %g7
-+ cmp %g1, %g7
-+ bne,pn %icc, 2f
-+ nop
-+ retl
-+ nop
-+2: BACKOFF_SPIN(%o2, %o3, 1b)
-+ .size atomic_add_unchecked, .-atomic_add_unchecked
-+
- .globl atomic_sub
- .type atomic_sub,#function
- atomic_sub: /* %o0 = decrement, %o1 = atomic_ptr */
- BACKOFF_SETUP(%o2)
- 1: lduw [%o1], %g1
-- sub %g1, %o0, %g7
-+ subcc %g1, %o0, %g7
-+
-+#ifdef CONFIG_PAX_REFCOUNT
-+ tvs %icc, 6
-+#endif
-+
- cas [%o1], %g1, %g7
- cmp %g1, %g7
- bne,pn %icc, BACKOFF_LABEL(2f, 1b)
-@@ -43,12 +68,32 @@ atomic_sub: /* %o0 = decrement, %o1 = at
- 2: BACKOFF_SPIN(%o2, %o3, 1b)
- .size atomic_sub, .-atomic_sub
-
-+ .globl atomic_sub_unchecked
-+ .type atomic_sub_unchecked,#function
-+atomic_sub_unchecked: /* %o0 = decrement, %o1 = atomic_ptr */
-+ BACKOFF_SETUP(%o2)
-+1: lduw [%o1], %g1
-+ sub %g1, %o0, %g7
-+ cas [%o1], %g1, %g7
-+ cmp %g1, %g7
-+ bne,pn %icc, 2f
-+ nop
-+ retl
-+ nop
-+2: BACKOFF_SPIN(%o2, %o3, 1b)
-+ .size atomic_sub_unchecked, .-atomic_sub_unchecked
-+
- .globl atomic_add_ret
- .type atomic_add_ret,#function
- atomic_add_ret: /* %o0 = increment, %o1 = atomic_ptr */
- BACKOFF_SETUP(%o2)
- 1: lduw [%o1], %g1
-- add %g1, %o0, %g7
-+ addcc %g1, %o0, %g7
-+
-+#ifdef CONFIG_PAX_REFCOUNT
-+ tvs %icc, 6
-+#endif
-+
- cas [%o1], %g1, %g7
- cmp %g1, %g7
- bne,pn %icc, BACKOFF_LABEL(2f, 1b)
-@@ -58,12 +103,33 @@ atomic_add_ret: /* %o0 = increment, %o1
- 2: BACKOFF_SPIN(%o2, %o3, 1b)
- .size atomic_add_ret, .-atomic_add_ret
-
-+ .globl atomic_add_ret_unchecked
-+ .type atomic_add_ret_unchecked,#function
-+atomic_add_ret_unchecked: /* %o0 = increment, %o1 = atomic_ptr */
-+ BACKOFF_SETUP(%o2)
-+1: lduw [%o1], %g1
-+ addcc %g1, %o0, %g7
-+ cas [%o1], %g1, %g7
-+ cmp %g1, %g7
-+ bne,pn %icc, 2f
-+ add %g7, %o0, %g7
-+ sra %g7, 0, %o0
-+ retl
-+ nop
-+2: BACKOFF_SPIN(%o2, %o3, 1b)
-+ .size atomic_add_ret_unchecked, .-atomic_add_ret_unchecked
-+
- .globl atomic_sub_ret
- .type atomic_sub_ret,#function
- atomic_sub_ret: /* %o0 = decrement, %o1 = atomic_ptr */
- BACKOFF_SETUP(%o2)
- 1: lduw [%o1], %g1
-- sub %g1, %o0, %g7
-+ subcc %g1, %o0, %g7
-+
-+#ifdef CONFIG_PAX_REFCOUNT
-+ tvs %icc, 6
-+#endif
-+
- cas [%o1], %g1, %g7
- cmp %g1, %g7
- bne,pn %icc, BACKOFF_LABEL(2f, 1b)
-@@ -78,7 +144,12 @@ atomic_sub_ret: /* %o0 = decrement, %o1
- atomic64_add: /* %o0 = increment, %o1 = atomic_ptr */
- BACKOFF_SETUP(%o2)
- 1: ldx [%o1], %g1
-- add %g1, %o0, %g7
-+ addcc %g1, %o0, %g7
-+
-+#ifdef CONFIG_PAX_REFCOUNT
-+ tvs %xcc, 6
-+#endif
-+
- casx [%o1], %g1, %g7
- cmp %g1, %g7
- bne,pn %xcc, BACKOFF_LABEL(2f, 1b)
-@@ -88,12 +159,32 @@ atomic64_add: /* %o0 = increment, %o1 =
- 2: BACKOFF_SPIN(%o2, %o3, 1b)
- .size atomic64_add, .-atomic64_add
-
-+ .globl atomic64_add_unchecked
-+ .type atomic64_add_unchecked,#function
-+atomic64_add_unchecked: /* %o0 = increment, %o1 = atomic_ptr */
-+ BACKOFF_SETUP(%o2)
-+1: ldx [%o1], %g1
-+ addcc %g1, %o0, %g7
-+ casx [%o1], %g1, %g7
-+ cmp %g1, %g7
-+ bne,pn %xcc, 2f
-+ nop
-+ retl
-+ nop
-+2: BACKOFF_SPIN(%o2, %o3, 1b)
-+ .size atomic64_add_unchecked, .-atomic64_add_unchecked
-+
- .globl atomic64_sub
- .type atomic64_sub,#function
- atomic64_sub: /* %o0 = decrement, %o1 = atomic_ptr */
- BACKOFF_SETUP(%o2)
- 1: ldx [%o1], %g1
-- sub %g1, %o0, %g7
-+ subcc %g1, %o0, %g7
-+
-+#ifdef CONFIG_PAX_REFCOUNT
-+ tvs %xcc, 6
-+#endif
-+
- casx [%o1], %g1, %g7
- cmp %g1, %g7
- bne,pn %xcc, BACKOFF_LABEL(2f, 1b)
-@@ -103,12 +194,32 @@ atomic64_sub: /* %o0 = decrement, %o1 =
- 2: BACKOFF_SPIN(%o2, %o3, 1b)
- .size atomic64_sub, .-atomic64_sub
-
-+ .globl atomic64_sub_unchecked
-+ .type atomic64_sub_unchecked,#function
-+atomic64_sub_unchecked: /* %o0 = decrement, %o1 = atomic_ptr */
-+ BACKOFF_SETUP(%o2)
-+1: ldx [%o1], %g1
-+ subcc %g1, %o0, %g7
-+ casx [%o1], %g1, %g7
-+ cmp %g1, %g7
-+ bne,pn %xcc, 2f
-+ nop
-+ retl
-+ nop
-+2: BACKOFF_SPIN(%o2, %o3, 1b)
-+ .size atomic64_sub_unchecked, .-atomic64_sub_unchecked
-+
- .globl atomic64_add_ret
- .type atomic64_add_ret,#function
- atomic64_add_ret: /* %o0 = increment, %o1 = atomic_ptr */
- BACKOFF_SETUP(%o2)
- 1: ldx [%o1], %g1
-- add %g1, %o0, %g7
-+ addcc %g1, %o0, %g7
-+
-+#ifdef CONFIG_PAX_REFCOUNT
-+ tvs %xcc, 6
-+#endif
-+
- casx [%o1], %g1, %g7
- cmp %g1, %g7
- bne,pn %xcc, BACKOFF_LABEL(2f, 1b)
-@@ -118,12 +229,33 @@ atomic64_add_ret: /* %o0 = increment, %o
- 2: BACKOFF_SPIN(%o2, %o3, 1b)
- .size atomic64_add_ret, .-atomic64_add_ret
-
-+ .globl atomic64_add_ret_unchecked
-+ .type atomic64_add_ret_unchecked,#function
-+atomic64_add_ret_unchecked: /* %o0 = increment, %o1 = atomic_ptr */
-+ BACKOFF_SETUP(%o2)
-+1: ldx [%o1], %g1
-+ addcc %g1, %o0, %g7
-+ casx [%o1], %g1, %g7
-+ cmp %g1, %g7
-+ bne,pn %xcc, 2f
-+ add %g7, %o0, %g7
-+ mov %g7, %o0
-+ retl
-+ nop
-+2: BACKOFF_SPIN(%o2, %o3, 1b)
-+ .size atomic64_add_ret_unchecked, .-atomic64_add_ret_unchecked
-+
- .globl atomic64_sub_ret
- .type atomic64_sub_ret,#function
- atomic64_sub_ret: /* %o0 = decrement, %o1 = atomic_ptr */
- BACKOFF_SETUP(%o2)
- 1: ldx [%o1], %g1
-- sub %g1, %o0, %g7
-+ subcc %g1, %o0, %g7
-+
-+#ifdef CONFIG_PAX_REFCOUNT
-+ tvs %xcc, 6
-+#endif
-+
- casx [%o1], %g1, %g7
- cmp %g1, %g7
- bne,pn %xcc, BACKOFF_LABEL(2f, 1b)
-diff -urNp linux-2.6.37.4/arch/sparc/lib/ksyms.c linux-2.6.37.4/arch/sparc/lib/ksyms.c
---- linux-2.6.37.4/arch/sparc/lib/ksyms.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/sparc/lib/ksyms.c 2011-01-17 02:41:00.000000000 -0500
-@@ -142,12 +142,17 @@ EXPORT_SYMBOL(__downgrade_write);
-
- /* Atomic counter implementation. */
- EXPORT_SYMBOL(atomic_add);
-+EXPORT_SYMBOL(atomic_add_unchecked);
- EXPORT_SYMBOL(atomic_add_ret);
- EXPORT_SYMBOL(atomic_sub);
-+EXPORT_SYMBOL(atomic_sub_unchecked);
- EXPORT_SYMBOL(atomic_sub_ret);
- EXPORT_SYMBOL(atomic64_add);
-+EXPORT_SYMBOL(atomic64_add_unchecked);
- EXPORT_SYMBOL(atomic64_add_ret);
-+EXPORT_SYMBOL(atomic64_add_ret_unchecked);
- EXPORT_SYMBOL(atomic64_sub);
-+EXPORT_SYMBOL(atomic64_sub_unchecked);
- EXPORT_SYMBOL(atomic64_sub_ret);
-
- /* Atomic bit operations. */
-diff -urNp linux-2.6.37.4/arch/sparc/Makefile linux-2.6.37.4/arch/sparc/Makefile
---- linux-2.6.37.4/arch/sparc/Makefile 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/sparc/Makefile 2011-01-17 02:41:00.000000000 -0500
-@@ -75,7 +75,7 @@ drivers-$(CONFIG_OPROFILE) += arch/sparc
- # Export what is needed by arch/sparc/boot/Makefile
- export VMLINUX_INIT VMLINUX_MAIN
- VMLINUX_INIT := $(head-y) $(init-y)
--VMLINUX_MAIN := $(core-y) kernel/ mm/ fs/ ipc/ security/ crypto/ block/
-+VMLINUX_MAIN := $(core-y) kernel/ mm/ fs/ ipc/ security/ crypto/ block/ grsecurity/
- VMLINUX_MAIN += $(patsubst %/, %/lib.a, $(libs-y)) $(libs-y)
- VMLINUX_MAIN += $(drivers-y) $(net-y)
-
-diff -urNp linux-2.6.37.4/arch/sparc/mm/fault_32.c linux-2.6.37.4/arch/sparc/mm/fault_32.c
---- linux-2.6.37.4/arch/sparc/mm/fault_32.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/sparc/mm/fault_32.c 2011-01-17 02:41:00.000000000 -0500
-@@ -22,6 +22,9 @@
- #include <linux/interrupt.h>
- #include <linux/module.h>
- #include <linux/kdebug.h>
-+#include <linux/slab.h>
-+#include <linux/pagemap.h>
-+#include <linux/compiler.h>
-
- #include <asm/system.h>
- #include <asm/page.h>
-@@ -209,6 +212,268 @@ static unsigned long compute_si_addr(str
- return safe_compute_effective_address(regs, insn);
- }
-
-+#ifdef CONFIG_PAX_PAGEEXEC
-+#ifdef CONFIG_PAX_DLRESOLVE
-+static void pax_emuplt_close(struct vm_area_struct *vma)
-+{
-+ vma->vm_mm->call_dl_resolve = 0UL;
-+}
-+
-+static int pax_emuplt_fault(struct vm_area_struct *vma, struct vm_fault *vmf)
-+{
-+ unsigned int *kaddr;
-+
-+ vmf->page = alloc_page(GFP_HIGHUSER);
-+ if (!vmf->page)
-+ return VM_FAULT_OOM;
-+
-+ kaddr = kmap(vmf->page);
-+ memset(kaddr, 0, PAGE_SIZE);
-+ kaddr[0] = 0x9DE3BFA8U; /* save */
-+ flush_dcache_page(vmf->page);
-+ kunmap(vmf->page);
-+ return VM_FAULT_MAJOR;
-+}
-+
-+static const struct vm_operations_struct pax_vm_ops = {
-+ .close = pax_emuplt_close,
-+ .fault = pax_emuplt_fault
-+};
-+
-+static int pax_insert_vma(struct vm_area_struct *vma, unsigned long addr)
-+{
-+ int ret;
-+
-+ INIT_LIST_HEAD(&vma->anon_vma_chain);
-+ vma->vm_mm = current->mm;
-+ vma->vm_start = addr;
-+ vma->vm_end = addr + PAGE_SIZE;
-+ vma->vm_flags = VM_READ | VM_EXEC | VM_MAYREAD | VM_MAYEXEC;
-+ vma->vm_page_prot = vm_get_page_prot(vma->vm_flags);
-+ vma->vm_ops = &pax_vm_ops;
-+
-+ ret = insert_vm_struct(current->mm, vma);
-+ if (ret)
-+ return ret;
-+
-+ ++current->mm->total_vm;
-+ return 0;
-+}
-+#endif
-+
-+/*
-+ * PaX: decide what to do with offenders (regs->pc = fault address)
-+ *
-+ * returns 1 when task should be killed
-+ * 2 when patched PLT trampoline was detected
-+ * 3 when unpatched PLT trampoline was detected
-+ */
-+static int pax_handle_fetch_fault(struct pt_regs *regs)
-+{
-+
-+#ifdef CONFIG_PAX_EMUPLT
-+ int err;
-+
-+ do { /* PaX: patched PLT emulation #1 */
-+ unsigned int sethi1, sethi2, jmpl;
-+
-+ err = get_user(sethi1, (unsigned int *)regs->pc);
-+ err |= get_user(sethi2, (unsigned int *)(regs->pc+4));
-+ err |= get_user(jmpl, (unsigned int *)(regs->pc+8));
-+
-+ if (err)
-+ break;
-+
-+ if ((sethi1 & 0xFFC00000U) == 0x03000000U &&
-+ (sethi2 & 0xFFC00000U) == 0x03000000U &&
-+ (jmpl & 0xFFFFE000U) == 0x81C06000U)
-+ {
-+ unsigned int addr;
-+
-+ regs->u_regs[UREG_G1] = (sethi2 & 0x003FFFFFU) << 10;
-+ addr = regs->u_regs[UREG_G1];
-+ addr += (((jmpl | 0xFFFFE000U) ^ 0x00001000U) + 0x00001000U);
-+ regs->pc = addr;
-+ regs->npc = addr+4;
-+ return 2;
-+ }
-+ } while (0);
-+
-+ { /* PaX: patched PLT emulation #2 */
-+ unsigned int ba;
-+
-+ err = get_user(ba, (unsigned int *)regs->pc);
-+
-+ if (!err && (ba & 0xFFC00000U) == 0x30800000U) {
-+ unsigned int addr;
-+
-+ addr = regs->pc + ((((ba | 0xFFC00000U) ^ 0x00200000U) + 0x00200000U) << 2);
-+ regs->pc = addr;
-+ regs->npc = addr+4;
-+ return 2;
-+ }
-+ }
-+
-+ do { /* PaX: patched PLT emulation #3 */
-+ unsigned int sethi, jmpl, nop;
-+
-+ err = get_user(sethi, (unsigned int *)regs->pc);
-+ err |= get_user(jmpl, (unsigned int *)(regs->pc+4));
-+ err |= get_user(nop, (unsigned int *)(regs->pc+8));
-+
-+ if (err)
-+ break;
-+
-+ if ((sethi & 0xFFC00000U) == 0x03000000U &&
-+ (jmpl & 0xFFFFE000U) == 0x81C06000U &&
-+ nop == 0x01000000U)
-+ {
-+ unsigned int addr;
-+
-+ addr = (sethi & 0x003FFFFFU) << 10;
-+ regs->u_regs[UREG_G1] = addr;
-+ addr += (((jmpl | 0xFFFFE000U) ^ 0x00001000U) + 0x00001000U);
-+ regs->pc = addr;
-+ regs->npc = addr+4;
-+ return 2;
-+ }
-+ } while (0);
-+
-+ do { /* PaX: unpatched PLT emulation step 1 */
-+ unsigned int sethi, ba, nop;
-+
-+ err = get_user(sethi, (unsigned int *)regs->pc);
-+ err |= get_user(ba, (unsigned int *)(regs->pc+4));
-+ err |= get_user(nop, (unsigned int *)(regs->pc+8));
-+
-+ if (err)
-+ break;
-+
-+ if ((sethi & 0xFFC00000U) == 0x03000000U &&
-+ ((ba & 0xFFC00000U) == 0x30800000U || (ba & 0xFFF80000U) == 0x30680000U) &&
-+ nop == 0x01000000U)
-+ {
-+ unsigned int addr, save, call;
-+
-+ if ((ba & 0xFFC00000U) == 0x30800000U)
-+ addr = regs->pc + 4 + ((((ba | 0xFFC00000U) ^ 0x00200000U) + 0x00200000U) << 2);
-+ else
-+ addr = regs->pc + 4 + ((((ba | 0xFFF80000U) ^ 0x00040000U) + 0x00040000U) << 2);
-+
-+ err = get_user(save, (unsigned int *)addr);
-+ err |= get_user(call, (unsigned int *)(addr+4));
-+ err |= get_user(nop, (unsigned int *)(addr+8));
-+ if (err)
-+ break;
-+
-+#ifdef CONFIG_PAX_DLRESOLVE
-+ if (save == 0x9DE3BFA8U &&
-+ (call & 0xC0000000U) == 0x40000000U &&
-+ nop == 0x01000000U)
-+ {
-+ struct vm_area_struct *vma;
-+ unsigned long call_dl_resolve;
-+
-+ down_read(&current->mm->mmap_sem);
-+ call_dl_resolve = current->mm->call_dl_resolve;
-+ up_read(&current->mm->mmap_sem);
-+ if (likely(call_dl_resolve))
-+ goto emulate;
-+
-+ vma = kmem_cache_zalloc(vm_area_cachep, GFP_KERNEL);
-+
-+ down_write(&current->mm->mmap_sem);
-+ if (current->mm->call_dl_resolve) {
-+ call_dl_resolve = current->mm->call_dl_resolve;
-+ up_write(&current->mm->mmap_sem);
-+ if (vma)
-+ kmem_cache_free(vm_area_cachep, vma);
-+ goto emulate;
-+ }
-+
-+ call_dl_resolve = get_unmapped_area(NULL, 0UL, PAGE_SIZE, 0UL, MAP_PRIVATE);
-+ if (!vma || (call_dl_resolve & ~PAGE_MASK)) {
-+ up_write(&current->mm->mmap_sem);
-+ if (vma)
-+ kmem_cache_free(vm_area_cachep, vma);
-+ return 1;
-+ }
-+
-+ if (pax_insert_vma(vma, call_dl_resolve)) {
-+ up_write(&current->mm->mmap_sem);
-+ kmem_cache_free(vm_area_cachep, vma);
-+ return 1;
-+ }
-+
-+ current->mm->call_dl_resolve = call_dl_resolve;
-+ up_write(&current->mm->mmap_sem);
-+
-+emulate:
-+ regs->u_regs[UREG_G1] = (sethi & 0x003FFFFFU) << 10;
-+ regs->pc = call_dl_resolve;
-+ regs->npc = addr+4;
-+ return 3;
-+ }
-+#endif
-+
-+ /* PaX: glibc 2.4+ generates sethi/jmpl instead of save/call */
-+ if ((save & 0xFFC00000U) == 0x05000000U &&
-+ (call & 0xFFFFE000U) == 0x85C0A000U &&
-+ nop == 0x01000000U)
-+ {
-+ regs->u_regs[UREG_G1] = (sethi & 0x003FFFFFU) << 10;
-+ regs->u_regs[UREG_G2] = addr + 4;
-+ addr = (save & 0x003FFFFFU) << 10;
-+ addr += (((call | 0xFFFFE000U) ^ 0x00001000U) + 0x00001000U);
-+ regs->pc = addr;
-+ regs->npc = addr+4;
-+ return 3;
-+ }
-+ }
-+ } while (0);
-+
-+ do { /* PaX: unpatched PLT emulation step 2 */
-+ unsigned int save, call, nop;
-+
-+ err = get_user(save, (unsigned int *)(regs->pc-4));
-+ err |= get_user(call, (unsigned int *)regs->pc);
-+ err |= get_user(nop, (unsigned int *)(regs->pc+4));
-+ if (err)
-+ break;
-+
-+ if (save == 0x9DE3BFA8U &&
-+ (call & 0xC0000000U) == 0x40000000U &&
-+ nop == 0x01000000U)
-+ {
-+ unsigned int dl_resolve = regs->pc + ((((call | 0xC0000000U) ^ 0x20000000U) + 0x20000000U) << 2);
-+
-+ regs->u_regs[UREG_RETPC] = regs->pc;
-+ regs->pc = dl_resolve;
-+ regs->npc = dl_resolve+4;
-+ return 3;
-+ }
-+ } while (0);
-+#endif
-+
-+ return 1;
-+}
-+
-+void pax_report_insns(void *pc, void *sp)
-+{
-+ unsigned long i;
-+
-+ printk(KERN_ERR "PAX: bytes at PC: ");
-+ for (i = 0; i < 8; i++) {
-+ unsigned int c;
-+ if (get_user(c, (unsigned int *)pc+i))
-+ printk(KERN_CONT "???????? ");
-+ else
-+ printk(KERN_CONT "%08x ", c);
-+ }
-+ printk("\n");
-+}
-+#endif
-+
- static noinline void do_fault_siginfo(int code, int sig, struct pt_regs *regs,
- int text_fault)
- {
-@@ -282,6 +547,24 @@ good_area:
- if(!(vma->vm_flags & VM_WRITE))
- goto bad_area;
- } else {
-+
-+#ifdef CONFIG_PAX_PAGEEXEC
-+ if ((mm->pax_flags & MF_PAX_PAGEEXEC) && text_fault && !(vma->vm_flags & VM_EXEC)) {
-+ up_read(&mm->mmap_sem);
-+ switch (pax_handle_fetch_fault(regs)) {
-+
-+#ifdef CONFIG_PAX_EMUPLT
-+ case 2:
-+ case 3:
-+ return;
-+#endif
-+
-+ }
-+ pax_report_fault(regs, (void *)regs->pc, (void *)regs->u_regs[UREG_FP]);
-+ do_group_exit(SIGKILL);
-+ }
-+#endif
-+
- /* Allow reads even for write-only mappings */
- if(!(vma->vm_flags & (VM_READ | VM_EXEC)))
- goto bad_area;
-diff -urNp linux-2.6.37.4/arch/sparc/mm/fault_64.c linux-2.6.37.4/arch/sparc/mm/fault_64.c
---- linux-2.6.37.4/arch/sparc/mm/fault_64.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/sparc/mm/fault_64.c 2011-01-17 02:41:00.000000000 -0500
-@@ -21,6 +21,9 @@
- #include <linux/kprobes.h>
- #include <linux/kdebug.h>
- #include <linux/percpu.h>
-+#include <linux/slab.h>
-+#include <linux/pagemap.h>
-+#include <linux/compiler.h>
-
- #include <asm/page.h>
- #include <asm/pgtable.h>
-@@ -74,7 +77,7 @@ static void __kprobes bad_kernel_pc(stru
- printk(KERN_CRIT "OOPS: Bogus kernel PC [%016lx] in fault handler\n",
- regs->tpc);
- printk(KERN_CRIT "OOPS: RPC [%016lx]\n", regs->u_regs[15]);
-- printk("OOPS: RPC <%pS>\n", (void *) regs->u_regs[15]);
-+ printk("OOPS: RPC <%pA>\n", (void *) regs->u_regs[15]);
- printk(KERN_CRIT "OOPS: Fault was to vaddr[%lx]\n", vaddr);
- dump_stack();
- unhandled_fault(regs->tpc, current, regs);
-@@ -272,6 +275,457 @@ static void noinline __kprobes bogus_32b
- show_regs(regs);
- }
-
-+#ifdef CONFIG_PAX_PAGEEXEC
-+#ifdef CONFIG_PAX_DLRESOLVE
-+static void pax_emuplt_close(struct vm_area_struct *vma)
-+{
-+ vma->vm_mm->call_dl_resolve = 0UL;
-+}
-+
-+static int pax_emuplt_fault(struct vm_area_struct *vma, struct vm_fault *vmf)
-+{
-+ unsigned int *kaddr;
-+
-+ vmf->page = alloc_page(GFP_HIGHUSER);
-+ if (!vmf->page)
-+ return VM_FAULT_OOM;
-+
-+ kaddr = kmap(vmf->page);
-+ memset(kaddr, 0, PAGE_SIZE);
-+ kaddr[0] = 0x9DE3BFA8U; /* save */
-+ flush_dcache_page(vmf->page);
-+ kunmap(vmf->page);
-+ return VM_FAULT_MAJOR;
-+}
-+
-+static const struct vm_operations_struct pax_vm_ops = {
-+ .close = pax_emuplt_close,
-+ .fault = pax_emuplt_fault
-+};
-+
-+static int pax_insert_vma(struct vm_area_struct *vma, unsigned long addr)
-+{
-+ int ret;
-+
-+ INIT_LIST_HEAD(&vma->anon_vma_chain);
-+ vma->vm_mm = current->mm;
-+ vma->vm_start = addr;
-+ vma->vm_end = addr + PAGE_SIZE;
-+ vma->vm_flags = VM_READ | VM_EXEC | VM_MAYREAD | VM_MAYEXEC;
-+ vma->vm_page_prot = vm_get_page_prot(vma->vm_flags);
-+ vma->vm_ops = &pax_vm_ops;
-+
-+ ret = insert_vm_struct(current->mm, vma);
-+ if (ret)
-+ return ret;
-+
-+ ++current->mm->total_vm;
-+ return 0;
-+}
-+#endif
-+
-+/*
-+ * PaX: decide what to do with offenders (regs->tpc = fault address)
-+ *
-+ * returns 1 when task should be killed
-+ * 2 when patched PLT trampoline was detected
-+ * 3 when unpatched PLT trampoline was detected
-+ */
-+static int pax_handle_fetch_fault(struct pt_regs *regs)
-+{
-+
-+#ifdef CONFIG_PAX_EMUPLT
-+ int err;
-+
-+ do { /* PaX: patched PLT emulation #1 */
-+ unsigned int sethi1, sethi2, jmpl;
-+
-+ err = get_user(sethi1, (unsigned int *)regs->tpc);
-+ err |= get_user(sethi2, (unsigned int *)(regs->tpc+4));
-+ err |= get_user(jmpl, (unsigned int *)(regs->tpc+8));
-+
-+ if (err)
-+ break;
-+
-+ if ((sethi1 & 0xFFC00000U) == 0x03000000U &&
-+ (sethi2 & 0xFFC00000U) == 0x03000000U &&
-+ (jmpl & 0xFFFFE000U) == 0x81C06000U)
-+ {
-+ unsigned long addr;
-+
-+ regs->u_regs[UREG_G1] = (sethi2 & 0x003FFFFFU) << 10;
-+ addr = regs->u_regs[UREG_G1];
-+ addr += (((jmpl | 0xFFFFFFFFFFFFE000UL) ^ 0x00001000UL) + 0x00001000UL);
-+
-+ if (test_thread_flag(TIF_32BIT))
-+ addr &= 0xFFFFFFFFUL;
-+
-+ regs->tpc = addr;
-+ regs->tnpc = addr+4;
-+ return 2;
-+ }
-+ } while (0);
-+
-+ { /* PaX: patched PLT emulation #2 */
-+ unsigned int ba;
-+
-+ err = get_user(ba, (unsigned int *)regs->tpc);
-+
-+ if (!err && (ba & 0xFFC00000U) == 0x30800000U) {
-+ unsigned long addr;
-+
-+ addr = regs->tpc + ((((ba | 0xFFFFFFFFFFC00000UL) ^ 0x00200000UL) + 0x00200000UL) << 2);
-+
-+ if (test_thread_flag(TIF_32BIT))
-+ addr &= 0xFFFFFFFFUL;
-+
-+ regs->tpc = addr;
-+ regs->tnpc = addr+4;
-+ return 2;
-+ }
-+ }
-+
-+ do { /* PaX: patched PLT emulation #3 */
-+ unsigned int sethi, jmpl, nop;
-+
-+ err = get_user(sethi, (unsigned int *)regs->tpc);
-+ err |= get_user(jmpl, (unsigned int *)(regs->tpc+4));
-+ err |= get_user(nop, (unsigned int *)(regs->tpc+8));
-+
-+ if (err)
-+ break;
-+
-+ if ((sethi & 0xFFC00000U) == 0x03000000U &&
-+ (jmpl & 0xFFFFE000U) == 0x81C06000U &&
-+ nop == 0x01000000U)
-+ {
-+ unsigned long addr;
-+
-+ addr = (sethi & 0x003FFFFFU) << 10;
-+ regs->u_regs[UREG_G1] = addr;
-+ addr += (((jmpl | 0xFFFFFFFFFFFFE000UL) ^ 0x00001000UL) + 0x00001000UL);
-+
-+ if (test_thread_flag(TIF_32BIT))
-+ addr &= 0xFFFFFFFFUL;
-+
-+ regs->tpc = addr;
-+ regs->tnpc = addr+4;
-+ return 2;
-+ }
-+ } while (0);
-+
-+ do { /* PaX: patched PLT emulation #4 */
-+ unsigned int sethi, mov1, call, mov2;
-+
-+ err = get_user(sethi, (unsigned int *)regs->tpc);
-+ err |= get_user(mov1, (unsigned int *)(regs->tpc+4));
-+ err |= get_user(call, (unsigned int *)(regs->tpc+8));
-+ err |= get_user(mov2, (unsigned int *)(regs->tpc+12));
-+
-+ if (err)
-+ break;
-+
-+ if ((sethi & 0xFFC00000U) == 0x03000000U &&
-+ mov1 == 0x8210000FU &&
-+ (call & 0xC0000000U) == 0x40000000U &&
-+ mov2 == 0x9E100001U)
-+ {
-+ unsigned long addr;
-+
-+ regs->u_regs[UREG_G1] = regs->u_regs[UREG_RETPC];
-+ addr = regs->tpc + 4 + ((((call | 0xFFFFFFFFC0000000UL) ^ 0x20000000UL) + 0x20000000UL) << 2);
-+
-+ if (test_thread_flag(TIF_32BIT))
-+ addr &= 0xFFFFFFFFUL;
-+
-+ regs->tpc = addr;
-+ regs->tnpc = addr+4;
-+ return 2;
-+ }
-+ } while (0);
-+
-+ do { /* PaX: patched PLT emulation #5 */
-+ unsigned int sethi, sethi1, sethi2, or1, or2, sllx, jmpl, nop;
-+
-+ err = get_user(sethi, (unsigned int *)regs->tpc);
-+ err |= get_user(sethi1, (unsigned int *)(regs->tpc+4));
-+ err |= get_user(sethi2, (unsigned int *)(regs->tpc+8));
-+ err |= get_user(or1, (unsigned int *)(regs->tpc+12));
-+ err |= get_user(or2, (unsigned int *)(regs->tpc+16));
-+ err |= get_user(sllx, (unsigned int *)(regs->tpc+20));
-+ err |= get_user(jmpl, (unsigned int *)(regs->tpc+24));
-+ err |= get_user(nop, (unsigned int *)(regs->tpc+28));
-+
-+ if (err)
-+ break;
-+
-+ if ((sethi & 0xFFC00000U) == 0x03000000U &&
-+ (sethi1 & 0xFFC00000U) == 0x03000000U &&
-+ (sethi2 & 0xFFC00000U) == 0x0B000000U &&
-+ (or1 & 0xFFFFE000U) == 0x82106000U &&
-+ (or2 & 0xFFFFE000U) == 0x8A116000U &&
-+ sllx == 0x83287020U &&
-+ jmpl == 0x81C04005U &&
-+ nop == 0x01000000U)
-+ {
-+ unsigned long addr;
-+
-+ regs->u_regs[UREG_G1] = ((sethi1 & 0x003FFFFFU) << 10) | (or1 & 0x000003FFU);
-+ regs->u_regs[UREG_G1] <<= 32;
-+ regs->u_regs[UREG_G5] = ((sethi2 & 0x003FFFFFU) << 10) | (or2 & 0x000003FFU);
-+ addr = regs->u_regs[UREG_G1] + regs->u_regs[UREG_G5];
-+ regs->tpc = addr;
-+ regs->tnpc = addr+4;
-+ return 2;
-+ }
-+ } while (0);
-+
-+ do { /* PaX: patched PLT emulation #6 */
-+ unsigned int sethi, sethi1, sethi2, sllx, or, jmpl, nop;
-+
-+ err = get_user(sethi, (unsigned int *)regs->tpc);
-+ err |= get_user(sethi1, (unsigned int *)(regs->tpc+4));
-+ err |= get_user(sethi2, (unsigned int *)(regs->tpc+8));
-+ err |= get_user(sllx, (unsigned int *)(regs->tpc+12));
-+ err |= get_user(or, (unsigned int *)(regs->tpc+16));
-+ err |= get_user(jmpl, (unsigned int *)(regs->tpc+20));
-+ err |= get_user(nop, (unsigned int *)(regs->tpc+24));
-+
-+ if (err)
-+ break;
-+
-+ if ((sethi & 0xFFC00000U) == 0x03000000U &&
-+ (sethi1 & 0xFFC00000U) == 0x03000000U &&
-+ (sethi2 & 0xFFC00000U) == 0x0B000000U &&
-+ sllx == 0x83287020U &&
-+ (or & 0xFFFFE000U) == 0x8A116000U &&
-+ jmpl == 0x81C04005U &&
-+ nop == 0x01000000U)
-+ {
-+ unsigned long addr;
-+
-+ regs->u_regs[UREG_G1] = (sethi1 & 0x003FFFFFU) << 10;
-+ regs->u_regs[UREG_G1] <<= 32;
-+ regs->u_regs[UREG_G5] = ((sethi2 & 0x003FFFFFU) << 10) | (or & 0x3FFU);
-+ addr = regs->u_regs[UREG_G1] + regs->u_regs[UREG_G5];
-+ regs->tpc = addr;
-+ regs->tnpc = addr+4;
-+ return 2;
-+ }
-+ } while (0);
-+
-+ do { /* PaX: unpatched PLT emulation step 1 */
-+ unsigned int sethi, ba, nop;
-+
-+ err = get_user(sethi, (unsigned int *)regs->tpc);
-+ err |= get_user(ba, (unsigned int *)(regs->tpc+4));
-+ err |= get_user(nop, (unsigned int *)(regs->tpc+8));
-+
-+ if (err)
-+ break;
-+
-+ if ((sethi & 0xFFC00000U) == 0x03000000U &&
-+ ((ba & 0xFFC00000U) == 0x30800000U || (ba & 0xFFF80000U) == 0x30680000U) &&
-+ nop == 0x01000000U)
-+ {
-+ unsigned long addr;
-+ unsigned int save, call;
-+ unsigned int sethi1, sethi2, or1, or2, sllx, add, jmpl;
-+
-+ if ((ba & 0xFFC00000U) == 0x30800000U)
-+ addr = regs->tpc + 4 + ((((ba | 0xFFFFFFFFFFC00000UL) ^ 0x00200000UL) + 0x00200000UL) << 2);
-+ else
-+ addr = regs->tpc + 4 + ((((ba | 0xFFFFFFFFFFF80000UL) ^ 0x00040000UL) + 0x00040000UL) << 2);
-+
-+ if (test_thread_flag(TIF_32BIT))
-+ addr &= 0xFFFFFFFFUL;
-+
-+ err = get_user(save, (unsigned int *)addr);
-+ err |= get_user(call, (unsigned int *)(addr+4));
-+ err |= get_user(nop, (unsigned int *)(addr+8));
-+ if (err)
-+ break;
-+
-+#ifdef CONFIG_PAX_DLRESOLVE
-+ if (save == 0x9DE3BFA8U &&
-+ (call & 0xC0000000U) == 0x40000000U &&
-+ nop == 0x01000000U)
-+ {
-+ struct vm_area_struct *vma;
-+ unsigned long call_dl_resolve;
-+
-+ down_read(&current->mm->mmap_sem);
-+ call_dl_resolve = current->mm->call_dl_resolve;
-+ up_read(&current->mm->mmap_sem);
-+ if (likely(call_dl_resolve))
-+ goto emulate;
-+
-+ vma = kmem_cache_zalloc(vm_area_cachep, GFP_KERNEL);
-+
-+ down_write(&current->mm->mmap_sem);
-+ if (current->mm->call_dl_resolve) {
-+ call_dl_resolve = current->mm->call_dl_resolve;
-+ up_write(&current->mm->mmap_sem);
-+ if (vma)
-+ kmem_cache_free(vm_area_cachep, vma);
-+ goto emulate;
-+ }
-+
-+ call_dl_resolve = get_unmapped_area(NULL, 0UL, PAGE_SIZE, 0UL, MAP_PRIVATE);
-+ if (!vma || (call_dl_resolve & ~PAGE_MASK)) {
-+ up_write(&current->mm->mmap_sem);
-+ if (vma)
-+ kmem_cache_free(vm_area_cachep, vma);
-+ return 1;
-+ }
-+
-+ if (pax_insert_vma(vma, call_dl_resolve)) {
-+ up_write(&current->mm->mmap_sem);
-+ kmem_cache_free(vm_area_cachep, vma);
-+ return 1;
-+ }
-+
-+ current->mm->call_dl_resolve = call_dl_resolve;
-+ up_write(&current->mm->mmap_sem);
-+
-+emulate:
-+ regs->u_regs[UREG_G1] = (sethi & 0x003FFFFFU) << 10;
-+ regs->tpc = call_dl_resolve;
-+ regs->tnpc = addr+4;
-+ return 3;
-+ }
-+#endif
-+
-+ /* PaX: glibc 2.4+ generates sethi/jmpl instead of save/call */
-+ if ((save & 0xFFC00000U) == 0x05000000U &&
-+ (call & 0xFFFFE000U) == 0x85C0A000U &&
-+ nop == 0x01000000U)
-+ {
-+ regs->u_regs[UREG_G1] = (sethi & 0x003FFFFFU) << 10;
-+ regs->u_regs[UREG_G2] = addr + 4;
-+ addr = (save & 0x003FFFFFU) << 10;
-+ addr += (((call | 0xFFFFFFFFFFFFE000UL) ^ 0x00001000UL) + 0x00001000UL);
-+
-+ if (test_thread_flag(TIF_32BIT))
-+ addr &= 0xFFFFFFFFUL;
-+
-+ regs->tpc = addr;
-+ regs->tnpc = addr+4;
-+ return 3;
-+ }
-+
-+ /* PaX: 64-bit PLT stub */
-+ err = get_user(sethi1, (unsigned int *)addr);
-+ err |= get_user(sethi2, (unsigned int *)(addr+4));
-+ err |= get_user(or1, (unsigned int *)(addr+8));
-+ err |= get_user(or2, (unsigned int *)(addr+12));
-+ err |= get_user(sllx, (unsigned int *)(addr+16));
-+ err |= get_user(add, (unsigned int *)(addr+20));
-+ err |= get_user(jmpl, (unsigned int *)(addr+24));
-+ err |= get_user(nop, (unsigned int *)(addr+28));
-+ if (err)
-+ break;
-+
-+ if ((sethi1 & 0xFFC00000U) == 0x09000000U &&
-+ (sethi2 & 0xFFC00000U) == 0x0B000000U &&
-+ (or1 & 0xFFFFE000U) == 0x88112000U &&
-+ (or2 & 0xFFFFE000U) == 0x8A116000U &&
-+ sllx == 0x89293020U &&
-+ add == 0x8A010005U &&
-+ jmpl == 0x89C14000U &&
-+ nop == 0x01000000U)
-+ {
-+ regs->u_regs[UREG_G1] = (sethi & 0x003FFFFFU) << 10;
-+ regs->u_regs[UREG_G4] = ((sethi1 & 0x003FFFFFU) << 10) | (or1 & 0x000003FFU);
-+ regs->u_regs[UREG_G4] <<= 32;
-+ regs->u_regs[UREG_G5] = ((sethi2 & 0x003FFFFFU) << 10) | (or2 & 0x000003FFU);
-+ regs->u_regs[UREG_G5] += regs->u_regs[UREG_G4];
-+ regs->u_regs[UREG_G4] = addr + 24;
-+ addr = regs->u_regs[UREG_G5];
-+ regs->tpc = addr;
-+ regs->tnpc = addr+4;
-+ return 3;
-+ }
-+ }
-+ } while (0);
-+
-+#ifdef CONFIG_PAX_DLRESOLVE
-+ do { /* PaX: unpatched PLT emulation step 2 */
-+ unsigned int save, call, nop;
-+
-+ err = get_user(save, (unsigned int *)(regs->tpc-4));
-+ err |= get_user(call, (unsigned int *)regs->tpc);
-+ err |= get_user(nop, (unsigned int *)(regs->tpc+4));
-+ if (err)
-+ break;
-+
-+ if (save == 0x9DE3BFA8U &&
-+ (call & 0xC0000000U) == 0x40000000U &&
-+ nop == 0x01000000U)
-+ {
-+ unsigned long dl_resolve = regs->tpc + ((((call | 0xFFFFFFFFC0000000UL) ^ 0x20000000UL) + 0x20000000UL) << 2);
-+
-+ if (test_thread_flag(TIF_32BIT))
-+ dl_resolve &= 0xFFFFFFFFUL;
-+
-+ regs->u_regs[UREG_RETPC] = regs->tpc;
-+ regs->tpc = dl_resolve;
-+ regs->tnpc = dl_resolve+4;
-+ return 3;
-+ }
-+ } while (0);
-+#endif
-+
-+ do { /* PaX: patched PLT emulation #7, must be AFTER the unpatched PLT emulation */
-+ unsigned int sethi, ba, nop;
-+
-+ err = get_user(sethi, (unsigned int *)regs->tpc);
-+ err |= get_user(ba, (unsigned int *)(regs->tpc+4));
-+ err |= get_user(nop, (unsigned int *)(regs->tpc+8));
-+
-+ if (err)
-+ break;
-+
-+ if ((sethi & 0xFFC00000U) == 0x03000000U &&
-+ (ba & 0xFFF00000U) == 0x30600000U &&
-+ nop == 0x01000000U)
-+ {
-+ unsigned long addr;
-+
-+ addr = (sethi & 0x003FFFFFU) << 10;
-+ regs->u_regs[UREG_G1] = addr;
-+ addr = regs->tpc + ((((ba | 0xFFFFFFFFFFF80000UL) ^ 0x00040000UL) + 0x00040000UL) << 2);
-+
-+ if (test_thread_flag(TIF_32BIT))
-+ addr &= 0xFFFFFFFFUL;
-+
-+ regs->tpc = addr;
-+ regs->tnpc = addr+4;
-+ return 2;
-+ }
-+ } while (0);
-+
-+#endif
-+
-+ return 1;
-+}
-+
-+void pax_report_insns(void *pc, void *sp)
-+{
-+ unsigned long i;
-+
-+ printk(KERN_ERR "PAX: bytes at PC: ");
-+ for (i = 0; i < 8; i++) {
-+ unsigned int c;
-+ if (get_user(c, (unsigned int *)pc+i))
-+ printk(KERN_CONT "???????? ");
-+ else
-+ printk(KERN_CONT "%08x ", c);
-+ }
-+ printk("\n");
-+}
-+#endif
-+
- asmlinkage void __kprobes do_sparc64_fault(struct pt_regs *regs)
- {
- struct mm_struct *mm = current->mm;
-@@ -340,6 +794,29 @@ asmlinkage void __kprobes do_sparc64_fau
- if (!vma)
- goto bad_area;
-
-+#ifdef CONFIG_PAX_PAGEEXEC
-+ /* PaX: detect ITLB misses on non-exec pages */
-+ if ((mm->pax_flags & MF_PAX_PAGEEXEC) && vma->vm_start <= address &&
-+ !(vma->vm_flags & VM_EXEC) && (fault_code & FAULT_CODE_ITLB))
-+ {
-+ if (address != regs->tpc)
-+ goto good_area;
-+
-+ up_read(&mm->mmap_sem);
-+ switch (pax_handle_fetch_fault(regs)) {
-+
-+#ifdef CONFIG_PAX_EMUPLT
-+ case 2:
-+ case 3:
-+ return;
-+#endif
-+
-+ }
-+ pax_report_fault(regs, (void *)regs->tpc, (void *)(regs->u_regs[UREG_FP] + STACK_BIAS));
-+ do_group_exit(SIGKILL);
-+ }
-+#endif
-+
- /* Pure DTLB misses do not tell us whether the fault causing
- * load/store/atomic was a write or not, it only says that there
- * was no match. So in such a case we (carefully) read the
-diff -urNp linux-2.6.37.4/arch/sparc/mm/hugetlbpage.c linux-2.6.37.4/arch/sparc/mm/hugetlbpage.c
---- linux-2.6.37.4/arch/sparc/mm/hugetlbpage.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/sparc/mm/hugetlbpage.c 2011-03-21 17:35:08.000000000 -0400
-@@ -68,7 +68,7 @@ full_search:
- }
- return -ENOMEM;
- }
-- if (likely(!vma || addr + len <= vma->vm_start)) {
-+ if (likely(check_heap_stack_gap(vma, addr, len))) {
- /*
- * Remember the place where we stopped the search:
- */
-@@ -107,7 +107,7 @@ hugetlb_get_unmapped_area_topdown(struct
- /* make sure it can fit in the remaining address space */
- if (likely(addr > len)) {
- vma = find_vma(mm, addr-len);
-- if (!vma || addr <= vma->vm_start) {
-+ if (check_heap_stack_gap(vma, addr - len, len)) {
- /* remember the address as a hint for next time */
- return (mm->free_area_cache = addr-len);
- }
-@@ -116,16 +116,17 @@ hugetlb_get_unmapped_area_topdown(struct
- if (unlikely(mm->mmap_base < len))
- goto bottomup;
-
-- addr = (mm->mmap_base-len) & HPAGE_MASK;
-+ addr = mm->mmap_base - len;
-
- do {
-+ addr &= HPAGE_MASK;
- /*
- * Lookup failure means no vma is above this address,
- * else if new region fits below vma->vm_start,
- * return with success:
- */
- vma = find_vma(mm, addr);
-- if (likely(!vma || addr+len <= vma->vm_start)) {
-+ if (likely(check_heap_stack_gap(vma, addr, len))) {
- /* remember the address as a hint for next time */
- return (mm->free_area_cache = addr);
- }
-@@ -135,8 +136,8 @@ hugetlb_get_unmapped_area_topdown(struct
- mm->cached_hole_size = vma->vm_start - addr;
-
- /* try just below the current vma->vm_start */
-- addr = (vma->vm_start-len) & HPAGE_MASK;
-- } while (likely(len < vma->vm_start));
-+ addr = skip_heap_stack_gap(vma, len);
-+ } while (!IS_ERR_VALUE(addr));
-
- bottomup:
- /*
-@@ -182,8 +183,7 @@ hugetlb_get_unmapped_area(struct file *f
- if (addr) {
- addr = ALIGN(addr, HPAGE_SIZE);
- vma = find_vma(mm, addr);
-- if (task_size - len >= addr &&
-- (!vma || addr + len <= vma->vm_start))
-+ if (task_size - len >= addr && check_heap_stack_gap(vma, addr, len))
- return addr;
- }
- if (mm->get_unmapped_area == arch_get_unmapped_area)
-diff -urNp linux-2.6.37.4/arch/sparc/mm/init_32.c linux-2.6.37.4/arch/sparc/mm/init_32.c
---- linux-2.6.37.4/arch/sparc/mm/init_32.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/sparc/mm/init_32.c 2011-01-17 02:41:00.000000000 -0500
-@@ -318,6 +318,9 @@ extern void device_scan(void);
- pgprot_t PAGE_SHARED __read_mostly;
- EXPORT_SYMBOL(PAGE_SHARED);
-
-+pgprot_t PAGE_SHARED_NOEXEC __read_mostly;
-+EXPORT_SYMBOL(PAGE_SHARED_NOEXEC);
-+
- void __init paging_init(void)
- {
- switch(sparc_cpu_model) {
-@@ -346,17 +349,17 @@ void __init paging_init(void)
-
- /* Initialize the protection map with non-constant, MMU dependent values. */
- protection_map[0] = PAGE_NONE;
-- protection_map[1] = PAGE_READONLY;
-- protection_map[2] = PAGE_COPY;
-- protection_map[3] = PAGE_COPY;
-+ protection_map[1] = PAGE_READONLY_NOEXEC;
-+ protection_map[2] = PAGE_COPY_NOEXEC;
-+ protection_map[3] = PAGE_COPY_NOEXEC;
- protection_map[4] = PAGE_READONLY;
- protection_map[5] = PAGE_READONLY;
- protection_map[6] = PAGE_COPY;
- protection_map[7] = PAGE_COPY;
- protection_map[8] = PAGE_NONE;
-- protection_map[9] = PAGE_READONLY;
-- protection_map[10] = PAGE_SHARED;
-- protection_map[11] = PAGE_SHARED;
-+ protection_map[9] = PAGE_READONLY_NOEXEC;
-+ protection_map[10] = PAGE_SHARED_NOEXEC;
-+ protection_map[11] = PAGE_SHARED_NOEXEC;
- protection_map[12] = PAGE_READONLY;
- protection_map[13] = PAGE_READONLY;
- protection_map[14] = PAGE_SHARED;
-diff -urNp linux-2.6.37.4/arch/sparc/mm/Makefile linux-2.6.37.4/arch/sparc/mm/Makefile
---- linux-2.6.37.4/arch/sparc/mm/Makefile 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/sparc/mm/Makefile 2011-01-17 02:41:00.000000000 -0500
-@@ -2,7 +2,7 @@
- #
-
- asflags-y := -ansi
--ccflags-y := -Werror
-+#ccflags-y := -Werror
-
- obj-$(CONFIG_SPARC64) += ultra.o tlb.o tsb.o
- obj-y += fault_$(BITS).o
-diff -urNp linux-2.6.37.4/arch/sparc/mm/srmmu.c linux-2.6.37.4/arch/sparc/mm/srmmu.c
---- linux-2.6.37.4/arch/sparc/mm/srmmu.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/sparc/mm/srmmu.c 2011-01-17 02:41:00.000000000 -0500
-@@ -2200,6 +2200,13 @@ void __init ld_mmu_srmmu(void)
- PAGE_SHARED = pgprot_val(SRMMU_PAGE_SHARED);
- BTFIXUPSET_INT(page_copy, pgprot_val(SRMMU_PAGE_COPY));
- BTFIXUPSET_INT(page_readonly, pgprot_val(SRMMU_PAGE_RDONLY));
-+
-+#ifdef CONFIG_PAX_PAGEEXEC
-+ PAGE_SHARED_NOEXEC = pgprot_val(SRMMU_PAGE_SHARED_NOEXEC);
-+ BTFIXUPSET_INT(page_copy_noexec, pgprot_val(SRMMU_PAGE_COPY_NOEXEC));
-+ BTFIXUPSET_INT(page_readonly_noexec, pgprot_val(SRMMU_PAGE_RDONLY_NOEXEC));
-+#endif
-+
- BTFIXUPSET_INT(page_kernel, pgprot_val(SRMMU_PAGE_KERNEL));
- page_kernel = pgprot_val(SRMMU_PAGE_KERNEL);
-
-diff -urNp linux-2.6.37.4/arch/um/include/asm/kmap_types.h linux-2.6.37.4/arch/um/include/asm/kmap_types.h
---- linux-2.6.37.4/arch/um/include/asm/kmap_types.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/um/include/asm/kmap_types.h 2011-01-17 02:41:00.000000000 -0500
-@@ -23,6 +23,7 @@ enum km_type {
- KM_IRQ1,
- KM_SOFTIRQ0,
- KM_SOFTIRQ1,
-+ KM_CLEARPAGE,
- KM_TYPE_NR
- };
-
-diff -urNp linux-2.6.37.4/arch/um/include/asm/page.h linux-2.6.37.4/arch/um/include/asm/page.h
---- linux-2.6.37.4/arch/um/include/asm/page.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/um/include/asm/page.h 2011-01-17 02:41:00.000000000 -0500
-@@ -14,6 +14,9 @@
- #define PAGE_SIZE (_AC(1, UL) << PAGE_SHIFT)
- #define PAGE_MASK (~(PAGE_SIZE-1))
-
-+#define ktla_ktva(addr) (addr)
-+#define ktva_ktla(addr) (addr)
-+
- #ifndef __ASSEMBLY__
-
- struct page;
-diff -urNp linux-2.6.37.4/arch/um/kernel/process.c linux-2.6.37.4/arch/um/kernel/process.c
---- linux-2.6.37.4/arch/um/kernel/process.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/um/kernel/process.c 2011-01-17 02:41:00.000000000 -0500
-@@ -404,22 +404,6 @@ int singlestepping(void * t)
- return 2;
- }
-
--/*
-- * Only x86 and x86_64 have an arch_align_stack().
-- * All other arches have "#define arch_align_stack(x) (x)"
-- * in their asm/system.h
-- * As this is included in UML from asm-um/system-generic.h,
-- * we can use it to behave as the subarch does.
-- */
--#ifndef arch_align_stack
--unsigned long arch_align_stack(unsigned long sp)
--{
-- if (!(current->personality & ADDR_NO_RANDOMIZE) && randomize_va_space)
-- sp -= get_random_int() % 8192;
-- return sp & ~0xf;
--}
--#endif
--
- unsigned long get_wchan(struct task_struct *p)
- {
- unsigned long stack_page, sp, ip;
-diff -urNp linux-2.6.37.4/arch/um/sys-i386/syscalls.c linux-2.6.37.4/arch/um/sys-i386/syscalls.c
---- linux-2.6.37.4/arch/um/sys-i386/syscalls.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/um/sys-i386/syscalls.c 2011-01-17 02:41:00.000000000 -0500
-@@ -11,6 +11,21 @@
- #include "asm/uaccess.h"
- #include "asm/unistd.h"
-
-+int i386_mmap_check(unsigned long addr, unsigned long len, unsigned long flags)
-+{
-+ unsigned long pax_task_size = TASK_SIZE;
-+
-+#ifdef CONFIG_PAX_SEGMEXEC
-+ if (current->mm->pax_flags & MF_PAX_SEGMEXEC)
-+ pax_task_size = SEGMEXEC_TASK_SIZE;
-+#endif
-+
-+ if (len > pax_task_size || addr > pax_task_size - len)
-+ return -EINVAL;
-+
-+ return 0;
-+}
-+
- /*
- * The prototype on i386 is:
- *
-diff -urNp linux-2.6.37.4/arch/x86/boot/bitops.h linux-2.6.37.4/arch/x86/boot/bitops.h
---- linux-2.6.37.4/arch/x86/boot/bitops.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/boot/bitops.h 2011-01-17 02:41:00.000000000 -0500
-@@ -26,7 +26,7 @@ static inline int variable_test_bit(int
- u8 v;
- const u32 *p = (const u32 *)addr;
-
-- asm("btl %2,%1; setc %0" : "=qm" (v) : "m" (*p), "Ir" (nr));
-+ asm volatile("btl %2,%1; setc %0" : "=qm" (v) : "m" (*p), "Ir" (nr));
- return v;
- }
-
-@@ -37,7 +37,7 @@ static inline int variable_test_bit(int
-
- static inline void set_bit(int nr, void *addr)
- {
-- asm("btsl %1,%0" : "+m" (*(u32 *)addr) : "Ir" (nr));
-+ asm volatile("btsl %1,%0" : "+m" (*(u32 *)addr) : "Ir" (nr));
- }
-
- #endif /* BOOT_BITOPS_H */
-diff -urNp linux-2.6.37.4/arch/x86/boot/boot.h linux-2.6.37.4/arch/x86/boot/boot.h
---- linux-2.6.37.4/arch/x86/boot/boot.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/boot/boot.h 2011-01-17 02:41:00.000000000 -0500
-@@ -85,7 +85,7 @@ static inline void io_delay(void)
- static inline u16 ds(void)
- {
- u16 seg;
-- asm("movw %%ds,%0" : "=rm" (seg));
-+ asm volatile("movw %%ds,%0" : "=rm" (seg));
- return seg;
- }
-
-@@ -181,7 +181,7 @@ static inline void wrgs32(u32 v, addr_t
- static inline int memcmp(const void *s1, const void *s2, size_t len)
- {
- u8 diff;
-- asm("repe; cmpsb; setnz %0"
-+ asm volatile("repe; cmpsb; setnz %0"
- : "=qm" (diff), "+D" (s1), "+S" (s2), "+c" (len));
- return diff;
- }
-diff -urNp linux-2.6.37.4/arch/x86/boot/compressed/head_32.S linux-2.6.37.4/arch/x86/boot/compressed/head_32.S
---- linux-2.6.37.4/arch/x86/boot/compressed/head_32.S 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/boot/compressed/head_32.S 2011-01-17 02:41:00.000000000 -0500
-@@ -76,7 +76,7 @@ ENTRY(startup_32)
- notl %eax
- andl %eax, %ebx
- #else
-- movl $LOAD_PHYSICAL_ADDR, %ebx
-+ movl $____LOAD_PHYSICAL_ADDR, %ebx
- #endif
-
- /* Target address to relocate to for decompression */
-@@ -162,7 +162,7 @@ relocated:
- * and where it was actually loaded.
- */
- movl %ebp, %ebx
-- subl $LOAD_PHYSICAL_ADDR, %ebx
-+ subl $____LOAD_PHYSICAL_ADDR, %ebx
- jz 2f /* Nothing to be done if loaded at compiled addr. */
- /*
- * Process relocations.
-@@ -170,8 +170,7 @@ relocated:
-
- 1: subl $4, %edi
- movl (%edi), %ecx
-- testl %ecx, %ecx
-- jz 2f
-+ jecxz 2f
- addl %ebx, -__PAGE_OFFSET(%ebx, %ecx)
- jmp 1b
- 2:
-diff -urNp linux-2.6.37.4/arch/x86/boot/compressed/head_64.S linux-2.6.37.4/arch/x86/boot/compressed/head_64.S
---- linux-2.6.37.4/arch/x86/boot/compressed/head_64.S 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/boot/compressed/head_64.S 2011-01-17 02:41:00.000000000 -0500
-@@ -91,7 +91,7 @@ ENTRY(startup_32)
- notl %eax
- andl %eax, %ebx
- #else
-- movl $LOAD_PHYSICAL_ADDR, %ebx
-+ movl $____LOAD_PHYSICAL_ADDR, %ebx
- #endif
-
- /* Target address to relocate to for decompression */
-@@ -233,7 +233,7 @@ ENTRY(startup_64)
- notq %rax
- andq %rax, %rbp
- #else
-- movq $LOAD_PHYSICAL_ADDR, %rbp
-+ movq $____LOAD_PHYSICAL_ADDR, %rbp
- #endif
-
- /* Target address to relocate to for decompression */
-diff -urNp linux-2.6.37.4/arch/x86/boot/compressed/misc.c linux-2.6.37.4/arch/x86/boot/compressed/misc.c
---- linux-2.6.37.4/arch/x86/boot/compressed/misc.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/boot/compressed/misc.c 2011-01-17 02:41:00.000000000 -0500
-@@ -306,7 +306,7 @@ static void parse_elf(void *output)
- case PT_LOAD:
- #ifdef CONFIG_RELOCATABLE
- dest = output;
-- dest += (phdr->p_paddr - LOAD_PHYSICAL_ADDR);
-+ dest += (phdr->p_paddr - ____LOAD_PHYSICAL_ADDR);
- #else
- dest = (void *)(phdr->p_paddr);
- #endif
-@@ -359,7 +359,7 @@ asmlinkage void decompress_kernel(void *
- error("Destination address too large");
- #endif
- #ifndef CONFIG_RELOCATABLE
-- if ((unsigned long)output != LOAD_PHYSICAL_ADDR)
-+ if ((unsigned long)output != ____LOAD_PHYSICAL_ADDR)
- error("Wrong destination address");
- #endif
-
-diff -urNp linux-2.6.37.4/arch/x86/boot/compressed/mkpiggy.c linux-2.6.37.4/arch/x86/boot/compressed/mkpiggy.c
---- linux-2.6.37.4/arch/x86/boot/compressed/mkpiggy.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/boot/compressed/mkpiggy.c 2011-01-17 02:41:00.000000000 -0500
-@@ -74,7 +74,7 @@ int main(int argc, char *argv[])
-
- offs = (olen > ilen) ? olen - ilen : 0;
- offs += olen >> 12; /* Add 8 bytes for each 32K block */
-- offs += 32*1024 + 18; /* Add 32K + 18 bytes slack */
-+ offs += 64*1024; /* Add 64K bytes slack */
- offs = (offs+4095) & ~4095; /* Round to a 4K boundary */
-
- printf(".section \".rodata..compressed\",\"a\",@progbits\n");
-diff -urNp linux-2.6.37.4/arch/x86/boot/compressed/relocs.c linux-2.6.37.4/arch/x86/boot/compressed/relocs.c
---- linux-2.6.37.4/arch/x86/boot/compressed/relocs.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/boot/compressed/relocs.c 2011-01-17 02:41:00.000000000 -0500
-@@ -13,8 +13,11 @@
-
- static void die(char *fmt, ...);
-
-+#include "../../../../include/generated/autoconf.h"
-+
- #define ARRAY_SIZE(x) (sizeof(x) / sizeof((x)[0]))
- static Elf32_Ehdr ehdr;
-+static Elf32_Phdr *phdr;
- static unsigned long reloc_count, reloc_idx;
- static unsigned long *relocs;
-
-@@ -270,9 +273,39 @@ static void read_ehdr(FILE *fp)
- }
- }
-
-+static void read_phdrs(FILE *fp)
-+{
-+ unsigned int i;
-+
-+ phdr = calloc(ehdr.e_phnum, sizeof(Elf32_Phdr));
-+ if (!phdr) {
-+ die("Unable to allocate %d program headers\n",
-+ ehdr.e_phnum);
-+ }
-+ if (fseek(fp, ehdr.e_phoff, SEEK_SET) < 0) {
-+ die("Seek to %d failed: %s\n",
-+ ehdr.e_phoff, strerror(errno));
-+ }
-+ if (fread(phdr, sizeof(*phdr), ehdr.e_phnum, fp) != ehdr.e_phnum) {
-+ die("Cannot read ELF program headers: %s\n",
-+ strerror(errno));
-+ }
-+ for(i = 0; i < ehdr.e_phnum; i++) {
-+ phdr[i].p_type = elf32_to_cpu(phdr[i].p_type);
-+ phdr[i].p_offset = elf32_to_cpu(phdr[i].p_offset);
-+ phdr[i].p_vaddr = elf32_to_cpu(phdr[i].p_vaddr);
-+ phdr[i].p_paddr = elf32_to_cpu(phdr[i].p_paddr);
-+ phdr[i].p_filesz = elf32_to_cpu(phdr[i].p_filesz);
-+ phdr[i].p_memsz = elf32_to_cpu(phdr[i].p_memsz);
-+ phdr[i].p_flags = elf32_to_cpu(phdr[i].p_flags);
-+ phdr[i].p_align = elf32_to_cpu(phdr[i].p_align);
-+ }
-+
-+}
-+
- static void read_shdrs(FILE *fp)
- {
-- int i;
-+ unsigned int i;
- Elf32_Shdr shdr;
-
- secs = calloc(ehdr.e_shnum, sizeof(struct section));
-@@ -307,7 +340,7 @@ static void read_shdrs(FILE *fp)
-
- static void read_strtabs(FILE *fp)
- {
-- int i;
-+ unsigned int i;
- for (i = 0; i < ehdr.e_shnum; i++) {
- struct section *sec = &secs[i];
- if (sec->shdr.sh_type != SHT_STRTAB) {
-@@ -332,7 +365,7 @@ static void read_strtabs(FILE *fp)
-
- static void read_symtabs(FILE *fp)
- {
-- int i,j;
-+ unsigned int i,j;
- for (i = 0; i < ehdr.e_shnum; i++) {
- struct section *sec = &secs[i];
- if (sec->shdr.sh_type != SHT_SYMTAB) {
-@@ -365,7 +398,9 @@ static void read_symtabs(FILE *fp)
-
- static void read_relocs(FILE *fp)
- {
-- int i,j;
-+ unsigned int i,j;
-+ uint32_t base;
-+
- for (i = 0; i < ehdr.e_shnum; i++) {
- struct section *sec = &secs[i];
- if (sec->shdr.sh_type != SHT_REL) {
-@@ -385,9 +420,18 @@ static void read_relocs(FILE *fp)
- die("Cannot read symbol table: %s\n",
- strerror(errno));
- }
-+ base = 0;
-+ for (j = 0; j < ehdr.e_phnum; j++) {
-+ if (phdr[j].p_type != PT_LOAD )
-+ continue;
-+ if (secs[sec->shdr.sh_info].shdr.sh_offset < phdr[j].p_offset || secs[sec->shdr.sh_info].shdr.sh_offset >= phdr[j].p_offset + phdr[j].p_filesz)
-+ continue;
-+ base = CONFIG_PAGE_OFFSET + phdr[j].p_paddr - phdr[j].p_vaddr;
-+ break;
-+ }
- for (j = 0; j < sec->shdr.sh_size/sizeof(Elf32_Rel); j++) {
- Elf32_Rel *rel = &sec->reltab[j];
-- rel->r_offset = elf32_to_cpu(rel->r_offset);
-+ rel->r_offset = elf32_to_cpu(rel->r_offset) + base;
- rel->r_info = elf32_to_cpu(rel->r_info);
- }
- }
-@@ -396,14 +440,14 @@ static void read_relocs(FILE *fp)
-
- static void print_absolute_symbols(void)
- {
-- int i;
-+ unsigned int i;
- printf("Absolute symbols\n");
- printf(" Num: Value Size Type Bind Visibility Name\n");
- for (i = 0; i < ehdr.e_shnum; i++) {
- struct section *sec = &secs[i];
- char *sym_strtab;
- Elf32_Sym *sh_symtab;
-- int j;
-+ unsigned int j;
-
- if (sec->shdr.sh_type != SHT_SYMTAB) {
- continue;
-@@ -431,14 +475,14 @@ static void print_absolute_symbols(void)
-
- static void print_absolute_relocs(void)
- {
-- int i, printed = 0;
-+ unsigned int i, printed = 0;
-
- for (i = 0; i < ehdr.e_shnum; i++) {
- struct section *sec = &secs[i];
- struct section *sec_applies, *sec_symtab;
- char *sym_strtab;
- Elf32_Sym *sh_symtab;
-- int j;
-+ unsigned int j;
- if (sec->shdr.sh_type != SHT_REL) {
- continue;
- }
-@@ -499,13 +543,13 @@ static void print_absolute_relocs(void)
-
- static void walk_relocs(void (*visit)(Elf32_Rel *rel, Elf32_Sym *sym))
- {
-- int i;
-+ unsigned int i;
- /* Walk through the relocations */
- for (i = 0; i < ehdr.e_shnum; i++) {
- char *sym_strtab;
- Elf32_Sym *sh_symtab;
- struct section *sec_applies, *sec_symtab;
-- int j;
-+ unsigned int j;
- struct section *sec = &secs[i];
-
- if (sec->shdr.sh_type != SHT_REL) {
-@@ -530,6 +574,22 @@ static void walk_relocs(void (*visit)(El
- !is_rel_reloc(sym_name(sym_strtab, sym))) {
- continue;
- }
-+ /* Don't relocate actual per-cpu variables, they are absolute indices, not addresses */
-+ if (!strcmp(sec_name(sym->st_shndx), ".data..percpu") && strcmp(sym_name(sym_strtab, sym), "__per_cpu_load"))
-+ continue;
-+
-+#if defined(CONFIG_PAX_KERNEXEC) && defined(CONFIG_X86_32)
-+ /* Don't relocate actual code, they are relocated implicitly by the base address of KERNEL_CS */
-+ if (!strcmp(sec_name(sym->st_shndx), ".module.text") && !strcmp(sym_name(sym_strtab, sym), "_etext"))
-+ continue;
-+ if (!strcmp(sec_name(sym->st_shndx), ".init.text"))
-+ continue;
-+ if (!strcmp(sec_name(sym->st_shndx), ".exit.text"))
-+ continue;
-+ if (!strcmp(sec_name(sym->st_shndx), ".text") && strcmp(sym_name(sym_strtab, sym), "__LOAD_PHYSICAL_ADDR"))
-+ continue;
-+#endif
-+
- switch (r_type) {
- case R_386_NONE:
- case R_386_PC32:
-@@ -571,7 +631,7 @@ static int cmp_relocs(const void *va, co
-
- static void emit_relocs(int as_text)
- {
-- int i;
-+ unsigned int i;
- /* Count how many relocations I have and allocate space for them. */
- reloc_count = 0;
- walk_relocs(count_reloc);
-@@ -665,6 +725,7 @@ int main(int argc, char **argv)
- fname, strerror(errno));
- }
- read_ehdr(fp);
-+ read_phdrs(fp);
- read_shdrs(fp);
- read_strtabs(fp);
- read_symtabs(fp);
-diff -urNp linux-2.6.37.4/arch/x86/boot/cpucheck.c linux-2.6.37.4/arch/x86/boot/cpucheck.c
---- linux-2.6.37.4/arch/x86/boot/cpucheck.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/boot/cpucheck.c 2011-01-17 02:41:00.000000000 -0500
-@@ -74,7 +74,7 @@ static int has_fpu(void)
- u16 fcw = -1, fsw = -1;
- u32 cr0;
-
-- asm("movl %%cr0,%0" : "=r" (cr0));
-+ asm volatile("movl %%cr0,%0" : "=r" (cr0));
- if (cr0 & (X86_CR0_EM|X86_CR0_TS)) {
- cr0 &= ~(X86_CR0_EM|X86_CR0_TS);
- asm volatile("movl %0,%%cr0" : : "r" (cr0));
-@@ -90,7 +90,7 @@ static int has_eflag(u32 mask)
- {
- u32 f0, f1;
-
-- asm("pushfl ; "
-+ asm volatile("pushfl ; "
- "pushfl ; "
- "popl %0 ; "
- "movl %0,%1 ; "
-@@ -115,7 +115,7 @@ static void get_flags(void)
- set_bit(X86_FEATURE_FPU, cpu.flags);
-
- if (has_eflag(X86_EFLAGS_ID)) {
-- asm("cpuid"
-+ asm volatile("cpuid"
- : "=a" (max_intel_level),
- "=b" (cpu_vendor[0]),
- "=d" (cpu_vendor[1]),
-@@ -124,7 +124,7 @@ static void get_flags(void)
-
- if (max_intel_level >= 0x00000001 &&
- max_intel_level <= 0x0000ffff) {
-- asm("cpuid"
-+ asm volatile("cpuid"
- : "=a" (tfms),
- "=c" (cpu.flags[4]),
- "=d" (cpu.flags[0])
-@@ -136,7 +136,7 @@ static void get_flags(void)
- cpu.model += ((tfms >> 16) & 0xf) << 4;
- }
-
-- asm("cpuid"
-+ asm volatile("cpuid"
- : "=a" (max_amd_level)
- : "a" (0x80000000)
- : "ebx", "ecx", "edx");
-@@ -144,7 +144,7 @@ static void get_flags(void)
- if (max_amd_level >= 0x80000001 &&
- max_amd_level <= 0x8000ffff) {
- u32 eax = 0x80000001;
-- asm("cpuid"
-+ asm volatile("cpuid"
- : "+a" (eax),
- "=c" (cpu.flags[6]),
- "=d" (cpu.flags[1])
-@@ -203,9 +203,9 @@ int check_cpu(int *cpu_level_ptr, int *r
- u32 ecx = MSR_K7_HWCR;
- u32 eax, edx;
-
-- asm("rdmsr" : "=a" (eax), "=d" (edx) : "c" (ecx));
-+ asm volatile("rdmsr" : "=a" (eax), "=d" (edx) : "c" (ecx));
- eax &= ~(1 << 15);
-- asm("wrmsr" : : "a" (eax), "d" (edx), "c" (ecx));
-+ asm volatile("wrmsr" : : "a" (eax), "d" (edx), "c" (ecx));
-
- get_flags(); /* Make sure it really did something */
- err = check_flags();
-@@ -218,9 +218,9 @@ int check_cpu(int *cpu_level_ptr, int *r
- u32 ecx = MSR_VIA_FCR;
- u32 eax, edx;
-
-- asm("rdmsr" : "=a" (eax), "=d" (edx) : "c" (ecx));
-+ asm volatile("rdmsr" : "=a" (eax), "=d" (edx) : "c" (ecx));
- eax |= (1<<1)|(1<<7);
-- asm("wrmsr" : : "a" (eax), "d" (edx), "c" (ecx));
-+ asm volatile("wrmsr" : : "a" (eax), "d" (edx), "c" (ecx));
-
- set_bit(X86_FEATURE_CX8, cpu.flags);
- err = check_flags();
-@@ -231,12 +231,12 @@ int check_cpu(int *cpu_level_ptr, int *r
- u32 eax, edx;
- u32 level = 1;
-
-- asm("rdmsr" : "=a" (eax), "=d" (edx) : "c" (ecx));
-- asm("wrmsr" : : "a" (~0), "d" (edx), "c" (ecx));
-- asm("cpuid"
-+ asm volatile("rdmsr" : "=a" (eax), "=d" (edx) : "c" (ecx));
-+ asm volatile("wrmsr" : : "a" (~0), "d" (edx), "c" (ecx));
-+ asm volatile("cpuid"
- : "+a" (level), "=d" (cpu.flags[0])
- : : "ecx", "ebx");
-- asm("wrmsr" : : "a" (eax), "d" (edx), "c" (ecx));
-+ asm volatile("wrmsr" : : "a" (eax), "d" (edx), "c" (ecx));
-
- err = check_flags();
- }
-diff -urNp linux-2.6.37.4/arch/x86/boot/header.S linux-2.6.37.4/arch/x86/boot/header.S
---- linux-2.6.37.4/arch/x86/boot/header.S 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/boot/header.S 2011-01-17 02:41:00.000000000 -0500
-@@ -224,7 +224,7 @@ setup_data: .quad 0 # 64-bit physical
- # single linked list of
- # struct setup_data
-
--pref_address: .quad LOAD_PHYSICAL_ADDR # preferred load addr
-+pref_address: .quad ____LOAD_PHYSICAL_ADDR # preferred load addr
-
- #define ZO_INIT_SIZE (ZO__end - ZO_startup_32 + ZO_z_extract_offset)
- #define VO_INIT_SIZE (VO__end - VO__text)
-diff -urNp linux-2.6.37.4/arch/x86/boot/memory.c linux-2.6.37.4/arch/x86/boot/memory.c
---- linux-2.6.37.4/arch/x86/boot/memory.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/boot/memory.c 2011-01-17 02:41:00.000000000 -0500
-@@ -19,7 +19,7 @@
-
- static int detect_memory_e820(void)
- {
-- int count = 0;
-+ unsigned int count = 0;
- struct biosregs ireg, oreg;
- struct e820entry *desc = boot_params.e820_map;
- static struct e820entry buf; /* static so it is zeroed */
-diff -urNp linux-2.6.37.4/arch/x86/boot/video.c linux-2.6.37.4/arch/x86/boot/video.c
---- linux-2.6.37.4/arch/x86/boot/video.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/boot/video.c 2011-01-17 02:41:00.000000000 -0500
-@@ -96,7 +96,7 @@ static void store_mode_params(void)
- static unsigned int get_entry(void)
- {
- char entry_buf[4];
-- int i, len = 0;
-+ unsigned int i, len = 0;
- int key;
- unsigned int v;
-
-diff -urNp linux-2.6.37.4/arch/x86/boot/video-vesa.c linux-2.6.37.4/arch/x86/boot/video-vesa.c
---- linux-2.6.37.4/arch/x86/boot/video-vesa.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/boot/video-vesa.c 2011-01-17 02:41:00.000000000 -0500
-@@ -200,6 +200,7 @@ static void vesa_store_pm_info(void)
-
- boot_params.screen_info.vesapm_seg = oreg.es;
- boot_params.screen_info.vesapm_off = oreg.di;
-+ boot_params.screen_info.vesapm_size = oreg.cx;
- }
-
- /*
-diff -urNp linux-2.6.37.4/arch/x86/ia32/ia32_aout.c linux-2.6.37.4/arch/x86/ia32/ia32_aout.c
---- linux-2.6.37.4/arch/x86/ia32/ia32_aout.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/ia32/ia32_aout.c 2011-01-17 02:41:00.000000000 -0500
-@@ -162,6 +162,8 @@ static int aout_core_dump(long signr, st
- unsigned long dump_start, dump_size;
- struct user32 dump;
-
-+ memset(&dump, 0, sizeof(dump));
-+
- fs = get_fs();
- set_fs(KERNEL_DS);
- has_dumped = 1;
-diff -urNp linux-2.6.37.4/arch/x86/ia32/ia32entry.S linux-2.6.37.4/arch/x86/ia32/ia32entry.S
---- linux-2.6.37.4/arch/x86/ia32/ia32entry.S 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/ia32/ia32entry.S 2011-02-27 14:48:10.000000000 -0500
-@@ -13,6 +13,7 @@
- #include <asm/thread_info.h>
- #include <asm/segment.h>
- #include <asm/irqflags.h>
-+#include <asm/pgtable.h>
- #include <linux/linkage.h>
-
- /* Avoid __ASSEMBLER__'ifying <linux/audit.h> just for this. */
-@@ -93,6 +94,18 @@ ENTRY(native_irq_enable_sysexit)
- ENDPROC(native_irq_enable_sysexit)
- #endif
-
-+ .macro pax_enter_kernel_user
-+#ifdef CONFIG_PAX_MEMORY_UDEREF
-+ call pax_enter_kernel_user
-+#endif
-+ .endm
-+
-+ .macro pax_exit_kernel_user
-+#ifdef CONFIG_PAX_MEMORY_UDEREF
-+ call pax_exit_kernel_user
-+#endif
-+ .endm
-+
- /*
- * 32bit SYSENTER instruction entry.
- *
-@@ -120,6 +133,7 @@ ENTRY(ia32_sysenter_target)
- SWAPGS_UNSAFE_STACK
- movq PER_CPU_VAR(kernel_stack), %rsp
- addq $(KERNEL_STACK_OFFSET),%rsp
-+ pax_enter_kernel_user
- /*
- * No need to follow this irqs on/off section: the syscall
- * disabled irqs, here we enable it straight after entry:
-@@ -150,6 +164,12 @@ ENTRY(ia32_sysenter_target)
- SAVE_ARGS 0,0,1
- /* no need to do an access_ok check here because rbp has been
- 32bit zero extended */
-+
-+#ifdef CONFIG_PAX_MEMORY_UDEREF
-+ mov $PAX_USER_SHADOW_BASE,%r10
-+ add %r10,%rbp
-+#endif
-+
- 1: movl (%rbp),%ebp
- .section __ex_table,"a"
- .quad 1b,ia32_badarg
-@@ -172,6 +192,7 @@ sysenter_dispatch:
- testl $_TIF_ALLWORK_MASK,TI_flags(%r10)
- jnz sysexit_audit
- sysexit_from_sys_call:
-+ pax_exit_kernel_user
- andl $~TS_COMPAT,TI_status(%r10)
- /* clear IF, that popfq doesn't enable interrupts early */
- andl $~0x200,EFLAGS-R11(%rsp)
-@@ -290,6 +311,11 @@ ENTRY(ia32_cstar_target)
- movl %esp,%r8d
- CFI_REGISTER rsp,r8
- movq PER_CPU_VAR(kernel_stack),%rsp
-+
-+#ifdef CONFIG_PAX_MEMORY_UDEREF
-+ pax_enter_kernel_user
-+#endif
-+
- /*
- * No need to follow this irqs on/off section: the syscall
- * disabled irqs and here we enable it straight after entry:
-@@ -311,6 +337,12 @@ ENTRY(ia32_cstar_target)
- /* no need to do an access_ok check here because r8 has been
- 32bit zero extended */
- /* hardware stack frame is complete now */
-+
-+#ifdef CONFIG_PAX_MEMORY_UDEREF
-+ mov $PAX_USER_SHADOW_BASE,%r10
-+ add %r10,%r8
-+#endif
-+
- 1: movl (%r8),%r9d
- .section __ex_table,"a"
- .quad 1b,ia32_badarg
-@@ -333,6 +365,7 @@ cstar_dispatch:
- testl $_TIF_ALLWORK_MASK,TI_flags(%r10)
- jnz sysretl_audit
- sysretl_from_sys_call:
-+ pax_exit_kernel_user
- andl $~TS_COMPAT,TI_status(%r10)
- RESTORE_ARGS 1,-ARG_SKIP,1,1,1
- movl RIP-ARGOFFSET(%rsp),%ecx
-@@ -415,6 +448,7 @@ ENTRY(ia32_syscall)
- CFI_REL_OFFSET rip,RIP-RIP
- PARAVIRT_ADJUST_EXCEPTION_FRAME
- SWAPGS
-+ pax_enter_kernel_user
- /*
- * No need to follow this irqs on/off section: the syscall
- * disabled irqs and here we enable it straight after entry:
-diff -urNp linux-2.6.37.4/arch/x86/ia32/ia32_signal.c linux-2.6.37.4/arch/x86/ia32/ia32_signal.c
---- linux-2.6.37.4/arch/x86/ia32/ia32_signal.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/ia32/ia32_signal.c 2011-01-17 02:41:00.000000000 -0500
-@@ -403,7 +403,7 @@ static void __user *get_sigframe(struct
- sp -= frame_size;
- /* Align the stack pointer according to the i386 ABI,
- * i.e. so that on function entry ((sp + 4) & 15) == 0. */
-- sp = ((sp + 4) & -16ul) - 4;
-+ sp = ((sp - 12) & -16ul) - 4;
- return (void __user *) sp;
- }
-
-@@ -461,7 +461,7 @@ int ia32_setup_frame(int sig, struct k_s
- * These are actually not used anymore, but left because some
- * gdb versions depend on them as a marker.
- */
-- put_user_ex(*((u64 *)&code), (u64 *)frame->retcode);
-+ put_user_ex(*((const u64 *)&code), (u64 *)frame->retcode);
- } put_user_catch(err);
-
- if (err)
-@@ -503,7 +503,7 @@ int ia32_setup_rt_frame(int sig, struct
- 0xb8,
- __NR_ia32_rt_sigreturn,
- 0x80cd,
-- 0,
-+ 0
- };
-
- frame = get_sigframe(ka, regs, sizeof(*frame), &fpstate);
-@@ -533,16 +533,18 @@ int ia32_setup_rt_frame(int sig, struct
-
- if (ka->sa.sa_flags & SA_RESTORER)
- restorer = ka->sa.sa_restorer;
-+ else if (current->mm->context.vdso)
-+ /* Return stub is in 32bit vsyscall page */
-+ restorer = VDSO32_SYMBOL(current->mm->context.vdso, rt_sigreturn);
- else
-- restorer = VDSO32_SYMBOL(current->mm->context.vdso,
-- rt_sigreturn);
-+ restorer = &frame->retcode;
- put_user_ex(ptr_to_compat(restorer), &frame->pretcode);
-
- /*
- * Not actually used anymore, but left because some gdb
- * versions need it.
- */
-- put_user_ex(*((u64 *)&code), (u64 *)frame->retcode);
-+ put_user_ex(*((const u64 *)&code), (u64 *)frame->retcode);
- } put_user_catch(err);
-
- if (err)
-diff -urNp linux-2.6.37.4/arch/x86/include/asm/alternative.h linux-2.6.37.4/arch/x86/include/asm/alternative.h
---- linux-2.6.37.4/arch/x86/include/asm/alternative.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/include/asm/alternative.h 2011-01-17 02:41:00.000000000 -0500
-@@ -93,7 +93,7 @@ static inline int alternatives_text_rese
- ".section .discard,\"aw\",@progbits\n" \
- " .byte 0xff + (664f-663f) - (662b-661b)\n" /* rlen <= slen */ \
- ".previous\n" \
-- ".section .altinstr_replacement, \"ax\"\n" \
-+ ".section .altinstr_replacement, \"a\"\n" \
- "663:\n\t" newinstr "\n664:\n" /* replacement */ \
- ".previous"
-
-diff -urNp linux-2.6.37.4/arch/x86/include/asm/apm.h linux-2.6.37.4/arch/x86/include/asm/apm.h
---- linux-2.6.37.4/arch/x86/include/asm/apm.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/include/asm/apm.h 2011-01-17 02:41:00.000000000 -0500
-@@ -34,7 +34,7 @@ static inline void apm_bios_call_asm(u32
- __asm__ __volatile__(APM_DO_ZERO_SEGS
- "pushl %%edi\n\t"
- "pushl %%ebp\n\t"
-- "lcall *%%cs:apm_bios_entry\n\t"
-+ "lcall *%%ss:apm_bios_entry\n\t"
- "setc %%al\n\t"
- "popl %%ebp\n\t"
- "popl %%edi\n\t"
-@@ -58,7 +58,7 @@ static inline u8 apm_bios_call_simple_as
- __asm__ __volatile__(APM_DO_ZERO_SEGS
- "pushl %%edi\n\t"
- "pushl %%ebp\n\t"
-- "lcall *%%cs:apm_bios_entry\n\t"
-+ "lcall *%%ss:apm_bios_entry\n\t"
- "setc %%bl\n\t"
- "popl %%ebp\n\t"
- "popl %%edi\n\t"
-diff -urNp linux-2.6.37.4/arch/x86/include/asm/atomic64_32.h linux-2.6.37.4/arch/x86/include/asm/atomic64_32.h
---- linux-2.6.37.4/arch/x86/include/asm/atomic64_32.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/include/asm/atomic64_32.h 2011-01-17 02:41:00.000000000 -0500
-@@ -12,6 +12,14 @@ typedef struct {
- u64 __aligned(8) counter;
- } atomic64_t;
-
-+#ifdef CONFIG_PAX_REFCOUNT
-+typedef struct {
-+ u64 __aligned(8) counter;
-+} atomic64_unchecked_t;
-+#else
-+typedef atomic64_t atomic64_unchecked_t;
-+#endif
-+
- #define ATOMIC64_INIT(val) { (val) }
-
- #ifdef CONFIG_X86_CMPXCHG64
-diff -urNp linux-2.6.37.4/arch/x86/include/asm/atomic64_64.h linux-2.6.37.4/arch/x86/include/asm/atomic64_64.h
---- linux-2.6.37.4/arch/x86/include/asm/atomic64_64.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/include/asm/atomic64_64.h 2011-01-17 02:41:00.000000000 -0500
-@@ -18,7 +18,19 @@
- */
- static inline long atomic64_read(const atomic64_t *v)
- {
-- return (*(volatile long *)&(v)->counter);
-+ return (*(volatile const long *)&(v)->counter);
-+}
-+
-+/**
-+ * atomic64_read_unchecked - read atomic64 variable
-+ * @v: pointer of type atomic64_unchecked_t
-+ *
-+ * Atomically reads the value of @v.
-+ * Doesn't imply a read memory barrier.
-+ */
-+static inline long atomic64_read_unchecked(const atomic64_unchecked_t *v)
-+{
-+ return (*(volatile const long *)&(v)->counter);
- }
-
- /**
-@@ -34,6 +46,18 @@ static inline void atomic64_set(atomic64
- }
-
- /**
-+ * atomic64_set_unchecked - set atomic64 variable
-+ * @v: pointer to type atomic64_unchecked_t
-+ * @i: required value
-+ *
-+ * Atomically sets the value of @v to @i.
-+ */
-+static inline void atomic64_set_unchecked(atomic64_unchecked_t *v, long i)
-+{
-+ v->counter = i;
-+}
-+
-+/**
- * atomic64_add - add integer to atomic64 variable
- * @i: integer value to add
- * @v: pointer to type atomic64_t
-@@ -42,6 +66,28 @@ static inline void atomic64_set(atomic64
- */
- static inline void atomic64_add(long i, atomic64_t *v)
- {
-+ asm volatile(LOCK_PREFIX "addq %1,%0\n"
-+
-+#ifdef CONFIG_PAX_REFCOUNT
-+ "jno 0f\n"
-+ LOCK_PREFIX "subq %1,%0\n"
-+ "int $4\n0:\n"
-+ _ASM_EXTABLE(0b, 0b)
-+#endif
-+
-+ : "=m" (v->counter)
-+ : "er" (i), "m" (v->counter));
-+}
-+
-+/**
-+ * atomic64_add_unchecked - add integer to atomic64 variable
-+ * @i: integer value to add
-+ * @v: pointer to type atomic64_unchecked_t
-+ *
-+ * Atomically adds @i to @v.
-+ */
-+static inline void atomic64_add_unchecked(long i, atomic64_unchecked_t *v)
-+{
- asm volatile(LOCK_PREFIX "addq %1,%0"
- : "=m" (v->counter)
- : "er" (i), "m" (v->counter));
-@@ -56,7 +102,29 @@ static inline void atomic64_add(long i,
- */
- static inline void atomic64_sub(long i, atomic64_t *v)
- {
-- asm volatile(LOCK_PREFIX "subq %1,%0"
-+ asm volatile(LOCK_PREFIX "subq %1,%0\n"
-+
-+#ifdef CONFIG_PAX_REFCOUNT
-+ "jno 0f\n"
-+ LOCK_PREFIX "addq %1,%0\n"
-+ "int $4\n0:\n"
-+ _ASM_EXTABLE(0b, 0b)
-+#endif
-+
-+ : "=m" (v->counter)
-+ : "er" (i), "m" (v->counter));
-+}
-+
-+/**
-+ * atomic64_sub_unchecked - subtract the atomic64 variable
-+ * @i: integer value to subtract
-+ * @v: pointer to type atomic64_unchecked_t
-+ *
-+ * Atomically subtracts @i from @v.
-+ */
-+static inline void atomic64_sub_unchecked(long i, atomic64_unchecked_t *v)
-+{
-+ asm volatile(LOCK_PREFIX "subq %1,%0\n"
- : "=m" (v->counter)
- : "er" (i), "m" (v->counter));
- }
-@@ -74,7 +142,16 @@ static inline int atomic64_sub_and_test(
- {
- unsigned char c;
-
-- asm volatile(LOCK_PREFIX "subq %2,%0; sete %1"
-+ asm volatile(LOCK_PREFIX "subq %2,%0\n"
-+
-+#ifdef CONFIG_PAX_REFCOUNT
-+ "jno 0f\n"
-+ LOCK_PREFIX "addq %2,%0\n"
-+ "int $4\n0:\n"
-+ _ASM_EXTABLE(0b, 0b)
-+#endif
-+
-+ "sete %1\n"
- : "=m" (v->counter), "=qm" (c)
- : "er" (i), "m" (v->counter) : "memory");
- return c;
-@@ -88,6 +165,27 @@ static inline int atomic64_sub_and_test(
- */
- static inline void atomic64_inc(atomic64_t *v)
- {
-+ asm volatile(LOCK_PREFIX "incq %0\n"
-+
-+#ifdef CONFIG_PAX_REFCOUNT
-+ "jno 0f\n"
-+ LOCK_PREFIX "decq %0\n"
-+ "int $4\n0:\n"
-+ _ASM_EXTABLE(0b, 0b)
-+#endif
-+
-+ : "=m" (v->counter)
-+ : "m" (v->counter));
-+}
-+
-+/**
-+ * atomic64_inc_unchecked - increment atomic64 variable
-+ * @v: pointer to type atomic64_unchecked_t
-+ *
-+ * Atomically increments @v by 1.
-+ */
-+static inline void atomic64_inc_unchecked(atomic64_unchecked_t *v)
-+{
- asm volatile(LOCK_PREFIX "incq %0"
- : "=m" (v->counter)
- : "m" (v->counter));
-@@ -101,7 +199,28 @@ static inline void atomic64_inc(atomic64
- */
- static inline void atomic64_dec(atomic64_t *v)
- {
-- asm volatile(LOCK_PREFIX "decq %0"
-+ asm volatile(LOCK_PREFIX "decq %0\n"
-+
-+#ifdef CONFIG_PAX_REFCOUNT
-+ "jno 0f\n"
-+ LOCK_PREFIX "incq %0\n"
-+ "int $4\n0:\n"
-+ _ASM_EXTABLE(0b, 0b)
-+#endif
-+
-+ : "=m" (v->counter)
-+ : "m" (v->counter));
-+}
-+
-+/**
-+ * atomic64_dec_unchecked - decrement atomic64 variable
-+ * @v: pointer to type atomic64_t
-+ *
-+ * Atomically decrements @v by 1.
-+ */
-+static inline void atomic64_dec_unchecked(atomic64_unchecked_t *v)
-+{
-+ asm volatile(LOCK_PREFIX "decq %0\n"
- : "=m" (v->counter)
- : "m" (v->counter));
- }
-@@ -118,7 +237,16 @@ static inline int atomic64_dec_and_test(
- {
- unsigned char c;
-
-- asm volatile(LOCK_PREFIX "decq %0; sete %1"
-+ asm volatile(LOCK_PREFIX "decq %0\n"
-+
-+#ifdef CONFIG_PAX_REFCOUNT
-+ "jno 0f\n"
-+ LOCK_PREFIX "incq %0\n"
-+ "int $4\n0:\n"
-+ _ASM_EXTABLE(0b, 0b)
-+#endif
-+
-+ "sete %1\n"
- : "=m" (v->counter), "=qm" (c)
- : "m" (v->counter) : "memory");
- return c != 0;
-@@ -136,7 +264,16 @@ static inline int atomic64_inc_and_test(
- {
- unsigned char c;
-
-- asm volatile(LOCK_PREFIX "incq %0; sete %1"
-+ asm volatile(LOCK_PREFIX "incq %0\n"
-+
-+#ifdef CONFIG_PAX_REFCOUNT
-+ "jno 0f\n"
-+ LOCK_PREFIX "decq %0\n"
-+ "int $4\n0:\n"
-+ _ASM_EXTABLE(0b, 0b)
-+#endif
-+
-+ "sete %1\n"
- : "=m" (v->counter), "=qm" (c)
- : "m" (v->counter) : "memory");
- return c != 0;
-@@ -155,7 +292,16 @@ static inline int atomic64_add_negative(
- {
- unsigned char c;
-
-- asm volatile(LOCK_PREFIX "addq %2,%0; sets %1"
-+ asm volatile(LOCK_PREFIX "addq %2,%0\n"
-+
-+#ifdef CONFIG_PAX_REFCOUNT
-+ "jno 0f\n"
-+ LOCK_PREFIX "subq %2,%0\n"
-+ "int $4\n0:\n"
-+ _ASM_EXTABLE(0b, 0b)
-+#endif
-+
-+ "sets %1\n"
- : "=m" (v->counter), "=qm" (c)
- : "er" (i), "m" (v->counter) : "memory");
- return c;
-@@ -171,7 +317,31 @@ static inline int atomic64_add_negative(
- static inline long atomic64_add_return(long i, atomic64_t *v)
- {
- long __i = i;
-- asm volatile(LOCK_PREFIX "xaddq %0, %1;"
-+ asm volatile(LOCK_PREFIX "xaddq %0, %1\n"
-+
-+#ifdef CONFIG_PAX_REFCOUNT
-+ "jno 0f\n"
-+ "movq %0, %1\n"
-+ "int $4\n0:\n"
-+ _ASM_EXTABLE(0b, 0b)
-+#endif
-+
-+ : "+r" (i), "+m" (v->counter)
-+ : : "memory");
-+ return i + __i;
-+}
-+
-+/**
-+ * atomic64_add_return_unchecked - add and return
-+ * @i: integer value to add
-+ * @v: pointer to type atomic64_unchecked_t
-+ *
-+ * Atomically adds @i to @v and returns @i + @v
-+ */
-+static inline long atomic64_add_return_unchecked(long i, atomic64_unchecked_t *v)
-+{
-+ long __i = i;
-+ asm volatile(LOCK_PREFIX "xaddq %0, %1"
- : "+r" (i), "+m" (v->counter)
- : : "memory");
- return i + __i;
-@@ -183,6 +353,10 @@ static inline long atomic64_sub_return(l
- }
-
- #define atomic64_inc_return(v) (atomic64_add_return(1, (v)))
-+static inline long atomic64_inc_return_unchecked(atomic64_unchecked_t *v)
-+{
-+ return atomic64_add_return_unchecked(1, v);
-+}
- #define atomic64_dec_return(v) (atomic64_sub_return(1, (v)))
-
- static inline long atomic64_cmpxchg(atomic64_t *v, long old, long new)
-@@ -206,17 +380,30 @@ static inline long atomic64_xchg(atomic6
- */
- static inline int atomic64_add_unless(atomic64_t *v, long a, long u)
- {
-- long c, old;
-+ long c, old, new;
- c = atomic64_read(v);
- for (;;) {
-- if (unlikely(c == (u)))
-+ if (unlikely(c == u))
- break;
-- old = atomic64_cmpxchg((v), c, c + (a));
-+
-+ asm volatile("add %2,%0\n"
-+
-+#ifdef CONFIG_PAX_REFCOUNT
-+ "jno 0f\n"
-+ "sub %2,%0\n"
-+ "int $4\n0:\n"
-+ _ASM_EXTABLE(0b, 0b)
-+#endif
-+
-+ : "=r" (new)
-+ : "0" (c), "ir" (a));
-+
-+ old = atomic64_cmpxchg(v, c, new);
- if (likely(old == c))
- break;
- c = old;
- }
-- return c != (u);
-+ return c != u;
- }
-
- #define atomic64_inc_not_zero(v) atomic64_add_unless((v), 1, 0)
-diff -urNp linux-2.6.37.4/arch/x86/include/asm/atomic.h linux-2.6.37.4/arch/x86/include/asm/atomic.h
---- linux-2.6.37.4/arch/x86/include/asm/atomic.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/include/asm/atomic.h 2011-02-27 14:48:10.000000000 -0500
-@@ -22,7 +22,18 @@
- */
- static inline int atomic_read(const atomic_t *v)
- {
-- return (*(volatile int *)&(v)->counter);
-+ return (*(volatile const int *)&(v)->counter);
-+}
-+
-+/**
-+ * atomic_read_unchecked - read atomic variable
-+ * @v: pointer of type atomic_unchecked_t
-+ *
-+ * Atomically reads the value of @v.
-+ */
-+static inline int atomic_read_unchecked(const atomic_unchecked_t *v)
-+{
-+ return (*(volatile const int *)&(v)->counter);
- }
-
- /**
-@@ -38,6 +49,18 @@ static inline void atomic_set(atomic_t *
- }
-
- /**
-+ * atomic_set_unchecked - set atomic variable
-+ * @v: pointer of type atomic_unchecked_t
-+ * @i: required value
-+ *
-+ * Atomically sets the value of @v to @i.
-+ */
-+static inline void atomic_set_unchecked(atomic_unchecked_t *v, int i)
-+{
-+ v->counter = i;
-+}
-+
-+/**
- * atomic_add - add integer to atomic variable
- * @i: integer value to add
- * @v: pointer of type atomic_t
-@@ -46,7 +69,29 @@ static inline void atomic_set(atomic_t *
- */
- static inline void atomic_add(int i, atomic_t *v)
- {
-- asm volatile(LOCK_PREFIX "addl %1,%0"
-+ asm volatile(LOCK_PREFIX "addl %1,%0\n"
-+
-+#ifdef CONFIG_PAX_REFCOUNT
-+ "jno 0f\n"
-+ LOCK_PREFIX "subl %1,%0\n"
-+ "int $4\n0:\n"
-+ _ASM_EXTABLE(0b, 0b)
-+#endif
-+
-+ : "+m" (v->counter)
-+ : "ir" (i));
-+}
-+
-+/**
-+ * atomic_add_unchecked - add integer to atomic variable
-+ * @i: integer value to add
-+ * @v: pointer of type atomic_unchecked_t
-+ *
-+ * Atomically adds @i to @v.
-+ */
-+static inline void atomic_add_unchecked(int i, atomic_unchecked_t *v)
-+{
-+ asm volatile(LOCK_PREFIX "addl %1,%0\n"
- : "+m" (v->counter)
- : "ir" (i));
- }
-@@ -60,7 +105,29 @@ static inline void atomic_add(int i, ato
- */
- static inline void atomic_sub(int i, atomic_t *v)
- {
-- asm volatile(LOCK_PREFIX "subl %1,%0"
-+ asm volatile(LOCK_PREFIX "subl %1,%0\n"
-+
-+#ifdef CONFIG_PAX_REFCOUNT
-+ "jno 0f\n"
-+ LOCK_PREFIX "addl %1,%0\n"
-+ "int $4\n0:\n"
-+ _ASM_EXTABLE(0b, 0b)
-+#endif
-+
-+ : "+m" (v->counter)
-+ : "ir" (i));
-+}
-+
-+/**
-+ * atomic_sub_unchecked - subtract integer from atomic variable
-+ * @i: integer value to subtract
-+ * @v: pointer of type atomic_t
-+ *
-+ * Atomically subtracts @i from @v.
-+ */
-+static inline void atomic_sub_unchecked(int i, atomic_unchecked_t *v)
-+{
-+ asm volatile(LOCK_PREFIX "subl %1,%0\n"
- : "+m" (v->counter)
- : "ir" (i));
- }
-@@ -78,7 +145,16 @@ static inline int atomic_sub_and_test(in
- {
- unsigned char c;
-
-- asm volatile(LOCK_PREFIX "subl %2,%0; sete %1"
-+ asm volatile(LOCK_PREFIX "subl %2,%0\n"
-+
-+#ifdef CONFIG_PAX_REFCOUNT
-+ "jno 0f\n"
-+ LOCK_PREFIX "addl %2,%0\n"
-+ "int $4\n0:\n"
-+ _ASM_EXTABLE(0b, 0b)
-+#endif
-+
-+ "sete %1\n"
- : "+m" (v->counter), "=qm" (c)
- : "ir" (i) : "memory");
- return c;
-@@ -92,7 +168,27 @@ static inline int atomic_sub_and_test(in
- */
- static inline void atomic_inc(atomic_t *v)
- {
-- asm volatile(LOCK_PREFIX "incl %0"
-+ asm volatile(LOCK_PREFIX "incl %0\n"
-+
-+#ifdef CONFIG_PAX_REFCOUNT
-+ "jno 0f\n"
-+ LOCK_PREFIX "decl %0\n"
-+ "int $4\n0:\n"
-+ _ASM_EXTABLE(0b, 0b)
-+#endif
-+
-+ : "+m" (v->counter));
-+}
-+
-+/**
-+ * atomic_inc_unchecked - increment atomic variable
-+ * @v: pointer of type atomic_unchecked_t
-+ *
-+ * Atomically increments @v by 1.
-+ */
-+static inline void atomic_inc_unchecked(atomic_unchecked_t *v)
-+{
-+ asm volatile(LOCK_PREFIX "incl %0\n"
- : "+m" (v->counter));
- }
-
-@@ -104,7 +200,27 @@ static inline void atomic_inc(atomic_t *
- */
- static inline void atomic_dec(atomic_t *v)
- {
-- asm volatile(LOCK_PREFIX "decl %0"
-+ asm volatile(LOCK_PREFIX "decl %0\n"
-+
-+#ifdef CONFIG_PAX_REFCOUNT
-+ "jno 0f\n"
-+ LOCK_PREFIX "incl %0\n"
-+ "int $4\n0:\n"
-+ _ASM_EXTABLE(0b, 0b)
-+#endif
-+
-+ : "+m" (v->counter));
-+}
-+
-+/**
-+ * atomic_dec_unchecked - decrement atomic variable
-+ * @v: pointer of type atomic_t
-+ *
-+ * Atomically decrements @v by 1.
-+ */
-+static inline void atomic_dec_unchecked(atomic_unchecked_t *v)
-+{
-+ asm volatile(LOCK_PREFIX "decl %0\n"
- : "+m" (v->counter));
- }
-
-@@ -120,7 +236,16 @@ static inline int atomic_dec_and_test(at
- {
- unsigned char c;
-
-- asm volatile(LOCK_PREFIX "decl %0; sete %1"
-+ asm volatile(LOCK_PREFIX "decl %0\n"
-+
-+#ifdef CONFIG_PAX_REFCOUNT
-+ "jno 0f\n"
-+ LOCK_PREFIX "incl %0\n"
-+ "int $4\n0:\n"
-+ _ASM_EXTABLE(0b, 0b)
-+#endif
-+
-+ "sete %1\n"
- : "+m" (v->counter), "=qm" (c)
- : : "memory");
- return c != 0;
-@@ -138,7 +263,16 @@ static inline int atomic_inc_and_test(at
- {
- unsigned char c;
-
-- asm volatile(LOCK_PREFIX "incl %0; sete %1"
-+ asm volatile(LOCK_PREFIX "incl %0\n"
-+
-+#ifdef CONFIG_PAX_REFCOUNT
-+ "jno 0f\n"
-+ LOCK_PREFIX "decl %0\n"
-+ "int $4\n0:\n"
-+ _ASM_EXTABLE(0b, 0b)
-+#endif
-+
-+ "sete %1\n"
- : "+m" (v->counter), "=qm" (c)
- : : "memory");
- return c != 0;
-@@ -157,7 +291,16 @@ static inline int atomic_add_negative(in
- {
- unsigned char c;
-
-- asm volatile(LOCK_PREFIX "addl %2,%0; sets %1"
-+ asm volatile(LOCK_PREFIX "addl %2,%0\n"
-+
-+#ifdef CONFIG_PAX_REFCOUNT
-+ "jno 0f\n"
-+ LOCK_PREFIX "subl %2,%0\n"
-+ "int $4\n0:\n"
-+ _ASM_EXTABLE(0b, 0b)
-+#endif
-+
-+ "sets %1\n"
- : "+m" (v->counter), "=qm" (c)
- : "ir" (i) : "memory");
- return c;
-@@ -180,6 +323,46 @@ static inline int atomic_add_return(int
- #endif
- /* Modern 486+ processor */
- __i = i;
-+ asm volatile(LOCK_PREFIX "xaddl %0, %1\n"
-+
-+#ifdef CONFIG_PAX_REFCOUNT
-+ "jno 0f\n"
-+ "movl %0, %1\n"
-+ "int $4\n0:\n"
-+ _ASM_EXTABLE(0b, 0b)
-+#endif
-+
-+ : "+r" (i), "+m" (v->counter)
-+ : : "memory");
-+ return i + __i;
-+
-+#ifdef CONFIG_M386
-+no_xadd: /* Legacy 386 processor */
-+ local_irq_save(flags);
-+ __i = atomic_read(v);
-+ atomic_set(v, i + __i);
-+ local_irq_restore(flags);
-+ return i + __i;
-+#endif
-+}
-+
-+/**
-+ * atomic_add_return_unchecked - add integer and return
-+ * @v: pointer of type atomic_unchecked_t
-+ * @i: integer value to add
-+ *
-+ * Atomically adds @i to @v and returns @i + @v
-+ */
-+static inline int atomic_add_return_unchecked(int i, atomic_unchecked_t *v)
-+{
-+ int __i;
-+#ifdef CONFIG_M386
-+ unsigned long flags;
-+ if (unlikely(boot_cpu_data.x86 <= 3))
-+ goto no_xadd;
-+#endif
-+ /* Modern 486+ processor */
-+ __i = i;
- asm volatile(LOCK_PREFIX "xaddl %0, %1"
- : "+r" (i), "+m" (v->counter)
- : : "memory");
-@@ -208,6 +391,10 @@ static inline int atomic_sub_return(int
- }
-
- #define atomic_inc_return(v) (atomic_add_return(1, v))
-+static inline int atomic_inc_return_unchecked(atomic_unchecked_t *v)
-+{
-+ return atomic_add_return_unchecked(1, v);
-+}
- #define atomic_dec_return(v) (atomic_sub_return(1, v))
-
- static inline int atomic_cmpxchg(atomic_t *v, int old, int new)
-@@ -231,21 +418,77 @@ static inline int atomic_xchg(atomic_t *
- */
- static inline int atomic_add_unless(atomic_t *v, int a, int u)
- {
-- int c, old;
-+ int c, old, new;
- c = atomic_read(v);
- for (;;) {
-- if (unlikely(c == (u)))
-+ if (unlikely(c == u))
- break;
-- old = atomic_cmpxchg((v), c, c + (a));
-+
-+ asm volatile("addl %2,%0\n"
-+
-+#ifdef CONFIG_PAX_REFCOUNT
-+ "jno 0f\n"
-+ "subl %2,%0\n"
-+ "int $4\n0:\n"
-+ _ASM_EXTABLE(0b, 0b)
-+#endif
-+
-+ : "=r" (new)
-+ : "0" (c), "ir" (a));
-+
-+ old = atomic_cmpxchg(v, c, new);
- if (likely(old == c))
- break;
- c = old;
- }
-- return c != (u);
-+ return c != u;
- }
-
- #define atomic_inc_not_zero(v) atomic_add_unless((v), 1, 0)
-
-+/**
-+ * atomic_inc_not_zero_hint - increment if not null
-+ * @v: pointer of type atomic_t
-+ * @hint: probable value of the atomic before the increment
-+ *
-+ * This version of atomic_inc_not_zero() gives a hint of probable
-+ * value of the atomic. This helps processor to not read the memory
-+ * before doing the atomic read/modify/write cycle, lowering
-+ * number of bus transactions on some arches.
-+ *
-+ * Returns: 0 if increment was not done, 1 otherwise.
-+ */
-+#define atomic_inc_not_zero_hint atomic_inc_not_zero_hint
-+static inline int atomic_inc_not_zero_hint(atomic_t *v, int hint)
-+{
-+ int val, c = hint, new;
-+
-+ /* sanity test, should be removed by compiler if hint is a constant */
-+ if (!hint)
-+ return atomic_inc_not_zero(v);
-+
-+ do {
-+ asm volatile("incl %0\n"
-+
-+#ifdef CONFIG_PAX_REFCOUNT
-+ "jno 0f\n"
-+ "decl %0\n"
-+ "int $4\n0:\n"
-+ _ASM_EXTABLE(0b, 0b)
-+#endif
-+
-+ : "=r" (new)
-+ : "0" (c));
-+
-+ val = atomic_cmpxchg(v, c, new);
-+ if (val == c)
-+ return 1;
-+ c = val;
-+ } while (c);
-+
-+ return 0;
-+}
-+
- /*
- * atomic_dec_if_positive - decrement by 1 if old value positive
- * @v: pointer of type atomic_t
-diff -urNp linux-2.6.37.4/arch/x86/include/asm/bitops.h linux-2.6.37.4/arch/x86/include/asm/bitops.h
---- linux-2.6.37.4/arch/x86/include/asm/bitops.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/include/asm/bitops.h 2011-01-17 02:41:00.000000000 -0500
-@@ -38,7 +38,7 @@
- * a mask operation on a byte.
- */
- #define IS_IMMEDIATE(nr) (__builtin_constant_p(nr))
--#define CONST_MASK_ADDR(nr, addr) BITOP_ADDR((void *)(addr) + ((nr)>>3))
-+#define CONST_MASK_ADDR(nr, addr) BITOP_ADDR((volatile void *)(addr) + ((nr)>>3))
- #define CONST_MASK(nr) (1 << ((nr) & 7))
-
- /**
-diff -urNp linux-2.6.37.4/arch/x86/include/asm/boot.h linux-2.6.37.4/arch/x86/include/asm/boot.h
---- linux-2.6.37.4/arch/x86/include/asm/boot.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/include/asm/boot.h 2011-01-17 02:41:00.000000000 -0500
-@@ -11,10 +11,15 @@
- #include <asm/pgtable_types.h>
-
- /* Physical address where kernel should be loaded. */
--#define LOAD_PHYSICAL_ADDR ((CONFIG_PHYSICAL_START \
-+#define ____LOAD_PHYSICAL_ADDR ((CONFIG_PHYSICAL_START \
- + (CONFIG_PHYSICAL_ALIGN - 1)) \
- & ~(CONFIG_PHYSICAL_ALIGN - 1))
-
-+#ifndef __ASSEMBLY__
-+extern unsigned char __LOAD_PHYSICAL_ADDR[];
-+#define LOAD_PHYSICAL_ADDR ((unsigned long)__LOAD_PHYSICAL_ADDR)
-+#endif
-+
- /* Minimum kernel alignment, as a power of two */
- #ifdef CONFIG_X86_64
- #define MIN_KERNEL_ALIGN_LG2 PMD_SHIFT
-diff -urNp linux-2.6.37.4/arch/x86/include/asm/cacheflush.h linux-2.6.37.4/arch/x86/include/asm/cacheflush.h
---- linux-2.6.37.4/arch/x86/include/asm/cacheflush.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/include/asm/cacheflush.h 2011-01-17 02:41:00.000000000 -0500
-@@ -66,7 +66,7 @@ static inline unsigned long get_page_mem
- unsigned long pg_flags = pg->flags & _PGMT_MASK;
-
- if (pg_flags == _PGMT_DEFAULT)
-- return -1;
-+ return ~0UL;
- else if (pg_flags == _PGMT_WC)
- return _PAGE_CACHE_WC;
- else if (pg_flags == _PGMT_UC_MINUS)
-diff -urNp linux-2.6.37.4/arch/x86/include/asm/cache.h linux-2.6.37.4/arch/x86/include/asm/cache.h
---- linux-2.6.37.4/arch/x86/include/asm/cache.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/include/asm/cache.h 2011-01-17 02:41:00.000000000 -0500
-@@ -8,6 +8,7 @@
- #define L1_CACHE_BYTES (1 << L1_CACHE_SHIFT)
-
- #define __read_mostly __attribute__((__section__(".data..read_mostly")))
-+#define __read_only __attribute__((__section__(".data..read_only")))
-
- #define INTERNODE_CACHE_SHIFT CONFIG_X86_INTERNODE_CACHE_SHIFT
- #define INTERNODE_CACHE_BYTES (1 << INTERNODE_CACHE_SHIFT)
-diff -urNp linux-2.6.37.4/arch/x86/include/asm/checksum_32.h linux-2.6.37.4/arch/x86/include/asm/checksum_32.h
---- linux-2.6.37.4/arch/x86/include/asm/checksum_32.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/include/asm/checksum_32.h 2011-01-17 02:41:00.000000000 -0500
-@@ -31,6 +31,14 @@ asmlinkage __wsum csum_partial_copy_gene
- int len, __wsum sum,
- int *src_err_ptr, int *dst_err_ptr);
-
-+asmlinkage __wsum csum_partial_copy_generic_to_user(const void *src, void *dst,
-+ int len, __wsum sum,
-+ int *src_err_ptr, int *dst_err_ptr);
-+
-+asmlinkage __wsum csum_partial_copy_generic_from_user(const void *src, void *dst,
-+ int len, __wsum sum,
-+ int *src_err_ptr, int *dst_err_ptr);
-+
- /*
- * Note: when you get a NULL pointer exception here this means someone
- * passed in an incorrect kernel address to one of these functions.
-@@ -50,7 +58,7 @@ static inline __wsum csum_partial_copy_f
- int *err_ptr)
- {
- might_sleep();
-- return csum_partial_copy_generic((__force void *)src, dst,
-+ return csum_partial_copy_generic_from_user((__force void *)src, dst,
- len, sum, err_ptr, NULL);
- }
-
-@@ -178,7 +186,7 @@ static inline __wsum csum_and_copy_to_us
- {
- might_sleep();
- if (access_ok(VERIFY_WRITE, dst, len))
-- return csum_partial_copy_generic(src, (__force void *)dst,
-+ return csum_partial_copy_generic_to_user(src, (__force void *)dst,
- len, sum, NULL, err_ptr);
-
- if (len)
-diff -urNp linux-2.6.37.4/arch/x86/include/asm/cpufeature.h linux-2.6.37.4/arch/x86/include/asm/cpufeature.h
---- linux-2.6.37.4/arch/x86/include/asm/cpufeature.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/include/asm/cpufeature.h 2011-01-17 02:41:00.000000000 -0500
-@@ -349,7 +349,7 @@ static __always_inline __pure bool __sta
- ".section .discard,\"aw\",@progbits\n"
- " .byte 0xff + (4f-3f) - (2b-1b)\n" /* size check */
- ".previous\n"
-- ".section .altinstr_replacement,\"ax\"\n"
-+ ".section .altinstr_replacement,\"a\"\n"
- "3: movb $1,%0\n"
- "4:\n"
- ".previous\n"
-diff -urNp linux-2.6.37.4/arch/x86/include/asm/desc_defs.h linux-2.6.37.4/arch/x86/include/asm/desc_defs.h
---- linux-2.6.37.4/arch/x86/include/asm/desc_defs.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/include/asm/desc_defs.h 2011-03-21 17:16:23.000000000 -0400
-@@ -31,6 +31,12 @@ struct desc_struct {
- unsigned base1: 8, type: 4, s: 1, dpl: 2, p: 1;
- unsigned limit: 4, avl: 1, l: 1, d: 1, g: 1, base2: 8;
- };
-+ struct {
-+ u16 offset_low;
-+ u16 seg;
-+ unsigned reserved: 8, type: 4, s: 1, dpl: 2, p: 1;
-+ unsigned offset_high: 16;
-+ } gate;
- };
- } __attribute__((packed));
-
-diff -urNp linux-2.6.37.4/arch/x86/include/asm/desc.h linux-2.6.37.4/arch/x86/include/asm/desc.h
---- linux-2.6.37.4/arch/x86/include/asm/desc.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/include/asm/desc.h 2011-03-21 17:16:23.000000000 -0400
-@@ -4,6 +4,7 @@
- #include <asm/desc_defs.h>
- #include <asm/ldt.h>
- #include <asm/mmu.h>
-+#include <asm/pgtable.h>
- #include <linux/smp.h>
-
- static inline void fill_ldt(struct desc_struct *desc,
-@@ -15,6 +16,7 @@ static inline void fill_ldt(struct desc_
- desc->base1 = (info->base_addr & 0x00ff0000) >> 16;
- desc->type = (info->read_exec_only ^ 1) << 1;
- desc->type |= info->contents << 2;
-+ desc->type |= info->seg_not_present ^ 1;
- desc->s = 1;
- desc->dpl = 0x3;
- desc->p = info->seg_not_present ^ 1;
-@@ -31,16 +33,12 @@ static inline void fill_ldt(struct desc_
- }
-
- extern struct desc_ptr idt_descr;
--extern gate_desc idt_table[];
--
--struct gdt_page {
-- struct desc_struct gdt[GDT_ENTRIES];
--} __attribute__((aligned(PAGE_SIZE)));
--DECLARE_PER_CPU_PAGE_ALIGNED(struct gdt_page, gdt_page);
-+extern gate_desc idt_table[256];
-
-+extern struct desc_struct cpu_gdt_table[NR_CPUS][PAGE_SIZE / sizeof(struct desc_struct)];
- static inline struct desc_struct *get_cpu_gdt_table(unsigned int cpu)
- {
-- return per_cpu(gdt_page, cpu).gdt;
-+ return cpu_gdt_table[cpu];
- }
-
- #ifdef CONFIG_X86_64
-@@ -65,9 +63,14 @@ static inline void pack_gate(gate_desc *
- unsigned long base, unsigned dpl, unsigned flags,
- unsigned short seg)
- {
-- gate->a = (seg << 16) | (base & 0xffff);
-- gate->b = (base & 0xffff0000) |
-- (((0x80 | type | (dpl << 5)) & 0xff) << 8);
-+ gate->gate.offset_low = base;
-+ gate->gate.seg = seg;
-+ gate->gate.reserved = 0;
-+ gate->gate.type = type;
-+ gate->gate.s = 0;
-+ gate->gate.dpl = dpl;
-+ gate->gate.p = 1;
-+ gate->gate.offset_high = base >> 16;
- }
-
- #endif
-@@ -115,19 +118,24 @@ static inline void paravirt_free_ldt(str
- static inline void native_write_idt_entry(gate_desc *idt, int entry,
- const gate_desc *gate)
- {
-+ pax_open_kernel();
- memcpy(&idt[entry], gate, sizeof(*gate));
-+ pax_close_kernel();
- }
-
- static inline void native_write_ldt_entry(struct desc_struct *ldt, int entry,
- const void *desc)
- {
-+ pax_open_kernel();
- memcpy(&ldt[entry], desc, 8);
-+ pax_close_kernel();
- }
-
- static inline void native_write_gdt_entry(struct desc_struct *gdt, int entry,
- const void *desc, int type)
- {
- unsigned int size;
-+
- switch (type) {
- case DESC_TSS:
- size = sizeof(tss_desc);
-@@ -139,7 +147,10 @@ static inline void native_write_gdt_entr
- size = sizeof(struct desc_struct);
- break;
- }
-+
-+ pax_open_kernel();
- memcpy(&gdt[entry], desc, size);
-+ pax_close_kernel();
- }
-
- static inline void pack_descriptor(struct desc_struct *desc, unsigned long base,
-@@ -211,7 +222,9 @@ static inline void native_set_ldt(const
-
- static inline void native_load_tr_desc(void)
- {
-+ pax_open_kernel();
- asm volatile("ltr %w0"::"q" (GDT_ENTRY_TSS*8));
-+ pax_close_kernel();
- }
-
- static inline void native_load_gdt(const struct desc_ptr *dtr)
-@@ -246,8 +259,10 @@ static inline void native_load_tls(struc
- unsigned int i;
- struct desc_struct *gdt = get_cpu_gdt_table(cpu);
-
-+ pax_open_kernel();
- for (i = 0; i < GDT_ENTRY_TLS_ENTRIES; i++)
- gdt[GDT_ENTRY_TLS_MIN + i] = t->tls_array[i];
-+ pax_close_kernel();
- }
-
- #define _LDT_empty(info) \
-@@ -309,7 +324,7 @@ static inline void set_desc_limit(struct
- desc->limit = (limit >> 16) & 0xf;
- }
-
--static inline void _set_gate(int gate, unsigned type, void *addr,
-+static inline void _set_gate(int gate, unsigned type, const void *addr,
- unsigned dpl, unsigned ist, unsigned seg)
- {
- gate_desc s;
-@@ -327,7 +342,7 @@ static inline void _set_gate(int gate, u
- * Pentium F0 0F bugfix can have resulted in the mapped
- * IDT being write-protected.
- */
--static inline void set_intr_gate(unsigned int n, void *addr)
-+static inline void set_intr_gate(unsigned int n, const void *addr)
- {
- BUG_ON((unsigned)n > 0xFF);
- _set_gate(n, GATE_INTERRUPT, addr, 0, 0, __KERNEL_CS);
-@@ -356,19 +371,19 @@ static inline void alloc_intr_gate(unsig
- /*
- * This routine sets up an interrupt gate at directory privilege level 3.
- */
--static inline void set_system_intr_gate(unsigned int n, void *addr)
-+static inline void set_system_intr_gate(unsigned int n, const void *addr)
- {
- BUG_ON((unsigned)n > 0xFF);
- _set_gate(n, GATE_INTERRUPT, addr, 0x3, 0, __KERNEL_CS);
- }
-
--static inline void set_system_trap_gate(unsigned int n, void *addr)
-+static inline void set_system_trap_gate(unsigned int n, const void *addr)
- {
- BUG_ON((unsigned)n > 0xFF);
- _set_gate(n, GATE_TRAP, addr, 0x3, 0, __KERNEL_CS);
- }
-
--static inline void set_trap_gate(unsigned int n, void *addr)
-+static inline void set_trap_gate(unsigned int n, const void *addr)
- {
- BUG_ON((unsigned)n > 0xFF);
- _set_gate(n, GATE_TRAP, addr, 0, 0, __KERNEL_CS);
-@@ -377,19 +392,31 @@ static inline void set_trap_gate(unsigne
- static inline void set_task_gate(unsigned int n, unsigned int gdt_entry)
- {
- BUG_ON((unsigned)n > 0xFF);
-- _set_gate(n, GATE_TASK, (void *)0, 0, 0, (gdt_entry<<3));
-+ _set_gate(n, GATE_TASK, (const void *)0, 0, 0, (gdt_entry<<3));
- }
-
--static inline void set_intr_gate_ist(int n, void *addr, unsigned ist)
-+static inline void set_intr_gate_ist(int n, const void *addr, unsigned ist)
- {
- BUG_ON((unsigned)n > 0xFF);
- _set_gate(n, GATE_INTERRUPT, addr, 0, ist, __KERNEL_CS);
- }
-
--static inline void set_system_intr_gate_ist(int n, void *addr, unsigned ist)
-+static inline void set_system_intr_gate_ist(int n, const void *addr, unsigned ist)
- {
- BUG_ON((unsigned)n > 0xFF);
- _set_gate(n, GATE_INTERRUPT, addr, 0x3, ist, __KERNEL_CS);
- }
-
-+#ifdef CONFIG_X86_32
-+static inline void set_user_cs(unsigned long base, unsigned long limit, int cpu)
-+{
-+ struct desc_struct d;
-+
-+ if (likely(limit))
-+ limit = (limit - 1UL) >> PAGE_SHIFT;
-+ pack_descriptor(&d, base, limit, 0xFB, 0xC);
-+ write_gdt_entry(get_cpu_gdt_table(cpu), GDT_ENTRY_DEFAULT_USER_CS, &d, DESCTYPE_S);
-+}
-+#endif
-+
- #endif /* _ASM_X86_DESC_H */
-diff -urNp linux-2.6.37.4/arch/x86/include/asm/device.h linux-2.6.37.4/arch/x86/include/asm/device.h
---- linux-2.6.37.4/arch/x86/include/asm/device.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/include/asm/device.h 2011-01-17 02:41:00.000000000 -0500
-@@ -6,7 +6,7 @@ struct dev_archdata {
- void *acpi_handle;
- #endif
- #ifdef CONFIG_X86_64
--struct dma_map_ops *dma_ops;
-+ const struct dma_map_ops *dma_ops;
- #endif
- #if defined(CONFIG_DMAR) || defined(CONFIG_AMD_IOMMU)
- void *iommu; /* hook for IOMMU specific extension */
-diff -urNp linux-2.6.37.4/arch/x86/include/asm/dma-mapping.h linux-2.6.37.4/arch/x86/include/asm/dma-mapping.h
---- linux-2.6.37.4/arch/x86/include/asm/dma-mapping.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/include/asm/dma-mapping.h 2011-01-17 02:41:00.000000000 -0500
-@@ -26,9 +26,9 @@ extern int iommu_merge;
- extern struct device x86_dma_fallback_dev;
- extern int panic_on_overflow;
-
--extern struct dma_map_ops *dma_ops;
-+extern const struct dma_map_ops *dma_ops;
-
--static inline struct dma_map_ops *get_dma_ops(struct device *dev)
-+static inline const struct dma_map_ops *get_dma_ops(struct device *dev)
- {
- #ifdef CONFIG_X86_32
- return dma_ops;
-@@ -45,7 +45,7 @@ static inline struct dma_map_ops *get_dm
- /* Make sure we keep the same behaviour */
- static inline int dma_mapping_error(struct device *dev, dma_addr_t dma_addr)
- {
-- struct dma_map_ops *ops = get_dma_ops(dev);
-+ const struct dma_map_ops *ops = get_dma_ops(dev);
- if (ops->mapping_error)
- return ops->mapping_error(dev, dma_addr);
-
-@@ -115,7 +115,7 @@ static inline void *
- dma_alloc_coherent(struct device *dev, size_t size, dma_addr_t *dma_handle,
- gfp_t gfp)
- {
-- struct dma_map_ops *ops = get_dma_ops(dev);
-+ const struct dma_map_ops *ops = get_dma_ops(dev);
- void *memory;
-
- gfp &= ~(__GFP_DMA | __GFP_HIGHMEM | __GFP_DMA32);
-@@ -142,7 +142,7 @@ dma_alloc_coherent(struct device *dev, s
- static inline void dma_free_coherent(struct device *dev, size_t size,
- void *vaddr, dma_addr_t bus)
- {
-- struct dma_map_ops *ops = get_dma_ops(dev);
-+ const struct dma_map_ops *ops = get_dma_ops(dev);
-
- WARN_ON(irqs_disabled()); /* for portability */
-
-diff -urNp linux-2.6.37.4/arch/x86/include/asm/e820.h linux-2.6.37.4/arch/x86/include/asm/e820.h
---- linux-2.6.37.4/arch/x86/include/asm/e820.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/include/asm/e820.h 2011-01-17 02:41:00.000000000 -0500
-@@ -69,7 +69,7 @@ struct e820map {
- #define ISA_START_ADDRESS 0xa0000
- #define ISA_END_ADDRESS 0x100000
-
--#define BIOS_BEGIN 0x000a0000
-+#define BIOS_BEGIN 0x000c0000
- #define BIOS_END 0x00100000
-
- #define BIOS_ROM_BASE 0xffe00000
-diff -urNp linux-2.6.37.4/arch/x86/include/asm/elf.h linux-2.6.37.4/arch/x86/include/asm/elf.h
---- linux-2.6.37.4/arch/x86/include/asm/elf.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/include/asm/elf.h 2011-01-17 02:41:00.000000000 -0500
-@@ -237,7 +237,25 @@ extern int force_personality32;
- the loader. We need to make sure that it is out of the way of the program
- that it will "exec", and that there is sufficient room for the brk. */
-
-+#ifdef CONFIG_PAX_SEGMEXEC
-+#define ELF_ET_DYN_BASE ((current->mm->pax_flags & MF_PAX_SEGMEXEC) ? SEGMEXEC_TASK_SIZE/3*2 : TASK_SIZE/3*2)
-+#else
- #define ELF_ET_DYN_BASE (TASK_SIZE / 3 * 2)
-+#endif
-+
-+#ifdef CONFIG_PAX_ASLR
-+#ifdef CONFIG_X86_32
-+#define PAX_ELF_ET_DYN_BASE 0x10000000UL
-+
-+#define PAX_DELTA_MMAP_LEN (current->mm->pax_flags & MF_PAX_SEGMEXEC ? 15 : 16)
-+#define PAX_DELTA_STACK_LEN (current->mm->pax_flags & MF_PAX_SEGMEXEC ? 15 : 16)
-+#else
-+#define PAX_ELF_ET_DYN_BASE 0x400000UL
-+
-+#define PAX_DELTA_MMAP_LEN ((test_thread_flag(TIF_IA32)) ? 16 : TASK_SIZE_MAX_SHIFT - PAGE_SHIFT - 3)
-+#define PAX_DELTA_STACK_LEN ((test_thread_flag(TIF_IA32)) ? 16 : TASK_SIZE_MAX_SHIFT - PAGE_SHIFT - 3)
-+#endif
-+#endif
-
- /* This yields a mask that user programs can use to figure out what
- instruction set this CPU supports. This could be done in user space,
-@@ -291,8 +309,7 @@ do { \
- #define ARCH_DLINFO \
- do { \
- if (vdso_enabled) \
-- NEW_AUX_ENT(AT_SYSINFO_EHDR, \
-- (unsigned long)current->mm->context.vdso); \
-+ NEW_AUX_ENT(AT_SYSINFO_EHDR, current->mm->context.vdso);\
- } while (0)
-
- #define AT_SYSINFO 32
-@@ -303,7 +320,7 @@ do { \
-
- #endif /* !CONFIG_X86_32 */
-
--#define VDSO_CURRENT_BASE ((unsigned long)current->mm->context.vdso)
-+#define VDSO_CURRENT_BASE (current->mm->context.vdso)
-
- #define VDSO_ENTRY \
- ((unsigned long)VDSO32_SYMBOL(VDSO_CURRENT_BASE, vsyscall))
-@@ -317,7 +334,4 @@ extern int arch_setup_additional_pages(s
- extern int syscall32_setup_pages(struct linux_binprm *, int exstack);
- #define compat_arch_setup_additional_pages syscall32_setup_pages
-
--extern unsigned long arch_randomize_brk(struct mm_struct *mm);
--#define arch_randomize_brk arch_randomize_brk
--
- #endif /* _ASM_X86_ELF_H */
-diff -urNp linux-2.6.37.4/arch/x86/include/asm/futex.h linux-2.6.37.4/arch/x86/include/asm/futex.h
---- linux-2.6.37.4/arch/x86/include/asm/futex.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/include/asm/futex.h 2011-03-12 22:50:25.000000000 -0500
-@@ -12,16 +12,18 @@
- #include <asm/system.h>
-
- #define __futex_atomic_op1(insn, ret, oldval, uaddr, oparg) \
-+ typecheck(u32 *, uaddr); \
- asm volatile("1:\t" insn "\n" \
- "2:\t.section .fixup,\"ax\"\n" \
- "3:\tmov\t%3, %1\n" \
- "\tjmp\t2b\n" \
- "\t.previous\n" \
- _ASM_EXTABLE(1b, 3b) \
-- : "=r" (oldval), "=r" (ret), "+m" (*uaddr) \
-+ : "=r" (oldval), "=r" (ret), "+m" (*(u32 *)____m(uaddr))\
- : "i" (-EFAULT), "0" (oparg), "1" (0))
-
- #define __futex_atomic_op2(insn, ret, oldval, uaddr, oparg) \
-+ typecheck(u32 *, uaddr); \
- asm volatile("1:\tmovl %2, %0\n" \
- "\tmovl\t%0, %3\n" \
- "\t" insn "\n" \
-@@ -34,10 +36,10 @@
- _ASM_EXTABLE(1b, 4b) \
- _ASM_EXTABLE(2b, 4b) \
- : "=&a" (oldval), "=&r" (ret), \
-- "+m" (*uaddr), "=&r" (tem) \
-+ "+m" (*(u32 *)____m(uaddr)), "=&r" (tem) \
- : "r" (oparg), "i" (-EFAULT), "1" (0))
-
--static inline int futex_atomic_op_inuser(int encoded_op, int __user *uaddr)
-+static inline int futex_atomic_op_inuser(int encoded_op, u32 __user *uaddr)
- {
- int op = (encoded_op >> 28) & 7;
- int cmp = (encoded_op >> 24) & 15;
-@@ -61,10 +63,10 @@ static inline int futex_atomic_op_inuser
-
- switch (op) {
- case FUTEX_OP_SET:
-- __futex_atomic_op1("xchgl %0, %2", ret, oldval, uaddr, oparg);
-+ __futex_atomic_op1(__copyuser_seg"xchgl %0, %2", ret, oldval, uaddr, oparg);
- break;
- case FUTEX_OP_ADD:
-- __futex_atomic_op1(LOCK_PREFIX "xaddl %0, %2", ret, oldval,
-+ __futex_atomic_op1(LOCK_PREFIX __copyuser_seg"xaddl %0, %2", ret, oldval,
- uaddr, oparg);
- break;
- case FUTEX_OP_OR:
-@@ -109,7 +111,7 @@ static inline int futex_atomic_op_inuser
- return ret;
- }
-
--static inline int futex_atomic_cmpxchg_inatomic(int __user *uaddr, int oldval,
-+static inline int futex_atomic_cmpxchg_inatomic(u32 __user *uaddr, int oldval,
- int newval)
- {
-
-@@ -119,16 +121,16 @@ static inline int futex_atomic_cmpxchg_i
- return -ENOSYS;
- #endif
-
-- if (!access_ok(VERIFY_WRITE, uaddr, sizeof(int)))
-+ if (!access_ok(VERIFY_WRITE, uaddr, sizeof(u32)))
- return -EFAULT;
-
-- asm volatile("1:\t" LOCK_PREFIX "cmpxchgl %3, %1\n"
-+ asm volatile("1:\t" LOCK_PREFIX __copyuser_seg"cmpxchgl %3, %1\n"
- "2:\t.section .fixup, \"ax\"\n"
- "3:\tmov %2, %0\n"
- "\tjmp 2b\n"
- "\t.previous\n"
- _ASM_EXTABLE(1b, 3b)
-- : "=a" (oldval), "+m" (*uaddr)
-+ : "=a" (oldval), "+m" (*(u32 *)____m(uaddr))
- : "i" (-EFAULT), "r" (newval), "0" (oldval)
- : "memory"
- );
-diff -urNp linux-2.6.37.4/arch/x86/include/asm/i387.h linux-2.6.37.4/arch/x86/include/asm/i387.h
---- linux-2.6.37.4/arch/x86/include/asm/i387.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/include/asm/i387.h 2011-01-17 02:41:00.000000000 -0500
-@@ -92,6 +92,11 @@ static inline int fxrstor_checking(struc
- {
- int err;
-
-+#if defined(CONFIG_X86_64) && defined(CONFIG_PAX_MEMORY_UDEREF)
-+ if ((unsigned long)fx < PAX_USER_SHADOW_BASE)
-+ fx = (struct i387_fxsave_struct *)((void *)fx + PAX_USER_SHADOW_BASE);
-+#endif
-+
- /* See comment in fxsave() below. */
- asm volatile("1: rex64/fxrstor (%[fx])\n\t"
- "2:\n"
-@@ -109,6 +114,11 @@ static inline int fxsave_user(struct i38
- {
- int err;
-
-+#if defined(CONFIG_X86_64) && defined(CONFIG_PAX_MEMORY_UDEREF)
-+ if ((unsigned long)fx < PAX_USER_SHADOW_BASE)
-+ fx = (struct i387_fxsave_struct __user *)((void __user *)fx + PAX_USER_SHADOW_BASE);
-+#endif
-+
- /*
- * Clear the bytes not touched by the fxsave and reserved
- * for the SW usage.
-@@ -189,13 +199,8 @@ static inline void fpu_fxsave(struct fpu
- #endif /* CONFIG_X86_64 */
-
- /* We need a safe address that is cheap to find and that is already
-- in L1 during context switch. The best choices are unfortunately
-- different for UP and SMP */
--#ifdef CONFIG_SMP
--#define safe_address (__per_cpu_offset[0])
--#else
--#define safe_address (kstat_cpu(0).cpustat.user)
--#endif
-+ in L1 during context switch. */
-+#define safe_address (init_tss[smp_processor_id()].x86_tss.sp0)
-
- /*
- * These must be called with preempt disabled
-diff -urNp linux-2.6.37.4/arch/x86/include/asm/io.h linux-2.6.37.4/arch/x86/include/asm/io.h
---- linux-2.6.37.4/arch/x86/include/asm/io.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/include/asm/io.h 2011-01-27 22:37:21.000000000 -0500
-@@ -216,6 +216,17 @@ extern void set_iounmap_nonlazy(void);
-
- #include <linux/vmalloc.h>
-
-+#define ARCH_HAS_VALID_PHYS_ADDR_RANGE
-+static inline int valid_phys_addr_range(unsigned long addr, size_t count)
-+{
-+ return ((addr + count + PAGE_SIZE - 1) >> PAGE_SHIFT) < (1ULL << (boot_cpu_data.x86_phys_bits - PAGE_SHIFT)) ? 1 : 0;
-+}
-+
-+static inline int valid_mmap_phys_addr_range(unsigned long pfn, size_t count)
-+{
-+ return (pfn + (count >> PAGE_SHIFT)) < (1ULL << (boot_cpu_data.x86_phys_bits - PAGE_SHIFT)) ? 1 : 0;
-+}
-+
- /*
- * Convert a virtual cached pointer to an uncached pointer
- */
-diff -urNp linux-2.6.37.4/arch/x86/include/asm/iommu.h linux-2.6.37.4/arch/x86/include/asm/iommu.h
---- linux-2.6.37.4/arch/x86/include/asm/iommu.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/include/asm/iommu.h 2011-01-17 02:41:00.000000000 -0500
-@@ -1,7 +1,7 @@
- #ifndef _ASM_X86_IOMMU_H
- #define _ASM_X86_IOMMU_H
-
--extern struct dma_map_ops nommu_dma_ops;
-+extern const struct dma_map_ops nommu_dma_ops;
- extern int force_iommu, no_iommu;
- extern int iommu_detected;
- extern int iommu_pass_through;
-diff -urNp linux-2.6.37.4/arch/x86/include/asm/irqflags.h linux-2.6.37.4/arch/x86/include/asm/irqflags.h
---- linux-2.6.37.4/arch/x86/include/asm/irqflags.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/include/asm/irqflags.h 2011-01-17 02:41:00.000000000 -0500
-@@ -140,6 +140,11 @@ static inline unsigned long arch_local_i
- sti; \
- sysexit
-
-+#define GET_CR0_INTO_RDI mov %cr0, %rdi
-+#define SET_RDI_INTO_CR0 mov %rdi, %cr0
-+#define GET_CR3_INTO_RDI mov %cr3, %rdi
-+#define SET_RDI_INTO_CR3 mov %rdi, %cr3
-+
- #else
- #define INTERRUPT_RETURN iret
- #define ENABLE_INTERRUPTS_SYSEXIT sti; sysexit
-diff -urNp linux-2.6.37.4/arch/x86/include/asm/kvm_host.h linux-2.6.37.4/arch/x86/include/asm/kvm_host.h
---- linux-2.6.37.4/arch/x86/include/asm/kvm_host.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/include/asm/kvm_host.h 2011-01-17 02:41:00.000000000 -0500
-@@ -585,7 +585,7 @@ struct kvm_x86_ops {
- const struct trace_print_flags *exit_reasons_str;
- };
-
--extern struct kvm_x86_ops *kvm_x86_ops;
-+extern const struct kvm_x86_ops *kvm_x86_ops;
-
- int kvm_mmu_module_init(void);
- void kvm_mmu_module_exit(void);
-diff -urNp linux-2.6.37.4/arch/x86/include/asm/local.h linux-2.6.37.4/arch/x86/include/asm/local.h
---- linux-2.6.37.4/arch/x86/include/asm/local.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/include/asm/local.h 2011-01-17 02:41:00.000000000 -0500
-@@ -18,26 +18,58 @@ typedef struct {
-
- static inline void local_inc(local_t *l)
- {
-- asm volatile(_ASM_INC "%0"
-+ asm volatile(_ASM_INC "%0\n"
-+
-+#ifdef CONFIG_PAX_REFCOUNT
-+ "jno 0f\n"
-+ _ASM_DEC "%0\n"
-+ "int $4\n0:\n"
-+ _ASM_EXTABLE(0b, 0b)
-+#endif
-+
- : "+m" (l->a.counter));
- }
-
- static inline void local_dec(local_t *l)
- {
-- asm volatile(_ASM_DEC "%0"
-+ asm volatile(_ASM_DEC "%0\n"
-+
-+#ifdef CONFIG_PAX_REFCOUNT
-+ "jno 0f\n"
-+ _ASM_INC "%0\n"
-+ "int $4\n0:\n"
-+ _ASM_EXTABLE(0b, 0b)
-+#endif
-+
- : "+m" (l->a.counter));
- }
-
- static inline void local_add(long i, local_t *l)
- {
-- asm volatile(_ASM_ADD "%1,%0"
-+ asm volatile(_ASM_ADD "%1,%0\n"
-+
-+#ifdef CONFIG_PAX_REFCOUNT
-+ "jno 0f\n"
-+ _ASM_SUB "%1,%0\n"
-+ "int $4\n0:\n"
-+ _ASM_EXTABLE(0b, 0b)
-+#endif
-+
- : "+m" (l->a.counter)
- : "ir" (i));
- }
-
- static inline void local_sub(long i, local_t *l)
- {
-- asm volatile(_ASM_SUB "%1,%0"
-+ asm volatile(_ASM_SUB "%1,%0\n"
-+
-+#ifdef CONFIG_PAX_REFCOUNT
-+ "jno 0f\n"
-+ _ASM_ADD "%1,%0\n"
-+ "int $4\n0:\n"
-+ _ASM_EXTABLE(0b, 0b)
-+#endif
-+
- : "+m" (l->a.counter)
- : "ir" (i));
- }
-@@ -55,7 +87,16 @@ static inline int local_sub_and_test(lon
- {
- unsigned char c;
-
-- asm volatile(_ASM_SUB "%2,%0; sete %1"
-+ asm volatile(_ASM_SUB "%2,%0\n"
-+
-+#ifdef CONFIG_PAX_REFCOUNT
-+ "jno 0f\n"
-+ _ASM_ADD "%2,%0\n"
-+ "int $4\n0:\n"
-+ _ASM_EXTABLE(0b, 0b)
-+#endif
-+
-+ "sete %1\n"
- : "+m" (l->a.counter), "=qm" (c)
- : "ir" (i) : "memory");
- return c;
-@@ -73,7 +114,16 @@ static inline int local_dec_and_test(loc
- {
- unsigned char c;
-
-- asm volatile(_ASM_DEC "%0; sete %1"
-+ asm volatile(_ASM_DEC "%0\n"
-+
-+#ifdef CONFIG_PAX_REFCOUNT
-+ "jno 0f\n"
-+ _ASM_INC "%0\n"
-+ "int $4\n0:\n"
-+ _ASM_EXTABLE(0b, 0b)
-+#endif
-+
-+ "sete %1\n"
- : "+m" (l->a.counter), "=qm" (c)
- : : "memory");
- return c != 0;
-@@ -91,7 +141,16 @@ static inline int local_inc_and_test(loc
- {
- unsigned char c;
-
-- asm volatile(_ASM_INC "%0; sete %1"
-+ asm volatile(_ASM_INC "%0\n"
-+
-+#ifdef CONFIG_PAX_REFCOUNT
-+ "jno 0f\n"
-+ _ASM_DEC "%0\n"
-+ "int $4\n0:\n"
-+ _ASM_EXTABLE(0b, 0b)
-+#endif
-+
-+ "sete %1\n"
- : "+m" (l->a.counter), "=qm" (c)
- : : "memory");
- return c != 0;
-@@ -110,7 +169,16 @@ static inline int local_add_negative(lon
- {
- unsigned char c;
-
-- asm volatile(_ASM_ADD "%2,%0; sets %1"
-+ asm volatile(_ASM_ADD "%2,%0\n"
-+
-+#ifdef CONFIG_PAX_REFCOUNT
-+ "jno 0f\n"
-+ _ASM_SUB "%2,%0\n"
-+ "int $4\n0:\n"
-+ _ASM_EXTABLE(0b, 0b)
-+#endif
-+
-+ "sets %1\n"
- : "+m" (l->a.counter), "=qm" (c)
- : "ir" (i) : "memory");
- return c;
-@@ -133,7 +201,15 @@ static inline long local_add_return(long
- #endif
- /* Modern 486+ processor */
- __i = i;
-- asm volatile(_ASM_XADD "%0, %1;"
-+ asm volatile(_ASM_XADD "%0, %1\n"
-+
-+#ifdef CONFIG_PAX_REFCOUNT
-+ "jno 0f\n"
-+ _ASM_MOV "%0,%1\n"
-+ "int $4\n0:\n"
-+ _ASM_EXTABLE(0b, 0b)
-+#endif
-+
- : "+r" (i), "+m" (l->a.counter)
- : : "memory");
- return i + __i;
-diff -urNp linux-2.6.37.4/arch/x86/include/asm/mc146818rtc.h linux-2.6.37.4/arch/x86/include/asm/mc146818rtc.h
---- linux-2.6.37.4/arch/x86/include/asm/mc146818rtc.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/include/asm/mc146818rtc.h 2011-01-17 02:41:00.000000000 -0500
-@@ -81,8 +81,8 @@ static inline unsigned char current_lock
- #else
- #define lock_cmos_prefix(reg) do {} while (0)
- #define lock_cmos_suffix(reg) do {} while (0)
--#define lock_cmos(reg)
--#define unlock_cmos()
-+#define lock_cmos(reg) do {} while (0)
-+#define unlock_cmos() do {} while (0)
- #define do_i_have_lock_cmos() 0
- #define current_lock_cmos_reg() 0
- #endif
-diff -urNp linux-2.6.37.4/arch/x86/include/asm/microcode.h linux-2.6.37.4/arch/x86/include/asm/microcode.h
---- linux-2.6.37.4/arch/x86/include/asm/microcode.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/include/asm/microcode.h 2011-01-17 02:41:00.000000000 -0500
-@@ -12,13 +12,13 @@ struct device;
- enum ucode_state { UCODE_ERROR, UCODE_OK, UCODE_NFOUND };
-
- struct microcode_ops {
-- enum ucode_state (*request_microcode_user) (int cpu,
-+ enum ucode_state (* const request_microcode_user) (int cpu,
- const void __user *buf, size_t size);
-
-- enum ucode_state (*request_microcode_fw) (int cpu,
-+ enum ucode_state (* const request_microcode_fw) (int cpu,
- struct device *device);
-
-- void (*microcode_fini_cpu) (int cpu);
-+ void (* const microcode_fini_cpu) (int cpu);
-
- /*
- * The generic 'microcode_core' part guarantees that
-@@ -38,18 +38,18 @@ struct ucode_cpu_info {
- extern struct ucode_cpu_info ucode_cpu_info[];
-
- #ifdef CONFIG_MICROCODE_INTEL
--extern struct microcode_ops * __init init_intel_microcode(void);
-+extern const struct microcode_ops * __init init_intel_microcode(void);
- #else
--static inline struct microcode_ops * __init init_intel_microcode(void)
-+static inline const struct microcode_ops * __init init_intel_microcode(void)
- {
- return NULL;
- }
- #endif /* CONFIG_MICROCODE_INTEL */
-
- #ifdef CONFIG_MICROCODE_AMD
--extern struct microcode_ops * __init init_amd_microcode(void);
-+extern const struct microcode_ops * __init init_amd_microcode(void);
- #else
--static inline struct microcode_ops * __init init_amd_microcode(void)
-+static inline const struct microcode_ops * __init init_amd_microcode(void)
- {
- return NULL;
- }
-diff -urNp linux-2.6.37.4/arch/x86/include/asm/mman.h linux-2.6.37.4/arch/x86/include/asm/mman.h
---- linux-2.6.37.4/arch/x86/include/asm/mman.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/include/asm/mman.h 2011-01-17 02:41:00.000000000 -0500
-@@ -5,4 +5,14 @@
-
- #include <asm-generic/mman.h>
-
-+#ifdef __KERNEL__
-+#ifndef __ASSEMBLY__
-+#ifdef CONFIG_X86_32
-+#define arch_mmap_check i386_mmap_check
-+int i386_mmap_check(unsigned long addr, unsigned long len,
-+ unsigned long flags);
-+#endif
-+#endif
-+#endif
-+
- #endif /* _ASM_X86_MMAN_H */
-diff -urNp linux-2.6.37.4/arch/x86/include/asm/mmu_context.h linux-2.6.37.4/arch/x86/include/asm/mmu_context.h
---- linux-2.6.37.4/arch/x86/include/asm/mmu_context.h 2011-02-22 16:05:30.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/include/asm/mmu_context.h 2011-02-22 16:05:42.000000000 -0500
-@@ -24,6 +24,21 @@ void destroy_context(struct mm_struct *m
-
- static inline void enter_lazy_tlb(struct mm_struct *mm, struct task_struct *tsk)
- {
-+
-+#if defined(CONFIG_X86_64) && defined(CONFIG_PAX_MEMORY_UDEREF)
-+ unsigned int i;
-+ pgd_t *pgd;
-+
-+ pax_open_kernel();
-+ pgd = get_cpu_pgd(smp_processor_id());
-+ for (i = USER_PGD_PTRS; i < 2 * USER_PGD_PTRS; ++i)
-+ if (paravirt_enabled())
-+ set_pgd(pgd+i, native_make_pgd(0));
-+ else
-+ pgd[i] = native_make_pgd(0);
-+ pax_close_kernel();
-+#endif
-+
- #ifdef CONFIG_SMP
- if (percpu_read(cpu_tlbstate.state) == TLBSTATE_OK)
- percpu_write(cpu_tlbstate.state, TLBSTATE_LAZY);
-@@ -34,17 +49,30 @@ static inline void switch_mm(struct mm_s
- struct task_struct *tsk)
- {
- unsigned cpu = smp_processor_id();
-+#if defined(CONFIG_X86_32) && defined(CONFIG_SMP)
-+ int tlbstate = TLBSTATE_OK;
-+#endif
-
- if (likely(prev != next)) {
- #ifdef CONFIG_SMP
-+#ifdef CONFIG_X86_32
-+ tlbstate = percpu_read(cpu_tlbstate.state);
-+#endif
- percpu_write(cpu_tlbstate.state, TLBSTATE_OK);
- percpu_write(cpu_tlbstate.active_mm, next);
- #endif
- cpumask_set_cpu(cpu, mm_cpumask(next));
-
- /* Re-load page tables */
-+#ifdef CONFIG_PAX_PER_CPU_PGD
-+ pax_open_kernel();
-+ __clone_user_pgds(get_cpu_pgd(cpu), next->pgd, USER_PGD_PTRS);
-+ __shadow_user_pgds(get_cpu_pgd(cpu) + USER_PGD_PTRS, next->pgd, USER_PGD_PTRS);
-+ pax_close_kernel();
-+ load_cr3(get_cpu_pgd(cpu));
-+#else
- load_cr3(next->pgd);
--
-+#endif
- /* stop flush ipis for the previous mm */
- cpumask_clear_cpu(cpu, mm_cpumask(prev));
-
-@@ -53,9 +81,38 @@ static inline void switch_mm(struct mm_s
- */
- if (unlikely(prev->context.ldt != next->context.ldt))
- load_LDT_nolock(&next->context);
-- }
-+
-+#if defined(CONFIG_X86_32) && defined(CONFIG_PAX_PAGEEXEC) && defined(CONFIG_SMP)
-+ if (!(__supported_pte_mask & _PAGE_NX)) {
-+ smp_mb__before_clear_bit();
-+ cpu_clear(cpu, prev->context.cpu_user_cs_mask);
-+ smp_mb__after_clear_bit();
-+ cpu_set(cpu, next->context.cpu_user_cs_mask);
-+ }
-+#endif
-+
-+#if defined(CONFIG_X86_32) && (defined(CONFIG_PAX_PAGEEXEC) || defined(CONFIG_PAX_SEGMEXEC))
-+ if (unlikely(prev->context.user_cs_base != next->context.user_cs_base ||
-+ prev->context.user_cs_limit != next->context.user_cs_limit))
-+ set_user_cs(next->context.user_cs_base, next->context.user_cs_limit, cpu);
- #ifdef CONFIG_SMP
-+ else if (unlikely(tlbstate != TLBSTATE_OK))
-+ set_user_cs(next->context.user_cs_base, next->context.user_cs_limit, cpu);
-+#endif
-+#endif
-+
-+ }
- else {
-+
-+#ifdef CONFIG_PAX_PER_CPU_PGD
-+ pax_open_kernel();
-+ __clone_user_pgds(get_cpu_pgd(cpu), next->pgd, USER_PGD_PTRS);
-+ __shadow_user_pgds(get_cpu_pgd(cpu) + USER_PGD_PTRS, next->pgd, USER_PGD_PTRS);
-+ pax_close_kernel();
-+ load_cr3(get_cpu_pgd(cpu));
-+#endif
-+
-+#ifdef CONFIG_SMP
- percpu_write(cpu_tlbstate.state, TLBSTATE_OK);
- BUG_ON(percpu_read(cpu_tlbstate.active_mm) != next);
-
-@@ -64,11 +121,28 @@ static inline void switch_mm(struct mm_s
- * tlb flush IPI delivery. We must reload CR3
- * to make sure to use no freed page tables.
- */
-+
-+#ifndef CONFIG_PAX_PER_CPU_PGD
- load_cr3(next->pgd);
-+#endif
-+
- load_LDT_nolock(&next->context);
-+
-+#if defined(CONFIG_X86_32) && defined(CONFIG_PAX_PAGEEXEC)
-+ if (!(__supported_pte_mask & _PAGE_NX))
-+ cpu_set(cpu, next->context.cpu_user_cs_mask);
-+#endif
-+
-+#if defined(CONFIG_X86_32) && (defined(CONFIG_PAX_PAGEEXEC) || defined(CONFIG_PAX_SEGMEXEC))
-+#ifdef CONFIG_PAX_PAGEEXEC
-+ if (!((next->pax_flags & MF_PAX_PAGEEXEC) && (__supported_pte_mask & _PAGE_NX)))
-+#endif
-+ set_user_cs(next->context.user_cs_base, next->context.user_cs_limit, cpu);
-+#endif
-+
- }
-- }
- #endif
-+ }
- }
-
- #define activate_mm(prev, next) \
-diff -urNp linux-2.6.37.4/arch/x86/include/asm/mmu.h linux-2.6.37.4/arch/x86/include/asm/mmu.h
---- linux-2.6.37.4/arch/x86/include/asm/mmu.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/include/asm/mmu.h 2011-01-17 02:41:00.000000000 -0500
-@@ -9,10 +9,23 @@
- * we put the segment information here.
- */
- typedef struct {
-- void *ldt;
-+ struct desc_struct *ldt;
- int size;
- struct mutex lock;
-- void *vdso;
-+ unsigned long vdso;
-+
-+#ifdef CONFIG_X86_32
-+#if defined(CONFIG_PAX_PAGEEXEC) || defined(CONFIG_PAX_SEGMEXEC)
-+ unsigned long user_cs_base;
-+ unsigned long user_cs_limit;
-+
-+#if defined(CONFIG_PAX_PAGEEXEC) && defined(CONFIG_SMP)
-+ cpumask_t cpu_user_cs_mask;
-+#endif
-+
-+#endif
-+#endif
-+
- } mm_context_t;
-
- #ifdef CONFIG_SMP
-diff -urNp linux-2.6.37.4/arch/x86/include/asm/module.h linux-2.6.37.4/arch/x86/include/asm/module.h
---- linux-2.6.37.4/arch/x86/include/asm/module.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/include/asm/module.h 2011-01-17 02:41:00.000000000 -0500
-@@ -59,8 +59,26 @@
- #error unknown processor family
- #endif
-
-+#ifdef CONFIG_PAX_MEMORY_UDEREF
-+#define MODULE_PAX_UDEREF "UDEREF "
-+#else
-+#define MODULE_PAX_UDEREF ""
-+#endif
-+
- #ifdef CONFIG_X86_32
--# define MODULE_ARCH_VERMAGIC MODULE_PROC_FAMILY
-+# ifdef CONFIG_PAX_KERNEXEC
-+# define MODULE_PAX_KERNEXEC "KERNEXEC "
-+# else
-+# define MODULE_PAX_KERNEXEC ""
-+# endif
-+# ifdef CONFIG_GRKERNSEC
-+# define MODULE_GRSEC "GRSECURITY "
-+# else
-+# define MODULE_GRSEC ""
-+# endif
-+# define MODULE_ARCH_VERMAGIC MODULE_PROC_FAMILY MODULE_GRSEC MODULE_PAX_KERNEXEC MODULE_PAX_UDEREF
-+#else
-+# define MODULE_ARCH_VERMAGIC MODULE_PAX_UDEREF
- #endif
-
- #endif /* _ASM_X86_MODULE_H */
-diff -urNp linux-2.6.37.4/arch/x86/include/asm/page_64_types.h linux-2.6.37.4/arch/x86/include/asm/page_64_types.h
---- linux-2.6.37.4/arch/x86/include/asm/page_64_types.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/include/asm/page_64_types.h 2011-01-17 02:41:00.000000000 -0500
-@@ -56,7 +56,7 @@ void copy_page(void *to, void *from);
-
- /* duplicated to the one in bootmem.h */
- extern unsigned long max_pfn;
--extern unsigned long phys_base;
-+extern const unsigned long phys_base;
-
- extern unsigned long __phys_addr(unsigned long);
- #define __phys_reloc_hide(x) (x)
-diff -urNp linux-2.6.37.4/arch/x86/include/asm/paravirt.h linux-2.6.37.4/arch/x86/include/asm/paravirt.h
---- linux-2.6.37.4/arch/x86/include/asm/paravirt.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/include/asm/paravirt.h 2011-01-17 02:41:00.000000000 -0500
-@@ -715,6 +715,21 @@ static inline void __set_fixmap(unsigned
- pv_mmu_ops.set_fixmap(idx, phys, flags);
- }
-
-+#ifdef CONFIG_PAX_KERNEXEC
-+static inline unsigned long pax_open_kernel(void)
-+{
-+ return PVOP_CALL0(unsigned long, pv_mmu_ops.pax_open_kernel);
-+}
-+
-+static inline unsigned long pax_close_kernel(void)
-+{
-+ return PVOP_CALL0(unsigned long, pv_mmu_ops.pax_close_kernel);
-+}
-+#else
-+static inline unsigned long pax_open_kernel(void) { return 0; }
-+static inline unsigned long pax_close_kernel(void) { return 0; }
-+#endif
-+
- #if defined(CONFIG_SMP) && defined(CONFIG_PARAVIRT_SPINLOCKS)
-
- static inline int arch_spin_is_locked(struct arch_spinlock *lock)
-@@ -931,7 +946,7 @@ extern void default_banner(void);
-
- #define PARA_PATCH(struct, off) ((PARAVIRT_PATCH_##struct + (off)) / 4)
- #define PARA_SITE(ptype, clobbers, ops) _PVSITE(ptype, clobbers, ops, .long, 4)
--#define PARA_INDIRECT(addr) *%cs:addr
-+#define PARA_INDIRECT(addr) *%ss:addr
- #endif
-
- #define INTERRUPT_RETURN \
-@@ -1008,6 +1023,21 @@ extern void default_banner(void);
- PARA_SITE(PARA_PATCH(pv_cpu_ops, PV_CPU_irq_enable_sysexit), \
- CLBR_NONE, \
- jmp PARA_INDIRECT(pv_cpu_ops+PV_CPU_irq_enable_sysexit))
-+
-+#define GET_CR0_INTO_RDI \
-+ call PARA_INDIRECT(pv_cpu_ops+PV_CPU_read_cr0); \
-+ mov %rax,%rdi
-+
-+#define SET_RDI_INTO_CR0 \
-+ call PARA_INDIRECT(pv_cpu_ops+PV_CPU_write_cr0)
-+
-+#define GET_CR3_INTO_RDI \
-+ call PARA_INDIRECT(pv_mmu_ops+PV_MMU_read_cr3); \
-+ mov %rax,%rdi
-+
-+#define SET_RDI_INTO_CR3 \
-+ call PARA_INDIRECT(pv_mmu_ops+PV_MMU_write_cr3)
-+
- #endif /* CONFIG_X86_32 */
-
- #endif /* __ASSEMBLY__ */
-diff -urNp linux-2.6.37.4/arch/x86/include/asm/paravirt_types.h linux-2.6.37.4/arch/x86/include/asm/paravirt_types.h
---- linux-2.6.37.4/arch/x86/include/asm/paravirt_types.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/include/asm/paravirt_types.h 2011-01-17 02:41:00.000000000 -0500
-@@ -311,6 +311,12 @@ struct pv_mmu_ops {
- an mfn. We can tell which is which from the index. */
- void (*set_fixmap)(unsigned /* enum fixed_addresses */ idx,
- phys_addr_t phys, pgprot_t flags);
-+
-+#ifdef CONFIG_PAX_KERNEXEC
-+ unsigned long (*pax_open_kernel)(void);
-+ unsigned long (*pax_close_kernel)(void);
-+#endif
-+
- };
-
- struct arch_spinlock;
-diff -urNp linux-2.6.37.4/arch/x86/include/asm/pci_x86.h linux-2.6.37.4/arch/x86/include/asm/pci_x86.h
---- linux-2.6.37.4/arch/x86/include/asm/pci_x86.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/include/asm/pci_x86.h 2011-01-17 02:41:01.000000000 -0500
-@@ -93,16 +93,16 @@ extern int (*pcibios_enable_irq)(struct
- extern void (*pcibios_disable_irq)(struct pci_dev *dev);
-
- struct pci_raw_ops {
-- int (*read)(unsigned int domain, unsigned int bus, unsigned int devfn,
-+ int (* const read)(unsigned int domain, unsigned int bus, unsigned int devfn,
- int reg, int len, u32 *val);
-- int (*write)(unsigned int domain, unsigned int bus, unsigned int devfn,
-+ int (* const write)(unsigned int domain, unsigned int bus, unsigned int devfn,
- int reg, int len, u32 val);
- };
-
--extern struct pci_raw_ops *raw_pci_ops;
--extern struct pci_raw_ops *raw_pci_ext_ops;
-+extern const struct pci_raw_ops *raw_pci_ops;
-+extern const struct pci_raw_ops *raw_pci_ext_ops;
-
--extern struct pci_raw_ops pci_direct_conf1;
-+extern const struct pci_raw_ops pci_direct_conf1;
- extern bool port_cf9_safe;
-
- /* arch_initcall level */
-diff -urNp linux-2.6.37.4/arch/x86/include/asm/pgalloc.h linux-2.6.37.4/arch/x86/include/asm/pgalloc.h
---- linux-2.6.37.4/arch/x86/include/asm/pgalloc.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/include/asm/pgalloc.h 2011-01-17 02:41:01.000000000 -0500
-@@ -63,6 +63,13 @@ static inline void pmd_populate_kernel(s
- pmd_t *pmd, pte_t *pte)
- {
- paravirt_alloc_pte(mm, __pa(pte) >> PAGE_SHIFT);
-+ set_pmd(pmd, __pmd(__pa(pte) | _KERNPG_TABLE));
-+}
-+
-+static inline void pmd_populate_user(struct mm_struct *mm,
-+ pmd_t *pmd, pte_t *pte)
-+{
-+ paravirt_alloc_pte(mm, __pa(pte) >> PAGE_SHIFT);
- set_pmd(pmd, __pmd(__pa(pte) | _PAGE_TABLE));
- }
-
-diff -urNp linux-2.6.37.4/arch/x86/include/asm/pgtable-2level.h linux-2.6.37.4/arch/x86/include/asm/pgtable-2level.h
---- linux-2.6.37.4/arch/x86/include/asm/pgtable-2level.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/include/asm/pgtable-2level.h 2011-01-17 02:41:01.000000000 -0500
-@@ -18,7 +18,9 @@ static inline void native_set_pte(pte_t
-
- static inline void native_set_pmd(pmd_t *pmdp, pmd_t pmd)
- {
-+ pax_open_kernel();
- *pmdp = pmd;
-+ pax_close_kernel();
- }
-
- static inline void native_set_pte_atomic(pte_t *ptep, pte_t pte)
-diff -urNp linux-2.6.37.4/arch/x86/include/asm/pgtable_32.h linux-2.6.37.4/arch/x86/include/asm/pgtable_32.h
---- linux-2.6.37.4/arch/x86/include/asm/pgtable_32.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/include/asm/pgtable_32.h 2011-01-17 02:41:01.000000000 -0500
-@@ -25,9 +25,6 @@
- struct mm_struct;
- struct vm_area_struct;
-
--extern pgd_t swapper_pg_dir[1024];
--extern pgd_t initial_page_table[1024];
--
- static inline void pgtable_cache_init(void) { }
- static inline void check_pgt_cache(void) { }
- void paging_init(void);
-@@ -48,6 +45,12 @@ extern void set_pmd_pfn(unsigned long, u
- # include <asm/pgtable-2level.h>
- #endif
-
-+extern pgd_t swapper_pg_dir[PTRS_PER_PGD];
-+extern pgd_t initial_page_table[PTRS_PER_PGD];
-+#ifdef CONFIG_X86_PAE
-+extern pmd_t swapper_pm_dir[PTRS_PER_PGD][PTRS_PER_PMD];
-+#endif
-+
- #if defined(CONFIG_HIGHPTE)
- #define pte_offset_map(dir, address) \
- ((pte_t *)kmap_atomic(pmd_page(*(dir))) + \
-@@ -62,7 +65,9 @@ extern void set_pmd_pfn(unsigned long, u
- /* Clear a kernel PTE and flush it from the TLB */
- #define kpte_clear_flush(ptep, vaddr) \
- do { \
-+ pax_open_kernel(); \
- pte_clear(&init_mm, (vaddr), (ptep)); \
-+ pax_close_kernel(); \
- __flush_tlb_one((vaddr)); \
- } while (0)
-
-@@ -74,6 +79,9 @@ do { \
-
- #endif /* !__ASSEMBLY__ */
-
-+#define HAVE_ARCH_UNMAPPED_AREA
-+#define HAVE_ARCH_UNMAPPED_AREA_TOPDOWN
-+
- /*
- * kern_addr_valid() is (1) for FLATMEM and (0) for
- * SPARSEMEM and DISCONTIGMEM
-diff -urNp linux-2.6.37.4/arch/x86/include/asm/pgtable_32_types.h linux-2.6.37.4/arch/x86/include/asm/pgtable_32_types.h
---- linux-2.6.37.4/arch/x86/include/asm/pgtable_32_types.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/include/asm/pgtable_32_types.h 2011-01-17 02:41:01.000000000 -0500
-@@ -8,7 +8,7 @@
- */
- #ifdef CONFIG_X86_PAE
- # include <asm/pgtable-3level_types.h>
--# define PMD_SIZE (1UL << PMD_SHIFT)
-+# define PMD_SIZE (_AC(1, UL) << PMD_SHIFT)
- # define PMD_MASK (~(PMD_SIZE - 1))
- #else
- # include <asm/pgtable-2level_types.h>
-@@ -46,6 +46,19 @@ extern bool __vmalloc_start_set; /* set
- # define VMALLOC_END (FIXADDR_START - 2 * PAGE_SIZE)
- #endif
-
-+#ifdef CONFIG_PAX_KERNEXEC
-+#ifndef __ASSEMBLY__
-+extern unsigned char MODULES_EXEC_VADDR[];
-+extern unsigned char MODULES_EXEC_END[];
-+#endif
-+#include <asm/boot.h>
-+#define ktla_ktva(addr) (addr + LOAD_PHYSICAL_ADDR + PAGE_OFFSET)
-+#define ktva_ktla(addr) (addr - LOAD_PHYSICAL_ADDR - PAGE_OFFSET)
-+#else
-+#define ktla_ktva(addr) (addr)
-+#define ktva_ktla(addr) (addr)
-+#endif
-+
- #define MODULES_VADDR VMALLOC_START
- #define MODULES_END VMALLOC_END
- #define MODULES_LEN (MODULES_VADDR - MODULES_END)
-diff -urNp linux-2.6.37.4/arch/x86/include/asm/pgtable-3level.h linux-2.6.37.4/arch/x86/include/asm/pgtable-3level.h
---- linux-2.6.37.4/arch/x86/include/asm/pgtable-3level.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/include/asm/pgtable-3level.h 2011-01-17 02:41:01.000000000 -0500
-@@ -38,12 +38,16 @@ static inline void native_set_pte_atomic
-
- static inline void native_set_pmd(pmd_t *pmdp, pmd_t pmd)
- {
-+ pax_open_kernel();
- set_64bit((unsigned long long *)(pmdp), native_pmd_val(pmd));
-+ pax_close_kernel();
- }
-
- static inline void native_set_pud(pud_t *pudp, pud_t pud)
- {
-+ pax_open_kernel();
- set_64bit((unsigned long long *)(pudp), native_pud_val(pud));
-+ pax_close_kernel();
- }
-
- /*
-diff -urNp linux-2.6.37.4/arch/x86/include/asm/pgtable_64.h linux-2.6.37.4/arch/x86/include/asm/pgtable_64.h
---- linux-2.6.37.4/arch/x86/include/asm/pgtable_64.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/include/asm/pgtable_64.h 2011-01-17 02:41:01.000000000 -0500
-@@ -16,10 +16,13 @@
-
- extern pud_t level3_kernel_pgt[512];
- extern pud_t level3_ident_pgt[512];
-+extern pud_t level3_vmalloc_pgt[512];
-+extern pud_t level3_vmemmap_pgt[512];
-+extern pud_t level2_vmemmap_pgt[512];
- extern pmd_t level2_kernel_pgt[512];
- extern pmd_t level2_fixmap_pgt[512];
--extern pmd_t level2_ident_pgt[512];
--extern pgd_t init_level4_pgt[];
-+extern pmd_t level2_ident_pgt[512*2];
-+extern pgd_t init_level4_pgt[512];
-
- #define swapper_pg_dir init_level4_pgt
-
-@@ -74,7 +77,9 @@ static inline pte_t native_ptep_get_and_
-
- static inline void native_set_pmd(pmd_t *pmdp, pmd_t pmd)
- {
-+ pax_open_kernel();
- *pmdp = pmd;
-+ pax_close_kernel();
- }
-
- static inline void native_pmd_clear(pmd_t *pmd)
-@@ -94,7 +99,9 @@ static inline void native_pud_clear(pud_
-
- static inline void native_set_pgd(pgd_t *pgdp, pgd_t pgd)
- {
-+ pax_open_kernel();
- *pgdp = pgd;
-+ pax_close_kernel();
- }
-
- static inline void native_pgd_clear(pgd_t *pgd)
-diff -urNp linux-2.6.37.4/arch/x86/include/asm/pgtable_64_types.h linux-2.6.37.4/arch/x86/include/asm/pgtable_64_types.h
---- linux-2.6.37.4/arch/x86/include/asm/pgtable_64_types.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/include/asm/pgtable_64_types.h 2011-01-17 02:41:01.000000000 -0500
-@@ -59,5 +59,10 @@ typedef struct { pteval_t pte; } pte_t;
- #define MODULES_VADDR _AC(0xffffffffa0000000, UL)
- #define MODULES_END _AC(0xffffffffff000000, UL)
- #define MODULES_LEN (MODULES_END - MODULES_VADDR)
-+#define MODULES_EXEC_VADDR MODULES_VADDR
-+#define MODULES_EXEC_END MODULES_END
-+
-+#define ktla_ktva(addr) (addr)
-+#define ktva_ktla(addr) (addr)
-
- #endif /* _ASM_X86_PGTABLE_64_DEFS_H */
-diff -urNp linux-2.6.37.4/arch/x86/include/asm/pgtable.h linux-2.6.37.4/arch/x86/include/asm/pgtable.h
---- linux-2.6.37.4/arch/x86/include/asm/pgtable.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/include/asm/pgtable.h 2011-01-17 02:41:01.000000000 -0500
-@@ -78,12 +78,51 @@ extern struct mm_struct *pgd_page_get_mm
-
- #define arch_end_context_switch(prev) do {} while(0)
-
-+#define pax_open_kernel() native_pax_open_kernel()
-+#define pax_close_kernel() native_pax_close_kernel()
- #endif /* CONFIG_PARAVIRT */
-
-+#define __HAVE_ARCH_PAX_OPEN_KERNEL
-+#define __HAVE_ARCH_PAX_CLOSE_KERNEL
-+
-+#ifdef CONFIG_PAX_KERNEXEC
-+static inline unsigned long native_pax_open_kernel(void)
-+{
-+ unsigned long cr0;
-+
-+ preempt_disable();
-+ barrier();
-+ cr0 = read_cr0() ^ X86_CR0_WP;
-+ BUG_ON(unlikely(cr0 & X86_CR0_WP));
-+ write_cr0(cr0);
-+ return cr0 ^ X86_CR0_WP;
-+}
-+
-+static inline unsigned long native_pax_close_kernel(void)
-+{
-+ unsigned long cr0;
-+
-+ cr0 = read_cr0() ^ X86_CR0_WP;
-+ BUG_ON(unlikely(!(cr0 & X86_CR0_WP)));
-+ write_cr0(cr0);
-+ barrier();
-+ preempt_enable_no_resched();
-+ return cr0 ^ X86_CR0_WP;
-+}
-+#else
-+static inline unsigned long native_pax_open_kernel(void) { return 0; }
-+static inline unsigned long native_pax_close_kernel(void) { return 0; }
-+#endif
-+
- /*
- * The following only work if pte_present() is true.
- * Undefined behaviour if not..
- */
-+static inline int pte_user(pte_t pte)
-+{
-+ return pte_val(pte) & _PAGE_USER;
-+}
-+
- static inline int pte_dirty(pte_t pte)
- {
- return pte_flags(pte) & _PAGE_DIRTY;
-@@ -171,9 +210,29 @@ static inline pte_t pte_wrprotect(pte_t
- return pte_clear_flags(pte, _PAGE_RW);
- }
-
-+static inline pte_t pte_mkread(pte_t pte)
-+{
-+ return __pte(pte_val(pte) | _PAGE_USER);
-+}
-+
- static inline pte_t pte_mkexec(pte_t pte)
- {
-- return pte_clear_flags(pte, _PAGE_NX);
-+#ifdef CONFIG_X86_PAE
-+ if (__supported_pte_mask & _PAGE_NX)
-+ return pte_clear_flags(pte, _PAGE_NX);
-+ else
-+#endif
-+ return pte_set_flags(pte, _PAGE_USER);
-+}
-+
-+static inline pte_t pte_exprotect(pte_t pte)
-+{
-+#ifdef CONFIG_X86_PAE
-+ if (__supported_pte_mask & _PAGE_NX)
-+ return pte_set_flags(pte, _PAGE_NX);
-+ else
-+#endif
-+ return pte_clear_flags(pte, _PAGE_USER);
- }
-
- static inline pte_t pte_mkdirty(pte_t pte)
-@@ -306,6 +365,15 @@ pte_t *populate_extra_pte(unsigned long
- #endif
-
- #ifndef __ASSEMBLY__
-+
-+#ifdef CONFIG_PAX_PER_CPU_PGD
-+extern pgd_t cpu_pgd[NR_CPUS][PTRS_PER_PGD];
-+static inline pgd_t *get_cpu_pgd(unsigned int cpu)
-+{
-+ return cpu_pgd[cpu];
-+}
-+#endif
-+
- #include <linux/mm_types.h>
-
- static inline int pte_none(pte_t pte)
-@@ -476,7 +544,7 @@ static inline pud_t *pud_offset(pgd_t *p
-
- static inline int pgd_bad(pgd_t pgd)
- {
-- return (pgd_flags(pgd) & ~_PAGE_USER) != _KERNPG_TABLE;
-+ return (pgd_flags(pgd) & ~(_PAGE_USER | _PAGE_NX)) != _KERNPG_TABLE;
- }
-
- static inline int pgd_none(pgd_t pgd)
-@@ -499,7 +567,12 @@ static inline int pgd_none(pgd_t pgd)
- * pgd_offset() returns a (pgd_t *)
- * pgd_index() is used get the offset into the pgd page's array of pgd_t's;
- */
--#define pgd_offset(mm, address) ((mm)->pgd + pgd_index((address)))
-+#define pgd_offset(mm, address) ((mm)->pgd + pgd_index(address))
-+
-+#ifdef CONFIG_PAX_PER_CPU_PGD
-+#define pgd_offset_cpu(cpu, address) (get_cpu_pgd(cpu) + pgd_index(address))
-+#endif
-+
- /*
- * a shortcut which implies the use of the kernel's pgd, instead
- * of a process's
-@@ -510,6 +583,20 @@ static inline int pgd_none(pgd_t pgd)
- #define KERNEL_PGD_BOUNDARY pgd_index(PAGE_OFFSET)
- #define KERNEL_PGD_PTRS (PTRS_PER_PGD - KERNEL_PGD_BOUNDARY)
-
-+#ifdef CONFIG_X86_32
-+#define USER_PGD_PTRS KERNEL_PGD_BOUNDARY
-+#else
-+#define TASK_SIZE_MAX_SHIFT CONFIG_TASK_SIZE_MAX_SHIFT
-+#define USER_PGD_PTRS (_AC(1,UL) << (TASK_SIZE_MAX_SHIFT - PGDIR_SHIFT))
-+
-+#ifdef CONFIG_PAX_MEMORY_UDEREF
-+#define PAX_USER_SHADOW_BASE (_AC(1,UL) << TASK_SIZE_MAX_SHIFT)
-+#else
-+#define PAX_USER_SHADOW_BASE (_AC(0,UL))
-+#endif
-+
-+#endif
-+
- #ifndef __ASSEMBLY__
-
- extern int direct_gbpages;
-@@ -617,11 +704,23 @@ static inline void ptep_set_wrprotect(st
- * dst and src can be on the same page, but the range must not overlap,
- * and must not cross a page boundary.
- */
--static inline void clone_pgd_range(pgd_t *dst, pgd_t *src, int count)
-+static inline void clone_pgd_range(pgd_t *dst, const pgd_t *src, int count)
- {
-- memcpy(dst, src, count * sizeof(pgd_t));
-+ pax_open_kernel();
-+ while (count--)
-+ *dst++ = *src++;
-+ pax_close_kernel();
- }
-
-+#ifdef CONFIG_PAX_PER_CPU_PGD
-+extern void __clone_user_pgds(pgd_t *dst, const pgd_t *src, int count);
-+#endif
-+
-+#if defined(CONFIG_X86_64) && defined(CONFIG_PAX_MEMORY_UDEREF)
-+extern void __shadow_user_pgds(pgd_t *dst, const pgd_t *src, int count);
-+#else
-+static inline void __shadow_user_pgds(pgd_t *dst, const pgd_t *src, int count) {}
-+#endif
-
- #include <asm-generic/pgtable.h>
- #endif /* __ASSEMBLY__ */
-diff -urNp linux-2.6.37.4/arch/x86/include/asm/pgtable_types.h linux-2.6.37.4/arch/x86/include/asm/pgtable_types.h
---- linux-2.6.37.4/arch/x86/include/asm/pgtable_types.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/include/asm/pgtable_types.h 2011-01-17 02:41:01.000000000 -0500
-@@ -16,12 +16,11 @@
- #define _PAGE_BIT_PSE 7 /* 4 MB (or 2MB) page */
- #define _PAGE_BIT_PAT 7 /* on 4KB pages */
- #define _PAGE_BIT_GLOBAL 8 /* Global TLB entry PPro+ */
--#define _PAGE_BIT_UNUSED1 9 /* available for programmer */
-+#define _PAGE_BIT_SPECIAL 9 /* special mappings, no associated struct page */
- #define _PAGE_BIT_IOMAP 10 /* flag used to indicate IO mapping */
- #define _PAGE_BIT_HIDDEN 11 /* hidden by kmemcheck */
- #define _PAGE_BIT_PAT_LARGE 12 /* On 2MB or 1GB pages */
--#define _PAGE_BIT_SPECIAL _PAGE_BIT_UNUSED1
--#define _PAGE_BIT_CPA_TEST _PAGE_BIT_UNUSED1
-+#define _PAGE_BIT_CPA_TEST _PAGE_BIT_SPECIAL
- #define _PAGE_BIT_NX 63 /* No execute: only valid after cpuid check */
-
- /* If _PAGE_BIT_PRESENT is clear, we use these: */
-@@ -39,7 +38,6 @@
- #define _PAGE_DIRTY (_AT(pteval_t, 1) << _PAGE_BIT_DIRTY)
- #define _PAGE_PSE (_AT(pteval_t, 1) << _PAGE_BIT_PSE)
- #define _PAGE_GLOBAL (_AT(pteval_t, 1) << _PAGE_BIT_GLOBAL)
--#define _PAGE_UNUSED1 (_AT(pteval_t, 1) << _PAGE_BIT_UNUSED1)
- #define _PAGE_IOMAP (_AT(pteval_t, 1) << _PAGE_BIT_IOMAP)
- #define _PAGE_PAT (_AT(pteval_t, 1) << _PAGE_BIT_PAT)
- #define _PAGE_PAT_LARGE (_AT(pteval_t, 1) << _PAGE_BIT_PAT_LARGE)
-@@ -55,8 +53,10 @@
-
- #if defined(CONFIG_X86_64) || defined(CONFIG_X86_PAE)
- #define _PAGE_NX (_AT(pteval_t, 1) << _PAGE_BIT_NX)
--#else
-+#elif defined(CONFIG_KMEMCHECK)
- #define _PAGE_NX (_AT(pteval_t, 0))
-+#else
-+#define _PAGE_NX (_AT(pteval_t, 1) << _PAGE_BIT_HIDDEN)
- #endif
-
- #define _PAGE_FILE (_AT(pteval_t, 1) << _PAGE_BIT_FILE)
-@@ -93,6 +93,9 @@
- #define PAGE_READONLY_EXEC __pgprot(_PAGE_PRESENT | _PAGE_USER | \
- _PAGE_ACCESSED)
-
-+#define PAGE_READONLY_NOEXEC PAGE_READONLY
-+#define PAGE_SHARED_NOEXEC PAGE_SHARED
-+
- #define __PAGE_KERNEL_EXEC \
- (_PAGE_PRESENT | _PAGE_RW | _PAGE_DIRTY | _PAGE_ACCESSED | _PAGE_GLOBAL)
- #define __PAGE_KERNEL (__PAGE_KERNEL_EXEC | _PAGE_NX)
-@@ -103,8 +106,8 @@
- #define __PAGE_KERNEL_WC (__PAGE_KERNEL | _PAGE_CACHE_WC)
- #define __PAGE_KERNEL_NOCACHE (__PAGE_KERNEL | _PAGE_PCD | _PAGE_PWT)
- #define __PAGE_KERNEL_UC_MINUS (__PAGE_KERNEL | _PAGE_PCD)
--#define __PAGE_KERNEL_VSYSCALL (__PAGE_KERNEL_RX | _PAGE_USER)
--#define __PAGE_KERNEL_VSYSCALL_NOCACHE (__PAGE_KERNEL_VSYSCALL | _PAGE_PCD | _PAGE_PWT)
-+#define __PAGE_KERNEL_VSYSCALL (__PAGE_KERNEL_RO | _PAGE_USER)
-+#define __PAGE_KERNEL_VSYSCALL_NOCACHE (__PAGE_KERNEL_RO | _PAGE_PCD | _PAGE_PWT | _PAGE_USER)
- #define __PAGE_KERNEL_LARGE (__PAGE_KERNEL | _PAGE_PSE)
- #define __PAGE_KERNEL_LARGE_NOCACHE (__PAGE_KERNEL | _PAGE_CACHE_UC | _PAGE_PSE)
- #define __PAGE_KERNEL_LARGE_EXEC (__PAGE_KERNEL_EXEC | _PAGE_PSE)
-@@ -163,8 +166,8 @@
- * bits are combined, this will alow user to access the high address mapped
- * VDSO in the presence of CONFIG_COMPAT_VDSO
- */
--#define PTE_IDENT_ATTR 0x003 /* PRESENT+RW */
--#define PDE_IDENT_ATTR 0x067 /* PRESENT+RW+USER+DIRTY+ACCESSED */
-+#define PTE_IDENT_ATTR 0x063 /* PRESENT+RW+DIRTY+ACCESSED */
-+#define PDE_IDENT_ATTR 0x063 /* PRESENT+RW+DIRTY+ACCESSED */
- #define PGD_IDENT_ATTR 0x001 /* PRESENT (no other attributes) */
- #endif
-
-@@ -202,7 +205,17 @@ static inline pgdval_t pgd_flags(pgd_t p
- {
- return native_pgd_val(pgd) & PTE_FLAGS_MASK;
- }
-+#endif
-
-+#if PAGETABLE_LEVELS == 3
-+#include <asm-generic/pgtable-nopud.h>
-+#endif
-+
-+#if PAGETABLE_LEVELS == 2
-+#include <asm-generic/pgtable-nopmd.h>
-+#endif
-+
-+#ifndef __ASSEMBLY__
- #if PAGETABLE_LEVELS > 3
- typedef struct { pudval_t pud; } pud_t;
-
-@@ -216,8 +229,6 @@ static inline pudval_t native_pud_val(pu
- return pud.pud;
- }
- #else
--#include <asm-generic/pgtable-nopud.h>
--
- static inline pudval_t native_pud_val(pud_t pud)
- {
- return native_pgd_val(pud.pgd);
-@@ -237,8 +248,6 @@ static inline pmdval_t native_pmd_val(pm
- return pmd.pmd;
- }
- #else
--#include <asm-generic/pgtable-nopmd.h>
--
- static inline pmdval_t native_pmd_val(pmd_t pmd)
- {
- return native_pgd_val(pmd.pud.pgd);
-@@ -278,7 +287,6 @@ typedef struct page *pgtable_t;
-
- extern pteval_t __supported_pte_mask;
- extern void set_nx(void);
--extern int nx_enabled;
-
- #define pgprot_writecombine pgprot_writecombine
- extern pgprot_t pgprot_writecombine(pgprot_t prot);
-diff -urNp linux-2.6.37.4/arch/x86/include/asm/processor.h linux-2.6.37.4/arch/x86/include/asm/processor.h
---- linux-2.6.37.4/arch/x86/include/asm/processor.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/include/asm/processor.h 2011-01-17 02:41:01.000000000 -0500
-@@ -271,7 +271,7 @@ struct tss_struct {
-
- } ____cacheline_aligned;
-
--DECLARE_PER_CPU_SHARED_ALIGNED(struct tss_struct, init_tss);
-+extern struct tss_struct init_tss[NR_CPUS];
-
- /*
- * Save the original ist values for checking stack pointers during debugging
-@@ -864,8 +864,15 @@ static inline void spin_lock_prefetch(co
- */
- #define TASK_SIZE PAGE_OFFSET
- #define TASK_SIZE_MAX TASK_SIZE
-+
-+#ifdef CONFIG_PAX_SEGMEXEC
-+#define SEGMEXEC_TASK_SIZE (TASK_SIZE / 2)
-+#define STACK_TOP ((current->mm->pax_flags & MF_PAX_SEGMEXEC)?SEGMEXEC_TASK_SIZE:TASK_SIZE)
-+#else
- #define STACK_TOP TASK_SIZE
--#define STACK_TOP_MAX STACK_TOP
-+#endif
-+
-+#define STACK_TOP_MAX TASK_SIZE
-
- #define INIT_THREAD { \
- .sp0 = sizeof(init_stack) + (long)&init_stack, \
-@@ -882,7 +889,7 @@ static inline void spin_lock_prefetch(co
- */
- #define INIT_TSS { \
- .x86_tss = { \
-- .sp0 = sizeof(init_stack) + (long)&init_stack, \
-+ .sp0 = sizeof(init_stack) + (long)&init_stack - 8, \
- .ss0 = __KERNEL_DS, \
- .ss1 = __KERNEL_CS, \
- .io_bitmap_base = INVALID_IO_BITMAP_OFFSET, \
-@@ -893,11 +900,7 @@ static inline void spin_lock_prefetch(co
- extern unsigned long thread_saved_pc(struct task_struct *tsk);
-
- #define THREAD_SIZE_LONGS (THREAD_SIZE/sizeof(unsigned long))
--#define KSTK_TOP(info) \
--({ \
-- unsigned long *__ptr = (unsigned long *)(info); \
-- (unsigned long)(&__ptr[THREAD_SIZE_LONGS]); \
--})
-+#define KSTK_TOP(info) ((info)->task.thread.sp0)
-
- /*
- * The below -8 is to reserve 8 bytes on top of the ring0 stack.
-@@ -912,7 +915,7 @@ extern unsigned long thread_saved_pc(str
- #define task_pt_regs(task) \
- ({ \
- struct pt_regs *__regs__; \
-- __regs__ = (struct pt_regs *)(KSTK_TOP(task_stack_page(task))-8); \
-+ __regs__ = (struct pt_regs *)((task)->thread.sp0); \
- __regs__ - 1; \
- })
-
-@@ -922,13 +925,13 @@ extern unsigned long thread_saved_pc(str
- /*
- * User space process size. 47bits minus one guard page.
- */
--#define TASK_SIZE_MAX ((1UL << 47) - PAGE_SIZE)
-+#define TASK_SIZE_MAX ((1UL << TASK_SIZE_MAX_SHIFT) - PAGE_SIZE)
-
- /* This decides where the kernel will search for a free chunk of vm
- * space during mmap's.
- */
- #define IA32_PAGE_OFFSET ((current->personality & ADDR_LIMIT_3GB) ? \
-- 0xc0000000 : 0xFFFFe000)
-+ 0xc0000000 : 0xFFFFf000)
-
- #define TASK_SIZE (test_thread_flag(TIF_IA32) ? \
- IA32_PAGE_OFFSET : TASK_SIZE_MAX)
-@@ -965,6 +968,10 @@ extern void start_thread(struct pt_regs
- */
- #define TASK_UNMAPPED_BASE (PAGE_ALIGN(TASK_SIZE / 3))
-
-+#ifdef CONFIG_PAX_SEGMEXEC
-+#define SEGMEXEC_TASK_UNMAPPED_BASE (PAGE_ALIGN(SEGMEXEC_TASK_SIZE / 3))
-+#endif
-+
- #define KSTK_EIP(task) (task_pt_regs(task)->ip)
-
- /* Get/set a process' ability to use the timestamp counter instruction */
-diff -urNp linux-2.6.37.4/arch/x86/include/asm/ptrace.h linux-2.6.37.4/arch/x86/include/asm/ptrace.h
---- linux-2.6.37.4/arch/x86/include/asm/ptrace.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/include/asm/ptrace.h 2011-01-17 02:41:01.000000000 -0500
-@@ -152,28 +152,29 @@ static inline unsigned long regs_return_
- }
-
- /*
-- * user_mode_vm(regs) determines whether a register set came from user mode.
-+ * user_mode(regs) determines whether a register set came from user mode.
- * This is true if V8086 mode was enabled OR if the register set was from
- * protected mode with RPL-3 CS value. This tricky test checks that with
- * one comparison. Many places in the kernel can bypass this full check
-- * if they have already ruled out V8086 mode, so user_mode(regs) can be used.
-+ * if they have already ruled out V8086 mode, so user_mode_novm(regs) can
-+ * be used.
- */
--static inline int user_mode(struct pt_regs *regs)
-+static inline int user_mode_novm(struct pt_regs *regs)
- {
- #ifdef CONFIG_X86_32
- return (regs->cs & SEGMENT_RPL_MASK) == USER_RPL;
- #else
-- return !!(regs->cs & 3);
-+ return !!(regs->cs & SEGMENT_RPL_MASK);
- #endif
- }
-
--static inline int user_mode_vm(struct pt_regs *regs)
-+static inline int user_mode(struct pt_regs *regs)
- {
- #ifdef CONFIG_X86_32
- return ((regs->cs & SEGMENT_RPL_MASK) | (regs->flags & X86_VM_MASK)) >=
- USER_RPL;
- #else
-- return user_mode(regs);
-+ return user_mode_novm(regs);
- #endif
- }
-
-diff -urNp linux-2.6.37.4/arch/x86/include/asm/reboot.h linux-2.6.37.4/arch/x86/include/asm/reboot.h
---- linux-2.6.37.4/arch/x86/include/asm/reboot.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/include/asm/reboot.h 2011-01-17 02:41:01.000000000 -0500
-@@ -18,7 +18,7 @@ extern struct machine_ops machine_ops;
-
- void native_machine_crash_shutdown(struct pt_regs *regs);
- void native_machine_shutdown(void);
--void machine_real_restart(const unsigned char *code, int length);
-+void machine_real_restart(const unsigned char *code, unsigned int length);
-
- typedef void (*nmi_shootdown_cb)(int, struct die_args*);
- void nmi_shootdown_cpus(nmi_shootdown_cb callback);
-diff -urNp linux-2.6.37.4/arch/x86/include/asm/rwsem.h linux-2.6.37.4/arch/x86/include/asm/rwsem.h
---- linux-2.6.37.4/arch/x86/include/asm/rwsem.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/include/asm/rwsem.h 2011-01-17 02:41:01.000000000 -0500
-@@ -118,6 +118,14 @@ static inline void __down_read(struct rw
- {
- asm volatile("# beginning down_read\n\t"
- LOCK_PREFIX _ASM_INC "(%1)\n\t"
-+
-+#ifdef CONFIG_PAX_REFCOUNT
-+ "jno 0f\n"
-+ LOCK_PREFIX _ASM_DEC "(%1)\n"
-+ "int $4\n0:\n"
-+ _ASM_EXTABLE(0b, 0b)
-+#endif
-+
- /* adds 0x00000001 */
- " jns 1f\n"
- " call call_rwsem_down_read_failed\n"
-@@ -139,6 +147,14 @@ static inline int __down_read_trylock(st
- "1:\n\t"
- " mov %1,%2\n\t"
- " add %3,%2\n\t"
-+
-+#ifdef CONFIG_PAX_REFCOUNT
-+ "jno 0f\n"
-+ "sub %3,%2\n"
-+ "int $4\n0:\n"
-+ _ASM_EXTABLE(0b, 0b)
-+#endif
-+
- " jle 2f\n\t"
- LOCK_PREFIX " cmpxchg %2,%0\n\t"
- " jnz 1b\n\t"
-@@ -158,6 +174,14 @@ static inline void __down_write_nested(s
- rwsem_count_t tmp;
- asm volatile("# beginning down_write\n\t"
- LOCK_PREFIX " xadd %1,(%2)\n\t"
-+
-+#ifdef CONFIG_PAX_REFCOUNT
-+ "jno 0f\n"
-+ "mov %1,(%2)\n"
-+ "int $4\n0:\n"
-+ _ASM_EXTABLE(0b, 0b)
-+#endif
-+
- /* adds 0xffff0001, returns the old value */
- " test %1,%1\n\t"
- /* was the count 0 before? */
-@@ -196,6 +220,14 @@ static inline void __up_read(struct rw_s
- rwsem_count_t tmp;
- asm volatile("# beginning __up_read\n\t"
- LOCK_PREFIX " xadd %1,(%2)\n\t"
-+
-+#ifdef CONFIG_PAX_REFCOUNT
-+ "jno 0f\n"
-+ "mov %1,(%2)\n"
-+ "int $4\n0:\n"
-+ _ASM_EXTABLE(0b, 0b)
-+#endif
-+
- /* subtracts 1, returns the old value */
- " jns 1f\n\t"
- " call call_rwsem_wake\n" /* expects old value in %edx */
-@@ -214,6 +246,14 @@ static inline void __up_write(struct rw_
- rwsem_count_t tmp;
- asm volatile("# beginning __up_write\n\t"
- LOCK_PREFIX " xadd %1,(%2)\n\t"
-+
-+#ifdef CONFIG_PAX_REFCOUNT
-+ "jno 0f\n"
-+ "mov %1,(%2)\n"
-+ "int $4\n0:\n"
-+ _ASM_EXTABLE(0b, 0b)
-+#endif
-+
- /* subtracts 0xffff0001, returns the old value */
- " jns 1f\n\t"
- " call call_rwsem_wake\n" /* expects old value in %edx */
-@@ -231,6 +271,14 @@ static inline void __downgrade_write(str
- {
- asm volatile("# beginning __downgrade_write\n\t"
- LOCK_PREFIX _ASM_ADD "%2,(%1)\n\t"
-+
-+#ifdef CONFIG_PAX_REFCOUNT
-+ "jno 0f\n"
-+ LOCK_PREFIX _ASM_SUB "%2,(%1)\n"
-+ "int $4\n0:\n"
-+ _ASM_EXTABLE(0b, 0b)
-+#endif
-+
- /*
- * transitions 0xZZZZ0001 -> 0xYYYY0001 (i386)
- * 0xZZZZZZZZ00000001 -> 0xYYYYYYYY00000001 (x86_64)
-@@ -250,7 +298,15 @@ static inline void __downgrade_write(str
- static inline void rwsem_atomic_add(rwsem_count_t delta,
- struct rw_semaphore *sem)
- {
-- asm volatile(LOCK_PREFIX _ASM_ADD "%1,%0"
-+ asm volatile(LOCK_PREFIX _ASM_ADD "%1,%0\n"
-+
-+#ifdef CONFIG_PAX_REFCOUNT
-+ "jno 0f\n"
-+ LOCK_PREFIX _ASM_SUB "%1,%0\n"
-+ "int $4\n0:\n"
-+ _ASM_EXTABLE(0b, 0b)
-+#endif
-+
- : "+m" (sem->count)
- : "er" (delta));
- }
-@@ -263,7 +319,15 @@ static inline rwsem_count_t rwsem_atomic
- {
- rwsem_count_t tmp = delta;
-
-- asm volatile(LOCK_PREFIX "xadd %0,%1"
-+ asm volatile(LOCK_PREFIX "xadd %0,%1\n"
-+
-+#ifdef CONFIG_PAX_REFCOUNT
-+ "jno 0f\n"
-+ "mov %0,%1\n"
-+ "int $4\n0:\n"
-+ _ASM_EXTABLE(0b, 0b)
-+#endif
-+
- : "+r" (tmp), "+m" (sem->count)
- : : "memory");
-
-diff -urNp linux-2.6.37.4/arch/x86/include/asm/segment.h linux-2.6.37.4/arch/x86/include/asm/segment.h
---- linux-2.6.37.4/arch/x86/include/asm/segment.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/include/asm/segment.h 2011-01-17 02:41:01.000000000 -0500
-@@ -62,8 +62,8 @@
- * 26 - ESPFIX small SS
- * 27 - per-cpu [ offset to per-cpu data area ]
- * 28 - stack_canary-20 [ for stack protector ]
-- * 29 - unused
-- * 30 - unused
-+ * 29 - PCI BIOS CS
-+ * 30 - PCI BIOS DS
- * 31 - TSS for double fault handler
- */
- #define GDT_ENTRY_TLS_MIN 6
-@@ -77,6 +77,8 @@
-
- #define GDT_ENTRY_KERNEL_CS (GDT_ENTRY_KERNEL_BASE+0)
-
-+#define GDT_ENTRY_KERNEXEC_KERNEL_CS (4)
-+
- #define GDT_ENTRY_KERNEL_DS (GDT_ENTRY_KERNEL_BASE+1)
-
- #define GDT_ENTRY_TSS (GDT_ENTRY_KERNEL_BASE+4)
-@@ -102,6 +104,12 @@
- #define __KERNEL_STACK_CANARY 0
- #endif
-
-+#define GDT_ENTRY_PCIBIOS_CS (GDT_ENTRY_KERNEL_BASE+17)
-+#define __PCIBIOS_CS (GDT_ENTRY_PCIBIOS_CS * 8)
-+
-+#define GDT_ENTRY_PCIBIOS_DS (GDT_ENTRY_KERNEL_BASE+18)
-+#define __PCIBIOS_DS (GDT_ENTRY_PCIBIOS_DS * 8)
-+
- #define GDT_ENTRY_DOUBLEFAULT_TSS 31
-
- /*
-@@ -139,7 +147,7 @@
- */
-
- /* Matches PNP_CS32 and PNP_CS16 (they must be consecutive) */
--#define SEGMENT_IS_PNP_CODE(x) (((x) & 0xf4) == GDT_ENTRY_PNPBIOS_BASE * 8)
-+#define SEGMENT_IS_PNP_CODE(x) (((x) & 0xFFFCU) == PNP_CS32 || ((x) & 0xFFFCU) == PNP_CS16)
-
-
- #else
-@@ -163,6 +171,8 @@
- #define __USER32_CS (GDT_ENTRY_DEFAULT_USER32_CS * 8 + 3)
- #define __USER32_DS __USER_DS
-
-+#define GDT_ENTRY_KERNEXEC_KERNEL_CS 7
-+
- #define GDT_ENTRY_TSS 8 /* needs two entries */
- #define GDT_ENTRY_LDT 10 /* needs two entries */
- #define GDT_ENTRY_TLS_MIN 12
-@@ -183,6 +193,7 @@
- #endif
-
- #define __KERNEL_CS (GDT_ENTRY_KERNEL_CS*8)
-+#define __KERNEXEC_KERNEL_CS (GDT_ENTRY_KERNEXEC_KERNEL_CS*8)
- #define __KERNEL_DS (GDT_ENTRY_KERNEL_DS*8)
- #define __USER_DS (GDT_ENTRY_DEFAULT_USER_DS*8+3)
- #define __USER_CS (GDT_ENTRY_DEFAULT_USER_CS*8+3)
-diff -urNp linux-2.6.37.4/arch/x86/include/asm/smp.h linux-2.6.37.4/arch/x86/include/asm/smp.h
---- linux-2.6.37.4/arch/x86/include/asm/smp.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/include/asm/smp.h 2011-01-17 02:41:01.000000000 -0500
-@@ -24,7 +24,7 @@ extern unsigned int num_processors;
- DECLARE_PER_CPU(cpumask_var_t, cpu_sibling_map);
- DECLARE_PER_CPU(cpumask_var_t, cpu_core_map);
- DECLARE_PER_CPU(u16, cpu_llc_id);
--DECLARE_PER_CPU(int, cpu_number);
-+DECLARE_PER_CPU(unsigned int, cpu_number);
-
- static inline struct cpumask *cpu_sibling_mask(int cpu)
- {
-diff -urNp linux-2.6.37.4/arch/x86/include/asm/spinlock.h linux-2.6.37.4/arch/x86/include/asm/spinlock.h
---- linux-2.6.37.4/arch/x86/include/asm/spinlock.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/include/asm/spinlock.h 2011-01-17 02:41:01.000000000 -0500
-@@ -249,6 +249,14 @@ static inline int arch_write_can_lock(ar
- static inline void arch_read_lock(arch_rwlock_t *rw)
- {
- asm volatile(LOCK_PREFIX " subl $1,(%0)\n\t"
-+
-+#ifdef CONFIG_PAX_REFCOUNT
-+ "jno 0f\n"
-+ LOCK_PREFIX " addl $1,(%0)\n"
-+ "int $4\n0:\n"
-+ _ASM_EXTABLE(0b, 0b)
-+#endif
-+
- "jns 1f\n"
- "call __read_lock_failed\n\t"
- "1:\n"
-@@ -258,6 +266,14 @@ static inline void arch_read_lock(arch_r
- static inline void arch_write_lock(arch_rwlock_t *rw)
- {
- asm volatile(LOCK_PREFIX " subl %1,(%0)\n\t"
-+
-+#ifdef CONFIG_PAX_REFCOUNT
-+ "jno 0f\n"
-+ LOCK_PREFIX " addl %1,(%0)\n"
-+ "int $4\n0:\n"
-+ _ASM_EXTABLE(0b, 0b)
-+#endif
-+
- "jz 1f\n"
- "call __write_lock_failed\n\t"
- "1:\n"
-@@ -286,12 +302,29 @@ static inline int arch_write_trylock(arc
-
- static inline void arch_read_unlock(arch_rwlock_t *rw)
- {
-- asm volatile(LOCK_PREFIX "incl %0" :"+m" (rw->lock) : : "memory");
-+ asm volatile(LOCK_PREFIX "incl %0\n"
-+
-+#ifdef CONFIG_PAX_REFCOUNT
-+ "jno 0f\n"
-+ LOCK_PREFIX "decl %0\n"
-+ "int $4\n0:\n"
-+ _ASM_EXTABLE(0b, 0b)
-+#endif
-+
-+ :"+m" (rw->lock) : : "memory");
- }
-
- static inline void arch_write_unlock(arch_rwlock_t *rw)
- {
-- asm volatile(LOCK_PREFIX "addl %1, %0"
-+ asm volatile(LOCK_PREFIX "addl %1, %0\n"
-+
-+#ifdef CONFIG_PAX_REFCOUNT
-+ "jno 0f\n"
-+ LOCK_PREFIX "subl %1, %0\n"
-+ "int $4\n0:\n"
-+ _ASM_EXTABLE(0b, 0b)
-+#endif
-+
- : "+m" (rw->lock) : "i" (RW_LOCK_BIAS) : "memory");
- }
-
-diff -urNp linux-2.6.37.4/arch/x86/include/asm/stackprotector.h linux-2.6.37.4/arch/x86/include/asm/stackprotector.h
---- linux-2.6.37.4/arch/x86/include/asm/stackprotector.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/include/asm/stackprotector.h 2011-01-17 02:41:01.000000000 -0500
-@@ -113,7 +113,7 @@ static inline void setup_stack_canary_se
-
- static inline void load_stack_canary_segment(void)
- {
--#ifdef CONFIG_X86_32
-+#if defined(CONFIG_X86_32) && !defined(CONFIG_PAX_MEMORY_UDEREF)
- asm volatile ("mov %0, %%gs" : : "r" (0));
- #endif
- }
-diff -urNp linux-2.6.37.4/arch/x86/include/asm/system.h linux-2.6.37.4/arch/x86/include/asm/system.h
---- linux-2.6.37.4/arch/x86/include/asm/system.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/include/asm/system.h 2011-01-17 02:41:01.000000000 -0500
-@@ -202,7 +202,7 @@ static inline unsigned long get_limit(un
- {
- unsigned long __limit;
- asm("lsll %1,%0" : "=r" (__limit) : "r" (segment));
-- return __limit + 1;
-+ return __limit;
- }
-
- static inline void native_clts(void)
-@@ -342,7 +342,7 @@ void enable_hlt(void);
-
- void cpu_idle_wait(void);
-
--extern unsigned long arch_align_stack(unsigned long sp);
-+#define arch_align_stack(x) ((x) & ~0xfUL)
- extern void free_init_pages(char *what, unsigned long begin, unsigned long end);
-
- void default_idle(void);
-diff -urNp linux-2.6.37.4/arch/x86/include/asm/uaccess_32.h linux-2.6.37.4/arch/x86/include/asm/uaccess_32.h
---- linux-2.6.37.4/arch/x86/include/asm/uaccess_32.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/include/asm/uaccess_32.h 2011-01-17 02:41:01.000000000 -0500
-@@ -44,6 +44,9 @@ unsigned long __must_check __copy_from_u
- static __always_inline unsigned long __must_check
- __copy_to_user_inatomic(void __user *to, const void *from, unsigned long n)
- {
-+ if ((long)n < 0)
-+ return n;
-+
- if (__builtin_constant_p(n)) {
- unsigned long ret;
-
-@@ -62,6 +65,8 @@ __copy_to_user_inatomic(void __user *to,
- return ret;
- }
- }
-+ if (!__builtin_constant_p(n))
-+ check_object_size(from, n, true);
- return __copy_to_user_ll(to, from, n);
- }
-
-@@ -89,6 +94,9 @@ __copy_to_user(void __user *to, const vo
- static __always_inline unsigned long
- __copy_from_user_inatomic(void *to, const void __user *from, unsigned long n)
- {
-+ if ((long)n < 0)
-+ return n;
-+
- /* Avoid zeroing the tail if the copy fails..
- * If 'n' is constant and 1, 2, or 4, we do still zero on a failure,
- * but as the zeroing behaviour is only significant when n is not
-@@ -138,6 +146,10 @@ static __always_inline unsigned long
- __copy_from_user(void *to, const void __user *from, unsigned long n)
- {
- might_fault();
-+
-+ if ((long)n < 0)
-+ return n;
-+
- if (__builtin_constant_p(n)) {
- unsigned long ret;
-
-@@ -153,6 +165,8 @@ __copy_from_user(void *to, const void __
- return ret;
- }
- }
-+ if (!__builtin_constant_p(n))
-+ check_object_size(to, n, false);
- return __copy_from_user_ll(to, from, n);
- }
-
-@@ -160,6 +174,10 @@ static __always_inline unsigned long __c
- const void __user *from, unsigned long n)
- {
- might_fault();
-+
-+ if ((long)n < 0)
-+ return n;
-+
- if (__builtin_constant_p(n)) {
- unsigned long ret;
-
-@@ -182,15 +200,19 @@ static __always_inline unsigned long
- __copy_from_user_inatomic_nocache(void *to, const void __user *from,
- unsigned long n)
- {
-- return __copy_from_user_ll_nocache_nozero(to, from, n);
--}
-+ if ((long)n < 0)
-+ return n;
-
--unsigned long __must_check copy_to_user(void __user *to,
-- const void *from, unsigned long n);
--unsigned long __must_check _copy_from_user(void *to,
-- const void __user *from,
-- unsigned long n);
-+ return __copy_from_user_ll_nocache_nozero(to, from, n);
-+}
-
-+extern void copy_to_user_overflow(void)
-+#ifdef CONFIG_DEBUG_STRICT_USER_COPY_CHECKS
-+ __compiletime_error("copy_to_user() buffer size is not provably correct")
-+#else
-+ __compiletime_warning("copy_to_user() buffer size is not provably correct")
-+#endif
-+;
-
- extern void copy_from_user_overflow(void)
- #ifdef CONFIG_DEBUG_STRICT_USER_COPY_CHECKS
-@@ -200,17 +222,61 @@ extern void copy_from_user_overflow(void
- #endif
- ;
-
--static inline unsigned long __must_check copy_from_user(void *to,
-- const void __user *from,
-- unsigned long n)
-+/**
-+ * copy_to_user: - Copy a block of data into user space.
-+ * @to: Destination address, in user space.
-+ * @from: Source address, in kernel space.
-+ * @n: Number of bytes to copy.
-+ *
-+ * Context: User context only. This function may sleep.
-+ *
-+ * Copy data from kernel space to user space.
-+ *
-+ * Returns number of bytes that could not be copied.
-+ * On success, this will be zero.
-+ */
-+static inline unsigned long __must_check
-+copy_to_user(void __user *to, const void *from, unsigned long n)
-+{
-+ int sz = __compiletime_object_size(from);
-+
-+ if (unlikely(sz != -1 && sz < n))
-+ copy_to_user_overflow();
-+ else if (access_ok(VERIFY_WRITE, to, n))
-+ n = __copy_to_user(to, from, n);
-+ return n;
-+}
-+
-+/**
-+ * copy_from_user: - Copy a block of data from user space.
-+ * @to: Destination address, in kernel space.
-+ * @from: Source address, in user space.
-+ * @n: Number of bytes to copy.
-+ *
-+ * Context: User context only. This function may sleep.
-+ *
-+ * Copy data from user space to kernel space.
-+ *
-+ * Returns number of bytes that could not be copied.
-+ * On success, this will be zero.
-+ *
-+ * If some data could not be copied, this function will pad the copied
-+ * data to the requested size using zero bytes.
-+ */
-+static inline unsigned long __must_check
-+copy_from_user(void *to, const void __user *from, unsigned long n)
- {
- int sz = __compiletime_object_size(to);
-
-- if (likely(sz == -1 || sz >= n))
-- n = _copy_from_user(to, from, n);
-- else
-+ if (unlikely(sz != -1 && sz < n))
- copy_from_user_overflow();
--
-+ else if (access_ok(VERIFY_READ, from, n))
-+ n = __copy_from_user(to, from, n);
-+ else if ((long)n > 0) {
-+ if (!__builtin_constant_p(n))
-+ check_object_size(to, n, false);
-+ memset(to, 0, n);
-+ }
- return n;
- }
-
-diff -urNp linux-2.6.37.4/arch/x86/include/asm/uaccess_64.h linux-2.6.37.4/arch/x86/include/asm/uaccess_64.h
---- linux-2.6.37.4/arch/x86/include/asm/uaccess_64.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/include/asm/uaccess_64.h 2011-01-17 02:41:01.000000000 -0500
-@@ -11,6 +11,9 @@
- #include <asm/alternative.h>
- #include <asm/cpufeature.h>
- #include <asm/page.h>
-+#include <asm/pgtable.h>
-+
-+#define set_fs(x) (current_thread_info()->addr_limit = (x))
-
- /*
- * Copy To/From Userspace
-@@ -37,26 +40,26 @@ copy_user_generic(void *to, const void *
- return ret;
- }
-
--__must_check unsigned long
--_copy_to_user(void __user *to, const void *from, unsigned len);
--__must_check unsigned long
--_copy_from_user(void *to, const void __user *from, unsigned len);
-+static __always_inline __must_check unsigned long
-+__copy_to_user(void __user *to, const void *from, unsigned len);
-+static __always_inline __must_check unsigned long
-+__copy_from_user(void *to, const void __user *from, unsigned len);
- __must_check unsigned long
- copy_in_user(void __user *to, const void __user *from, unsigned len);
-
- static inline unsigned long __must_check copy_from_user(void *to,
- const void __user *from,
-- unsigned long n)
-+ unsigned n)
- {
-- int sz = __compiletime_object_size(to);
--
- might_fault();
-- if (likely(sz == -1 || sz >= n))
-- n = _copy_from_user(to, from, n);
--#ifdef CONFIG_DEBUG_VM
-- else
-- WARN(1, "Buffer overflow detected!\n");
--#endif
-+
-+ if (access_ok(VERIFY_READ, from, n))
-+ n = __copy_from_user(to, from, n);
-+ else if ((int)n > 0) {
-+ if (!__builtin_constant_p(n))
-+ check_object_size(to, n, false);
-+ memset(to, 0, n);
-+ }
- return n;
- }
-
-@@ -65,110 +68,174 @@ int copy_to_user(void __user *dst, const
- {
- might_fault();
-
-- return _copy_to_user(dst, src, size);
-+ if (access_ok(VERIFY_WRITE, dst, size))
-+ size = __copy_to_user(dst, src, size);
-+ return size;
- }
-
- static __always_inline __must_check
--int __copy_from_user(void *dst, const void __user *src, unsigned size)
-+unsigned long __copy_from_user(void *dst, const void __user *src, unsigned size)
- {
-- int ret = 0;
-+ int sz = __compiletime_object_size(dst);
-+ unsigned ret = 0;
-
- might_fault();
-- if (!__builtin_constant_p(size))
-- return copy_user_generic(dst, (__force void *)src, size);
-+
-+ if ((int)size < 0)
-+ return size;
-+
-+#ifdef CONFIG_PAX_MEMORY_UDEREF
-+ if (!__access_ok(VERIFY_READ, src, size))
-+ return size;
-+#endif
-+
-+ if (unlikely(sz != -1 && sz < size)) {
-+#ifdef CONFIG_DEBUG_VM
-+ WARN(1, "Buffer overflow detected!\n");
-+#endif
-+ return size;
-+ }
-+
-+ if (!__builtin_constant_p(size)) {
-+ check_object_size(dst, size, false);
-+ if ((unsigned long)src < PAX_USER_SHADOW_BASE)
-+ src += PAX_USER_SHADOW_BASE;
-+ return copy_user_generic(dst, (__force const void *)src, size);
-+ }
- switch (size) {
-- case 1:__get_user_asm(*(u8 *)dst, (u8 __user *)src,
-+ case 1:__get_user_asm(*(u8 *)dst, (const u8 __user *)src,
- ret, "b", "b", "=q", 1);
- return ret;
-- case 2:__get_user_asm(*(u16 *)dst, (u16 __user *)src,
-+ case 2:__get_user_asm(*(u16 *)dst, (const u16 __user *)src,
- ret, "w", "w", "=r", 2);
- return ret;
-- case 4:__get_user_asm(*(u32 *)dst, (u32 __user *)src,
-+ case 4:__get_user_asm(*(u32 *)dst, (const u32 __user *)src,
- ret, "l", "k", "=r", 4);
- return ret;
-- case 8:__get_user_asm(*(u64 *)dst, (u64 __user *)src,
-+ case 8:__get_user_asm(*(u64 *)dst, (const u64 __user *)src,
- ret, "q", "", "=r", 8);
- return ret;
- case 10:
-- __get_user_asm(*(u64 *)dst, (u64 __user *)src,
-+ __get_user_asm(*(u64 *)dst, (const u64 __user *)src,
- ret, "q", "", "=r", 10);
- if (unlikely(ret))
- return ret;
- __get_user_asm(*(u16 *)(8 + (char *)dst),
-- (u16 __user *)(8 + (char __user *)src),
-+ (const u16 __user *)(8 + (const char __user *)src),
- ret, "w", "w", "=r", 2);
- return ret;
- case 16:
-- __get_user_asm(*(u64 *)dst, (u64 __user *)src,
-+ __get_user_asm(*(u64 *)dst, (const u64 __user *)src,
- ret, "q", "", "=r", 16);
- if (unlikely(ret))
- return ret;
- __get_user_asm(*(u64 *)(8 + (char *)dst),
-- (u64 __user *)(8 + (char __user *)src),
-+ (const u64 __user *)(8 + (const char __user *)src),
- ret, "q", "", "=r", 8);
- return ret;
- default:
-- return copy_user_generic(dst, (__force void *)src, size);
-+ if ((unsigned long)src < PAX_USER_SHADOW_BASE)
-+ src += PAX_USER_SHADOW_BASE;
-+ return copy_user_generic(dst, (__force const void *)src, size);
- }
- }
-
- static __always_inline __must_check
--int __copy_to_user(void __user *dst, const void *src, unsigned size)
-+unsigned long __copy_to_user(void __user *dst, const void *src, unsigned size)
- {
-- int ret = 0;
-+ int sz = __compiletime_object_size(src);
-+ unsigned ret = 0;
-
- might_fault();
-- if (!__builtin_constant_p(size))
-+
-+ if ((int)size < 0)
-+ return size;
-+
-+#ifdef CONFIG_PAX_MEMORY_UDEREF
-+ if (!__access_ok(VERIFY_WRITE, dst, size))
-+ return size;
-+#endif
-+
-+ if (unlikely(sz != -1 && sz < size)) {
-+#ifdef CONFIG_DEBUG_VM
-+ WARN(1, "Buffer overflow detected!\n");
-+#endif
-+ return size;
-+ }
-+
-+ if (!__builtin_constant_p(size)) {
-+ check_object_size(src, size, true);
-+ if ((unsigned long)dst < PAX_USER_SHADOW_BASE)
-+ dst += PAX_USER_SHADOW_BASE;
- return copy_user_generic((__force void *)dst, src, size);
-+ }
- switch (size) {
-- case 1:__put_user_asm(*(u8 *)src, (u8 __user *)dst,
-+ case 1:__put_user_asm(*(const u8 *)src, (u8 __user *)dst,
- ret, "b", "b", "iq", 1);
- return ret;
-- case 2:__put_user_asm(*(u16 *)src, (u16 __user *)dst,
-+ case 2:__put_user_asm(*(const u16 *)src, (u16 __user *)dst,
- ret, "w", "w", "ir", 2);
- return ret;
-- case 4:__put_user_asm(*(u32 *)src, (u32 __user *)dst,
-+ case 4:__put_user_asm(*(const u32 *)src, (u32 __user *)dst,
- ret, "l", "k", "ir", 4);
- return ret;
-- case 8:__put_user_asm(*(u64 *)src, (u64 __user *)dst,
-+ case 8:__put_user_asm(*(const u64 *)src, (u64 __user *)dst,
- ret, "q", "", "er", 8);
- return ret;
- case 10:
-- __put_user_asm(*(u64 *)src, (u64 __user *)dst,
-+ __put_user_asm(*(const u64 *)src, (u64 __user *)dst,
- ret, "q", "", "er", 10);
- if (unlikely(ret))
- return ret;
- asm("":::"memory");
-- __put_user_asm(4[(u16 *)src], 4 + (u16 __user *)dst,
-+ __put_user_asm(4[(const u16 *)src], 4 + (u16 __user *)dst,
- ret, "w", "w", "ir", 2);
- return ret;
- case 16:
-- __put_user_asm(*(u64 *)src, (u64 __user *)dst,
-+ __put_user_asm(*(const u64 *)src, (u64 __user *)dst,
- ret, "q", "", "er", 16);
- if (unlikely(ret))
- return ret;
- asm("":::"memory");
-- __put_user_asm(1[(u64 *)src], 1 + (u64 __user *)dst,
-+ __put_user_asm(1[(const u64 *)src], 1 + (u64 __user *)dst,
- ret, "q", "", "er", 8);
- return ret;
- default:
-+ if ((unsigned long)dst < PAX_USER_SHADOW_BASE)
-+ dst += PAX_USER_SHADOW_BASE;
- return copy_user_generic((__force void *)dst, src, size);
- }
- }
-
- static __always_inline __must_check
--int __copy_in_user(void __user *dst, const void __user *src, unsigned size)
-+unsigned long __copy_in_user(void __user *dst, const void __user *src, unsigned size)
- {
-- int ret = 0;
-+ unsigned ret = 0;
-
- might_fault();
-- if (!__builtin_constant_p(size))
-+
-+ if ((int)size < 0)
-+ return size;
-+
-+#ifdef CONFIG_PAX_MEMORY_UDEREF
-+ if (!__access_ok(VERIFY_READ, src, size))
-+ return size;
-+ if (!__access_ok(VERIFY_WRITE, dst, size))
-+ return size;
-+#endif
-+
-+ if (!__builtin_constant_p(size)) {
-+ if ((unsigned long)src < PAX_USER_SHADOW_BASE)
-+ src += PAX_USER_SHADOW_BASE;
-+ if ((unsigned long)dst < PAX_USER_SHADOW_BASE)
-+ dst += PAX_USER_SHADOW_BASE;
- return copy_user_generic((__force void *)dst,
-- (__force void *)src, size);
-+ (__force const void *)src, size);
-+ }
- switch (size) {
- case 1: {
- u8 tmp;
-- __get_user_asm(tmp, (u8 __user *)src,
-+ __get_user_asm(tmp, (const u8 __user *)src,
- ret, "b", "b", "=q", 1);
- if (likely(!ret))
- __put_user_asm(tmp, (u8 __user *)dst,
-@@ -177,7 +244,7 @@ int __copy_in_user(void __user *dst, con
- }
- case 2: {
- u16 tmp;
-- __get_user_asm(tmp, (u16 __user *)src,
-+ __get_user_asm(tmp, (const u16 __user *)src,
- ret, "w", "w", "=r", 2);
- if (likely(!ret))
- __put_user_asm(tmp, (u16 __user *)dst,
-@@ -187,7 +254,7 @@ int __copy_in_user(void __user *dst, con
-
- case 4: {
- u32 tmp;
-- __get_user_asm(tmp, (u32 __user *)src,
-+ __get_user_asm(tmp, (const u32 __user *)src,
- ret, "l", "k", "=r", 4);
- if (likely(!ret))
- __put_user_asm(tmp, (u32 __user *)dst,
-@@ -196,7 +263,7 @@ int __copy_in_user(void __user *dst, con
- }
- case 8: {
- u64 tmp;
-- __get_user_asm(tmp, (u64 __user *)src,
-+ __get_user_asm(tmp, (const u64 __user *)src,
- ret, "q", "", "=r", 8);
- if (likely(!ret))
- __put_user_asm(tmp, (u64 __user *)dst,
-@@ -204,8 +271,12 @@ int __copy_in_user(void __user *dst, con
- return ret;
- }
- default:
-+ if ((unsigned long)src < PAX_USER_SHADOW_BASE)
-+ src += PAX_USER_SHADOW_BASE;
-+ if ((unsigned long)dst < PAX_USER_SHADOW_BASE)
-+ dst += PAX_USER_SHADOW_BASE;
- return copy_user_generic((__force void *)dst,
-- (__force void *)src, size);
-+ (__force const void *)src, size);
- }
- }
-
-@@ -222,33 +293,68 @@ __must_check unsigned long __clear_user(
- static __must_check __always_inline int
- __copy_from_user_inatomic(void *dst, const void __user *src, unsigned size)
- {
-+ if ((int)size < 0)
-+ return size;
-+
-+#ifdef CONFIG_PAX_MEMORY_UDEREF
-+ if (!__access_ok(VERIFY_READ, src, size))
-+ return size;
-+#endif
-+
-+ if ((unsigned long)src < PAX_USER_SHADOW_BASE)
-+ src += PAX_USER_SHADOW_BASE;
- return copy_user_generic(dst, (__force const void *)src, size);
- }
-
--static __must_check __always_inline int
-+static __must_check __always_inline unsigned long
- __copy_to_user_inatomic(void __user *dst, const void *src, unsigned size)
- {
-+ if ((int)size < 0)
-+ return size;
-+
-+#ifdef CONFIG_PAX_MEMORY_UDEREF
-+ if (!__access_ok(VERIFY_WRITE, dst, size))
-+ return size;
-+#endif
-+
-+ if ((unsigned long)dst < PAX_USER_SHADOW_BASE)
-+ dst += PAX_USER_SHADOW_BASE;
- return copy_user_generic((__force void *)dst, src, size);
- }
-
--extern long __copy_user_nocache(void *dst, const void __user *src,
-+extern unsigned long __copy_user_nocache(void *dst, const void __user *src,
- unsigned size, int zerorest);
-
--static inline int
--__copy_from_user_nocache(void *dst, const void __user *src, unsigned size)
-+static inline unsigned long __copy_from_user_nocache(void *dst, const void __user *src, unsigned size)
- {
- might_sleep();
-+
-+ if ((int)size < 0)
-+ return size;
-+
-+#ifdef CONFIG_PAX_MEMORY_UDEREF
-+ if (!__access_ok(VERIFY_READ, src, size))
-+ return size;
-+#endif
-+
- return __copy_user_nocache(dst, src, size, 1);
- }
-
--static inline int
--__copy_from_user_inatomic_nocache(void *dst, const void __user *src,
-+static inline unsigned long __copy_from_user_inatomic_nocache(void *dst, const void __user *src,
- unsigned size)
- {
-+ if ((int)size < 0)
-+ return size;
-+
-+#ifdef CONFIG_PAX_MEMORY_UDEREF
-+ if (!__access_ok(VERIFY_READ, src, size))
-+ return size;
-+#endif
-+
- return __copy_user_nocache(dst, src, size, 0);
- }
-
--unsigned long
-+extern unsigned long
- copy_user_handle_tail(char *to, char *from, unsigned len, unsigned zerorest);
-
- #endif /* _ASM_X86_UACCESS_64_H */
-diff -urNp linux-2.6.37.4/arch/x86/include/asm/uaccess.h linux-2.6.37.4/arch/x86/include/asm/uaccess.h
---- linux-2.6.37.4/arch/x86/include/asm/uaccess.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/include/asm/uaccess.h 2011-03-12 22:50:25.000000000 -0500
-@@ -8,12 +8,15 @@
- #include <linux/thread_info.h>
- #include <linux/prefetch.h>
- #include <linux/string.h>
-+#include <linux/sched.h>
- #include <asm/asm.h>
- #include <asm/page.h>
-
- #define VERIFY_READ 0
- #define VERIFY_WRITE 1
-
-+extern void check_object_size(const void *ptr, unsigned long n, bool to);
-+
- /*
- * The fs value determines whether argument validity checking should be
- * performed or not. If get_fs() == USER_DS, checking is performed, with
-@@ -29,7 +32,12 @@
-
- #define get_ds() (KERNEL_DS)
- #define get_fs() (current_thread_info()->addr_limit)
-+#if defined(CONFIG_X86_32) && defined(CONFIG_PAX_MEMORY_UDEREF)
-+void __set_fs(mm_segment_t x);
-+void set_fs(mm_segment_t x);
-+#else
- #define set_fs(x) (current_thread_info()->addr_limit = (x))
-+#endif
-
- #define segment_eq(a, b) ((a).seg == (b).seg)
-
-@@ -77,7 +85,33 @@
- * checks that the pointer is in the user space range - after calling
- * this function, memory access functions may still return -EFAULT.
- */
--#define access_ok(type, addr, size) (likely(__range_not_ok(addr, size) == 0))
-+#define __access_ok(type, addr, size) (likely(__range_not_ok(addr, size) == 0))
-+#define access_ok(type, addr, size) \
-+({ \
-+ long __size = size; \
-+ unsigned long __addr = (unsigned long)addr; \
-+ unsigned long __addr_ao = __addr & PAGE_MASK; \
-+ unsigned long __end_ao = __addr + __size - 1; \
-+ bool __ret_ao = __range_not_ok(__addr, __size) == 0; \
-+ if (__ret_ao && unlikely((__end_ao ^ __addr_ao) & PAGE_MASK)) { \
-+ while(__addr_ao <= __end_ao) { \
-+ char __c_ao; \
-+ __addr_ao += PAGE_SIZE; \
-+ if (__size > PAGE_SIZE) \
-+ cond_resched(); \
-+ if (__get_user(__c_ao, (char __user *)__addr)) \
-+ break; \
-+ if (type != VERIFY_WRITE) { \
-+ __addr = __addr_ao; \
-+ continue; \
-+ } \
-+ if (__put_user(__c_ao, (char __user *)__addr)) \
-+ break; \
-+ __addr = __addr_ao; \
-+ } \
-+ } \
-+ __ret_ao; \
-+})
-
- /*
- * The exception table consists of pairs of addresses: the first is the
-@@ -183,12 +217,20 @@ extern int __get_user_bad(void);
- asm volatile("call __put_user_" #size : "=a" (__ret_pu) \
- : "0" ((typeof(*(ptr)))(x)), "c" (ptr) : "ebx")
-
--
-+#if defined(CONFIG_X86_32) && defined(CONFIG_PAX_MEMORY_UDEREF)
-+#define __copyuser_seg "gs;"
-+#define __COPYUSER_SET_ES "pushl %%gs; popl %%es\n"
-+#define __COPYUSER_RESTORE_ES "pushl %%ss; popl %%es\n"
-+#else
-+#define __copyuser_seg
-+#define __COPYUSER_SET_ES
-+#define __COPYUSER_RESTORE_ES
-+#endif
-
- #ifdef CONFIG_X86_32
- #define __put_user_asm_u64(x, addr, err, errret) \
-- asm volatile("1: movl %%eax,0(%2)\n" \
-- "2: movl %%edx,4(%2)\n" \
-+ asm volatile("1: "__copyuser_seg"movl %%eax,0(%2)\n" \
-+ "2: "__copyuser_seg"movl %%edx,4(%2)\n" \
- "3:\n" \
- ".section .fixup,\"ax\"\n" \
- "4: movl %3,%0\n" \
-@@ -200,8 +242,8 @@ extern int __get_user_bad(void);
- : "A" (x), "r" (addr), "i" (errret), "0" (err))
-
- #define __put_user_asm_ex_u64(x, addr) \
-- asm volatile("1: movl %%eax,0(%1)\n" \
-- "2: movl %%edx,4(%1)\n" \
-+ asm volatile("1: "__copyuser_seg"movl %%eax,0(%1)\n" \
-+ "2: "__copyuser_seg"movl %%edx,4(%1)\n" \
- "3:\n" \
- _ASM_EXTABLE(1b, 2b - 1b) \
- _ASM_EXTABLE(2b, 3b - 2b) \
-@@ -374,7 +416,7 @@ do { \
- } while (0)
-
- #define __get_user_asm(x, addr, err, itype, rtype, ltype, errret) \
-- asm volatile("1: mov"itype" %2,%"rtype"1\n" \
-+ asm volatile("1: "__copyuser_seg"mov"itype" %2,%"rtype"1\n"\
- "2:\n" \
- ".section .fixup,\"ax\"\n" \
- "3: mov %3,%0\n" \
-@@ -382,7 +424,7 @@ do { \
- " jmp 2b\n" \
- ".previous\n" \
- _ASM_EXTABLE(1b, 3b) \
-- : "=r" (err), ltype(x) \
-+ : "=r" (err), ltype (x) \
- : "m" (__m(addr)), "i" (errret), "0" (err))
-
- #define __get_user_size_ex(x, ptr, size) \
-@@ -407,7 +449,7 @@ do { \
- } while (0)
-
- #define __get_user_asm_ex(x, addr, itype, rtype, ltype) \
-- asm volatile("1: mov"itype" %1,%"rtype"0\n" \
-+ asm volatile("1: "__copyuser_seg"mov"itype" %1,%"rtype"0\n"\
- "2:\n" \
- _ASM_EXTABLE(1b, 2b - 1b) \
- : ltype(x) : "m" (__m(addr)))
-@@ -424,13 +466,24 @@ do { \
- int __gu_err; \
- unsigned long __gu_val; \
- __get_user_size(__gu_val, (ptr), (size), __gu_err, -EFAULT); \
-- (x) = (__force __typeof__(*(ptr)))__gu_val; \
-+ (x) = (__typeof__(*(ptr)))__gu_val; \
- __gu_err; \
- })
-
- /* FIXME: this hack is definitely wrong -AK */
- struct __large_struct { unsigned long buf[100]; };
--#define __m(x) (*(struct __large_struct __user *)(x))
-+#if defined(CONFIG_X86_64) && defined(CONFIG_PAX_MEMORY_UDEREF)
-+#define ____m(x) \
-+({ \
-+ unsigned long ____x = (unsigned long)(x); \
-+ if (____x < PAX_USER_SHADOW_BASE) \
-+ ____x += PAX_USER_SHADOW_BASE; \
-+ (void __user *)____x; \
-+})
-+#else
-+#define ____m(x) (x)
-+#endif
-+#define __m(x) (*(struct __large_struct __user *)____m(x))
-
- /*
- * Tell gcc we read from memory instead of writing: this is because
-@@ -438,7 +491,7 @@ struct __large_struct { unsigned long bu
- * aliasing issues.
- */
- #define __put_user_asm(x, addr, err, itype, rtype, ltype, errret) \
-- asm volatile("1: mov"itype" %"rtype"1,%2\n" \
-+ asm volatile("1: "__copyuser_seg"mov"itype" %"rtype"1,%2\n"\
- "2:\n" \
- ".section .fixup,\"ax\"\n" \
- "3: mov %3,%0\n" \
-@@ -446,10 +499,10 @@ struct __large_struct { unsigned long bu
- ".previous\n" \
- _ASM_EXTABLE(1b, 3b) \
- : "=r"(err) \
-- : ltype(x), "m" (__m(addr)), "i" (errret), "0" (err))
-+ : ltype (x), "m" (__m(addr)), "i" (errret), "0" (err))
-
- #define __put_user_asm_ex(x, addr, itype, rtype, ltype) \
-- asm volatile("1: mov"itype" %"rtype"0,%1\n" \
-+ asm volatile("1: "__copyuser_seg"mov"itype" %"rtype"0,%1\n"\
- "2:\n" \
- _ASM_EXTABLE(1b, 2b - 1b) \
- : : ltype(x), "m" (__m(addr)))
-@@ -488,8 +541,12 @@ struct __large_struct { unsigned long bu
- * On error, the variable @x is set to zero.
- */
-
-+#if defined(CONFIG_X86_64) && defined(CONFIG_PAX_MEMORY_UDEREF)
-+#define __get_user(x, ptr) get_user((x), (ptr))
-+#else
- #define __get_user(x, ptr) \
- __get_user_nocheck((x), (ptr), sizeof(*(ptr)))
-+#endif
-
- /**
- * __put_user: - Write a simple value into user space, with less checking.
-@@ -511,8 +568,12 @@ struct __large_struct { unsigned long bu
- * Returns zero on success, or -EFAULT on error.
- */
-
-+#if defined(CONFIG_X86_64) && defined(CONFIG_PAX_MEMORY_UDEREF)
-+#define __put_user(x, ptr) put_user((x), (ptr))
-+#else
- #define __put_user(x, ptr) \
- __put_user_nocheck((__typeof__(*(ptr)))(x), (ptr), sizeof(*(ptr)))
-+#endif
-
- #define __get_user_unaligned __get_user
- #define __put_user_unaligned __put_user
-@@ -530,7 +591,7 @@ struct __large_struct { unsigned long bu
- #define get_user_ex(x, ptr) do { \
- unsigned long __gue_val; \
- __get_user_size_ex((__gue_val), (ptr), (sizeof(*(ptr)))); \
-- (x) = (__force __typeof__(*(ptr)))__gue_val; \
-+ (x) = (__typeof__(*(ptr)))__gue_val; \
- } while (0)
-
- #ifdef CONFIG_X86_WP_WORKS_OK
-@@ -567,6 +628,7 @@ extern struct movsl_mask {
-
- #define ARCH_HAS_NOCACHE_UACCESS 1
-
-+#define ARCH_HAS_SORT_EXTABLE
- #ifdef CONFIG_X86_32
- # include "uaccess_32.h"
- #else
-diff -urNp linux-2.6.37.4/arch/x86/include/asm/vgtod.h linux-2.6.37.4/arch/x86/include/asm/vgtod.h
---- linux-2.6.37.4/arch/x86/include/asm/vgtod.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/include/asm/vgtod.h 2011-01-17 02:41:01.000000000 -0500
-@@ -14,6 +14,7 @@ struct vsyscall_gtod_data {
- int sysctl_enabled;
- struct timezone sys_tz;
- struct { /* extract of a clocksource struct */
-+ char name[8];
- cycle_t (*vread)(void);
- cycle_t cycle_last;
- cycle_t mask;
-diff -urNp linux-2.6.37.4/arch/x86/include/asm/vsyscall.h linux-2.6.37.4/arch/x86/include/asm/vsyscall.h
---- linux-2.6.37.4/arch/x86/include/asm/vsyscall.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/include/asm/vsyscall.h 2011-01-17 02:41:01.000000000 -0500
-@@ -15,9 +15,10 @@ enum vsyscall_num {
-
- #ifdef __KERNEL__
- #include <linux/seqlock.h>
-+#include <linux/getcpu.h>
-+#include <linux/time.h>
-
- #define __section_vgetcpu_mode __attribute__ ((unused, __section__ (".vgetcpu_mode"), aligned(16)))
--#define __section_jiffies __attribute__ ((unused, __section__ (".jiffies"), aligned(16)))
-
- /* Definitions for CONFIG_GENERIC_TIME definitions */
- #define __section_vsyscall_gtod_data __attribute__ \
-@@ -31,7 +32,6 @@ enum vsyscall_num {
- #define VGETCPU_LSL 2
-
- extern int __vgetcpu_mode;
--extern volatile unsigned long __jiffies;
-
- /* kernel space (writeable) */
- extern int vgetcpu_mode;
-@@ -39,6 +39,9 @@ extern struct timezone sys_tz;
-
- extern void map_vsyscall(void);
-
-+extern int vgettimeofday(struct timeval * tv, struct timezone * tz);
-+extern time_t vtime(time_t *t);
-+extern long vgetcpu(unsigned *cpu, unsigned *node, struct getcpu_cache *tcache);
- #endif /* __KERNEL__ */
-
- #endif /* _ASM_X86_VSYSCALL_H */
-diff -urNp linux-2.6.37.4/arch/x86/include/asm/xsave.h linux-2.6.37.4/arch/x86/include/asm/xsave.h
---- linux-2.6.37.4/arch/x86/include/asm/xsave.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/include/asm/xsave.h 2011-01-17 02:41:01.000000000 -0500
-@@ -65,6 +65,11 @@ static inline int xsave_user(struct xsav
- {
- int err;
-
-+#if defined(CONFIG_X86_64) && defined(CONFIG_PAX_MEMORY_UDEREF)
-+ if ((unsigned long)buf < PAX_USER_SHADOW_BASE)
-+ buf = (struct xsave_struct __user *)((void __user*)buf + PAX_USER_SHADOW_BASE);
-+#endif
-+
- /*
- * Clear the xsave header first, so that reserved fields are
- * initialized to zero.
-@@ -100,6 +105,11 @@ static inline int xrestore_user(struct x
- u32 lmask = mask;
- u32 hmask = mask >> 32;
-
-+#if defined(CONFIG_X86_64) && defined(CONFIG_PAX_MEMORY_UDEREF)
-+ if ((unsigned long)xstate < PAX_USER_SHADOW_BASE)
-+ xstate = (struct xsave_struct *)((void *)xstate + PAX_USER_SHADOW_BASE);
-+#endif
-+
- __asm__ __volatile__("1: .byte " REX_PREFIX "0x0f,0xae,0x2f\n"
- "2:\n"
- ".section .fixup,\"ax\"\n"
-diff -urNp linux-2.6.37.4/arch/x86/Kconfig linux-2.6.37.4/arch/x86/Kconfig
---- linux-2.6.37.4/arch/x86/Kconfig 2011-02-22 16:05:30.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/Kconfig 2011-02-22 16:05:42.000000000 -0500
-@@ -225,7 +225,7 @@ config X86_TRAMPOLINE
-
- config X86_32_LAZY_GS
- def_bool y
-- depends on X86_32 && !CC_STACKPROTECTOR
-+ depends on X86_32 && !CC_STACKPROTECTOR && !PAX_MEMORY_UDEREF
-
- config ARCH_HWEIGHT_CFLAGS
- string
-@@ -992,7 +992,7 @@ choice
-
- config NOHIGHMEM
- bool "off"
-- depends on !X86_NUMAQ
-+ depends on !X86_NUMAQ && !(PAX_PAGEEXEC && PAX_ENABLE_PAE)
- ---help---
- Linux can use up to 64 Gigabytes of physical memory on x86 systems.
- However, the address space of 32-bit x86 processors is only 4
-@@ -1029,7 +1029,7 @@ config NOHIGHMEM
-
- config HIGHMEM4G
- bool "4GB"
-- depends on !X86_NUMAQ
-+ depends on !X86_NUMAQ && !(PAX_PAGEEXEC && PAX_ENABLE_PAE)
- ---help---
- Select this if you have a 32-bit processor and between 1 and 4
- gigabytes of physical RAM.
-@@ -1083,7 +1083,7 @@ config PAGE_OFFSET
- hex
- default 0xB0000000 if VMSPLIT_3G_OPT
- default 0x80000000 if VMSPLIT_2G
-- default 0x78000000 if VMSPLIT_2G_OPT
-+ default 0x70000000 if VMSPLIT_2G_OPT
- default 0x40000000 if VMSPLIT_1G
- default 0xC0000000
- depends on X86_32
-@@ -1427,7 +1427,7 @@ config ARCH_USES_PG_UNCACHED
-
- config EFI
- bool "EFI runtime service support"
-- depends on ACPI
-+ depends on ACPI && !PAX_KERNEXEC
- ---help---
- This enables the kernel to use EFI runtime services that are
- available (such as the EFI variable services).
-@@ -1457,6 +1457,7 @@ config SECCOMP
-
- config CC_STACKPROTECTOR
- bool "Enable -fstack-protector buffer overflow detection (EXPERIMENTAL)"
-+ depends on X86_64 || !PAX_MEMORY_UDEREF
- ---help---
- This option turns on the -fstack-protector GCC feature. This
- feature puts, at the beginning of functions, a canary value on
-@@ -1514,6 +1515,7 @@ config KEXEC_JUMP
- config PHYSICAL_START
- hex "Physical address where the kernel is loaded" if (EMBEDDED || CRASH_DUMP)
- default "0x1000000"
-+ range 0x400000 0x40000000
- ---help---
- This gives the physical address where the kernel is loaded.
-
-@@ -1577,6 +1579,7 @@ config X86_NEED_RELOCS
- config PHYSICAL_ALIGN
- hex "Alignment value to which kernel should be aligned" if X86_32
- default "0x1000000"
-+ range 0x400000 0x1000000 if PAX_KERNEXEC
- range 0x2000 0x1000000
- ---help---
- This value puts the alignment restrictions on physical address
-@@ -1608,9 +1611,10 @@ config HOTPLUG_CPU
- Say N if you want to disable CPU hotplug.
-
- config COMPAT_VDSO
-- def_bool y
-+ def_bool n
- prompt "Compat VDSO support"
- depends on X86_32 || IA32_EMULATION
-+ depends on !PAX_NOEXEC && !PAX_MEMORY_UDEREF
- ---help---
- Map the 32-bit VDSO to the predictable old-style address too.
-
-diff -urNp linux-2.6.37.4/arch/x86/Kconfig.cpu linux-2.6.37.4/arch/x86/Kconfig.cpu
---- linux-2.6.37.4/arch/x86/Kconfig.cpu 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/Kconfig.cpu 2011-01-17 02:41:01.000000000 -0500
-@@ -336,7 +336,7 @@ config X86_PPRO_FENCE
-
- config X86_F00F_BUG
- def_bool y
-- depends on M586MMX || M586TSC || M586 || M486 || M386
-+ depends on (M586MMX || M586TSC || M586 || M486 || M386) && !PAX_KERNEXEC
-
- config X86_INVD_BUG
- def_bool y
-@@ -360,7 +360,7 @@ config X86_POPAD_OK
-
- config X86_ALIGNMENT_16
- def_bool y
-- depends on MWINCHIP3D || MWINCHIPC6 || MCYRIXIII || X86_ELAN || MK6 || M586MMX || M586TSC || M586 || M486 || MVIAC3_2 || MGEODEGX1
-+ depends on MWINCHIP3D || MWINCHIPC6 || MCYRIXIII || X86_ELAN || MK8 || MK7 || MK6 || MCORE2 || MPENTIUM4 || MPENTIUMIII || MPENTIUMII || M686 || M586MMX || M586TSC || M586 || M486 || MVIAC3_2 || MGEODEGX1
-
- config X86_INTEL_USERCOPY
- def_bool y
-@@ -406,7 +406,7 @@ config X86_CMPXCHG64
- # generates cmov.
- config X86_CMOV
- def_bool y
-- depends on (MK8 || MK7 || MCORE2 || MPENTIUM4 || MPENTIUMM || MPENTIUMIII || MPENTIUMII || M686 || MVIAC3_2 || MVIAC7 || MCRUSOE || MEFFICEON || X86_64 || MATOM || MGEODE_LX)
-+ depends on (MK8 || MK7 || MCORE2 || MPSC || MPENTIUM4 || MPENTIUMM || MPENTIUMIII || MPENTIUMII || M686 || MVIAC3_2 || MVIAC7 || MCRUSOE || MEFFICEON || X86_64 || MATOM || MGEODE_LX)
-
- config X86_MINIMUM_CPU_FAMILY
- int
-diff -urNp linux-2.6.37.4/arch/x86/Kconfig.debug linux-2.6.37.4/arch/x86/Kconfig.debug
---- linux-2.6.37.4/arch/x86/Kconfig.debug 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/Kconfig.debug 2011-01-17 02:41:01.000000000 -0500
-@@ -101,7 +101,7 @@ config X86_PTDUMP
- config DEBUG_RODATA
- bool "Write protect kernel read-only data structures"
- default y
-- depends on DEBUG_KERNEL
-+ depends on DEBUG_KERNEL && BROKEN
- ---help---
- Mark the kernel read-only data as write-protected in the pagetables,
- in order to catch accidental (and incorrect) writes to such const
-diff -urNp linux-2.6.37.4/arch/x86/kernel/acpi/boot.c linux-2.6.37.4/arch/x86/kernel/acpi/boot.c
---- linux-2.6.37.4/arch/x86/kernel/acpi/boot.c 2011-03-11 17:22:42.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/acpi/boot.c 2011-03-11 17:22:59.000000000 -0500
-@@ -1504,7 +1504,7 @@ static struct dmi_system_id __initdata a
- DMI_MATCH(DMI_PRODUCT_NAME, "HP Compaq 6715b"),
- },
- },
-- {}
-+ { NULL, NULL, {{0, {0}}}, NULL}
- };
-
- /*
-diff -urNp linux-2.6.37.4/arch/x86/kernel/acpi/sleep.c linux-2.6.37.4/arch/x86/kernel/acpi/sleep.c
---- linux-2.6.37.4/arch/x86/kernel/acpi/sleep.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/acpi/sleep.c 2011-01-17 02:41:01.000000000 -0500
-@@ -12,6 +12,7 @@
- #include <linux/cpumask.h>
- #include <asm/segment.h>
- #include <asm/desc.h>
-+#include <asm/e820.h>
-
- #ifdef CONFIG_X86_32
- #include <asm/pgtable.h>
-@@ -20,7 +21,7 @@
- #include "realmode/wakeup.h"
- #include "sleep.h"
-
--unsigned long acpi_wakeup_address;
-+unsigned long acpi_wakeup_address = 0x2000;
- unsigned long acpi_realmode_flags;
-
- /* address in low memory of the wakeup routine. */
-@@ -101,8 +102,12 @@ int acpi_save_state_mem(void)
- header->trampoline_segment = setup_trampoline() >> 4;
- #ifdef CONFIG_SMP
- stack_start.sp = temp_stack + sizeof(temp_stack);
-+
-+ pax_open_kernel();
- early_gdt_descr.address =
- (unsigned long)get_cpu_gdt_table(smp_processor_id());
-+ pax_close_kernel();
-+
- initial_gs = per_cpu_offset(smp_processor_id());
- #endif
- initial_code = (unsigned long)wakeup_long64;
-diff -urNp linux-2.6.37.4/arch/x86/kernel/acpi/wakeup_32.S linux-2.6.37.4/arch/x86/kernel/acpi/wakeup_32.S
---- linux-2.6.37.4/arch/x86/kernel/acpi/wakeup_32.S 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/acpi/wakeup_32.S 2011-01-17 02:41:01.000000000 -0500
-@@ -30,13 +30,11 @@ wakeup_pmode_return:
- # and restore the stack ... but you need gdt for this to work
- movl saved_context_esp, %esp
-
-- movl %cs:saved_magic, %eax
-- cmpl $0x12345678, %eax
-+ cmpl $0x12345678, saved_magic
- jne bogus_magic
-
- # jump to place where we left off
-- movl saved_eip, %eax
-- jmp *%eax
-+ jmp *(saved_eip)
-
- bogus_magic:
- jmp bogus_magic
-diff -urNp linux-2.6.37.4/arch/x86/kernel/alternative.c linux-2.6.37.4/arch/x86/kernel/alternative.c
---- linux-2.6.37.4/arch/x86/kernel/alternative.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/alternative.c 2011-03-21 17:35:08.000000000 -0400
-@@ -248,7 +248,7 @@ static void alternatives_smp_lock(const
- if (!*poff || ptr < text || ptr >= text_end)
- continue;
- /* turn DS segment override prefix into lock prefix */
-- if (*ptr == 0x3e)
-+ if (*ktla_ktva(ptr) == 0x3e)
- text_poke(ptr, ((unsigned char []){0xf0}), 1);
- };
- mutex_unlock(&text_mutex);
-@@ -269,7 +269,7 @@ static void alternatives_smp_unlock(cons
- if (!*poff || ptr < text || ptr >= text_end)
- continue;
- /* turn lock prefix into DS segment override prefix */
-- if (*ptr == 0xf0)
-+ if (*ktla_ktva(ptr) == 0xf0)
- text_poke(ptr, ((unsigned char []){0x3E}), 1);
- };
- mutex_unlock(&text_mutex);
-@@ -437,7 +437,7 @@ void __init_or_module apply_paravirt(str
-
- BUG_ON(p->len > MAX_PATCH_LEN);
- /* prep the buffer with the original instructions */
-- memcpy(insnbuf, p->instr, p->len);
-+ memcpy(insnbuf, ktla_ktva(p->instr), p->len);
- used = pv_init_ops.patch(p->instrtype, p->clobbers, insnbuf,
- (unsigned long)p->instr, p->len);
-
-@@ -505,7 +505,7 @@ void __init alternative_instructions(voi
- if (smp_alt_once)
- free_init_pages("SMP alternatives",
- (unsigned long)__smp_locks,
-- (unsigned long)__smp_locks_end);
-+ PAGE_ALIGN((unsigned long)__smp_locks_end));
-
- restart_nmi();
- }
-@@ -522,13 +522,17 @@ void __init alternative_instructions(voi
- * instructions. And on the local CPU you need to be protected again NMI or MCE
- * handlers seeing an inconsistent instruction while you patch.
- */
--void *__init_or_module text_poke_early(void *addr, const void *opcode,
-+void *__kprobes text_poke_early(void *addr, const void *opcode,
- size_t len)
- {
- unsigned long flags;
- local_irq_save(flags);
-- memcpy(addr, opcode, len);
-+
-+ pax_open_kernel();
-+ memcpy(ktla_ktva(addr), opcode, len);
- sync_core();
-+ pax_close_kernel();
-+
- local_irq_restore(flags);
- /* Could also do a CLFLUSH here to speed up CPU recovery; but
- that causes hangs on some VIA CPUs. */
-@@ -550,36 +554,22 @@ void *__init_or_module text_poke_early(v
- */
- void *__kprobes text_poke(void *addr, const void *opcode, size_t len)
- {
-- unsigned long flags;
-- char *vaddr;
-+ unsigned char *vaddr = ktla_ktva(addr);
- struct page *pages[2];
-- int i;
-+ size_t i;
-
- if (!core_kernel_text((unsigned long)addr)) {
-- pages[0] = vmalloc_to_page(addr);
-- pages[1] = vmalloc_to_page(addr + PAGE_SIZE);
-+ pages[0] = vmalloc_to_page(vaddr);
-+ pages[1] = vmalloc_to_page(vaddr + PAGE_SIZE);
- } else {
-- pages[0] = virt_to_page(addr);
-+ pages[0] = virt_to_page(vaddr);
- WARN_ON(!PageReserved(pages[0]));
-- pages[1] = virt_to_page(addr + PAGE_SIZE);
-+ pages[1] = virt_to_page(vaddr + PAGE_SIZE);
- }
- BUG_ON(!pages[0]);
-- local_irq_save(flags);
-- set_fixmap(FIX_TEXT_POKE0, page_to_phys(pages[0]));
-- if (pages[1])
-- set_fixmap(FIX_TEXT_POKE1, page_to_phys(pages[1]));
-- vaddr = (char *)fix_to_virt(FIX_TEXT_POKE0);
-- memcpy(&vaddr[(unsigned long)addr & ~PAGE_MASK], opcode, len);
-- clear_fixmap(FIX_TEXT_POKE0);
-- if (pages[1])
-- clear_fixmap(FIX_TEXT_POKE1);
-- local_flush_tlb();
-- sync_core();
-- /* Could also do a CLFLUSH here to speed up CPU recovery; but
-- that causes hangs on some VIA CPUs. */
-+ text_poke_early(addr, opcode, len);
- for (i = 0; i < len; i++)
-- BUG_ON(((char *)addr)[i] != ((char *)opcode)[i]);
-- local_irq_restore(flags);
-+ BUG_ON((vaddr)[i] != ((const unsigned char *)opcode)[i]);
- return addr;
- }
-
-@@ -612,6 +602,13 @@ static int __kprobes stop_machine_text_p
-
- flush_icache_range((unsigned long)tpp->addr,
- (unsigned long)tpp->addr + tpp->len);
-+
-+ /*
-+ * Intel Archiecture Software Developer's Manual section 7.1.3 specifies
-+ * that a core serializing instruction such as "cpuid" should be
-+ * executed on _each_ core before the new instruction is made visible.
-+ */
-+ sync_core();
- return 0;
- }
-
-@@ -645,9 +642,9 @@ void *__kprobes text_poke_smp(void *addr
- #if defined(CONFIG_DYNAMIC_FTRACE) || defined(HAVE_JUMP_LABEL)
-
- #ifdef CONFIG_X86_64
--unsigned char ideal_nop5[5] = { 0x66, 0x66, 0x66, 0x66, 0x90 };
-+unsigned char ideal_nop5[5] __read_only = { 0x66, 0x66, 0x66, 0x66, 0x90 };
- #else
--unsigned char ideal_nop5[5] = { 0x3e, 0x8d, 0x74, 0x26, 0x00 };
-+unsigned char ideal_nop5[5] __read_only = { 0x3e, 0x8d, 0x74, 0x26, 0x00 };
- #endif
-
- void __init arch_init_ideal_nop5(void)
-diff -urNp linux-2.6.37.4/arch/x86/kernel/amd_iommu.c linux-2.6.37.4/arch/x86/kernel/amd_iommu.c
---- linux-2.6.37.4/arch/x86/kernel/amd_iommu.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/amd_iommu.c 2011-01-17 02:41:01.000000000 -0500
-@@ -2286,7 +2286,7 @@ static void prealloc_protection_domains(
- }
- }
-
--static struct dma_map_ops amd_iommu_dma_ops = {
-+static const struct dma_map_ops amd_iommu_dma_ops = {
- .alloc_coherent = alloc_coherent,
- .free_coherent = free_coherent,
- .map_page = map_page,
-diff -urNp linux-2.6.37.4/arch/x86/kernel/apic/io_apic.c linux-2.6.37.4/arch/x86/kernel/apic/io_apic.c
---- linux-2.6.37.4/arch/x86/kernel/apic/io_apic.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/apic/io_apic.c 2011-01-17 02:41:01.000000000 -0500
-@@ -597,7 +597,7 @@ struct IO_APIC_route_entry **alloc_ioapi
- ioapic_entries = kzalloc(sizeof(*ioapic_entries) * nr_ioapics,
- GFP_KERNEL);
- if (!ioapic_entries)
-- return 0;
-+ return NULL;
-
- for (apic = 0; apic < nr_ioapics; apic++) {
- ioapic_entries[apic] =
-@@ -614,7 +614,7 @@ nomem:
- kfree(ioapic_entries[apic]);
- kfree(ioapic_entries);
-
-- return 0;
-+ return NULL;
- }
-
- /*
-@@ -1024,7 +1024,7 @@ int IO_APIC_get_PCI_irq_vector(int bus,
- }
- EXPORT_SYMBOL(IO_APIC_get_PCI_irq_vector);
-
--void lock_vector_lock(void)
-+void lock_vector_lock(void) __acquires(vector_lock)
- {
- /* Used to the online set of cpus does not change
- * during assign_irq_vector.
-@@ -1032,7 +1032,7 @@ void lock_vector_lock(void)
- raw_spin_lock(&vector_lock);
- }
-
--void unlock_vector_lock(void)
-+void unlock_vector_lock(void) __releases(vector_lock)
- {
- raw_spin_unlock(&vector_lock);
- }
-diff -urNp linux-2.6.37.4/arch/x86/kernel/apm_32.c linux-2.6.37.4/arch/x86/kernel/apm_32.c
---- linux-2.6.37.4/arch/x86/kernel/apm_32.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/apm_32.c 2011-01-17 02:41:01.000000000 -0500
-@@ -410,7 +410,7 @@ static DEFINE_MUTEX(apm_mutex);
- * This is for buggy BIOS's that refer to (real mode) segment 0x40
- * even though they are called in protected mode.
- */
--static struct desc_struct bad_bios_desc = GDT_ENTRY_INIT(0x4092,
-+static const struct desc_struct bad_bios_desc = GDT_ENTRY_INIT(0x4093,
- (unsigned long)__va(0x400UL), PAGE_SIZE - 0x400 - 1);
-
- static const char driver_version[] = "1.16ac"; /* no spaces */
-@@ -588,7 +588,10 @@ static long __apm_bios_call(void *_call)
- BUG_ON(cpu != 0);
- gdt = get_cpu_gdt_table(cpu);
- save_desc_40 = gdt[0x40 / 8];
-+
-+ pax_open_kernel();
- gdt[0x40 / 8] = bad_bios_desc;
-+ pax_close_kernel();
-
- apm_irq_save(flags);
- APM_DO_SAVE_SEGS;
-@@ -597,7 +600,11 @@ static long __apm_bios_call(void *_call)
- &call->esi);
- APM_DO_RESTORE_SEGS;
- apm_irq_restore(flags);
-+
-+ pax_open_kernel();
- gdt[0x40 / 8] = save_desc_40;
-+ pax_close_kernel();
-+
- put_cpu();
-
- return call->eax & 0xff;
-@@ -664,7 +671,10 @@ static long __apm_bios_call_simple(void
- BUG_ON(cpu != 0);
- gdt = get_cpu_gdt_table(cpu);
- save_desc_40 = gdt[0x40 / 8];
-+
-+ pax_open_kernel();
- gdt[0x40 / 8] = bad_bios_desc;
-+ pax_close_kernel();
-
- apm_irq_save(flags);
- APM_DO_SAVE_SEGS;
-@@ -672,7 +682,11 @@ static long __apm_bios_call_simple(void
- &call->eax);
- APM_DO_RESTORE_SEGS;
- apm_irq_restore(flags);
-+
-+ pax_open_kernel();
- gdt[0x40 / 8] = save_desc_40;
-+ pax_close_kernel();
-+
- put_cpu();
- return error;
- }
-@@ -975,7 +989,7 @@ recalc:
-
- static void apm_power_off(void)
- {
-- unsigned char po_bios_call[] = {
-+ const unsigned char po_bios_call[] = {
- 0xb8, 0x00, 0x10, /* movw $0x1000,ax */
- 0x8e, 0xd0, /* movw ax,ss */
- 0xbc, 0x00, 0xf0, /* movw $0xf000,sp */
-@@ -1932,7 +1946,10 @@ static const struct file_operations apm_
- static struct miscdevice apm_device = {
- APM_MINOR_DEV,
- "apm_bios",
-- &apm_bios_fops
-+ &apm_bios_fops,
-+ {NULL, NULL},
-+ NULL,
-+ NULL
- };
-
-
-@@ -2253,7 +2270,7 @@ static struct dmi_system_id __initdata a
- { DMI_MATCH(DMI_SYS_VENDOR, "IBM"), },
- },
-
-- { }
-+ { NULL, NULL, {DMI_MATCH(DMI_NONE, {0})}, NULL}
- };
-
- /*
-@@ -2356,12 +2373,15 @@ static int __init apm_init(void)
- * code to that CPU.
- */
- gdt = get_cpu_gdt_table(0);
-+
-+ pax_open_kernel();
- set_desc_base(&gdt[APM_CS >> 3],
- (unsigned long)__va((unsigned long)apm_info.bios.cseg << 4));
- set_desc_base(&gdt[APM_CS_16 >> 3],
- (unsigned long)__va((unsigned long)apm_info.bios.cseg_16 << 4));
- set_desc_base(&gdt[APM_DS >> 3],
- (unsigned long)__va((unsigned long)apm_info.bios.dseg << 4));
-+ pax_close_kernel();
-
- proc_create("apm", 0, NULL, &apm_file_ops);
-
-diff -urNp linux-2.6.37.4/arch/x86/kernel/asm-offsets_32.c linux-2.6.37.4/arch/x86/kernel/asm-offsets_32.c
---- linux-2.6.37.4/arch/x86/kernel/asm-offsets_32.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/asm-offsets_32.c 2011-01-17 02:41:01.000000000 -0500
-@@ -113,6 +113,11 @@ void foo(void)
- OFFSET(PV_CPU_iret, pv_cpu_ops, iret);
- OFFSET(PV_CPU_irq_enable_sysexit, pv_cpu_ops, irq_enable_sysexit);
- OFFSET(PV_CPU_read_cr0, pv_cpu_ops, read_cr0);
-+
-+#ifdef CONFIG_PAX_KERNEXEC
-+ OFFSET(PV_CPU_write_cr0, pv_cpu_ops, write_cr0);
-+#endif
-+
- #endif
-
- #ifdef CONFIG_XEN
-diff -urNp linux-2.6.37.4/arch/x86/kernel/asm-offsets_64.c linux-2.6.37.4/arch/x86/kernel/asm-offsets_64.c
---- linux-2.6.37.4/arch/x86/kernel/asm-offsets_64.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/asm-offsets_64.c 2011-01-17 02:41:01.000000000 -0500
-@@ -63,6 +63,18 @@ int main(void)
- OFFSET(PV_CPU_irq_enable_sysexit, pv_cpu_ops, irq_enable_sysexit);
- OFFSET(PV_CPU_swapgs, pv_cpu_ops, swapgs);
- OFFSET(PV_MMU_read_cr2, pv_mmu_ops, read_cr2);
-+
-+#ifdef CONFIG_PAX_KERNEXEC
-+ OFFSET(PV_CPU_read_cr0, pv_cpu_ops, read_cr0);
-+ OFFSET(PV_CPU_write_cr0, pv_cpu_ops, write_cr0);
-+#endif
-+
-+#ifdef CONFIG_PAX_MEMORY_UDEREF
-+ OFFSET(PV_MMU_read_cr3, pv_mmu_ops, read_cr3);
-+ OFFSET(PV_MMU_write_cr3, pv_mmu_ops, write_cr3);
-+ OFFSET(PV_MMU_set_pgd, pv_mmu_ops, set_pgd);
-+#endif
-+
- #endif
-
-
-@@ -115,6 +127,7 @@ int main(void)
- ENTRY(cr8);
- BLANK();
- #undef ENTRY
-+ DEFINE(TSS_size, sizeof(struct tss_struct));
- DEFINE(TSS_ist, offsetof(struct tss_struct, x86_tss.ist));
- BLANK();
- DEFINE(crypto_tfm_ctx_offset, offsetof(struct crypto_tfm, __crt_ctx));
-diff -urNp linux-2.6.37.4/arch/x86/kernel/cpu/common.c linux-2.6.37.4/arch/x86/kernel/cpu/common.c
---- linux-2.6.37.4/arch/x86/kernel/cpu/common.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/cpu/common.c 2011-01-17 02:41:01.000000000 -0500
-@@ -83,60 +83,6 @@ static const struct cpu_dev __cpuinitcon
-
- static const struct cpu_dev *this_cpu __cpuinitdata = &default_cpu;
-
--DEFINE_PER_CPU_PAGE_ALIGNED(struct gdt_page, gdt_page) = { .gdt = {
--#ifdef CONFIG_X86_64
-- /*
-- * We need valid kernel segments for data and code in long mode too
-- * IRET will check the segment types kkeil 2000/10/28
-- * Also sysret mandates a special GDT layout
-- *
-- * TLS descriptors are currently at a different place compared to i386.
-- * Hopefully nobody expects them at a fixed place (Wine?)
-- */
-- [GDT_ENTRY_KERNEL32_CS] = GDT_ENTRY_INIT(0xc09b, 0, 0xfffff),
-- [GDT_ENTRY_KERNEL_CS] = GDT_ENTRY_INIT(0xa09b, 0, 0xfffff),
-- [GDT_ENTRY_KERNEL_DS] = GDT_ENTRY_INIT(0xc093, 0, 0xfffff),
-- [GDT_ENTRY_DEFAULT_USER32_CS] = GDT_ENTRY_INIT(0xc0fb, 0, 0xfffff),
-- [GDT_ENTRY_DEFAULT_USER_DS] = GDT_ENTRY_INIT(0xc0f3, 0, 0xfffff),
-- [GDT_ENTRY_DEFAULT_USER_CS] = GDT_ENTRY_INIT(0xa0fb, 0, 0xfffff),
--#else
-- [GDT_ENTRY_KERNEL_CS] = GDT_ENTRY_INIT(0xc09a, 0, 0xfffff),
-- [GDT_ENTRY_KERNEL_DS] = GDT_ENTRY_INIT(0xc092, 0, 0xfffff),
-- [GDT_ENTRY_DEFAULT_USER_CS] = GDT_ENTRY_INIT(0xc0fa, 0, 0xfffff),
-- [GDT_ENTRY_DEFAULT_USER_DS] = GDT_ENTRY_INIT(0xc0f2, 0, 0xfffff),
-- /*
-- * Segments used for calling PnP BIOS have byte granularity.
-- * They code segments and data segments have fixed 64k limits,
-- * the transfer segment sizes are set at run time.
-- */
-- /* 32-bit code */
-- [GDT_ENTRY_PNPBIOS_CS32] = GDT_ENTRY_INIT(0x409a, 0, 0xffff),
-- /* 16-bit code */
-- [GDT_ENTRY_PNPBIOS_CS16] = GDT_ENTRY_INIT(0x009a, 0, 0xffff),
-- /* 16-bit data */
-- [GDT_ENTRY_PNPBIOS_DS] = GDT_ENTRY_INIT(0x0092, 0, 0xffff),
-- /* 16-bit data */
-- [GDT_ENTRY_PNPBIOS_TS1] = GDT_ENTRY_INIT(0x0092, 0, 0),
-- /* 16-bit data */
-- [GDT_ENTRY_PNPBIOS_TS2] = GDT_ENTRY_INIT(0x0092, 0, 0),
-- /*
-- * The APM segments have byte granularity and their bases
-- * are set at run time. All have 64k limits.
-- */
-- /* 32-bit code */
-- [GDT_ENTRY_APMBIOS_BASE] = GDT_ENTRY_INIT(0x409a, 0, 0xffff),
-- /* 16-bit code */
-- [GDT_ENTRY_APMBIOS_BASE+1] = GDT_ENTRY_INIT(0x009a, 0, 0xffff),
-- /* data */
-- [GDT_ENTRY_APMBIOS_BASE+2] = GDT_ENTRY_INIT(0x4092, 0, 0xffff),
--
-- [GDT_ENTRY_ESPFIX_SS] = GDT_ENTRY_INIT(0xc092, 0, 0xfffff),
-- [GDT_ENTRY_PERCPU] = GDT_ENTRY_INIT(0xc092, 0, 0xfffff),
-- GDT_STACK_CANARY_INIT
--#endif
--} };
--EXPORT_PER_CPU_SYMBOL_GPL(gdt_page);
--
- static int __init x86_xsave_setup(char *s)
- {
- setup_clear_cpu_cap(X86_FEATURE_XSAVE);
-@@ -352,7 +298,7 @@ void switch_to_new_gdt(int cpu)
- {
- struct desc_ptr gdt_descr;
-
-- gdt_descr.address = (long)get_cpu_gdt_table(cpu);
-+ gdt_descr.address = (unsigned long)get_cpu_gdt_table(cpu);
- gdt_descr.size = GDT_SIZE - 1;
- load_gdt(&gdt_descr);
- /* Reload the per-cpu base */
-@@ -825,6 +771,10 @@ static void __cpuinit identify_cpu(struc
- /* Filter out anything that depends on CPUID levels we don't have */
- filter_cpuid_features(c, true);
-
-+#if defined(CONFIG_PAX_SEGMEXEC) || defined(CONFIG_PAX_KERNEXEC) || (defined(CONFIG_PAX_MEMORY_UDEREF) && defined(CONFIG_X86_32))
-+ setup_clear_cpu_cap(X86_FEATURE_SEP);
-+#endif
-+
- /* If the model name is still unset, do table lookup. */
- if (!c->x86_model_id[0]) {
- const char *p;
-@@ -1085,7 +1035,7 @@ struct pt_regs * __cpuinit idle_regs(str
- {
- memset(regs, 0, sizeof(struct pt_regs));
- regs->fs = __KERNEL_PERCPU;
-- regs->gs = __KERNEL_STACK_CANARY;
-+ savesegment(gs, regs->gs);
-
- return regs;
- }
-@@ -1140,7 +1090,7 @@ void __cpuinit cpu_init(void)
- int i;
-
- cpu = stack_smp_processor_id();
-- t = &per_cpu(init_tss, cpu);
-+ t = init_tss + cpu;
- oist = &per_cpu(orig_ist, cpu);
-
- #ifdef CONFIG_NUMA
-@@ -1166,7 +1116,7 @@ void __cpuinit cpu_init(void)
- switch_to_new_gdt(cpu);
- loadsegment(fs, 0);
-
-- load_idt((const struct desc_ptr *)&idt_descr);
-+ load_idt(&idt_descr);
-
- memset(me->thread.tls_array, 0, GDT_ENTRY_TLS_ENTRIES * 8);
- syscall_init();
-@@ -1175,7 +1125,6 @@ void __cpuinit cpu_init(void)
- wrmsrl(MSR_KERNEL_GS_BASE, 0);
- barrier();
-
-- x86_configure_nx();
- if (cpu != 0)
- enable_x2apic();
-
-@@ -1229,7 +1178,7 @@ void __cpuinit cpu_init(void)
- {
- int cpu = smp_processor_id();
- struct task_struct *curr = current;
-- struct tss_struct *t = &per_cpu(init_tss, cpu);
-+ struct tss_struct *t = init_tss + cpu;
- struct thread_struct *thread = &curr->thread;
-
- if (cpumask_test_and_set_cpu(cpu, cpu_initialized_mask)) {
-diff -urNp linux-2.6.37.4/arch/x86/kernel/cpu/cpufreq/acpi-cpufreq.c linux-2.6.37.4/arch/x86/kernel/cpu/cpufreq/acpi-cpufreq.c
---- linux-2.6.37.4/arch/x86/kernel/cpu/cpufreq/acpi-cpufreq.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/cpu/cpufreq/acpi-cpufreq.c 2011-01-17 02:41:01.000000000 -0500
-@@ -481,7 +481,7 @@ static const struct dmi_system_id sw_any
- DMI_MATCH(DMI_PRODUCT_NAME, "X6DLP"),
- },
- },
-- { }
-+ { NULL, NULL, {DMI_MATCH(DMI_NONE, {0})}, NULL }
- };
-
- static int acpi_cpufreq_blacklist(struct cpuinfo_x86 *c)
-diff -urNp linux-2.6.37.4/arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c linux-2.6.37.4/arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c
---- linux-2.6.37.4/arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c 2011-01-17 02:41:01.000000000 -0500
-@@ -226,7 +226,7 @@ static struct cpu_model models[] =
- { &cpu_ids[CPU_MP4HT_D0], NULL, 0, NULL },
- { &cpu_ids[CPU_MP4HT_E0], NULL, 0, NULL },
-
-- { NULL, }
-+ { NULL, NULL, 0, NULL}
- };
- #undef _BANIAS
- #undef BANIAS
-diff -urNp linux-2.6.37.4/arch/x86/kernel/cpu/intel.c linux-2.6.37.4/arch/x86/kernel/cpu/intel.c
---- linux-2.6.37.4/arch/x86/kernel/cpu/intel.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/cpu/intel.c 2011-01-17 02:41:01.000000000 -0500
-@@ -161,7 +161,7 @@ static void __cpuinit trap_init_f00f_bug
- * Update the IDT descriptor and reload the IDT so that
- * it uses the read-only mapped virtual address.
- */
-- idt_descr.address = fix_to_virt(FIX_F00F_IDT);
-+ idt_descr.address = (struct desc_struct *)fix_to_virt(FIX_F00F_IDT);
- load_idt(&idt_descr);
- }
- #endif
-diff -urNp linux-2.6.37.4/arch/x86/kernel/cpu/Makefile linux-2.6.37.4/arch/x86/kernel/cpu/Makefile
---- linux-2.6.37.4/arch/x86/kernel/cpu/Makefile 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/cpu/Makefile 2011-01-17 02:41:01.000000000 -0500
-@@ -8,10 +8,6 @@ CFLAGS_REMOVE_common.o = -pg
- CFLAGS_REMOVE_perf_event.o = -pg
- endif
-
--# Make sure load_percpu_segment has no stackprotector
--nostackp := $(call cc-option, -fno-stack-protector)
--CFLAGS_common.o := $(nostackp)
--
- obj-y := intel_cacheinfo.o scattered.o topology.o
- obj-y += proc.o capflags.o powerflags.o common.o
- obj-y += vmware.o hypervisor.o sched.o mshyperv.o
-diff -urNp linux-2.6.37.4/arch/x86/kernel/cpu/mcheck/mce.c linux-2.6.37.4/arch/x86/kernel/cpu/mcheck/mce.c
---- linux-2.6.37.4/arch/x86/kernel/cpu/mcheck/mce.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/cpu/mcheck/mce.c 2011-01-25 20:24:56.000000000 -0500
-@@ -45,6 +45,7 @@
- #include <asm/ipi.h>
- #include <asm/mce.h>
- #include <asm/msr.h>
-+#include <asm/local.h>
-
- #include "mce-internal.h"
-
-@@ -219,7 +220,7 @@ static void print_mce(struct mce *m)
- !(m->mcgstatus & MCG_STATUS_EIPV) ? " !INEXACT!" : "",
- m->cs, m->ip);
-
-- if (m->cs == __KERNEL_CS)
-+ if (m->cs == __KERNEL_CS || m->cs == __KERNEXEC_KERNEL_CS)
- print_symbol("{%s}", m->ip);
- pr_cont("\n");
- }
-@@ -1460,14 +1461,14 @@ void __cpuinit mcheck_cpu_init(struct cp
- */
-
- static DEFINE_SPINLOCK(mce_state_lock);
--static int open_count; /* #times opened */
-+static local_t open_count; /* #times opened */
- static int open_exclu; /* already open exclusive? */
-
- static int mce_open(struct inode *inode, struct file *file)
- {
- spin_lock(&mce_state_lock);
-
-- if (open_exclu || (open_count && (file->f_flags & O_EXCL))) {
-+ if (open_exclu || (local_read(&open_count) && (file->f_flags & O_EXCL))) {
- spin_unlock(&mce_state_lock);
-
- return -EBUSY;
-@@ -1475,7 +1476,7 @@ static int mce_open(struct inode *inode,
-
- if (file->f_flags & O_EXCL)
- open_exclu = 1;
-- open_count++;
-+ local_inc(&open_count);
-
- spin_unlock(&mce_state_lock);
-
-@@ -1486,7 +1487,7 @@ static int mce_release(struct inode *ino
- {
- spin_lock(&mce_state_lock);
-
-- open_count--;
-+ local_dec(&open_count);
- open_exclu = 0;
-
- spin_unlock(&mce_state_lock);
-@@ -1673,6 +1674,7 @@ static struct miscdevice mce_log_device
- MISC_MCELOG_MINOR,
- "mcelog",
- &mce_chrdev_ops,
-+ {NULL, NULL}, NULL, NULL
- };
-
- /*
-diff -urNp linux-2.6.37.4/arch/x86/kernel/cpu/mtrr/generic.c linux-2.6.37.4/arch/x86/kernel/cpu/mtrr/generic.c
---- linux-2.6.37.4/arch/x86/kernel/cpu/mtrr/generic.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/cpu/mtrr/generic.c 2011-01-17 02:41:01.000000000 -0500
-@@ -28,7 +28,7 @@ static struct fixed_range_block fixed_ra
- { MSR_MTRRfix64K_00000, 1 }, /* one 64k MTRR */
- { MSR_MTRRfix16K_80000, 2 }, /* two 16k MTRRs */
- { MSR_MTRRfix4K_C0000, 8 }, /* eight 4k MTRRs */
-- {}
-+ { 0, 0 }
- };
-
- static unsigned long smp_changes_mask;
-diff -urNp linux-2.6.37.4/arch/x86/kernel/cpu/mtrr/main.c linux-2.6.37.4/arch/x86/kernel/cpu/mtrr/main.c
---- linux-2.6.37.4/arch/x86/kernel/cpu/mtrr/main.c 2011-02-22 16:05:30.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/cpu/mtrr/main.c 2011-02-22 16:05:42.000000000 -0500
-@@ -61,7 +61,7 @@ static DEFINE_MUTEX(mtrr_mutex);
- u64 size_or_mask, size_and_mask;
- static bool mtrr_aps_delayed_init;
-
--static const struct mtrr_ops *mtrr_ops[X86_VENDOR_NUM];
-+static const struct mtrr_ops *mtrr_ops[X86_VENDOR_NUM] __read_only;
-
- const struct mtrr_ops *mtrr_if;
-
-diff -urNp linux-2.6.37.4/arch/x86/kernel/cpu/mtrr/mtrr.h linux-2.6.37.4/arch/x86/kernel/cpu/mtrr/mtrr.h
---- linux-2.6.37.4/arch/x86/kernel/cpu/mtrr/mtrr.h 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/cpu/mtrr/mtrr.h 2011-01-17 02:41:01.000000000 -0500
-@@ -12,19 +12,19 @@
- extern unsigned int mtrr_usage_table[MTRR_MAX_VAR_RANGES];
-
- struct mtrr_ops {
-- u32 vendor;
-- u32 use_intel_if;
-- void (*set)(unsigned int reg, unsigned long base,
-+ const u32 vendor;
-+ const u32 use_intel_if;
-+ void (* const set)(unsigned int reg, unsigned long base,
- unsigned long size, mtrr_type type);
-- void (*set_all)(void);
-+ void (* const set_all)(void);
-
-- void (*get)(unsigned int reg, unsigned long *base,
-+ void (* const get)(unsigned int reg, unsigned long *base,
- unsigned long *size, mtrr_type *type);
-- int (*get_free_region)(unsigned long base, unsigned long size,
-+ int (* const get_free_region)(unsigned long base, unsigned long size,
- int replace_reg);
-- int (*validate_add_page)(unsigned long base, unsigned long size,
-+ int (* const validate_add_page)(unsigned long base, unsigned long size,
- unsigned int type);
-- int (*have_wrcomb)(void);
-+ int (* const have_wrcomb)(void);
- };
-
- extern int generic_get_free_region(unsigned long base, unsigned long size,
-diff -urNp linux-2.6.37.4/arch/x86/kernel/cpu/perfctr-watchdog.c linux-2.6.37.4/arch/x86/kernel/cpu/perfctr-watchdog.c
---- linux-2.6.37.4/arch/x86/kernel/cpu/perfctr-watchdog.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/cpu/perfctr-watchdog.c 2011-01-17 02:41:01.000000000 -0500
-@@ -30,11 +30,11 @@ struct nmi_watchdog_ctlblk {
-
- /* Interface defining a CPU specific perfctr watchdog */
- struct wd_ops {
-- int (*reserve)(void);
-- void (*unreserve)(void);
-- int (*setup)(unsigned nmi_hz);
-- void (*rearm)(struct nmi_watchdog_ctlblk *wd, unsigned nmi_hz);
-- void (*stop)(void);
-+ int (* const reserve)(void);
-+ void (* const unreserve)(void);
-+ int (* const setup)(unsigned nmi_hz);
-+ void (* const rearm)(struct nmi_watchdog_ctlblk *wd, unsigned nmi_hz);
-+ void (* const stop)(void);
- unsigned perfctr;
- unsigned evntsel;
- u64 checkbit;
-@@ -634,6 +634,7 @@ static const struct wd_ops p4_wd_ops = {
- #define ARCH_PERFMON_NMI_EVENT_SEL ARCH_PERFMON_UNHALTED_CORE_CYCLES_SEL
- #define ARCH_PERFMON_NMI_EVENT_UMASK ARCH_PERFMON_UNHALTED_CORE_CYCLES_UMASK
-
-+/* cannot be const, see probe_nmi_watchdog */
- static struct wd_ops intel_arch_wd_ops;
-
- static int setup_intel_arch_watchdog(unsigned nmi_hz)
-@@ -686,6 +687,7 @@ static int setup_intel_arch_watchdog(uns
- return 1;
- }
-
-+/* cannot be const */
- static struct wd_ops intel_arch_wd_ops __read_mostly = {
- .reserve = single_msr_reserve,
- .unreserve = single_msr_unreserve,
-diff -urNp linux-2.6.37.4/arch/x86/kernel/cpu/perf_event.c linux-2.6.37.4/arch/x86/kernel/cpu/perf_event.c
---- linux-2.6.37.4/arch/x86/kernel/cpu/perf_event.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/cpu/perf_event.c 2011-01-17 02:41:01.000000000 -0500
-@@ -1757,7 +1757,7 @@ perf_callchain_user(struct perf_callchai
- break;
-
- perf_callchain_store(entry, frame.return_address);
-- fp = frame.next_frame;
-+ fp = (__force const void __user *)frame.next_frame;
- }
- }
-
-diff -urNp linux-2.6.37.4/arch/x86/kernel/crash.c linux-2.6.37.4/arch/x86/kernel/crash.c
---- linux-2.6.37.4/arch/x86/kernel/crash.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/crash.c 2011-01-17 02:41:01.000000000 -0500
-@@ -42,7 +42,7 @@ static void kdump_nmi_callback(int cpu,
- regs = args->regs;
-
- #ifdef CONFIG_X86_32
-- if (!user_mode_vm(regs)) {
-+ if (!user_mode(regs)) {
- crash_fixup_ss_esp(&fixed_regs, regs);
- regs = &fixed_regs;
- }
-diff -urNp linux-2.6.37.4/arch/x86/kernel/doublefault_32.c linux-2.6.37.4/arch/x86/kernel/doublefault_32.c
---- linux-2.6.37.4/arch/x86/kernel/doublefault_32.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/doublefault_32.c 2011-01-17 02:41:01.000000000 -0500
-@@ -11,7 +11,7 @@
-
- #define DOUBLEFAULT_STACKSIZE (1024)
- static unsigned long doublefault_stack[DOUBLEFAULT_STACKSIZE];
--#define STACK_START (unsigned long)(doublefault_stack+DOUBLEFAULT_STACKSIZE)
-+#define STACK_START (unsigned long)(doublefault_stack+DOUBLEFAULT_STACKSIZE-2)
-
- #define ptr_ok(x) ((x) > PAGE_OFFSET && (x) < PAGE_OFFSET + MAXMEM)
-
-@@ -21,7 +21,7 @@ static void doublefault_fn(void)
- unsigned long gdt, tss;
-
- store_gdt(&gdt_desc);
-- gdt = gdt_desc.address;
-+ gdt = (unsigned long)gdt_desc.address;
-
- printk(KERN_EMERG "PANIC: double fault, gdt at %08lx [%d bytes]\n", gdt, gdt_desc.size);
-
-@@ -58,10 +58,10 @@ struct tss_struct doublefault_tss __cach
- /* 0x2 bit is always set */
- .flags = X86_EFLAGS_SF | 0x2,
- .sp = STACK_START,
-- .es = __USER_DS,
-+ .es = __KERNEL_DS,
- .cs = __KERNEL_CS,
- .ss = __KERNEL_DS,
-- .ds = __USER_DS,
-+ .ds = __KERNEL_DS,
- .fs = __KERNEL_PERCPU,
-
- .__cr3 = __pa_nodebug(swapper_pg_dir),
-diff -urNp linux-2.6.37.4/arch/x86/kernel/dumpstack_32.c linux-2.6.37.4/arch/x86/kernel/dumpstack_32.c
---- linux-2.6.37.4/arch/x86/kernel/dumpstack_32.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/dumpstack_32.c 2011-01-17 02:41:01.000000000 -0500
-@@ -105,11 +105,12 @@ void show_registers(struct pt_regs *regs
- * When in-kernel, we also print out the stack and code at the
- * time of the fault..
- */
-- if (!user_mode_vm(regs)) {
-+ if (!user_mode(regs)) {
- unsigned int code_prologue = code_bytes * 43 / 64;
- unsigned int code_len = code_bytes;
- unsigned char c;
- u8 *ip;
-+ unsigned long cs_base = get_desc_base(&get_cpu_gdt_table(smp_processor_id())[(0xffff & regs->cs) >> 3]);
-
- printk(KERN_EMERG "Stack:\n");
- show_stack_log_lvl(NULL, regs, &regs->sp,
-@@ -117,10 +118,10 @@ void show_registers(struct pt_regs *regs
-
- printk(KERN_EMERG "Code: ");
-
-- ip = (u8 *)regs->ip - code_prologue;
-+ ip = (u8 *)regs->ip - code_prologue + cs_base;
- if (ip < (u8 *)PAGE_OFFSET || probe_kernel_address(ip, c)) {
- /* try starting at IP */
-- ip = (u8 *)regs->ip;
-+ ip = (u8 *)regs->ip + cs_base;
- code_len = code_len - code_prologue + 1;
- }
- for (i = 0; i < code_len; i++, ip++) {
-@@ -129,7 +130,7 @@ void show_registers(struct pt_regs *regs
- printk(" Bad EIP value.");
- break;
- }
-- if (ip == (u8 *)regs->ip)
-+ if (ip == (u8 *)regs->ip + cs_base)
- printk("<%02x> ", c);
- else
- printk("%02x ", c);
-@@ -142,6 +143,7 @@ int is_valid_bugaddr(unsigned long ip)
- {
- unsigned short ud2;
-
-+ ip = ktla_ktva(ip);
- if (ip < PAGE_OFFSET)
- return 0;
- if (probe_kernel_address((unsigned short *)ip, ud2))
-diff -urNp linux-2.6.37.4/arch/x86/kernel/dumpstack.c linux-2.6.37.4/arch/x86/kernel/dumpstack.c
---- linux-2.6.37.4/arch/x86/kernel/dumpstack.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/dumpstack.c 2011-01-17 02:41:01.000000000 -0500
-@@ -2,6 +2,9 @@
- * Copyright (C) 1991, 1992 Linus Torvalds
- * Copyright (C) 2000, 2001, 2002 Andi Kleen, SuSE Labs
- */
-+#ifdef CONFIG_GRKERNSEC_HIDESYM
-+#define __INCLUDED_BY_HIDESYM 1
-+#endif
- #include <linux/kallsyms.h>
- #include <linux/kprobes.h>
- #include <linux/uaccess.h>
-@@ -27,7 +30,7 @@ static int die_counter;
-
- void printk_address(unsigned long address, int reliable)
- {
-- printk(" [<%p>] %s%pS\n", (void *) address,
-+ printk(" [<%p>] %s%pA\n", (void *) address,
- reliable ? "" : "? ", (void *) address);
- }
-
-@@ -206,7 +209,7 @@ void dump_stack(void)
- #endif
-
- printk("Pid: %d, comm: %.20s %s %s %.*s\n",
-- current->pid, current->comm, print_tainted(),
-+ task_pid_nr(current), current->comm, print_tainted(),
- init_utsname()->release,
- (int)strcspn(init_utsname()->version, " "),
- init_utsname()->version);
-@@ -262,7 +265,7 @@ void __kprobes oops_end(unsigned long fl
- panic("Fatal exception in interrupt");
- if (panic_on_oops)
- panic("Fatal exception");
-- do_exit(signr);
-+ do_group_exit(signr);
- }
-
- int __kprobes __die(const char *str, struct pt_regs *regs, long err)
-@@ -289,7 +292,7 @@ int __kprobes __die(const char *str, str
-
- show_registers(regs);
- #ifdef CONFIG_X86_32
-- if (user_mode_vm(regs)) {
-+ if (user_mode(regs)) {
- sp = regs->sp;
- ss = regs->ss & 0xffff;
- } else {
-@@ -317,7 +320,7 @@ void die(const char *str, struct pt_regs
- unsigned long flags = oops_begin();
- int sig = SIGSEGV;
-
-- if (!user_mode_vm(regs))
-+ if (!user_mode(regs))
- report_bug(regs->ip, regs);
-
- if (__die(str, regs, err))
-diff -urNp linux-2.6.37.4/arch/x86/kernel/entry_32.S linux-2.6.37.4/arch/x86/kernel/entry_32.S
---- linux-2.6.37.4/arch/x86/kernel/entry_32.S 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/entry_32.S 2011-01-17 02:41:01.000000000 -0500
-@@ -183,13 +183,81 @@
- /*CFI_REL_OFFSET gs, PT_GS*/
- .endm
- .macro SET_KERNEL_GS reg
-+
-+#ifdef CONFIG_CC_STACKPROTECTOR
- movl $(__KERNEL_STACK_CANARY), \reg
-+#elif defined(CONFIG_PAX_MEMORY_UDEREF)
-+ movl $(__USER_DS), \reg
-+#else
-+ xorl \reg, \reg
-+#endif
-+
- movl \reg, %gs
- .endm
-
- #endif /* CONFIG_X86_32_LAZY_GS */
-
--.macro SAVE_ALL
-+.macro PAX_EXIT_KERNEL
-+#ifdef CONFIG_PAX_KERNEXEC
-+#ifdef CONFIG_PARAVIRT
-+ push %eax; push %ecx
-+#endif
-+ mov %cs, %esi
-+ cmp $__KERNEXEC_KERNEL_CS, %esi
-+ jnz 2f
-+#ifdef CONFIG_PARAVIRT
-+ call PARA_INDIRECT(pv_cpu_ops+PV_CPU_read_cr0);
-+ mov %eax, %esi
-+#else
-+ mov %cr0, %esi
-+#endif
-+ btr $16, %esi
-+ ljmp $__KERNEL_CS, $1f
-+1:
-+#ifdef CONFIG_PARAVIRT
-+ mov %esi, %eax
-+ call PARA_INDIRECT(pv_cpu_ops+PV_CPU_write_cr0);
-+#else
-+ mov %esi, %cr0
-+#endif
-+2:
-+#ifdef CONFIG_PARAVIRT
-+ pop %ecx; pop %eax
-+#endif
-+#endif
-+.endm
-+
-+.macro PAX_ENTER_KERNEL
-+#ifdef CONFIG_PAX_KERNEXEC
-+#ifdef CONFIG_PARAVIRT
-+ push %eax; push %ecx
-+ call PARA_INDIRECT(pv_cpu_ops+PV_CPU_read_cr0)
-+ mov %eax, %esi
-+#else
-+ mov %cr0, %esi
-+#endif
-+ bts $16, %esi
-+ jnc 1f
-+ mov %cs, %esi
-+ cmp $__KERNEL_CS, %esi
-+ jz 3f
-+ ljmp $__KERNEL_CS, $3f
-+1: ljmp $__KERNEXEC_KERNEL_CS, $2f
-+2:
-+#ifdef CONFIG_PARAVIRT
-+ mov %esi, %eax
-+ call PARA_INDIRECT(pv_cpu_ops+PV_CPU_write_cr0)
-+#else
-+ mov %esi, %cr0
-+#endif
-+3:
-+#ifdef CONFIG_PARAVIRT
-+ pop %ecx; pop %eax
-+#endif
-+#endif
-+.endm
-+
-+.macro __SAVE_ALL _DS
- cld
- PUSH_GS
- pushl_cfi %fs
-@@ -212,7 +280,7 @@
- CFI_REL_OFFSET ecx, 0
- pushl_cfi %ebx
- CFI_REL_OFFSET ebx, 0
-- movl $(__USER_DS), %edx
-+ movl $\_DS, %edx
- movl %edx, %ds
- movl %edx, %es
- movl $(__KERNEL_PERCPU), %edx
-@@ -220,6 +288,15 @@
- SET_KERNEL_GS %edx
- .endm
-
-+.macro SAVE_ALL
-+#if defined(CONFIG_PAX_KERNEXEC) || defined(CONFIG_PAX_PAGEEXEC) || defined(CONFIG_PAX_SEGMEXEC) || defined(CONFIG_PAX_MEMORY_UDEREF)
-+ __SAVE_ALL __KERNEL_DS
-+ PAX_ENTER_KERNEL
-+#else
-+ __SAVE_ALL __USER_DS
-+#endif
-+.endm
-+
- .macro RESTORE_INT_REGS
- popl_cfi %ebx
- CFI_RESTORE ebx
-@@ -330,7 +407,15 @@ check_userspace:
- movb PT_CS(%esp), %al
- andl $(X86_EFLAGS_VM | SEGMENT_RPL_MASK), %eax
- cmpl $USER_RPL, %eax
-+
-+#ifdef CONFIG_PAX_KERNEXEC
-+ jae resume_userspace
-+
-+ PAX_EXIT_KERNEL
-+ jmp resume_kernel
-+#else
- jb resume_kernel # not returning to v8086 or userspace
-+#endif
-
- ENTRY(resume_userspace)
- LOCKDEP_SYS_EXIT
-@@ -392,10 +477,9 @@ sysenter_past_esp:
- /*CFI_REL_OFFSET cs, 0*/
- /*
- * Push current_thread_info()->sysenter_return to the stack.
-- * A tiny bit of offset fixup is necessary - 4*4 means the 4 words
-- * pushed above; +8 corresponds to copy_thread's esp0 setting.
- */
-- pushl_cfi ((TI_sysenter_return)-THREAD_SIZE_asm+8+4*4)(%esp)
-+ GET_THREAD_INFO(%ebp)
-+ pushl_cfi TI_sysenter_return(%ebp)
- CFI_REL_OFFSET eip, 0
-
- pushl_cfi %eax
-@@ -406,9 +490,19 @@ sysenter_past_esp:
- * Load the potential sixth argument from user stack.
- * Careful about security.
- */
-+ movl PT_OLDESP(%esp),%ebp
-+
-+#ifdef CONFIG_PAX_MEMORY_UDEREF
-+ mov PT_OLDSS(%esp),%ds
-+1: movl %ds:(%ebp),%ebp
-+ push %ss
-+ pop %ds
-+#else
- cmpl $__PAGE_OFFSET-3,%ebp
- jae syscall_fault
- 1: movl (%ebp),%ebp
-+#endif
-+
- movl %ebp,PT_EBP(%esp)
- .section __ex_table,"a"
- .align 4
-@@ -431,12 +525,23 @@ sysenter_do_call:
- testl $_TIF_ALLWORK_MASK, %ecx
- jne sysexit_audit
- sysenter_exit:
-+
-+#ifdef CONFIG_PAX_RANDKSTACK
-+ pushl %eax
-+ CFI_ADJUST_CFA_OFFSET 4
-+ call pax_randomize_kstack
-+ popl %eax
-+ CFI_ADJUST_CFA_OFFSET -4
-+#endif
-+
- /* if something modifies registers it must also disable sysexit */
- movl PT_EIP(%esp), %edx
- movl PT_OLDESP(%esp), %ecx
- xorl %ebp,%ebp
- TRACE_IRQS_ON
- 1: mov PT_FS(%esp), %fs
-+2: mov PT_DS(%esp), %ds
-+3: mov PT_ES(%esp), %es
- PTGS_TO_GS
- ENABLE_INTERRUPTS_SYSEXIT
-
-@@ -479,11 +584,17 @@ sysexit_audit:
-
- CFI_ENDPROC
- .pushsection .fixup,"ax"
--2: movl $0,PT_FS(%esp)
-+4: movl $0,PT_FS(%esp)
-+ jmp 1b
-+5: movl $0,PT_DS(%esp)
-+ jmp 1b
-+6: movl $0,PT_ES(%esp)
- jmp 1b
- .section __ex_table,"a"
- .align 4
-- .long 1b,2b
-+ .long 1b,4b
-+ .long 2b,5b
-+ .long 3b,6b
- .popsection
- PTGS_TO_GS_EX
- ENDPROC(ia32_sysenter_target)
-@@ -516,6 +627,10 @@ syscall_exit:
- testl $_TIF_ALLWORK_MASK, %ecx # current->work
- jne syscall_exit_work
-
-+#ifdef CONFIG_PAX_RANDKSTACK
-+ call pax_randomize_kstack
-+#endif
-+
- restore_all:
- TRACE_IRQS_IRET
- restore_all_notrace:
-@@ -575,14 +690,21 @@ ldt_ss:
- * compensating for the offset by changing to the ESPFIX segment with
- * a base address that matches for the difference.
- */
--#define GDT_ESPFIX_SS PER_CPU_VAR(gdt_page) + (GDT_ENTRY_ESPFIX_SS * 8)
-+#define GDT_ESPFIX_SS (GDT_ENTRY_ESPFIX_SS * 8)(%ebx)
- mov %esp, %edx /* load kernel esp */
- mov PT_OLDESP(%esp), %eax /* load userspace esp */
- mov %dx, %ax /* eax: new kernel esp */
- sub %eax, %edx /* offset (low word is 0) */
-+#ifdef CONFIG_SMP
-+ movl PER_CPU_VAR(cpu_number), %ebx
-+ shll $PAGE_SHIFT_asm, %ebx
-+ addl $cpu_gdt_table, %ebx
-+#else
-+ movl $cpu_gdt_table, %ebx
-+#endif
- shr $16, %edx
-- mov %dl, GDT_ESPFIX_SS + 4 /* bits 16..23 */
-- mov %dh, GDT_ESPFIX_SS + 7 /* bits 24..31 */
-+ mov %dl, 4 + GDT_ESPFIX_SS /* bits 16..23 */
-+ mov %dh, 7 + GDT_ESPFIX_SS /* bits 24..31 */
- pushl_cfi $__ESPFIX_SS
- pushl_cfi %eax /* new kernel esp */
- /* Disable interrupts, but do not irqtrace this section: we
-@@ -617,23 +739,17 @@ work_resched:
-
- work_notifysig: # deal with pending signals and
- # notify-resume requests
-+ movl %esp, %eax
- #ifdef CONFIG_VM86
- testl $X86_EFLAGS_VM, PT_EFLAGS(%esp)
-- movl %esp, %eax
-- jne work_notifysig_v86 # returning to kernel-space or
-+ jz 1f # returning to kernel-space or
- # vm86-space
-- xorl %edx, %edx
-- call do_notify_resume
-- jmp resume_userspace_sig
-
-- ALIGN
--work_notifysig_v86:
- pushl_cfi %ecx # save ti_flags for do_notify_resume
- call save_v86_state # %eax contains pt_regs pointer
- popl_cfi %ecx
- movl %eax, %esp
--#else
-- movl %esp, %eax
-+1:
- #endif
- xorl %edx, %edx
- call do_notify_resume
-@@ -668,6 +784,10 @@ END(syscall_exit_work)
-
- RING0_INT_FRAME # can't unwind into user space anyway
- syscall_fault:
-+#ifdef CONFIG_PAX_MEMORY_UDEREF
-+ push %ss
-+ pop %ds
-+#endif
- GET_THREAD_INFO(%ebp)
- movl $-EFAULT,PT_EAX(%esp)
- jmp resume_userspace
-@@ -750,6 +870,36 @@ ptregs_clone:
- CFI_ENDPROC
- ENDPROC(ptregs_clone)
-
-+ ALIGN;
-+ENTRY(kernel_execve)
-+ CFI_STARTPROC
-+ pushl_cfi %ebp
-+ sub $PT_OLDSS+4,%esp
-+ pushl_cfi %edi
-+ pushl_cfi %ecx
-+ pushl_cfi %eax
-+ lea 3*4(%esp),%edi
-+ mov $PT_OLDSS/4+1,%ecx
-+ xorl %eax,%eax
-+ rep stosl
-+ popl_cfi %eax
-+ popl_cfi %ecx
-+ popl_cfi %edi
-+ movl $X86_EFLAGS_IF,PT_EFLAGS(%esp)
-+ pushl_cfi %esp
-+ call sys_execve
-+ add $4,%esp
-+ CFI_ADJUST_CFA_OFFSET -4
-+ GET_THREAD_INFO(%ebp)
-+ test %eax,%eax
-+ jz syscall_exit
-+ add $PT_OLDSS+4,%esp
-+ CFI_ADJUST_CFA_OFFSET -PT_OLDSS-4
-+ popl_cfi %ebp
-+ ret
-+ CFI_ENDPROC
-+ENDPROC(kernel_execve)
-+
- .macro FIXUP_ESPFIX_STACK
- /*
- * Switch back for ESPFIX stack to the normal zerobased stack
-@@ -759,8 +909,15 @@ ENDPROC(ptregs_clone)
- * normal stack and adjusts ESP with the matching offset.
- */
- /* fixup the stack */
-- mov GDT_ESPFIX_SS + 4, %al /* bits 16..23 */
-- mov GDT_ESPFIX_SS + 7, %ah /* bits 24..31 */
-+#ifdef CONFIG_SMP
-+ movl PER_CPU_VAR(cpu_number), %ebx
-+ shll $PAGE_SHIFT_asm, %ebx
-+ addl $cpu_gdt_table, %ebx
-+#else
-+ movl $cpu_gdt_table, %ebx
-+#endif
-+ mov 4 + GDT_ESPFIX_SS, %al /* bits 16..23 */
-+ mov 7 + GDT_ESPFIX_SS, %ah /* bits 24..31 */
- shl $16, %eax
- addl %esp, %eax /* the adjusted stack pointer */
- pushl_cfi $__KERNEL_DS
-@@ -1211,7 +1368,6 @@ return_to_handler:
- jmp *%ecx
- #endif
-
--.section .rodata,"a"
- #include "syscall_table_32.S"
-
- syscall_table_size=(.-sys_call_table)
-@@ -1257,9 +1413,12 @@ error_code:
- movl $-1, PT_ORIG_EAX(%esp) # no syscall to restart
- REG_TO_PTGS %ecx
- SET_KERNEL_GS %ecx
-- movl $(__USER_DS), %ecx
-+ movl $(__KERNEL_DS), %ecx
- movl %ecx, %ds
- movl %ecx, %es
-+
-+ PAX_ENTER_KERNEL
-+
- TRACE_IRQS_OFF
- movl %esp,%eax # pt_regs pointer
- call *%edi
-@@ -1344,6 +1503,9 @@ nmi_stack_correct:
- xorl %edx,%edx # zero error code
- movl %esp,%eax # pt_regs pointer
- call do_nmi
-+
-+ PAX_EXIT_KERNEL
-+
- jmp restore_all_notrace
- CFI_ENDPROC
-
-@@ -1380,6 +1542,9 @@ nmi_espfix_stack:
- FIXUP_ESPFIX_STACK # %eax == %esp
- xorl %edx,%edx # zero error code
- call do_nmi
-+
-+ PAX_EXIT_KERNEL
-+
- RESTORE_REGS
- lss 12+4(%esp), %esp # back to espfix stack
- CFI_ADJUST_CFA_OFFSET -24
-diff -urNp linux-2.6.37.4/arch/x86/kernel/entry_64.S linux-2.6.37.4/arch/x86/kernel/entry_64.S
---- linux-2.6.37.4/arch/x86/kernel/entry_64.S 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/entry_64.S 2011-02-27 14:48:10.000000000 -0500
-@@ -53,6 +53,7 @@
- #include <asm/paravirt.h>
- #include <asm/ftrace.h>
- #include <asm/percpu.h>
-+#include <asm/pgtable.h>
-
- /* Avoid __ASSEMBLER__'ifying <linux/audit.h> just for this. */
- #include <linux/elf-em.h>
-@@ -174,6 +175,201 @@ ENTRY(native_usergs_sysret64)
- ENDPROC(native_usergs_sysret64)
- #endif /* CONFIG_PARAVIRT */
-
-+ .macro ljmpq sel, off
-+#if defined(CONFIG_MCORE2) || defined (CONFIG_MATOM)
-+ .byte 0x48; ljmp *1234f(%rip)
-+ .pushsection .rodata
-+ .align 16
-+ 1234: .quad \off; .word \sel
-+ .popsection
-+#else
-+ push $\sel
-+ push $\off
-+ lretq
-+#endif
-+ .endm
-+
-+ .macro pax_enter_kernel
-+#ifdef CONFIG_PAX_KERNEXEC
-+ call pax_enter_kernel
-+#endif
-+ .endm
-+
-+ .macro pax_exit_kernel
-+#ifdef CONFIG_PAX_KERNEXEC
-+ call pax_exit_kernel
-+#endif
-+ .endm
-+
-+#ifdef CONFIG_PAX_KERNEXEC
-+ENTRY(pax_enter_kernel)
-+ push %rdi
-+
-+#ifdef CONFIG_PARAVIRT
-+ PV_SAVE_REGS(CLBR_RDI)
-+#endif
-+
-+ GET_CR0_INTO_RDI
-+ bts $16,%rdi
-+ jnc 1f
-+ mov %cs,%edi
-+ cmp $__KERNEL_CS,%edi
-+ jz 3f
-+ ljmpq __KERNEL_CS,3f
-+1: ljmpq __KERNEXEC_KERNEL_CS,2f
-+2: SET_RDI_INTO_CR0
-+3:
-+
-+#ifdef CONFIG_PARAVIRT
-+ PV_RESTORE_REGS(CLBR_RDI)
-+#endif
-+
-+ pop %rdi
-+ retq
-+ENDPROC(pax_enter_kernel)
-+
-+ENTRY(pax_exit_kernel)
-+ push %rdi
-+
-+#ifdef CONFIG_PARAVIRT
-+ PV_SAVE_REGS(CLBR_RDI)
-+#endif
-+
-+ mov %cs,%rdi
-+ cmp $__KERNEXEC_KERNEL_CS,%edi
-+ jnz 2f
-+ GET_CR0_INTO_RDI
-+ btr $16,%rdi
-+ ljmpq __KERNEL_CS,1f
-+1: SET_RDI_INTO_CR0
-+2:
-+
-+#ifdef CONFIG_PARAVIRT
-+ PV_RESTORE_REGS(CLBR_RDI);
-+#endif
-+
-+ pop %rdi
-+ retq
-+ENDPROC(pax_exit_kernel)
-+#endif
-+
-+ .macro pax_enter_kernel_user
-+#ifdef CONFIG_PAX_MEMORY_UDEREF
-+ call pax_enter_kernel_user
-+#endif
-+ .endm
-+
-+ .macro pax_exit_kernel_user
-+#ifdef CONFIG_PAX_MEMORY_UDEREF
-+ call pax_exit_kernel_user
-+#endif
-+ .endm
-+
-+#ifdef CONFIG_PAX_MEMORY_UDEREF
-+ENTRY(pax_enter_kernel_user)
-+ push %rdi
-+ push %rbx
-+
-+#ifdef CONFIG_PARAVIRT
-+ PV_SAVE_REGS(CLBR_RDI)
-+#endif
-+
-+ GET_CR3_INTO_RDI
-+ mov %rdi,%rbx
-+ add $__START_KERNEL_map,%rbx
-+ sub phys_base(%rip),%rbx
-+
-+#ifdef CONFIG_PARAVIRT
-+ push %rdi
-+ cmpl $0, pv_info+PARAVIRT_enabled
-+ jz 1f
-+ i = 0
-+ .rept USER_PGD_PTRS
-+ mov i*8(%rbx),%rsi
-+ mov $0,%sil
-+ lea i*8(%rbx),%rdi
-+ call PARA_INDIRECT(pv_mmu_ops+PV_MMU_set_pgd)
-+ i = i + 1
-+ .endr
-+ jmp 2f
-+1:
-+#endif
-+
-+ i = 0
-+ .rept USER_PGD_PTRS
-+ movb $0,i*8(%rbx)
-+ i = i + 1
-+ .endr
-+
-+#ifdef CONFIG_PARAVIRT
-+2: pop %rdi
-+#endif
-+ SET_RDI_INTO_CR3
-+
-+#ifdef CONFIG_PAX_KERNEXEC
-+ GET_CR0_INTO_RDI
-+ bts $16,%rdi
-+ SET_RDI_INTO_CR0
-+#endif
-+
-+#ifdef CONFIG_PARAVIRT
-+ PV_RESTORE_REGS(CLBR_RDI)
-+#endif
-+
-+ pop %rbx
-+ pop %rdi
-+ retq
-+ENDPROC(pax_enter_kernel_user)
-+
-+ENTRY(pax_exit_kernel_user)
-+ push %rdi
-+
-+#ifdef CONFIG_PARAVIRT
-+ push %rbx
-+ PV_SAVE_REGS(CLBR_RDI)
-+#endif
-+
-+#ifdef CONFIG_PAX_KERNEXEC
-+ GET_CR0_INTO_RDI
-+ btr $16,%rdi
-+ SET_RDI_INTO_CR0
-+#endif
-+
-+ GET_CR3_INTO_RDI
-+ add $__START_KERNEL_map,%rdi
-+ sub phys_base(%rip),%rdi
-+
-+#ifdef CONFIG_PARAVIRT
-+ cmpl $0, pv_info+PARAVIRT_enabled
-+ jz 1f
-+ mov %rdi,%rbx
-+ i = 0
-+ .rept USER_PGD_PTRS
-+ mov i*8(%rbx),%rsi
-+ mov $0x67,%sil
-+ lea i*8(%rbx),%rdi
-+ call PARA_INDIRECT(pv_mmu_ops+PV_MMU_set_pgd)
-+ i = i + 1
-+ .endr
-+ jmp 2f
-+1:
-+#endif
-+
-+ i = 0
-+ .rept USER_PGD_PTRS
-+ movb $0x67,i*8(%rdi)
-+ i = i + 1
-+ .endr
-+
-+#ifdef CONFIG_PARAVIRT
-+2: PV_RESTORE_REGS(CLBR_RDI)
-+ pop %rbx
-+#endif
-+
-+ pop %rdi
-+ retq
-+ENDPROC(pax_exit_kernel_user)
-+#endif
-
- .macro TRACE_IRQS_IRETQ offset=ARGOFFSET
- #ifdef CONFIG_TRACE_IRQFLAGS
-@@ -312,7 +508,7 @@ ENTRY(save_args)
- leaq -ARGOFFSET+16(%rsp),%rdi /* arg1 for handler */
- movq_cfi rbp, 8 /* push %rbp */
- leaq 8(%rsp), %rbp /* mov %rsp, %ebp */
-- testl $3, CS(%rdi)
-+ testb $3, CS(%rdi)
- je 1f
- SWAPGS
- /*
-@@ -403,7 +599,7 @@ ENTRY(ret_from_fork)
-
- RESTORE_REST
-
-- testl $3, CS-ARGOFFSET(%rsp) # from kernel_thread?
-+ testb $3, CS-ARGOFFSET(%rsp) # from kernel_thread?
- je int_ret_from_sys_call
-
- testl $_TIF_IA32, TI_flags(%rcx) # 32-bit compat task needs IRET
-@@ -462,6 +658,7 @@ ENTRY(system_call_after_swapgs)
-
- movq %rsp,PER_CPU_VAR(old_rsp)
- movq PER_CPU_VAR(kernel_stack),%rsp
-+ pax_enter_kernel_user
- /*
- * No need to follow this irqs off/on section - it's straight
- * and short:
-@@ -496,6 +693,7 @@ sysret_check:
- andl %edi,%edx
- jnz sysret_careful
- CFI_REMEMBER_STATE
-+ pax_exit_kernel_user
- /*
- * sysretq will re-enable interrupts:
- */
-@@ -605,7 +803,7 @@ tracesys:
- GLOBAL(int_ret_from_sys_call)
- DISABLE_INTERRUPTS(CLBR_NONE)
- TRACE_IRQS_OFF
-- testl $3,CS-ARGOFFSET(%rsp)
-+ testb $3,CS-ARGOFFSET(%rsp)
- je retint_restore_args
- movl $_TIF_ALLWORK_MASK,%edi
- /* edi: mask to check */
-@@ -786,6 +984,16 @@ END(interrupt)
- CFI_ADJUST_CFA_OFFSET ORIG_RAX-ARGOFFSET+8
- call save_args
- PARTIAL_FRAME 0
-+#ifdef CONFIG_PAX_MEMORY_UDEREF
-+ testb $3, CS(%rdi)
-+ jnz 1f
-+ pax_enter_kernel
-+ jmp 2f
-+1: pax_enter_kernel_user
-+2:
-+#else
-+ pax_enter_kernel
-+#endif
- call \func
- .endm
-
-@@ -813,7 +1021,7 @@ ret_from_intr:
- CFI_ADJUST_CFA_OFFSET -8
- exit_intr:
- GET_THREAD_INFO(%rcx)
-- testl $3,CS-ARGOFFSET(%rsp)
-+ testb $3,CS-ARGOFFSET(%rsp)
- je retint_kernel
-
- /* Interrupt came from user space */
-@@ -835,12 +1043,14 @@ retint_swapgs: /* return to user-space
- * The iretq could re-enable interrupts:
- */
- DISABLE_INTERRUPTS(CLBR_ANY)
-+ pax_exit_kernel_user
- TRACE_IRQS_IRETQ
- SWAPGS
- jmp restore_args
-
- retint_restore_args: /* return to kernel space */
- DISABLE_INTERRUPTS(CLBR_ANY)
-+ pax_exit_kernel
- /*
- * The iretq could re-enable interrupts:
- */
-@@ -1012,6 +1222,16 @@ ENTRY(\sym)
- CFI_ADJUST_CFA_OFFSET ORIG_RAX-R15
- call error_entry
- DEFAULT_FRAME 0
-+#ifdef CONFIG_PAX_MEMORY_UDEREF
-+ testb $3, CS(%rsp)
-+ jnz 1f
-+ pax_enter_kernel
-+ jmp 2f
-+1: pax_enter_kernel_user
-+2:
-+#else
-+ pax_enter_kernel
-+#endif
- movq %rsp,%rdi /* pt_regs pointer */
- xorl %esi,%esi /* no error code */
- call \do_sym
-@@ -1029,6 +1249,16 @@ ENTRY(\sym)
- CFI_ADJUST_CFA_OFFSET ORIG_RAX-R15
- call save_paranoid
- TRACE_IRQS_OFF
-+#ifdef CONFIG_PAX_MEMORY_UDEREF
-+ testb $3, CS(%rsp)
-+ jnz 1f
-+ pax_enter_kernel
-+ jmp 2f
-+1: pax_enter_kernel_user
-+2:
-+#else
-+ pax_enter_kernel
-+#endif
- movq %rsp,%rdi /* pt_regs pointer */
- xorl %esi,%esi /* no error code */
- call \do_sym
-@@ -1037,7 +1267,7 @@ ENTRY(\sym)
- END(\sym)
- .endm
-
--#define INIT_TSS_IST(x) PER_CPU_VAR(init_tss) + (TSS_ist + ((x) - 1) * 8)
-+#define INIT_TSS_IST(x) (TSS_ist + ((x) - 1) * 8)(%r12)
- .macro paranoidzeroentry_ist sym do_sym ist
- ENTRY(\sym)
- INTR_FRAME
-@@ -1047,8 +1277,24 @@ ENTRY(\sym)
- CFI_ADJUST_CFA_OFFSET ORIG_RAX-R15
- call save_paranoid
- TRACE_IRQS_OFF
-+#ifdef CONFIG_PAX_MEMORY_UDEREF
-+ testb $3, CS(%rsp)
-+ jnz 1f
-+ pax_enter_kernel
-+ jmp 2f
-+1: pax_enter_kernel_user
-+2:
-+#else
-+ pax_enter_kernel
-+#endif
- movq %rsp,%rdi /* pt_regs pointer */
- xorl %esi,%esi /* no error code */
-+#ifdef CONFIG_SMP
-+ imul $TSS_size, PER_CPU_VAR(cpu_number), %r12d
-+ lea init_tss(%r12), %r12
-+#else
-+ lea init_tss(%rip), %r12
-+#endif
- subq $EXCEPTION_STKSZ, INIT_TSS_IST(\ist)
- call \do_sym
- addq $EXCEPTION_STKSZ, INIT_TSS_IST(\ist)
-@@ -1065,6 +1311,16 @@ ENTRY(\sym)
- CFI_ADJUST_CFA_OFFSET ORIG_RAX-R15
- call error_entry
- DEFAULT_FRAME 0
-+#ifdef CONFIG_PAX_MEMORY_UDEREF
-+ testb $3, CS(%rsp)
-+ jnz 1f
-+ pax_enter_kernel
-+ jmp 2f
-+1: pax_enter_kernel_user
-+2:
-+#else
-+ pax_enter_kernel
-+#endif
- movq %rsp,%rdi /* pt_regs pointer */
- movq ORIG_RAX(%rsp),%rsi /* get error code */
- movq $-1,ORIG_RAX(%rsp) /* no syscall to restart */
-@@ -1084,6 +1340,16 @@ ENTRY(\sym)
- call save_paranoid
- DEFAULT_FRAME 0
- TRACE_IRQS_OFF
-+#ifdef CONFIG_PAX_MEMORY_UDEREF
-+ testb $3, CS(%rsp)
-+ jnz 1f
-+ pax_enter_kernel
-+ jmp 2f
-+1: pax_enter_kernel_user
-+2:
-+#else
-+ pax_enter_kernel
-+#endif
- movq %rsp,%rdi /* pt_regs pointer */
- movq ORIG_RAX(%rsp),%rsi /* get error code */
- movq $-1,ORIG_RAX(%rsp) /* no syscall to restart */
-@@ -1343,14 +1609,27 @@ ENTRY(paranoid_exit)
- TRACE_IRQS_OFF
- testl %ebx,%ebx /* swapgs needed? */
- jnz paranoid_restore
-- testl $3,CS(%rsp)
-+ testb $3,CS(%rsp)
- jnz paranoid_userspace
-+#ifdef CONFIG_PAX_MEMORY_UDEREF
-+ pax_exit_kernel
-+ TRACE_IRQS_IRETQ 0
-+ SWAPGS_UNSAFE_STACK
-+ RESTORE_ALL 8
-+ jmp irq_return
-+#endif
- paranoid_swapgs:
-+#ifdef CONFIG_PAX_MEMORY_UDEREF
-+ pax_exit_kernel_user
-+#else
-+ pax_exit_kernel
-+#endif
- TRACE_IRQS_IRETQ 0
- SWAPGS_UNSAFE_STACK
- RESTORE_ALL 8
- jmp irq_return
- paranoid_restore:
-+ pax_exit_kernel
- TRACE_IRQS_IRETQ 0
- RESTORE_ALL 8
- jmp irq_return
-@@ -1408,7 +1687,7 @@ ENTRY(error_entry)
- movq_cfi r14, R14+8
- movq_cfi r15, R15+8
- xorl %ebx,%ebx
-- testl $3,CS+8(%rsp)
-+ testb $3,CS+8(%rsp)
- je error_kernelspace
- error_swapgs:
- SWAPGS
-@@ -1472,6 +1751,16 @@ ENTRY(nmi)
- CFI_ADJUST_CFA_OFFSET ORIG_RAX-R15
- call save_paranoid
- DEFAULT_FRAME 0
-+#ifdef CONFIG_PAX_MEMORY_UDEREF
-+ testb $3, CS(%rsp)
-+ jnz 1f
-+ pax_enter_kernel
-+ jmp 2f
-+1: pax_enter_kernel_user
-+2:
-+#else
-+ pax_enter_kernel
-+#endif
- /* paranoidentry do_nmi, 0; without TRACE_IRQS_OFF */
- movq %rsp,%rdi
- movq $-1,%rsi
-@@ -1482,11 +1771,25 @@ ENTRY(nmi)
- DISABLE_INTERRUPTS(CLBR_NONE)
- testl %ebx,%ebx /* swapgs needed? */
- jnz nmi_restore
-- testl $3,CS(%rsp)
-+ testb $3,CS(%rsp)
- jnz nmi_userspace
-+#ifdef CONFIG_PAX_MEMORY_UDEREF
-+ pax_exit_kernel
-+ SWAPGS_UNSAFE_STACK
-+ RESTORE_ALL 8
-+ jmp irq_return
-+#endif
- nmi_swapgs:
-+#ifdef CONFIG_PAX_MEMORY_UDEREF
-+ pax_exit_kernel_user
-+#else
-+ pax_exit_kernel
-+#endif
- SWAPGS_UNSAFE_STACK
-+ RESTORE_ALL 8
-+ jmp irq_return
- nmi_restore:
-+ pax_exit_kernel
- RESTORE_ALL 8
- jmp irq_return
- nmi_userspace:
-diff -urNp linux-2.6.37.4/arch/x86/kernel/ftrace.c linux-2.6.37.4/arch/x86/kernel/ftrace.c
---- linux-2.6.37.4/arch/x86/kernel/ftrace.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/ftrace.c 2011-01-17 02:41:01.000000000 -0500
-@@ -174,7 +174,9 @@ void ftrace_nmi_enter(void)
-
- if (atomic_inc_return(&nmi_running) & MOD_CODE_WRITE_FLAG) {
- smp_rmb();
-+ pax_open_kernel();
- ftrace_mod_code();
-+ pax_close_kernel();
- atomic_inc(&nmi_update_count);
- }
- /* Must have previous changes seen before executions */
-@@ -268,6 +270,8 @@ ftrace_modify_code(unsigned long ip, uns
- {
- unsigned char replaced[MCOUNT_INSN_SIZE];
-
-+ ip = ktla_ktva(ip);
-+
- /*
- * Note: Due to modules and __init, code can
- * disappear and change, we need to protect against faulting
-@@ -324,7 +328,7 @@ int ftrace_update_ftrace_func(ftrace_fun
- unsigned char old[MCOUNT_INSN_SIZE], *new;
- int ret;
-
-- memcpy(old, &ftrace_call, MCOUNT_INSN_SIZE);
-+ memcpy(old, (void *)ktla_ktva((unsigned long)ftrace_call), MCOUNT_INSN_SIZE);
- new = ftrace_call_replace(ip, (unsigned long)func);
- ret = ftrace_modify_code(ip, old, new);
-
-@@ -350,6 +354,8 @@ static int ftrace_mod_jmp(unsigned long
- {
- unsigned char code[MCOUNT_INSN_SIZE];
-
-+ ip = ktla_ktva(ip);
-+
- if (probe_kernel_read(code, (void *)ip, MCOUNT_INSN_SIZE))
- return -EFAULT;
-
-diff -urNp linux-2.6.37.4/arch/x86/kernel/head32.c linux-2.6.37.4/arch/x86/kernel/head32.c
---- linux-2.6.37.4/arch/x86/kernel/head32.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/head32.c 2011-01-17 02:41:01.000000000 -0500
-@@ -19,6 +19,7 @@
- #include <asm/io_apic.h>
- #include <asm/bios_ebda.h>
- #include <asm/tlbflush.h>
-+#include <asm/boot.h>
-
- static void __init i386_default_early_setup(void)
- {
-@@ -43,7 +44,7 @@ void __init i386_start_kernel(void)
- memblock_x86_reserve_range(PAGE_SIZE, PAGE_SIZE + PAGE_SIZE, "EX TRAMPOLINE");
- #endif
-
-- memblock_x86_reserve_range(__pa_symbol(&_text), __pa_symbol(&__bss_stop), "TEXT DATA BSS");
-+ memblock_x86_reserve_range(LOAD_PHYSICAL_ADDR, __pa_symbol(&__bss_stop), "TEXT DATA BSS");
-
- #ifdef CONFIG_BLK_DEV_INITRD
- /* Reserve INITRD */
-diff -urNp linux-2.6.37.4/arch/x86/kernel/head_32.S linux-2.6.37.4/arch/x86/kernel/head_32.S
---- linux-2.6.37.4/arch/x86/kernel/head_32.S 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/head_32.S 2011-01-25 20:24:56.000000000 -0500
-@@ -25,6 +25,12 @@
- /* Physical address */
- #define pa(X) ((X) - __PAGE_OFFSET)
-
-+#ifdef CONFIG_PAX_KERNEXEC
-+#define ta(X) (X)
-+#else
-+#define ta(X) ((X) - __PAGE_OFFSET)
-+#endif
-+
- /*
- * References to members of the new_cpu_data structure.
- */
-@@ -54,11 +60,7 @@
- * and small than max_low_pfn, otherwise will waste some page table entries
- */
-
--#if PTRS_PER_PMD > 1
--#define PAGE_TABLE_SIZE(pages) (((pages) / PTRS_PER_PMD) + PTRS_PER_PGD)
--#else
--#define PAGE_TABLE_SIZE(pages) ((pages) / PTRS_PER_PGD)
--#endif
-+#define PAGE_TABLE_SIZE(pages) ((pages) / PTRS_PER_PTE)
-
- /* Number of possible pages in the lowmem region */
- LOWMEM_PAGES = (((1<<32) - __PAGE_OFFSET) >> PAGE_SHIFT)
-@@ -77,6 +79,12 @@ INIT_MAP_SIZE = PAGE_TABLE_SIZE(KERNEL_P
- RESERVE_BRK(pagetables, INIT_MAP_SIZE)
-
- /*
-+ * Real beginning of normal "text" segment
-+ */
-+ENTRY(stext)
-+ENTRY(_stext)
-+
-+/*
- * 32-bit kernel entrypoint; only used by the boot CPU. On entry,
- * %esi points to the real-mode code as a 32-bit pointer.
- * CS and DS must be 4 GB flat segments, but we don't depend on
-@@ -84,6 +92,13 @@ RESERVE_BRK(pagetables, INIT_MAP_SIZE)
- * can.
- */
- __HEAD
-+
-+#ifdef CONFIG_PAX_KERNEXEC
-+ jmp startup_32
-+/* PaX: fill first page in .text with int3 to catch NULL derefs in kernel mode */
-+.fill PAGE_SIZE-5,1,0xcc
-+#endif
-+
- ENTRY(startup_32)
- /* test KEEP_SEGMENTS flag to see if the bootloader is asking
- us to not reload segments */
-@@ -101,6 +116,57 @@ ENTRY(startup_32)
- movl %eax,%gs
- 2:
-
-+#ifdef CONFIG_SMP
-+ movl $pa(cpu_gdt_table),%edi
-+ movl $__per_cpu_load,%eax
-+ movw %ax,__KERNEL_PERCPU + 2(%edi)
-+ rorl $16,%eax
-+ movb %al,__KERNEL_PERCPU + 4(%edi)
-+ movb %ah,__KERNEL_PERCPU + 7(%edi)
-+ movl $__per_cpu_end - 1,%eax
-+ subl $__per_cpu_start,%eax
-+ movw %ax,__KERNEL_PERCPU + 0(%edi)
-+#endif
-+
-+#ifdef CONFIG_PAX_MEMORY_UDEREF
-+ movl $NR_CPUS,%ecx
-+ movl $pa(cpu_gdt_table),%edi
-+1:
-+ movl $((((__PAGE_OFFSET-1) & 0xf0000000) >> 12) | 0x00c09700),GDT_ENTRY_KERNEL_DS * 8 + 4(%edi)
-+ movl $((((__PAGE_OFFSET-1) & 0xf0000000) >> 12) | 0x00c0fb00),GDT_ENTRY_DEFAULT_USER_CS * 8 + 4(%edi)
-+ movl $((((__PAGE_OFFSET-1) & 0xf0000000) >> 12) | 0x00c0f300),GDT_ENTRY_DEFAULT_USER_DS * 8 + 4(%edi)
-+ addl $PAGE_SIZE_asm,%edi
-+ loop 1b
-+#endif
-+
-+#ifdef CONFIG_PAX_KERNEXEC
-+ movl $pa(boot_gdt),%edi
-+ movl $__LOAD_PHYSICAL_ADDR,%eax
-+ movw %ax,__BOOT_CS + 2(%edi)
-+ rorl $16,%eax
-+ movb %al,__BOOT_CS + 4(%edi)
-+ movb %ah,__BOOT_CS + 7(%edi)
-+ rorl $16,%eax
-+
-+ ljmp $(__BOOT_CS),$1f
-+1:
-+
-+ movl $NR_CPUS,%ecx
-+ movl $pa(cpu_gdt_table),%edi
-+ addl $__PAGE_OFFSET,%eax
-+1:
-+ movw %ax,__KERNEL_CS + 2(%edi)
-+ movw %ax,__KERNEXEC_KERNEL_CS + 2(%edi)
-+ rorl $16,%eax
-+ movb %al,__KERNEL_CS + 4(%edi)
-+ movb %al,__KERNEXEC_KERNEL_CS + 4(%edi)
-+ movb %ah,__KERNEL_CS + 7(%edi)
-+ movb %ah,__KERNEXEC_KERNEL_CS + 7(%edi)
-+ rorl $16,%eax
-+ addl $PAGE_SIZE_asm,%edi
-+ loop 1b
-+#endif
-+
- /*
- * Clear BSS first so that there are no surprises...
- */
-@@ -150,9 +216,7 @@ ENTRY(startup_32)
- cmpl $num_subarch_entries, %eax
- jae bad_subarch
-
-- movl pa(subarch_entries)(,%eax,4), %eax
-- subl $__PAGE_OFFSET, %eax
-- jmp *%eax
-+ jmp *pa(subarch_entries)(,%eax,4)
-
- bad_subarch:
- WEAK(lguest_entry)
-@@ -164,10 +228,10 @@ WEAK(xen_entry)
- __INITDATA
-
- subarch_entries:
-- .long default_entry /* normal x86/PC */
-- .long lguest_entry /* lguest hypervisor */
-- .long xen_entry /* Xen hypervisor */
-- .long default_entry /* Moorestown MID */
-+ .long ta(default_entry) /* normal x86/PC */
-+ .long ta(lguest_entry) /* lguest hypervisor */
-+ .long ta(xen_entry) /* Xen hypervisor */
-+ .long ta(default_entry) /* Moorestown MID */
- num_subarch_entries = (. - subarch_entries) / 4
- .previous
- #endif /* CONFIG_PARAVIRT */
-@@ -227,8 +291,11 @@ default_entry:
- movl %eax, pa(max_pfn_mapped)
-
- /* Do early initialization of the fixmap area */
-- movl $pa(initial_pg_fixmap)+PDE_IDENT_ATTR,%eax
-- movl %eax,pa(initial_pg_pmd+0x1000*KPMDS-8)
-+#ifdef CONFIG_COMPAT_VDSO
-+ movl $pa(initial_pg_fixmap)+PDE_IDENT_ATTR+_PAGE_USER,pa(initial_pg_pmd+0x1000*KPMDS-8)
-+#else
-+ movl $pa(initial_pg_fixmap)+PDE_IDENT_ATTR,pa(initial_pg_pmd+0x1000*KPMDS-8)
-+#endif
- #else /* Not PAE */
-
- page_pde_offset = (__PAGE_OFFSET >> 20);
-@@ -258,8 +325,11 @@ page_pde_offset = (__PAGE_OFFSET >> 20);
- movl %eax, pa(max_pfn_mapped)
-
- /* Do early initialization of the fixmap area */
-- movl $pa(initial_pg_fixmap)+PDE_IDENT_ATTR,%eax
-- movl %eax,pa(initial_page_table+0xffc)
-+#ifdef CONFIG_COMPAT_VDSO
-+ movl $pa(initial_pg_fixmap)+PDE_IDENT_ATTR+_PAGE_USER,pa(initial_page_table+0xffc)
-+#else
-+ movl $pa(initial_pg_fixmap)+PDE_IDENT_ATTR,pa(initial_page_table+0xffc)
-+#endif
- #endif
- jmp 3f
- /*
-@@ -306,6 +376,7 @@ ENTRY(startup_32_smp)
- orl %edx,%eax
- movl %eax,%cr4
-
-+#ifdef CONFIG_X86_PAE
- testb $X86_CR4_PAE, %al # check if PAE is enabled
- jz 6f
-
-@@ -330,6 +401,9 @@ ENTRY(startup_32_smp)
- /* Make changes effective */
- wrmsr
-
-+ btsl $_PAGE_BIT_NX-32,pa(__supported_pte_mask+4)
-+#endif
-+
- 6:
-
- /*
-@@ -355,9 +429,7 @@ ENTRY(startup_32_smp)
-
- #ifdef CONFIG_SMP
- cmpb $0, ready
-- jz 1f /* Initial CPU cleans BSS */
-- jmp checkCPUtype
--1:
-+ jnz checkCPUtype /* Initial CPU cleans BSS */
- #endif /* CONFIG_SMP */
-
- /*
-@@ -435,7 +507,7 @@ is386: movl $2,%ecx # set MP
- 1: movl $(__KERNEL_DS),%eax # reload all the segment registers
- movl %eax,%ss # after changing gdt.
-
-- movl $(__USER_DS),%eax # DS/ES contains default USER segment
-+# movl $(__KERNEL_DS),%eax # DS/ES contains default KERNEL segment
- movl %eax,%ds
- movl %eax,%es
-
-@@ -449,15 +521,22 @@ is386: movl $2,%ecx # set MP
- */
- cmpb $0,ready
- jne 1f
-- movl $gdt_page,%eax
-+ movl $cpu_gdt_table,%eax
- movl $stack_canary,%ecx
-+#ifdef CONFIG_SMP
-+ addl $__per_cpu_load,%ecx
-+#endif
- movw %cx, 8 * GDT_ENTRY_STACK_CANARY + 2(%eax)
- shrl $16, %ecx
- movb %cl, 8 * GDT_ENTRY_STACK_CANARY + 4(%eax)
- movb %ch, 8 * GDT_ENTRY_STACK_CANARY + 7(%eax)
- 1:
--#endif
- movl $(__KERNEL_STACK_CANARY),%eax
-+#elif defined(CONFIG_PAX_MEMORY_UDEREF)
-+ movl $(__USER_DS),%eax
-+#else
-+ xorl %eax,%eax
-+#endif
- movl %eax,%gs
-
- xorl %eax,%eax # Clear LDT
-@@ -468,10 +547,6 @@ is386: movl $2,%ecx # set MP
- #ifdef CONFIG_SMP
- movb ready, %cl
- movb $1, ready
-- cmpb $0,%cl # the first CPU calls start_kernel
-- je 1f
-- movl (stack_start), %esp
--1:
- #endif /* CONFIG_SMP */
- jmp *(initial_code)
-
-@@ -557,22 +632,22 @@ early_page_fault:
- jmp early_fault
-
- early_fault:
-- cld
- #ifdef CONFIG_PRINTK
-+ cmpl $1,%ss:early_recursion_flag
-+ je hlt_loop
-+ incl %ss:early_recursion_flag
-+ cld
- pusha
- movl $(__KERNEL_DS),%eax
- movl %eax,%ds
- movl %eax,%es
-- cmpl $2,early_recursion_flag
-- je hlt_loop
-- incl early_recursion_flag
- movl %cr2,%eax
- pushl %eax
- pushl %edx /* trapno */
- pushl $fault_msg
- call printk
-+; call dump_stack
- #endif
-- call dump_stack
- hlt_loop:
- hlt
- jmp hlt_loop
-@@ -580,8 +655,11 @@ hlt_loop:
- /* This is the default interrupt "handler" :-) */
- ALIGN
- ignore_int:
-- cld
- #ifdef CONFIG_PRINTK
-+ cmpl $2,%ss:early_recursion_flag
-+ je hlt_loop
-+ incl %ss:early_recursion_flag
-+ cld
- pushl %eax
- pushl %ecx
- pushl %edx
-@@ -590,9 +668,6 @@ ignore_int:
- movl $(__KERNEL_DS),%eax
- movl %eax,%ds
- movl %eax,%es
-- cmpl $2,early_recursion_flag
-- je hlt_loop
-- incl early_recursion_flag
- pushl 16(%esp)
- pushl 24(%esp)
- pushl 32(%esp)
-@@ -619,29 +694,43 @@ ENTRY(initial_code)
- /*
- * BSS section
- */
--__PAGE_ALIGNED_BSS
-- .align PAGE_SIZE_asm
- #ifdef CONFIG_X86_PAE
-+.section .initial_pg_pmd,"a",@progbits
- ENTRY(initial_pg_pmd)
- .fill 1024*KPMDS,4,0
- #else
-+.section .initial_page_table,"a",@progbits
- ENTRY(initial_page_table)
- .fill 1024,4,0
- #endif
-+.section .initial_pg_fixmap,"a",@progbits
- ENTRY(initial_pg_fixmap)
- .fill 1024,4,0
-+.section .empty_zero_page,"a",@progbits
- ENTRY(empty_zero_page)
- .fill 4096,1,0
-+.section .swapper_pg_dir,"a",@progbits
- ENTRY(swapper_pg_dir)
-+#ifdef CONFIG_X86_PAE
-+ .fill 4,8,0
-+#else
- .fill 1024,4,0
-+#endif
-+
-+/*
-+ * The IDT has to be page-aligned to simplify the Pentium
-+ * F0 0F bug workaround.. We have a special link segment
-+ * for this.
-+ */
-+.section .idt,"a",@progbits
-+ENTRY(idt_table)
-+ .fill 256,8,0
-
- /*
- * This starts the data section.
- */
- #ifdef CONFIG_X86_PAE
--__PAGE_ALIGNED_DATA
-- /* Page-aligned for the benefit of paravirt? */
-- .align PAGE_SIZE_asm
-+.section .initial_page_table,"a",@progbits
- ENTRY(initial_page_table)
- .long pa(initial_pg_pmd+PGD_IDENT_ATTR),0 /* low identity map */
- # if KPMDS == 3
-@@ -660,15 +749,24 @@ ENTRY(initial_page_table)
- # error "Kernel PMDs should be 1, 2 or 3"
- # endif
- .align PAGE_SIZE_asm /* needs to be page-sized too */
-+
-+#ifdef CONFIG_PAX_PER_CPU_PGD
-+ENTRY(cpu_pgd)
-+ .rept NR_CPUS
-+ .fill 4,8,0
-+ .endr
-+#endif
-+
- #endif
-
- .data
- ENTRY(stack_start)
-- .long init_thread_union+THREAD_SIZE
-+ .long init_thread_union+THREAD_SIZE-8
- .long __BOOT_DS
-
- ready: .byte 0
-
-+.section .rodata,"a",@progbits
- early_recursion_flag:
- .long 0
-
-@@ -704,7 +802,7 @@ fault_msg:
- .word 0 # 32 bit align gdt_desc.address
- boot_gdt_descr:
- .word __BOOT_DS+7
-- .long boot_gdt - __PAGE_OFFSET
-+ .long pa(boot_gdt)
-
- .word 0 # 32-bit align idt_desc.address
- idt_descr:
-@@ -715,7 +813,7 @@ idt_descr:
- .word 0 # 32 bit align gdt_desc.address
- ENTRY(early_gdt_descr)
- .word GDT_ENTRIES*8-1
-- .long gdt_page /* Overwritten for secondary CPUs */
-+ .long cpu_gdt_table /* Overwritten for secondary CPUs */
-
- /*
- * The boot_gdt must mirror the equivalent in setup.S and is
-@@ -724,5 +822,65 @@ ENTRY(early_gdt_descr)
- .align L1_CACHE_BYTES
- ENTRY(boot_gdt)
- .fill GDT_ENTRY_BOOT_CS,8,0
-- .quad 0x00cf9a000000ffff /* kernel 4GB code at 0x00000000 */
-- .quad 0x00cf92000000ffff /* kernel 4GB data at 0x00000000 */
-+ .quad 0x00cf9b000000ffff /* kernel 4GB code at 0x00000000 */
-+ .quad 0x00cf93000000ffff /* kernel 4GB data at 0x00000000 */
-+
-+ .align PAGE_SIZE_asm
-+ENTRY(cpu_gdt_table)
-+ .rept NR_CPUS
-+ .quad 0x0000000000000000 /* NULL descriptor */
-+ .quad 0x0000000000000000 /* 0x0b reserved */
-+ .quad 0x0000000000000000 /* 0x13 reserved */
-+ .quad 0x0000000000000000 /* 0x1b reserved */
-+
-+#ifdef CONFIG_PAX_KERNEXEC
-+ .quad 0x00cf9b000000ffff /* 0x20 alternate kernel 4GB code at 0x00000000 */
-+#else
-+ .quad 0x0000000000000000 /* 0x20 unused */
-+#endif
-+
-+ .quad 0x0000000000000000 /* 0x28 unused */
-+ .quad 0x0000000000000000 /* 0x33 TLS entry 1 */
-+ .quad 0x0000000000000000 /* 0x3b TLS entry 2 */
-+ .quad 0x0000000000000000 /* 0x43 TLS entry 3 */
-+ .quad 0x0000000000000000 /* 0x4b reserved */
-+ .quad 0x0000000000000000 /* 0x53 reserved */
-+ .quad 0x0000000000000000 /* 0x5b reserved */
-+
-+ .quad 0x00cf9b000000ffff /* 0x60 kernel 4GB code at 0x00000000 */
-+ .quad 0x00cf93000000ffff /* 0x68 kernel 4GB data at 0x00000000 */
-+ .quad 0x00cffb000000ffff /* 0x73 user 4GB code at 0x00000000 */
-+ .quad 0x00cff3000000ffff /* 0x7b user 4GB data at 0x00000000 */
-+
-+ .quad 0x0000000000000000 /* 0x80 TSS descriptor */
-+ .quad 0x0000000000000000 /* 0x88 LDT descriptor */
-+
-+ /*
-+ * Segments used for calling PnP BIOS have byte granularity.
-+ * The code segments and data segments have fixed 64k limits,
-+ * the transfer segment sizes are set at run time.
-+ */
-+ .quad 0x00409b000000ffff /* 0x90 32-bit code */
-+ .quad 0x00009b000000ffff /* 0x98 16-bit code */
-+ .quad 0x000093000000ffff /* 0xa0 16-bit data */
-+ .quad 0x0000930000000000 /* 0xa8 16-bit data */
-+ .quad 0x0000930000000000 /* 0xb0 16-bit data */
-+
-+ /*
-+ * The APM segments have byte granularity and their bases
-+ * are set at run time. All have 64k limits.
-+ */
-+ .quad 0x00409b000000ffff /* 0xb8 APM CS code */
-+ .quad 0x00009b000000ffff /* 0xc0 APM CS 16 code (16 bit) */
-+ .quad 0x004093000000ffff /* 0xc8 APM DS data */
-+
-+ .quad 0x00c0930000000000 /* 0xd0 - ESPFIX SS */
-+ .quad 0x0040930000000000 /* 0xd8 - PERCPU */
-+ .quad 0x0040910000000018 /* 0xe0 - STACK_CANARY */
-+ .quad 0x0000000000000000 /* 0xe8 - PCIBIOS_CS */
-+ .quad 0x0000000000000000 /* 0xf0 - PCIBIOS_DS */
-+ .quad 0x0000000000000000 /* 0xf8 - GDT entry 31: double-fault TSS */
-+
-+ /* Be sure this is zeroed to avoid false validations in Xen */
-+ .fill PAGE_SIZE_asm - GDT_SIZE,1,0
-+ .endr
-diff -urNp linux-2.6.37.4/arch/x86/kernel/head_64.S linux-2.6.37.4/arch/x86/kernel/head_64.S
---- linux-2.6.37.4/arch/x86/kernel/head_64.S 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/head_64.S 2011-01-17 02:41:01.000000000 -0500
-@@ -19,6 +19,7 @@
- #include <asm/cache.h>
- #include <asm/processor-flags.h>
- #include <asm/percpu.h>
-+#include <asm/cpufeature.h>
-
- #ifdef CONFIG_PARAVIRT
- #include <asm/asm-offsets.h>
-@@ -38,6 +39,10 @@ L4_PAGE_OFFSET = pgd_index(__PAGE_OFFSET
- L3_PAGE_OFFSET = pud_index(__PAGE_OFFSET)
- L4_START_KERNEL = pgd_index(__START_KERNEL_map)
- L3_START_KERNEL = pud_index(__START_KERNEL_map)
-+L4_VMALLOC_START = pgd_index(VMALLOC_START)
-+L3_VMALLOC_START = pud_index(VMALLOC_START)
-+L4_VMEMMAP_START = pgd_index(VMEMMAP_START)
-+L3_VMEMMAP_START = pud_index(VMEMMAP_START)
-
- .text
- __HEAD
-@@ -85,35 +90,22 @@ startup_64:
- */
- addq %rbp, init_level4_pgt + 0(%rip)
- addq %rbp, init_level4_pgt + (L4_PAGE_OFFSET*8)(%rip)
-+ addq %rbp, init_level4_pgt + (L4_VMALLOC_START*8)(%rip)
-+ addq %rbp, init_level4_pgt + (L4_VMEMMAP_START*8)(%rip)
- addq %rbp, init_level4_pgt + (L4_START_KERNEL*8)(%rip)
-
- addq %rbp, level3_ident_pgt + 0(%rip)
-+#ifndef CONFIG_XEN
-+ addq %rbp, level3_ident_pgt + 8(%rip)
-+#endif
-
-- addq %rbp, level3_kernel_pgt + (510*8)(%rip)
-- addq %rbp, level3_kernel_pgt + (511*8)(%rip)
-+ addq %rbp, level3_vmemmap_pgt + (L3_VMEMMAP_START*8)(%rip)
-
-- addq %rbp, level2_fixmap_pgt + (506*8)(%rip)
-+ addq %rbp, level3_kernel_pgt + (L3_START_KERNEL*8)(%rip)
-+ addq %rbp, level3_kernel_pgt + (L3_START_KERNEL*8+8)(%rip)
-
-- /* Add an Identity mapping if I am above 1G */
-- leaq _text(%rip), %rdi
-- andq $PMD_PAGE_MASK, %rdi
--
-- movq %rdi, %rax
-- shrq $PUD_SHIFT, %rax
-- andq $(PTRS_PER_PUD - 1), %rax
-- jz ident_complete
--
-- leaq (level2_spare_pgt - __START_KERNEL_map + _KERNPG_TABLE)(%rbp), %rdx
-- leaq level3_ident_pgt(%rip), %rbx
-- movq %rdx, 0(%rbx, %rax, 8)
--
-- movq %rdi, %rax
-- shrq $PMD_SHIFT, %rax
-- andq $(PTRS_PER_PMD - 1), %rax
-- leaq __PAGE_KERNEL_IDENT_LARGE_EXEC(%rdi), %rdx
-- leaq level2_spare_pgt(%rip), %rbx
-- movq %rdx, 0(%rbx, %rax, 8)
--ident_complete:
-+ addq %rbp, level2_fixmap_pgt + (506*8)(%rip)
-+ addq %rbp, level2_fixmap_pgt + (507*8)(%rip)
-
- /*
- * Fixup the kernel text+data virtual addresses. Note that
-@@ -161,8 +153,8 @@ ENTRY(secondary_startup_64)
- * after the boot processor executes this code.
- */
-
-- /* Enable PAE mode and PGE */
-- movl $(X86_CR4_PAE | X86_CR4_PGE), %eax
-+ /* Enable PAE mode and PSE/PGE */
-+ movl $(X86_CR4_PSE | X86_CR4_PAE | X86_CR4_PGE), %eax
- movq %rax, %cr4
-
- /* Setup early boot stage 4 level pagetables. */
-@@ -184,9 +176,14 @@ ENTRY(secondary_startup_64)
- movl $MSR_EFER, %ecx
- rdmsr
- btsl $_EFER_SCE, %eax /* Enable System Call */
-- btl $20,%edi /* No Execute supported? */
-+ btl $(X86_FEATURE_NX & 31),%edi /* No Execute supported? */
- jnc 1f
- btsl $_EFER_NX, %eax
-+ leaq init_level4_pgt(%rip), %rdi
-+ btsq $_PAGE_BIT_NX, 8*L4_PAGE_OFFSET(%rdi)
-+ btsq $_PAGE_BIT_NX, 8*L4_VMALLOC_START(%rdi)
-+ btsq $_PAGE_BIT_NX, 8*L4_VMEMMAP_START(%rdi)
-+ btsq $_PAGE_BIT_NX, __supported_pte_mask(%rip)
- 1: wrmsr /* Make changes effective */
-
- /* Setup cr0 */
-@@ -270,7 +267,7 @@ ENTRY(secondary_startup_64)
- bad_address:
- jmp bad_address
-
-- .section ".init.text","ax"
-+ __INIT
- #ifdef CONFIG_EARLY_PRINTK
- .globl early_idt_handlers
- early_idt_handlers:
-@@ -315,18 +312,23 @@ ENTRY(early_idt_handler)
- #endif /* EARLY_PRINTK */
- 1: hlt
- jmp 1b
-+ .previous
-
- #ifdef CONFIG_EARLY_PRINTK
-+ __INITDATA
- early_recursion_flag:
- .long 0
-+ .previous
-
-+ .section .rodata,"a",@progbits
- early_idt_msg:
- .asciz "PANIC: early exception %02lx rip %lx:%lx error %lx cr2 %lx\n"
- early_idt_ripmsg:
- .asciz "RIP %s\n"
--#endif /* CONFIG_EARLY_PRINTK */
- .previous
-+#endif /* CONFIG_EARLY_PRINTK */
-
-+ .section .rodata,"a",@progbits
- #define NEXT_PAGE(name) \
- .balign PAGE_SIZE; \
- ENTRY(name)
-@@ -339,7 +341,6 @@ ENTRY(name)
- i = i + 1 ; \
- .endr
-
-- .data
- /*
- * This default setting generates an ident mapping at address 0x100000
- * and a mapping for the kernel that precisely maps virtual address
-@@ -350,13 +351,36 @@ NEXT_PAGE(init_level4_pgt)
- .quad level3_ident_pgt - __START_KERNEL_map + _KERNPG_TABLE
- .org init_level4_pgt + L4_PAGE_OFFSET*8, 0
- .quad level3_ident_pgt - __START_KERNEL_map + _KERNPG_TABLE
-+ .org init_level4_pgt + L4_VMALLOC_START*8, 0
-+ .quad level3_vmalloc_pgt - __START_KERNEL_map + _KERNPG_TABLE
-+ .org init_level4_pgt + L4_VMEMMAP_START*8, 0
-+ .quad level3_vmemmap_pgt - __START_KERNEL_map + _KERNPG_TABLE
- .org init_level4_pgt + L4_START_KERNEL*8, 0
- /* (2^48-(2*1024*1024*1024))/(2^39) = 511 */
- .quad level3_kernel_pgt - __START_KERNEL_map + _PAGE_TABLE
-
-+#ifdef CONFIG_PAX_PER_CPU_PGD
-+NEXT_PAGE(cpu_pgd)
-+ .rept NR_CPUS
-+ .fill 512,8,0
-+ .endr
-+#endif
-+
- NEXT_PAGE(level3_ident_pgt)
- .quad level2_ident_pgt - __START_KERNEL_map + _KERNPG_TABLE
-+#ifdef CONFIG_XEN
- .fill 511,8,0
-+#else
-+ .quad level2_ident_pgt + PAGE_SIZE - __START_KERNEL_map + _KERNPG_TABLE
-+ .fill 510,8,0
-+#endif
-+
-+NEXT_PAGE(level3_vmalloc_pgt)
-+ .fill 512,8,0
-+
-+NEXT_PAGE(level3_vmemmap_pgt)
-+ .fill L3_VMEMMAP_START,8,0
-+ .quad level2_vmemmap_pgt - __START_KERNEL_map + _KERNPG_TABLE
-
- NEXT_PAGE(level3_kernel_pgt)
- .fill L3_START_KERNEL,8,0
-@@ -364,20 +388,23 @@ NEXT_PAGE(level3_kernel_pgt)
- .quad level2_kernel_pgt - __START_KERNEL_map + _KERNPG_TABLE
- .quad level2_fixmap_pgt - __START_KERNEL_map + _PAGE_TABLE
-
-+NEXT_PAGE(level2_vmemmap_pgt)
-+ .fill 512,8,0
-+
- NEXT_PAGE(level2_fixmap_pgt)
-- .fill 506,8,0
-- .quad level1_fixmap_pgt - __START_KERNEL_map + _PAGE_TABLE
-- /* 8MB reserved for vsyscalls + a 2MB hole = 4 + 1 entries */
-- .fill 5,8,0
-+ .fill 507,8,0
-+ .quad level1_vsyscall_pgt - __START_KERNEL_map + _PAGE_TABLE
-+ /* 6MB reserved for vsyscalls + a 2MB hole = 3 + 1 entries */
-+ .fill 4,8,0
-
--NEXT_PAGE(level1_fixmap_pgt)
-+NEXT_PAGE(level1_vsyscall_pgt)
- .fill 512,8,0
-
--NEXT_PAGE(level2_ident_pgt)
-- /* Since I easily can, map the first 1G.
-+ /* Since I easily can, map the first 2G.
- * Don't set NX because code runs from these pages.
- */
-- PMDS(0, __PAGE_KERNEL_IDENT_LARGE_EXEC, PTRS_PER_PMD)
-+NEXT_PAGE(level2_ident_pgt)
-+ PMDS(0, __PAGE_KERNEL_IDENT_LARGE_EXEC, 2*PTRS_PER_PMD)
-
- NEXT_PAGE(level2_kernel_pgt)
- /*
-@@ -390,33 +417,55 @@ NEXT_PAGE(level2_kernel_pgt)
- * If you want to increase this then increase MODULES_VADDR
- * too.)
- */
-- PMDS(0, __PAGE_KERNEL_LARGE_EXEC,
-- KERNEL_IMAGE_SIZE/PMD_SIZE)
--
--NEXT_PAGE(level2_spare_pgt)
-- .fill 512, 8, 0
-+ PMDS(0, __PAGE_KERNEL_LARGE_EXEC, KERNEL_IMAGE_SIZE/PMD_SIZE)
-
- #undef PMDS
- #undef NEXT_PAGE
-
-- .data
-+ .align PAGE_SIZE
-+ENTRY(cpu_gdt_table)
-+ .rept NR_CPUS
-+ .quad 0x0000000000000000 /* NULL descriptor */
-+ .quad 0x00cf9b000000ffff /* __KERNEL32_CS */
-+ .quad 0x00af9b000000ffff /* __KERNEL_CS */
-+ .quad 0x00cf93000000ffff /* __KERNEL_DS */
-+ .quad 0x00cffb000000ffff /* __USER32_CS */
-+ .quad 0x00cff3000000ffff /* __USER_DS, __USER32_DS */
-+ .quad 0x00affb000000ffff /* __USER_CS */
-+
-+#ifdef CONFIG_PAX_KERNEXEC
-+ .quad 0x00af9b000000ffff /* __KERNEXEC_KERNEL_CS */
-+#else
-+ .quad 0x0 /* unused */
-+#endif
-+
-+ .quad 0,0 /* TSS */
-+ .quad 0,0 /* LDT */
-+ .quad 0,0,0 /* three TLS descriptors */
-+ .quad 0x0000f40000000000 /* node/CPU stored in limit */
-+ /* asm/segment.h:GDT_ENTRIES must match this */
-+
-+ /* zero the remaining page */
-+ .fill PAGE_SIZE / 8 - GDT_ENTRIES,8,0
-+ .endr
-+
- .align 16
- .globl early_gdt_descr
- early_gdt_descr:
- .word GDT_ENTRIES*8-1
- early_gdt_descr_base:
-- .quad INIT_PER_CPU_VAR(gdt_page)
-+ .quad cpu_gdt_table
-
- ENTRY(phys_base)
- /* This must match the first entry in level2_kernel_pgt */
- .quad 0x0000000000000000
-
- #include "../../x86/xen/xen-head.S"
--
-- .section .bss, "aw", @nobits
-+
-+ .section .rodata,"a",@progbits
- .align L1_CACHE_BYTES
- ENTRY(idt_table)
-- .skip IDT_ENTRIES * 16
-+ .fill 512,8,0
-
- __PAGE_ALIGNED_BSS
- .align PAGE_SIZE
-diff -urNp linux-2.6.37.4/arch/x86/kernel/i386_ksyms_32.c linux-2.6.37.4/arch/x86/kernel/i386_ksyms_32.c
---- linux-2.6.37.4/arch/x86/kernel/i386_ksyms_32.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/i386_ksyms_32.c 2011-01-17 02:41:01.000000000 -0500
-@@ -20,8 +20,12 @@ extern void cmpxchg8b_emu(void);
- EXPORT_SYMBOL(cmpxchg8b_emu);
- #endif
-
-+EXPORT_SYMBOL_GPL(cpu_gdt_table);
-+
- /* Networking helper routines. */
- EXPORT_SYMBOL(csum_partial_copy_generic);
-+EXPORT_SYMBOL(csum_partial_copy_generic_to_user);
-+EXPORT_SYMBOL(csum_partial_copy_generic_from_user);
-
- EXPORT_SYMBOL(__get_user_1);
- EXPORT_SYMBOL(__get_user_2);
-@@ -36,3 +40,7 @@ EXPORT_SYMBOL(strstr);
-
- EXPORT_SYMBOL(csum_partial);
- EXPORT_SYMBOL(empty_zero_page);
-+
-+#ifdef CONFIG_PAX_KERNEXEC
-+EXPORT_SYMBOL(__LOAD_PHYSICAL_ADDR);
-+#endif
-diff -urNp linux-2.6.37.4/arch/x86/kernel/init_task.c linux-2.6.37.4/arch/x86/kernel/init_task.c
---- linux-2.6.37.4/arch/x86/kernel/init_task.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/init_task.c 2011-01-17 02:41:01.000000000 -0500
-@@ -38,5 +38,5 @@ EXPORT_SYMBOL(init_task);
- * section. Since TSS's are completely CPU-local, we want them
- * on exact cacheline boundaries, to eliminate cacheline ping-pong.
- */
--DEFINE_PER_CPU_SHARED_ALIGNED(struct tss_struct, init_tss) = INIT_TSS;
--
-+struct tss_struct init_tss[NR_CPUS] ____cacheline_internodealigned_in_smp = { [0 ... NR_CPUS-1] = INIT_TSS };
-+EXPORT_SYMBOL(init_tss);
-diff -urNp linux-2.6.37.4/arch/x86/kernel/ioport.c linux-2.6.37.4/arch/x86/kernel/ioport.c
---- linux-2.6.37.4/arch/x86/kernel/ioport.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/ioport.c 2011-01-17 02:41:01.000000000 -0500
-@@ -6,6 +6,7 @@
- #include <linux/sched.h>
- #include <linux/kernel.h>
- #include <linux/capability.h>
-+#include <linux/security.h>
- #include <linux/errno.h>
- #include <linux/types.h>
- #include <linux/ioport.h>
-@@ -41,6 +42,12 @@ asmlinkage long sys_ioperm(unsigned long
-
- if ((from + num <= from) || (from + num > IO_BITMAP_BITS))
- return -EINVAL;
-+#ifdef CONFIG_GRKERNSEC_IO
-+ if (turn_on && grsec_disable_privio) {
-+ gr_handle_ioperm();
-+ return -EPERM;
-+ }
-+#endif
- if (turn_on && !capable(CAP_SYS_RAWIO))
- return -EPERM;
-
-@@ -67,7 +74,7 @@ asmlinkage long sys_ioperm(unsigned long
- * because the ->io_bitmap_max value must match the bitmap
- * contents:
- */
-- tss = &per_cpu(init_tss, get_cpu());
-+ tss = init_tss + get_cpu();
-
- set_bitmap(t->io_bitmap_ptr, from, num, !turn_on);
-
-@@ -112,6 +119,12 @@ long sys_iopl(unsigned int level, struct
- return -EINVAL;
- /* Trying to gain more privileges? */
- if (level > old) {
-+#ifdef CONFIG_GRKERNSEC_IO
-+ if (grsec_disable_privio) {
-+ gr_handle_iopl();
-+ return -EPERM;
-+ }
-+#endif
- if (!capable(CAP_SYS_RAWIO))
- return -EPERM;
- }
-diff -urNp linux-2.6.37.4/arch/x86/kernel/irq_32.c linux-2.6.37.4/arch/x86/kernel/irq_32.c
---- linux-2.6.37.4/arch/x86/kernel/irq_32.c 2011-02-27 14:49:17.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/irq_32.c 2011-01-24 18:04:15.000000000 -0500
-@@ -91,7 +91,7 @@ execute_on_irq_stack(int overflow, struc
- return 0;
-
- /* build the stack frame on the IRQ stack */
-- isp = (u32 *) ((char *)irqctx + sizeof(*irqctx));
-+ isp = (u32 *) ((char *)irqctx + sizeof(*irqctx) - 8);
- irqctx->tinfo.task = curctx->tinfo.task;
- irqctx->tinfo.previous_esp = current_stack_pointer;
-
-@@ -103,6 +103,10 @@ execute_on_irq_stack(int overflow, struc
- (irqctx->tinfo.preempt_count & ~SOFTIRQ_MASK) |
- (curctx->tinfo.preempt_count & SOFTIRQ_MASK);
-
-+#ifdef CONFIG_PAX_MEMORY_UDEREF
-+ __set_fs(irqctx->tinfo.addr_limit);
-+#endif
-+
- if (unlikely(overflow))
- call_on_stack(print_stack_overflow, isp);
-
-@@ -113,6 +117,11 @@ execute_on_irq_stack(int overflow, struc
- : "0" (irq), "1" (desc), "2" (isp),
- "D" (desc->handle_irq)
- : "memory", "cc", "ecx");
-+
-+#ifdef CONFIG_PAX_MEMORY_UDEREF
-+ __set_fs(curctx->tinfo.addr_limit);
-+#endif
-+
- return 1;
- }
-
-@@ -168,9 +177,18 @@ asmlinkage void do_softirq(void)
- irqctx->tinfo.previous_esp = current_stack_pointer;
-
- /* build the stack frame on the softirq stack */
-- isp = (u32 *) ((char *)irqctx + sizeof(*irqctx));
-+ isp = (u32 *) ((char *)irqctx + sizeof(*irqctx) - 8);
-+
-+#ifdef CONFIG_PAX_MEMORY_UDEREF
-+ __set_fs(irqctx->tinfo.addr_limit);
-+#endif
-
- call_on_stack(__do_softirq, isp);
-+
-+#ifdef CONFIG_PAX_MEMORY_UDEREF
-+ __set_fs(curctx->addr_limit);
-+#endif
-+
- /*
- * Shouldnt happen, we returned above if in_interrupt():
- */
-diff -urNp linux-2.6.37.4/arch/x86/kernel/kgdb.c linux-2.6.37.4/arch/x86/kernel/kgdb.c
---- linux-2.6.37.4/arch/x86/kernel/kgdb.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/kgdb.c 2011-01-17 02:41:01.000000000 -0500
-@@ -123,11 +123,11 @@ char *dbg_get_reg(int regno, void *mem,
- switch (regno) {
- #ifdef CONFIG_X86_32
- case GDB_SS:
-- if (!user_mode_vm(regs))
-+ if (!user_mode(regs))
- *(unsigned long *)mem = __KERNEL_DS;
- break;
- case GDB_SP:
-- if (!user_mode_vm(regs))
-+ if (!user_mode(regs))
- *(unsigned long *)mem = kernel_stack_pointer(regs);
- break;
- case GDB_GS:
-@@ -722,7 +722,7 @@ void kgdb_arch_set_pc(struct pt_regs *re
- regs->ip = ip;
- }
-
--struct kgdb_arch arch_kgdb_ops = {
-+const struct kgdb_arch arch_kgdb_ops = {
- /* Breakpoint instruction: */
- .gdb_bpt_instr = { 0xcc },
- .flags = KGDB_HW_BREAKPOINT,
-diff -urNp linux-2.6.37.4/arch/x86/kernel/kprobes.c linux-2.6.37.4/arch/x86/kernel/kprobes.c
---- linux-2.6.37.4/arch/x86/kernel/kprobes.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/kprobes.c 2011-01-17 02:41:01.000000000 -0500
-@@ -115,8 +115,11 @@ static void __kprobes __synthesize_relat
- } __attribute__((packed)) *insn;
-
- insn = (struct __arch_relative_insn *)from;
-+
-+ pax_open_kernel();
- insn->raddr = (s32)((long)(to) - ((long)(from) + 5));
- insn->op = op;
-+ pax_close_kernel();
- }
-
- /* Insert a jump instruction at address 'from', which jumps to address 'to'.*/
-@@ -153,7 +156,7 @@ static int __kprobes can_boost(kprobe_op
- kprobe_opcode_t opcode;
- kprobe_opcode_t *orig_opcodes = opcodes;
-
-- if (search_exception_tables((unsigned long)opcodes))
-+ if (search_exception_tables(ktva_ktla((unsigned long)opcodes)))
- return 0; /* Page fault may occur on this address. */
-
- retry:
-@@ -314,7 +317,9 @@ static int __kprobes __copy_instruction(
- }
- }
- insn_get_length(&insn);
-+ pax_open_kernel();
- memcpy(dest, insn.kaddr, insn.length);
-+ pax_close_kernel();
-
- #ifdef CONFIG_X86_64
- if (insn_rip_relative(&insn)) {
-@@ -338,7 +343,9 @@ static int __kprobes __copy_instruction(
- (u8 *) dest;
- BUG_ON((s64) (s32) newdisp != newdisp); /* Sanity check. */
- disp = (u8 *) dest + insn_offset_displacement(&insn);
-+ pax_open_kernel();
- *(s32 *) disp = (s32) newdisp;
-+ pax_close_kernel();
- }
- #endif
- return insn.length;
-@@ -352,12 +359,12 @@ static void __kprobes arch_copy_kprobe(s
- */
- __copy_instruction(p->ainsn.insn, p->addr, 0);
-
-- if (can_boost(p->addr))
-+ if (can_boost(ktla_ktva(p->addr)))
- p->ainsn.boostable = 0;
- else
- p->ainsn.boostable = -1;
-
-- p->opcode = *p->addr;
-+ p->opcode = *(ktla_ktva(p->addr));
- }
-
- int __kprobes arch_prepare_kprobe(struct kprobe *p)
-@@ -474,7 +481,7 @@ static void __kprobes setup_singlestep(s
- * nor set current_kprobe, because it doesn't use single
- * stepping.
- */
-- regs->ip = (unsigned long)p->ainsn.insn;
-+ regs->ip = ktva_ktla((unsigned long)p->ainsn.insn);
- preempt_enable_no_resched();
- return;
- }
-@@ -493,7 +500,7 @@ static void __kprobes setup_singlestep(s
- if (p->opcode == BREAKPOINT_INSTRUCTION)
- regs->ip = (unsigned long)p->addr;
- else
-- regs->ip = (unsigned long)p->ainsn.insn;
-+ regs->ip = ktva_ktla((unsigned long)p->ainsn.insn);
- }
-
- /*
-@@ -572,7 +579,7 @@ static int __kprobes kprobe_handler(stru
- setup_singlestep(p, regs, kcb, 0);
- return 1;
- }
-- } else if (*addr != BREAKPOINT_INSTRUCTION) {
-+ } else if (*(kprobe_opcode_t *)ktla_ktva((unsigned long)addr) != BREAKPOINT_INSTRUCTION) {
- /*
- * The breakpoint instruction was removed right
- * after we hit it. Another cpu has removed
-@@ -817,7 +824,7 @@ static void __kprobes resume_execution(s
- struct pt_regs *regs, struct kprobe_ctlblk *kcb)
- {
- unsigned long *tos = stack_addr(regs);
-- unsigned long copy_ip = (unsigned long)p->ainsn.insn;
-+ unsigned long copy_ip = ktva_ktla((unsigned long)p->ainsn.insn);
- unsigned long orig_ip = (unsigned long)p->addr;
- kprobe_opcode_t *insn = p->ainsn.insn;
-
-@@ -999,7 +1006,7 @@ int __kprobes kprobe_exceptions_notify(s
- struct die_args *args = data;
- int ret = NOTIFY_DONE;
-
-- if (args->regs && user_mode_vm(args->regs))
-+ if (args->regs && user_mode(args->regs))
- return ret;
-
- switch (val) {
-@@ -1368,7 +1375,7 @@ int __kprobes arch_prepare_optimized_kpr
- * Verify if the address gap is in 2GB range, because this uses
- * a relative jump.
- */
-- rel = (long)op->optinsn.insn - (long)op->kp.addr + RELATIVEJUMP_SIZE;
-+ rel = (long)op->optinsn.insn - ktla_ktva((long)op->kp.addr) + RELATIVEJUMP_SIZE;
- if (abs(rel) > 0x7fffffff)
- return -ERANGE;
-
-@@ -1389,11 +1396,11 @@ int __kprobes arch_prepare_optimized_kpr
- synthesize_set_arg1(buf + TMPL_MOVE_IDX, (unsigned long)op);
-
- /* Set probe function call */
-- synthesize_relcall(buf + TMPL_CALL_IDX, optimized_callback);
-+ synthesize_relcall(buf + TMPL_CALL_IDX, ktla_ktva(optimized_callback));
-
- /* Set returning jmp instruction at the tail of out-of-line buffer */
- synthesize_reljump(buf + TMPL_END_IDX + op->optinsn.size,
-- (u8 *)op->kp.addr + op->optinsn.size);
-+ (u8 *)ktla_ktva(op->kp.addr) + op->optinsn.size);
-
- flush_icache_range((unsigned long) buf,
- (unsigned long) buf + TMPL_END_IDX +
-@@ -1409,7 +1416,7 @@ int __kprobes arch_optimize_kprobe(struc
- ((long)op->kp.addr + RELATIVEJUMP_SIZE));
-
- /* Backup instructions which will be replaced by jump address */
-- memcpy(op->optinsn.copied_insn, op->kp.addr + INT3_SIZE,
-+ memcpy(op->optinsn.copied_insn, ktla_ktva(op->kp.addr) + INT3_SIZE,
- RELATIVE_ADDR_SIZE);
-
- jmp_code[0] = RELATIVEJUMP_OPCODE;
-diff -urNp linux-2.6.37.4/arch/x86/kernel/ldt.c linux-2.6.37.4/arch/x86/kernel/ldt.c
---- linux-2.6.37.4/arch/x86/kernel/ldt.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/ldt.c 2011-01-17 02:41:01.000000000 -0500
-@@ -67,13 +67,13 @@ static int alloc_ldt(mm_context_t *pc, i
- if (reload) {
- #ifdef CONFIG_SMP
- preempt_disable();
-- load_LDT(pc);
-+ load_LDT_nolock(pc);
- if (!cpumask_equal(mm_cpumask(current->mm),
- cpumask_of(smp_processor_id())))
- smp_call_function(flush_ldt, current->mm, 1);
- preempt_enable();
- #else
-- load_LDT(pc);
-+ load_LDT_nolock(pc);
- #endif
- }
- if (oldsize) {
-@@ -95,7 +95,7 @@ static inline int copy_ldt(mm_context_t
- return err;
-
- for (i = 0; i < old->size; i++)
-- write_ldt_entry(new->ldt, i, old->ldt + i * LDT_ENTRY_SIZE);
-+ write_ldt_entry(new->ldt, i, old->ldt + i);
- return 0;
- }
-
-@@ -116,6 +116,24 @@ int init_new_context(struct task_struct
- retval = copy_ldt(&mm->context, &old_mm->context);
- mutex_unlock(&old_mm->context.lock);
- }
-+
-+ if (tsk == current) {
-+ mm->context.vdso = 0;
-+
-+#ifdef CONFIG_X86_32
-+#if defined(CONFIG_PAX_PAGEEXEC) || defined(CONFIG_PAX_SEGMEXEC)
-+ mm->context.user_cs_base = 0UL;
-+ mm->context.user_cs_limit = ~0UL;
-+
-+#if defined(CONFIG_PAX_PAGEEXEC) && defined(CONFIG_SMP)
-+ cpus_clear(mm->context.cpu_user_cs_mask);
-+#endif
-+
-+#endif
-+#endif
-+
-+ }
-+
- return retval;
- }
-
-@@ -230,6 +248,13 @@ static int write_ldt(void __user *ptr, u
- }
- }
-
-+#ifdef CONFIG_PAX_SEGMEXEC
-+ if ((mm->pax_flags & MF_PAX_SEGMEXEC) && (ldt_info.contents & MODIFY_LDT_CONTENTS_CODE)) {
-+ error = -EINVAL;
-+ goto out_unlock;
-+ }
-+#endif
-+
- fill_ldt(&ldt, &ldt_info);
- if (oldmode)
- ldt.avl = 0;
-diff -urNp linux-2.6.37.4/arch/x86/kernel/machine_kexec_32.c linux-2.6.37.4/arch/x86/kernel/machine_kexec_32.c
---- linux-2.6.37.4/arch/x86/kernel/machine_kexec_32.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/machine_kexec_32.c 2011-01-17 02:41:01.000000000 -0500
-@@ -27,7 +27,7 @@
- #include <asm/cacheflush.h>
- #include <asm/debugreg.h>
-
--static void set_idt(void *newidt, __u16 limit)
-+static void set_idt(struct desc_struct *newidt, __u16 limit)
- {
- struct desc_ptr curidt;
-
-@@ -39,7 +39,7 @@ static void set_idt(void *newidt, __u16
- }
-
-
--static void set_gdt(void *newgdt, __u16 limit)
-+static void set_gdt(struct desc_struct *newgdt, __u16 limit)
- {
- struct desc_ptr curgdt;
-
-@@ -217,7 +217,7 @@ void machine_kexec(struct kimage *image)
- }
-
- control_page = page_address(image->control_code_page);
-- memcpy(control_page, relocate_kernel, KEXEC_CONTROL_CODE_MAX_SIZE);
-+ memcpy(control_page, (void *)ktla_ktva((unsigned long)relocate_kernel), KEXEC_CONTROL_CODE_MAX_SIZE);
-
- relocate_kernel_ptr = control_page;
- page_list[PA_CONTROL_PAGE] = __pa(control_page);
-diff -urNp linux-2.6.37.4/arch/x86/kernel/microcode_amd.c linux-2.6.37.4/arch/x86/kernel/microcode_amd.c
---- linux-2.6.37.4/arch/x86/kernel/microcode_amd.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/microcode_amd.c 2011-01-17 02:41:01.000000000 -0500
-@@ -331,7 +331,7 @@ static void microcode_fini_cpu_amd(int c
- uci->mc = NULL;
- }
-
--static struct microcode_ops microcode_amd_ops = {
-+static const struct microcode_ops microcode_amd_ops = {
- .request_microcode_user = request_microcode_user,
- .request_microcode_fw = request_microcode_fw,
- .collect_cpu_info = collect_cpu_info_amd,
-@@ -339,7 +339,7 @@ static struct microcode_ops microcode_am
- .microcode_fini_cpu = microcode_fini_cpu_amd,
- };
-
--struct microcode_ops * __init init_amd_microcode(void)
-+const struct microcode_ops * __init init_amd_microcode(void)
- {
- return &microcode_amd_ops;
- }
-diff -urNp linux-2.6.37.4/arch/x86/kernel/microcode_core.c linux-2.6.37.4/arch/x86/kernel/microcode_core.c
---- linux-2.6.37.4/arch/x86/kernel/microcode_core.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/microcode_core.c 2011-01-17 02:41:01.000000000 -0500
-@@ -92,7 +92,7 @@ MODULE_LICENSE("GPL");
-
- #define MICROCODE_VERSION "2.00"
-
--static struct microcode_ops *microcode_ops;
-+static const struct microcode_ops *microcode_ops;
-
- /*
- * Synchronization.
-diff -urNp linux-2.6.37.4/arch/x86/kernel/microcode_intel.c linux-2.6.37.4/arch/x86/kernel/microcode_intel.c
---- linux-2.6.37.4/arch/x86/kernel/microcode_intel.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/microcode_intel.c 2011-01-17 02:41:01.000000000 -0500
-@@ -440,13 +440,13 @@ static enum ucode_state request_microcod
-
- static int get_ucode_user(void *to, const void *from, size_t n)
- {
-- return copy_from_user(to, from, n);
-+ return copy_from_user(to, (__force const void __user *)from, n);
- }
-
- static enum ucode_state
- request_microcode_user(int cpu, const void __user *buf, size_t size)
- {
-- return generic_load_microcode(cpu, (void *)buf, size, &get_ucode_user);
-+ return generic_load_microcode(cpu, (__force void *)buf, size, &get_ucode_user);
- }
-
- static void microcode_fini_cpu(int cpu)
-@@ -457,7 +457,7 @@ static void microcode_fini_cpu(int cpu)
- uci->mc = NULL;
- }
-
--static struct microcode_ops microcode_intel_ops = {
-+static const struct microcode_ops microcode_intel_ops = {
- .request_microcode_user = request_microcode_user,
- .request_microcode_fw = request_microcode_fw,
- .collect_cpu_info = collect_cpu_info,
-@@ -465,7 +465,7 @@ static struct microcode_ops microcode_in
- .microcode_fini_cpu = microcode_fini_cpu,
- };
-
--struct microcode_ops * __init init_intel_microcode(void)
-+const struct microcode_ops * __init init_intel_microcode(void)
- {
- return &microcode_intel_ops;
- }
-diff -urNp linux-2.6.37.4/arch/x86/kernel/module.c linux-2.6.37.4/arch/x86/kernel/module.c
---- linux-2.6.37.4/arch/x86/kernel/module.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/module.c 2011-01-17 02:41:01.000000000 -0500
-@@ -35,7 +35,7 @@
- #define DEBUGP(fmt...)
- #endif
-
--void *module_alloc(unsigned long size)
-+static void *__module_alloc(unsigned long size, pgprot_t prot)
- {
- struct vm_struct *area;
-
-@@ -49,8 +49,18 @@ void *module_alloc(unsigned long size)
- if (!area)
- return NULL;
-
-- return __vmalloc_area(area, GFP_KERNEL | __GFP_HIGHMEM,
-- PAGE_KERNEL_EXEC);
-+ return __vmalloc_area(area, GFP_KERNEL | __GFP_HIGHMEM | __GFP_ZERO, prot);
-+}
-+
-+void *module_alloc(unsigned long size)
-+{
-+
-+#ifdef CONFIG_PAX_KERNEXEC
-+ return __module_alloc(size, PAGE_KERNEL);
-+#else
-+ return __module_alloc(size, PAGE_KERNEL_EXEC);
-+#endif
-+
- }
-
- /* Free memory returned from module_alloc */
-@@ -59,6 +69,40 @@ void module_free(struct module *mod, voi
- vfree(module_region);
- }
-
-+#ifdef CONFIG_PAX_KERNEXEC
-+#ifdef CONFIG_X86_32
-+void *module_alloc_exec(unsigned long size)
-+{
-+ struct vm_struct *area;
-+
-+ if (size == 0)
-+ return NULL;
-+
-+ area = __get_vm_area(size, VM_ALLOC, (unsigned long)&MODULES_EXEC_VADDR, (unsigned long)&MODULES_EXEC_END);
-+ return area ? area->addr : NULL;
-+}
-+EXPORT_SYMBOL(module_alloc_exec);
-+
-+void module_free_exec(struct module *mod, void *module_region)
-+{
-+ vunmap(module_region);
-+}
-+EXPORT_SYMBOL(module_free_exec);
-+#else
-+void module_free_exec(struct module *mod, void *module_region)
-+{
-+ module_free(mod, module_region);
-+}
-+EXPORT_SYMBOL(module_free_exec);
-+
-+void *module_alloc_exec(unsigned long size)
-+{
-+ return __module_alloc(size, PAGE_KERNEL_RX);
-+}
-+EXPORT_SYMBOL(module_alloc_exec);
-+#endif
-+#endif
-+
- /* We don't need anything special. */
- int module_frob_arch_sections(Elf_Ehdr *hdr,
- Elf_Shdr *sechdrs,
-@@ -78,14 +122,16 @@ int apply_relocate(Elf32_Shdr *sechdrs,
- unsigned int i;
- Elf32_Rel *rel = (void *)sechdrs[relsec].sh_addr;
- Elf32_Sym *sym;
-- uint32_t *location;
-+ uint32_t *plocation, location;
-
- DEBUGP("Applying relocate section %u to %u\n", relsec,
- sechdrs[relsec].sh_info);
- for (i = 0; i < sechdrs[relsec].sh_size / sizeof(*rel); i++) {
- /* This is where to make the change */
-- location = (void *)sechdrs[sechdrs[relsec].sh_info].sh_addr
-- + rel[i].r_offset;
-+ plocation = (void *)sechdrs[sechdrs[relsec].sh_info].sh_addr + rel[i].r_offset;
-+ location = (uint32_t)plocation;
-+ if (sechdrs[sechdrs[relsec].sh_info].sh_flags & SHF_EXECINSTR)
-+ plocation = ktla_ktva((void *)plocation);
- /* This is the symbol it is referring to. Note that all
- undefined symbols have been resolved. */
- sym = (Elf32_Sym *)sechdrs[symindex].sh_addr
-@@ -94,11 +140,15 @@ int apply_relocate(Elf32_Shdr *sechdrs,
- switch (ELF32_R_TYPE(rel[i].r_info)) {
- case R_386_32:
- /* We add the value into the location given */
-- *location += sym->st_value;
-+ pax_open_kernel();
-+ *plocation += sym->st_value;
-+ pax_close_kernel();
- break;
- case R_386_PC32:
- /* Add the value, subtract its postition */
-- *location += sym->st_value - (uint32_t)location;
-+ pax_open_kernel();
-+ *plocation += sym->st_value - location;
-+ pax_close_kernel();
- break;
- default:
- printk(KERN_ERR "module %s: Unknown relocation: %u\n",
-@@ -154,21 +204,30 @@ int apply_relocate_add(Elf64_Shdr *sechd
- case R_X86_64_NONE:
- break;
- case R_X86_64_64:
-+ pax_open_kernel();
- *(u64 *)loc = val;
-+ pax_close_kernel();
- break;
- case R_X86_64_32:
-+ pax_open_kernel();
- *(u32 *)loc = val;
-+ pax_close_kernel();
- if (val != *(u32 *)loc)
- goto overflow;
- break;
- case R_X86_64_32S:
-+ pax_open_kernel();
- *(s32 *)loc = val;
-+ pax_close_kernel();
- if ((s64)val != *(s32 *)loc)
- goto overflow;
- break;
- case R_X86_64_PC32:
- val -= (u64)loc;
-+ pax_open_kernel();
- *(u32 *)loc = val;
-+ pax_close_kernel();
-+
- #if 0
- if ((s64)val != *(s32 *)loc)
- goto overflow;
-diff -urNp linux-2.6.37.4/arch/x86/kernel/paravirt.c linux-2.6.37.4/arch/x86/kernel/paravirt.c
---- linux-2.6.37.4/arch/x86/kernel/paravirt.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/paravirt.c 2011-01-17 02:41:01.000000000 -0500
-@@ -122,7 +122,7 @@ unsigned paravirt_patch_jmp(void *insnbu
- * corresponding structure. */
- static void *get_call_destination(u8 type)
- {
-- struct paravirt_patch_template tmpl = {
-+ const struct paravirt_patch_template tmpl = {
- .pv_init_ops = pv_init_ops,
- .pv_time_ops = pv_time_ops,
- .pv_cpu_ops = pv_cpu_ops,
-@@ -145,14 +145,14 @@ unsigned paravirt_patch_default(u8 type,
- if (opfunc == NULL)
- /* If there's no function, patch it with a ud2a (BUG) */
- ret = paravirt_patch_insns(insnbuf, len, ud2a, ud2a+sizeof(ud2a));
-- else if (opfunc == _paravirt_nop)
-+ else if (opfunc == (void *)_paravirt_nop)
- /* If the operation is a nop, then nop the callsite */
- ret = paravirt_patch_nop();
-
- /* identity functions just return their single argument */
-- else if (opfunc == _paravirt_ident_32)
-+ else if (opfunc == (void *)_paravirt_ident_32)
- ret = paravirt_patch_ident_32(insnbuf, len);
-- else if (opfunc == _paravirt_ident_64)
-+ else if (opfunc == (void *)_paravirt_ident_64)
- ret = paravirt_patch_ident_64(insnbuf, len);
-
- else if (type == PARAVIRT_PATCH(pv_cpu_ops.iret) ||
-@@ -178,7 +178,7 @@ unsigned paravirt_patch_insns(void *insn
- if (insn_len > len || start == NULL)
- insn_len = len;
- else
-- memcpy(insnbuf, start, insn_len);
-+ memcpy(insnbuf, ktla_ktva(start), insn_len);
-
- return insn_len;
- }
-@@ -294,22 +294,22 @@ void arch_flush_lazy_mmu_mode(void)
- preempt_enable();
- }
-
--struct pv_info pv_info = {
-+struct pv_info pv_info __read_only = {
- .name = "bare hardware",
- .paravirt_enabled = 0,
- .kernel_rpl = 0,
- .shared_kernel_pmd = 1, /* Only used when CONFIG_X86_PAE is set */
- };
-
--struct pv_init_ops pv_init_ops = {
-+struct pv_init_ops pv_init_ops __read_only = {
- .patch = native_patch,
- };
-
--struct pv_time_ops pv_time_ops = {
-+struct pv_time_ops pv_time_ops __read_only = {
- .sched_clock = native_sched_clock,
- };
-
--struct pv_irq_ops pv_irq_ops = {
-+struct pv_irq_ops pv_irq_ops __read_only = {
- .save_fl = __PV_IS_CALLEE_SAVE(native_save_fl),
- .restore_fl = __PV_IS_CALLEE_SAVE(native_restore_fl),
- .irq_disable = __PV_IS_CALLEE_SAVE(native_irq_disable),
-@@ -321,7 +321,7 @@ struct pv_irq_ops pv_irq_ops = {
- #endif
- };
-
--struct pv_cpu_ops pv_cpu_ops = {
-+struct pv_cpu_ops pv_cpu_ops __read_only = {
- .cpuid = native_cpuid,
- .get_debugreg = native_get_debugreg,
- .set_debugreg = native_set_debugreg,
-@@ -382,7 +382,7 @@ struct pv_cpu_ops pv_cpu_ops = {
- .end_context_switch = paravirt_nop,
- };
-
--struct pv_apic_ops pv_apic_ops = {
-+struct pv_apic_ops pv_apic_ops __read_only = {
- #ifdef CONFIG_X86_LOCAL_APIC
- .startup_ipi_hook = paravirt_nop,
- #endif
-@@ -396,7 +396,7 @@ struct pv_apic_ops pv_apic_ops = {
- #define PTE_IDENT __PV_IS_CALLEE_SAVE(_paravirt_ident_64)
- #endif
-
--struct pv_mmu_ops pv_mmu_ops = {
-+struct pv_mmu_ops pv_mmu_ops __read_only = {
-
- .read_cr2 = native_read_cr2,
- .write_cr2 = native_write_cr2,
-@@ -462,6 +462,12 @@ struct pv_mmu_ops pv_mmu_ops = {
- },
-
- .set_fixmap = native_set_fixmap,
-+
-+#ifdef CONFIG_PAX_KERNEXEC
-+ .pax_open_kernel = native_pax_open_kernel,
-+ .pax_close_kernel = native_pax_close_kernel,
-+#endif
-+
- };
-
- EXPORT_SYMBOL_GPL(pv_time_ops);
-diff -urNp linux-2.6.37.4/arch/x86/kernel/paravirt-spinlocks.c linux-2.6.37.4/arch/x86/kernel/paravirt-spinlocks.c
---- linux-2.6.37.4/arch/x86/kernel/paravirt-spinlocks.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/paravirt-spinlocks.c 2011-01-17 02:41:01.000000000 -0500
-@@ -13,7 +13,7 @@ default_spin_lock_flags(arch_spinlock_t
- arch_spin_lock(lock);
- }
-
--struct pv_lock_ops pv_lock_ops = {
-+struct pv_lock_ops pv_lock_ops __read_only = {
- #ifdef CONFIG_SMP
- .spin_is_locked = __ticket_spin_is_locked,
- .spin_is_contended = __ticket_spin_is_contended,
-diff -urNp linux-2.6.37.4/arch/x86/kernel/pci-calgary_64.c linux-2.6.37.4/arch/x86/kernel/pci-calgary_64.c
---- linux-2.6.37.4/arch/x86/kernel/pci-calgary_64.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/pci-calgary_64.c 2011-01-17 02:41:01.000000000 -0500
-@@ -476,7 +476,7 @@ static void calgary_free_coherent(struct
- free_pages((unsigned long)vaddr, get_order(size));
- }
-
--static struct dma_map_ops calgary_dma_ops = {
-+static const struct dma_map_ops calgary_dma_ops = {
- .alloc_coherent = calgary_alloc_coherent,
- .free_coherent = calgary_free_coherent,
- .map_sg = calgary_map_sg,
-diff -urNp linux-2.6.37.4/arch/x86/kernel/pci-dma.c linux-2.6.37.4/arch/x86/kernel/pci-dma.c
---- linux-2.6.37.4/arch/x86/kernel/pci-dma.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/pci-dma.c 2011-01-17 02:41:01.000000000 -0500
-@@ -16,7 +16,7 @@
-
- static int forbid_dac __read_mostly;
-
--struct dma_map_ops *dma_ops = &nommu_dma_ops;
-+const struct dma_map_ops *dma_ops = &nommu_dma_ops;
- EXPORT_SYMBOL(dma_ops);
-
- static int iommu_sac_force __read_mostly;
-@@ -250,7 +250,7 @@ early_param("iommu", iommu_setup);
-
- int dma_supported(struct device *dev, u64 mask)
- {
-- struct dma_map_ops *ops = get_dma_ops(dev);
-+ const struct dma_map_ops *ops = get_dma_ops(dev);
-
- #ifdef CONFIG_PCI
- if (mask > 0xffffffff && forbid_dac > 0) {
-diff -urNp linux-2.6.37.4/arch/x86/kernel/pci-gart_64.c linux-2.6.37.4/arch/x86/kernel/pci-gart_64.c
---- linux-2.6.37.4/arch/x86/kernel/pci-gart_64.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/pci-gart_64.c 2011-01-17 02:41:01.000000000 -0500
-@@ -706,7 +706,7 @@ static __init int init_k8_gatt(struct ag
- return -1;
- }
-
--static struct dma_map_ops gart_dma_ops = {
-+static const struct dma_map_ops gart_dma_ops = {
- .map_sg = gart_map_sg,
- .unmap_sg = gart_unmap_sg,
- .map_page = gart_map_page,
-diff -urNp linux-2.6.37.4/arch/x86/kernel/pci-nommu.c linux-2.6.37.4/arch/x86/kernel/pci-nommu.c
---- linux-2.6.37.4/arch/x86/kernel/pci-nommu.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/pci-nommu.c 2011-01-17 02:41:01.000000000 -0500
-@@ -95,7 +95,7 @@ static void nommu_sync_sg_for_device(str
- flush_write_buffers();
- }
-
--struct dma_map_ops nommu_dma_ops = {
-+const struct dma_map_ops nommu_dma_ops = {
- .alloc_coherent = dma_generic_alloc_coherent,
- .free_coherent = nommu_free_coherent,
- .map_sg = nommu_map_sg,
-diff -urNp linux-2.6.37.4/arch/x86/kernel/pci-swiotlb.c linux-2.6.37.4/arch/x86/kernel/pci-swiotlb.c
---- linux-2.6.37.4/arch/x86/kernel/pci-swiotlb.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/pci-swiotlb.c 2011-01-17 02:41:01.000000000 -0500
-@@ -26,7 +26,7 @@ static void *x86_swiotlb_alloc_coherent(
- return swiotlb_alloc_coherent(hwdev, size, dma_handle, flags);
- }
-
--static struct dma_map_ops swiotlb_dma_ops = {
-+static const struct dma_map_ops swiotlb_dma_ops = {
- .mapping_error = swiotlb_dma_mapping_error,
- .alloc_coherent = x86_swiotlb_alloc_coherent,
- .free_coherent = swiotlb_free_coherent,
-diff -urNp linux-2.6.37.4/arch/x86/kernel/process_32.c linux-2.6.37.4/arch/x86/kernel/process_32.c
---- linux-2.6.37.4/arch/x86/kernel/process_32.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/process_32.c 2011-01-17 02:41:01.000000000 -0500
-@@ -67,6 +67,7 @@ asmlinkage void ret_from_fork(void) __as
- unsigned long thread_saved_pc(struct task_struct *tsk)
- {
- return ((unsigned long *)tsk->thread.sp)[3];
-+//XXX return tsk->thread.eip;
- }
-
- #ifndef CONFIG_SMP
-@@ -130,15 +131,14 @@ void __show_regs(struct pt_regs *regs, i
- unsigned long sp;
- unsigned short ss, gs;
-
-- if (user_mode_vm(regs)) {
-+ if (user_mode(regs)) {
- sp = regs->sp;
- ss = regs->ss & 0xffff;
-- gs = get_user_gs(regs);
- } else {
- sp = kernel_stack_pointer(regs);
- savesegment(ss, ss);
-- savesegment(gs, gs);
- }
-+ gs = get_user_gs(regs);
-
- show_regs_common();
-
-@@ -200,7 +200,7 @@ int copy_thread(unsigned long clone_flag
- struct task_struct *tsk;
- int err;
-
-- childregs = task_pt_regs(p);
-+ childregs = task_stack_page(p) + THREAD_SIZE - sizeof(struct pt_regs) - 8;
- *childregs = *regs;
- childregs->ax = 0;
- childregs->sp = sp;
-@@ -297,7 +297,7 @@ __switch_to(struct task_struct *prev_p,
- struct thread_struct *prev = &prev_p->thread,
- *next = &next_p->thread;
- int cpu = smp_processor_id();
-- struct tss_struct *tss = &per_cpu(init_tss, cpu);
-+ struct tss_struct *tss = init_tss + cpu;
- bool preload_fpu;
-
- /* never put a printk in __switch_to... printk() calls wake_up*() indirectly */
-@@ -332,6 +332,10 @@ __switch_to(struct task_struct *prev_p,
- */
- lazy_save_gs(prev->gs);
-
-+#ifdef CONFIG_PAX_MEMORY_UDEREF
-+ __set_fs(task_thread_info(next_p)->addr_limit);
-+#endif
-+
- /*
- * Load the per-thread Thread-Local Storage descriptor.
- */
-@@ -408,3 +412,27 @@ unsigned long get_wchan(struct task_stru
- return 0;
- }
-
-+#ifdef CONFIG_PAX_RANDKSTACK
-+asmlinkage void pax_randomize_kstack(void)
-+{
-+ struct thread_struct *thread = &current->thread;
-+ unsigned long time;
-+
-+ if (!randomize_va_space)
-+ return;
-+
-+ rdtscl(time);
-+
-+ /* P4 seems to return a 0 LSB, ignore it */
-+#ifdef CONFIG_MPENTIUM4
-+ time &= 0x1EUL;
-+ time <<= 2;
-+#else
-+ time &= 0xFUL;
-+ time <<= 3;
-+#endif
-+
-+ thread->sp0 ^= time;
-+ load_sp0(init_tss + smp_processor_id(), thread);
-+}
-+#endif
-diff -urNp linux-2.6.37.4/arch/x86/kernel/process_64.c linux-2.6.37.4/arch/x86/kernel/process_64.c
---- linux-2.6.37.4/arch/x86/kernel/process_64.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/process_64.c 2011-01-17 02:41:01.000000000 -0500
-@@ -89,7 +89,7 @@ static void __exit_idle(void)
- void exit_idle(void)
- {
- /* idle loop has pid 0 */
-- if (current->pid)
-+ if (task_pid_nr(current))
- return;
- __exit_idle();
- }
-@@ -380,7 +380,7 @@ __switch_to(struct task_struct *prev_p,
- struct thread_struct *prev = &prev_p->thread;
- struct thread_struct *next = &next_p->thread;
- int cpu = smp_processor_id();
-- struct tss_struct *tss = &per_cpu(init_tss, cpu);
-+ struct tss_struct *tss = init_tss + cpu;
- unsigned fsindex, gsindex;
- bool preload_fpu;
-
-@@ -533,12 +533,11 @@ unsigned long get_wchan(struct task_stru
- if (!p || p == current || p->state == TASK_RUNNING)
- return 0;
- stack = (unsigned long)task_stack_page(p);
-- if (p->thread.sp < stack || p->thread.sp >= stack+THREAD_SIZE)
-+ if (p->thread.sp < stack || p->thread.sp > stack+THREAD_SIZE-8-sizeof(u64))
- return 0;
- fp = *(u64 *)(p->thread.sp);
- do {
-- if (fp < (unsigned long)stack ||
-- fp >= (unsigned long)stack+THREAD_SIZE)
-+ if (fp < stack || fp > stack+THREAD_SIZE-8-sizeof(u64))
- return 0;
- ip = *(u64 *)(fp+8);
- if (!in_sched_functions(ip))
-diff -urNp linux-2.6.37.4/arch/x86/kernel/process.c linux-2.6.37.4/arch/x86/kernel/process.c
---- linux-2.6.37.4/arch/x86/kernel/process.c 2011-02-27 14:49:17.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/process.c 2011-02-27 14:49:45.000000000 -0500
-@@ -75,7 +75,7 @@ void exit_thread(void)
- unsigned long *bp = t->io_bitmap_ptr;
-
- if (bp) {
-- struct tss_struct *tss = &per_cpu(init_tss, get_cpu());
-+ struct tss_struct *tss = init_tss + get_cpu();
-
- t->io_bitmap_ptr = NULL;
- clear_thread_flag(TIF_IO_BITMAP);
-@@ -103,6 +103,7 @@ void show_regs_common(void)
- vendor = dmi_get_system_info(DMI_SYS_VENDOR);
- if (!vendor)
- vendor = "";
-+
- product = dmi_get_system_info(DMI_PRODUCT_NAME);
- if (!product)
- product = "";
-@@ -111,8 +112,8 @@ void show_regs_common(void)
- board = dmi_get_system_info(DMI_BOARD_NAME);
-
- printk(KERN_CONT "\n");
-- printk(KERN_DEFAULT "Pid: %d, comm: %.20s %s %s %.*s",
-- current->pid, current->comm, print_tainted(),
-+ printk(KERN_DEFAULT "Pid: %d, comm: %.20s %s %s %.*s\n",
-+ task_pid_nr(current), current->comm, print_tainted(),
- init_utsname()->release,
- (int)strcspn(init_utsname()->version, " "),
- init_utsname()->version);
-@@ -129,6 +130,9 @@ void flush_thread(void)
- {
- struct task_struct *tsk = current;
-
-+#if defined(CONFIG_X86_32) && !defined(CONFIG_CC_STACKPROTECTOR) && !defined(CONFIG_PAX_MEMORY_UDEREF)
-+ loadsegment(gs, 0);
-+#endif
- flush_ptrace_hw_breakpoint(tsk);
- memset(tsk->thread.tls_array, 0, sizeof(tsk->thread.tls_array));
- /*
-@@ -291,10 +295,10 @@ int kernel_thread(int (*fn)(void *), voi
- regs.di = (unsigned long) arg;
-
- #ifdef CONFIG_X86_32
-- regs.ds = __USER_DS;
-- regs.es = __USER_DS;
-+ regs.ds = __KERNEL_DS;
-+ regs.es = __KERNEL_DS;
- regs.fs = __KERNEL_PERCPU;
-- regs.gs = __KERNEL_STACK_CANARY;
-+ savesegment(gs, regs.gs);
- #else
- regs.ss = __KERNEL_DS;
- #endif
-@@ -669,17 +673,3 @@ static int __init idle_setup(char *str)
- return 0;
- }
- early_param("idle", idle_setup);
--
--unsigned long arch_align_stack(unsigned long sp)
--{
-- if (!(current->personality & ADDR_NO_RANDOMIZE) && randomize_va_space)
-- sp -= get_random_int() % 8192;
-- return sp & ~0xf;
--}
--
--unsigned long arch_randomize_brk(struct mm_struct *mm)
--{
-- unsigned long range_end = mm->brk + 0x02000000;
-- return randomize_range(mm->brk, range_end, 0) ? : mm->brk;
--}
--
-diff -urNp linux-2.6.37.4/arch/x86/kernel/ptrace.c linux-2.6.37.4/arch/x86/kernel/ptrace.c
---- linux-2.6.37.4/arch/x86/kernel/ptrace.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/ptrace.c 2011-03-21 17:16:23.000000000 -0400
-@@ -805,7 +805,7 @@ long arch_ptrace(struct task_struct *chi
- unsigned long addr, unsigned long data)
- {
- int ret;
-- unsigned long __user *datap = (unsigned long __user *)data;
-+ unsigned long __user *datap = (__force unsigned long __user *)data;
-
- switch (request) {
- /* read the word at location addr in the USER area. */
-@@ -890,14 +890,14 @@ long arch_ptrace(struct task_struct *chi
- if ((int) addr < 0)
- return -EIO;
- ret = do_get_thread_area(child, addr,
-- (struct user_desc __user *)data);
-+ (__force struct user_desc __user *) data);
- break;
-
- case PTRACE_SET_THREAD_AREA:
- if ((int) addr < 0)
- return -EIO;
- ret = do_set_thread_area(child, addr,
-- (struct user_desc __user *)data, 0);
-+ (__force struct user_desc __user *) data, 0);
- break;
- #endif
-
-@@ -1314,7 +1314,7 @@ static void fill_sigtrap_info(struct tas
- memset(info, 0, sizeof(*info));
- info->si_signo = SIGTRAP;
- info->si_code = si_code;
-- info->si_addr = user_mode_vm(regs) ? (void __user *)regs->ip : NULL;
-+ info->si_addr = user_mode(regs) ? (__force void __user *)regs->ip : NULL;
- }
-
- void user_single_step_siginfo(struct task_struct *tsk,
-@@ -1347,7 +1347,7 @@ void send_sigtrap(struct task_struct *ts
- * We must return the syscall number to actually look up in the table.
- * This can be -1L to skip running any syscall at all.
- */
--asmregparm long syscall_trace_enter(struct pt_regs *regs)
-+long syscall_trace_enter(struct pt_regs *regs)
- {
- long ret = 0;
-
-@@ -1392,7 +1392,7 @@ asmregparm long syscall_trace_enter(stru
- return ret ?: regs->orig_ax;
- }
-
--asmregparm void syscall_trace_leave(struct pt_regs *regs)
-+void syscall_trace_leave(struct pt_regs *regs)
- {
- bool step;
-
-diff -urNp linux-2.6.37.4/arch/x86/kernel/reboot.c linux-2.6.37.4/arch/x86/kernel/reboot.c
---- linux-2.6.37.4/arch/x86/kernel/reboot.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/reboot.c 2011-01-17 02:41:01.000000000 -0500
-@@ -33,7 +33,7 @@ void (*pm_power_off)(void);
- EXPORT_SYMBOL(pm_power_off);
-
- static const struct desc_ptr no_idt = {};
--static int reboot_mode;
-+static unsigned short reboot_mode;
- enum reboot_type reboot_type = BOOT_KBD;
- int reboot_force;
-
-@@ -284,7 +284,7 @@ static struct dmi_system_id __initdata r
- DMI_MATCH(DMI_BOARD_NAME, "P4S800"),
- },
- },
-- { }
-+ { NULL, NULL, {{0, {0}}}, NULL}
- };
-
- static int __init reboot_init(void)
-@@ -300,12 +300,12 @@ core_initcall(reboot_init);
- controller to pulse the CPU reset line, which is more thorough, but
- doesn't work with at least one type of 486 motherboard. It is easy
- to stop this code working; hence the copious comments. */
--static const unsigned long long
--real_mode_gdt_entries [3] =
-+static struct desc_struct
-+real_mode_gdt_entries [3] __read_only =
- {
-- 0x0000000000000000ULL, /* Null descriptor */
-- 0x00009b000000ffffULL, /* 16-bit real-mode 64k code at 0x00000000 */
-- 0x000093000100ffffULL /* 16-bit real-mode 64k data at 0x00000100 */
-+ GDT_ENTRY_INIT(0, 0, 0), /* Null descriptor */
-+ GDT_ENTRY_INIT(0x9b, 0, 0xffff), /* 16-bit real-mode 64k code at 0x00000000 */
-+ GDT_ENTRY_INIT(0x93, 0x100, 0xffff) /* 16-bit real-mode 64k data at 0x00000100 */
- };
-
- static const struct desc_ptr
-@@ -354,7 +354,7 @@ static const unsigned char jump_to_bios
- * specified by the code and length parameters.
- * We assume that length will aways be less that 100!
- */
--void machine_real_restart(const unsigned char *code, int length)
-+void machine_real_restart(const unsigned char *code, unsigned int length)
- {
- local_irq_disable();
-
-@@ -381,16 +381,15 @@ void machine_real_restart(const unsigned
- boot)". This seems like a fairly standard thing that gets set by
- REBOOT.COM programs, and the previous reset routine did this
- too. */
-- *((unsigned short *)0x472) = reboot_mode;
-+ *(unsigned short *)(__va(0x472)) = reboot_mode;
-
- /* For the switch to real mode, copy some code to low memory. It has
- to be in the first 64k because it is running in 16-bit mode, and it
- has to have the same physical and virtual address, because it turns
- off paging. Copy it near the end of the first page, out of the way
- of BIOS variables. */
-- memcpy((void *)(0x1000 - sizeof(real_mode_switch) - 100),
-- real_mode_switch, sizeof (real_mode_switch));
-- memcpy((void *)(0x1000 - 100), code, length);
-+ memcpy(__va(0x1000 - sizeof (real_mode_switch) - 100), real_mode_switch, sizeof (real_mode_switch));
-+ memcpy(__va(0x1000 - 100), code, length);
-
- /* Set up the IDT for real mode. */
- load_idt(&real_mode_idt);
-diff -urNp linux-2.6.37.4/arch/x86/kernel/setup.c linux-2.6.37.4/arch/x86/kernel/setup.c
---- linux-2.6.37.4/arch/x86/kernel/setup.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/setup.c 2011-01-17 02:41:01.000000000 -0500
-@@ -654,7 +654,7 @@ static void __init trim_bios_range(void)
- * area (640->1Mb) as ram even though it is not.
- * take them out.
- */
-- e820_remove_range(BIOS_BEGIN, BIOS_END - BIOS_BEGIN, E820_RAM, 1);
-+ e820_remove_range(ISA_START_ADDRESS, ISA_END_ADDRESS - ISA_START_ADDRESS, E820_RAM, 1);
- sanitize_e820_map(e820.map, ARRAY_SIZE(e820.map), &e820.nr_map);
- }
-
-@@ -790,14 +790,14 @@ void __init setup_arch(char **cmdline_p)
-
- if (!boot_params.hdr.root_flags)
- root_mountflags &= ~MS_RDONLY;
-- init_mm.start_code = (unsigned long) _text;
-- init_mm.end_code = (unsigned long) _etext;
-+ init_mm.start_code = ktla_ktva((unsigned long) _text);
-+ init_mm.end_code = ktla_ktva((unsigned long) _etext);
- init_mm.end_data = (unsigned long) _edata;
- init_mm.brk = _brk_end;
-
-- code_resource.start = virt_to_phys(_text);
-- code_resource.end = virt_to_phys(_etext)-1;
-- data_resource.start = virt_to_phys(_etext);
-+ code_resource.start = virt_to_phys(ktla_ktva(_text));
-+ code_resource.end = virt_to_phys(ktla_ktva(_etext))-1;
-+ data_resource.start = virt_to_phys(_sdata);
- data_resource.end = virt_to_phys(_edata)-1;
- bss_resource.start = virt_to_phys(&__bss_start);
- bss_resource.end = virt_to_phys(&__bss_stop)-1;
-diff -urNp linux-2.6.37.4/arch/x86/kernel/setup_percpu.c linux-2.6.37.4/arch/x86/kernel/setup_percpu.c
---- linux-2.6.37.4/arch/x86/kernel/setup_percpu.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/setup_percpu.c 2011-01-17 02:41:01.000000000 -0500
-@@ -21,19 +21,17 @@
- #include <asm/cpu.h>
- #include <asm/stackprotector.h>
-
--DEFINE_PER_CPU(int, cpu_number);
-+#ifdef CONFIG_SMP
-+DEFINE_PER_CPU(unsigned int, cpu_number);
- EXPORT_PER_CPU_SYMBOL(cpu_number);
-+#endif
-
--#ifdef CONFIG_X86_64
- #define BOOT_PERCPU_OFFSET ((unsigned long)__per_cpu_load)
--#else
--#define BOOT_PERCPU_OFFSET 0
--#endif
-
- DEFINE_PER_CPU(unsigned long, this_cpu_off) = BOOT_PERCPU_OFFSET;
- EXPORT_PER_CPU_SYMBOL(this_cpu_off);
-
--unsigned long __per_cpu_offset[NR_CPUS] __read_mostly = {
-+unsigned long __per_cpu_offset[NR_CPUS] __read_only = {
- [0 ... NR_CPUS-1] = BOOT_PERCPU_OFFSET,
- };
- EXPORT_SYMBOL(__per_cpu_offset);
-@@ -155,10 +153,10 @@ static inline void setup_percpu_segment(
- {
- #ifdef CONFIG_X86_32
- struct desc_struct gdt;
-+ unsigned long base = per_cpu_offset(cpu);
-
-- pack_descriptor(&gdt, per_cpu_offset(cpu), 0xFFFFF,
-- 0x2 | DESCTYPE_S, 0x8);
-- gdt.s = 1;
-+ pack_descriptor(&gdt, base, (VMALLOC_END - base - 1) >> PAGE_SHIFT,
-+ 0x83 | DESCTYPE_S, 0xC);
- write_gdt_entry(get_cpu_gdt_table(cpu),
- GDT_ENTRY_PERCPU, &gdt, DESCTYPE_S);
- #endif
-@@ -207,6 +205,11 @@ void __init setup_per_cpu_areas(void)
- /* alrighty, percpu areas up and running */
- delta = (unsigned long)pcpu_base_addr - (unsigned long)__per_cpu_start;
- for_each_possible_cpu(cpu) {
-+#ifdef CONFIG_CC_STACKPROTECTOR
-+#ifdef CONFIG_x86_32
-+ unsigned long canary = per_cpu(stack_canary, cpu);
-+#endif
-+#endif
- per_cpu_offset(cpu) = delta + pcpu_unit_offsets[cpu];
- per_cpu(this_cpu_off, cpu) = per_cpu_offset(cpu);
- per_cpu(cpu_number, cpu) = cpu;
-@@ -243,6 +246,12 @@ void __init setup_per_cpu_areas(void)
- set_cpu_numa_node(cpu, early_cpu_to_node(cpu));
- #endif
- #endif
-+#ifdef CONFIG_CC_STACKPROTECTOR
-+#ifdef CONFIG_x86_32
-+ if (cpu == boot_cpu_id)
-+ per_cpu(stack_canary, cpu) = canary;
-+#endif
-+#endif
- /*
- * Up to this point, the boot CPU has been using .init.data
- * area. Reload any changed state for the boot CPU.
-diff -urNp linux-2.6.37.4/arch/x86/kernel/signal.c linux-2.6.37.4/arch/x86/kernel/signal.c
---- linux-2.6.37.4/arch/x86/kernel/signal.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/signal.c 2011-01-17 02:41:01.000000000 -0500
-@@ -198,7 +198,7 @@ static unsigned long align_sigframe(unsi
- * Align the stack pointer according to the i386 ABI,
- * i.e. so that on function entry ((sp + 4) & 15) == 0.
- */
-- sp = ((sp + 4) & -16ul) - 4;
-+ sp = ((sp - 12) & -16ul) - 4;
- #else /* !CONFIG_X86_32 */
- sp = round_down(sp, 16) - 8;
- #endif
-@@ -249,11 +249,11 @@ get_sigframe(struct k_sigaction *ka, str
- * Return an always-bogus address instead so we will die with SIGSEGV.
- */
- if (onsigstack && !likely(on_sig_stack(sp)))
-- return (void __user *)-1L;
-+ return (__force void __user *)-1L;
-
- /* save i387 state */
- if (used_math() && save_i387_xstate(*fpstate) < 0)
-- return (void __user *)-1L;
-+ return (__force void __user *)-1L;
-
- return (void __user *)sp;
- }
-@@ -308,9 +308,9 @@ __setup_frame(int sig, struct k_sigactio
- }
-
- if (current->mm->context.vdso)
-- restorer = VDSO32_SYMBOL(current->mm->context.vdso, sigreturn);
-+ restorer = (__force void __user *)VDSO32_SYMBOL(current->mm->context.vdso, sigreturn);
- else
-- restorer = &frame->retcode;
-+ restorer = (void __user *)&frame->retcode;
- if (ka->sa.sa_flags & SA_RESTORER)
- restorer = ka->sa.sa_restorer;
-
-@@ -324,7 +324,7 @@ __setup_frame(int sig, struct k_sigactio
- * reasons and because gdb uses it as a signature to notice
- * signal handler stack frames.
- */
-- err |= __put_user(*((u64 *)&retcode), (u64 *)frame->retcode);
-+ err |= __put_user(*((u64 *)&retcode), (u64 __user *)frame->retcode);
-
- if (err)
- return -EFAULT;
-@@ -378,7 +378,10 @@ static int __setup_rt_frame(int sig, str
- err |= __copy_to_user(&frame->uc.uc_sigmask, set, sizeof(*set));
-
- /* Set up to return from userspace. */
-- restorer = VDSO32_SYMBOL(current->mm->context.vdso, rt_sigreturn);
-+ if (current->mm->context.vdso)
-+ restorer = (__force void __user *)VDSO32_SYMBOL(current->mm->context.vdso, rt_sigreturn);
-+ else
-+ restorer = (void __user *)&frame->retcode;
- if (ka->sa.sa_flags & SA_RESTORER)
- restorer = ka->sa.sa_restorer;
- put_user_ex(restorer, &frame->pretcode);
-@@ -390,7 +393,7 @@ static int __setup_rt_frame(int sig, str
- * reasons and because gdb uses it as a signature to notice
- * signal handler stack frames.
- */
-- put_user_ex(*((u64 *)&rt_retcode), (u64 *)frame->retcode);
-+ put_user_ex(*((u64 *)&rt_retcode), (u64 __user *)frame->retcode);
- } put_user_catch(err);
-
- if (err)
-@@ -780,7 +783,7 @@ static void do_signal(struct pt_regs *re
- * X86_32: vm86 regs switched out by assembly code before reaching
- * here, so testing against kernel CS suffices.
- */
-- if (!user_mode(regs))
-+ if (!user_mode_novm(regs))
- return;
-
- if (current_thread_info()->status & TS_RESTORE_SIGMASK)
-diff -urNp linux-2.6.37.4/arch/x86/kernel/smpboot.c linux-2.6.37.4/arch/x86/kernel/smpboot.c
---- linux-2.6.37.4/arch/x86/kernel/smpboot.c 2011-02-27 14:49:17.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/smpboot.c 2011-02-27 14:49:45.000000000 -0500
-@@ -786,7 +786,11 @@ do_rest:
- (unsigned long)task_stack_page(c_idle.idle) -
- KERNEL_STACK_OFFSET + THREAD_SIZE;
- #endif
-+
-+ pax_open_kernel();
- early_gdt_descr.address = (unsigned long)get_cpu_gdt_table(cpu);
-+ pax_close_kernel();
-+
- initial_code = (unsigned long)start_secondary;
- stack_start.sp = (void *) c_idle.idle->thread.sp;
-
-@@ -926,6 +930,12 @@ int __cpuinit native_cpu_up(unsigned int
-
- per_cpu(cpu_state, cpu) = CPU_UP_PREPARE;
-
-+#ifdef CONFIG_PAX_PER_CPU_PGD
-+ clone_pgd_range(get_cpu_pgd(cpu) + KERNEL_PGD_BOUNDARY,
-+ swapper_pg_dir + KERNEL_PGD_BOUNDARY,
-+ KERNEL_PGD_PTRS);
-+#endif
-+
- err = do_boot_cpu(apicid, cpu);
- if (err) {
- pr_debug("do_boot_cpu failed %d\n", err);
-diff -urNp linux-2.6.37.4/arch/x86/kernel/step.c linux-2.6.37.4/arch/x86/kernel/step.c
---- linux-2.6.37.4/arch/x86/kernel/step.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/step.c 2011-01-17 02:41:01.000000000 -0500
-@@ -27,10 +27,10 @@ unsigned long convert_ip_to_linear(struc
- struct desc_struct *desc;
- unsigned long base;
-
-- seg &= ~7UL;
-+ seg >>= 3;
-
- mutex_lock(&child->mm->context.lock);
-- if (unlikely((seg >> 3) >= child->mm->context.size))
-+ if (unlikely(seg >= child->mm->context.size))
- addr = -1L; /* bogus selector, access would fault */
- else {
- desc = child->mm->context.ldt + seg;
-@@ -42,7 +42,8 @@ unsigned long convert_ip_to_linear(struc
- addr += base;
- }
- mutex_unlock(&child->mm->context.lock);
-- }
-+ } else if (seg == __KERNEL_CS || seg == __KERNEXEC_KERNEL_CS)
-+ addr = ktla_ktva(addr);
-
- return addr;
- }
-@@ -53,6 +54,9 @@ static int is_setting_trap_flag(struct t
- unsigned char opcode[15];
- unsigned long addr = convert_ip_to_linear(child, regs);
-
-+ if (addr == -EINVAL)
-+ return 0;
-+
- copied = access_process_vm(child, addr, opcode, sizeof(opcode), 0);
- for (i = 0; i < copied; i++) {
- switch (opcode[i]) {
-@@ -74,7 +78,7 @@ static int is_setting_trap_flag(struct t
-
- #ifdef CONFIG_X86_64
- case 0x40 ... 0x4f:
-- if (regs->cs != __USER_CS)
-+ if ((regs->cs & 0xffff) != __USER_CS)
- /* 32-bit mode: register increment */
- return 0;
- /* 64-bit mode: REX prefix */
-diff -urNp linux-2.6.37.4/arch/x86/kernel/syscall_table_32.S linux-2.6.37.4/arch/x86/kernel/syscall_table_32.S
---- linux-2.6.37.4/arch/x86/kernel/syscall_table_32.S 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/syscall_table_32.S 2011-01-17 02:41:01.000000000 -0500
-@@ -1,3 +1,4 @@
-+.section .rodata,"a",@progbits
- ENTRY(sys_call_table)
- .long sys_restart_syscall /* 0 - old "setup()" system call, used for restarting */
- .long sys_exit
-diff -urNp linux-2.6.37.4/arch/x86/kernel/sys_i386_32.c linux-2.6.37.4/arch/x86/kernel/sys_i386_32.c
---- linux-2.6.37.4/arch/x86/kernel/sys_i386_32.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/sys_i386_32.c 2011-03-21 17:35:08.000000000 -0400
-@@ -24,17 +24,224 @@
-
- #include <asm/syscalls.h>
-
--/*
-- * Do a system call from kernel instead of calling sys_execve so we
-- * end up with proper pt_regs.
-- */
--int kernel_execve(const char *filename,
-- const char *const argv[],
-- const char *const envp[])
-+int i386_mmap_check(unsigned long addr, unsigned long len, unsigned long flags)
- {
-- long __res;
-- asm volatile ("int $0x80"
-- : "=a" (__res)
-- : "0" (__NR_execve), "b" (filename), "c" (argv), "d" (envp) : "memory");
-- return __res;
-+ unsigned long pax_task_size = TASK_SIZE;
-+
-+#ifdef CONFIG_PAX_SEGMEXEC
-+ if (current->mm->pax_flags & MF_PAX_SEGMEXEC)
-+ pax_task_size = SEGMEXEC_TASK_SIZE;
-+#endif
-+
-+ if (len > pax_task_size || addr > pax_task_size - len)
-+ return -EINVAL;
-+
-+ return 0;
-+}
-+
-+unsigned long
-+arch_get_unmapped_area(struct file *filp, unsigned long addr,
-+ unsigned long len, unsigned long pgoff, unsigned long flags)
-+{
-+ struct mm_struct *mm = current->mm;
-+ struct vm_area_struct *vma;
-+ unsigned long start_addr, pax_task_size = TASK_SIZE;
-+
-+#ifdef CONFIG_PAX_SEGMEXEC
-+ if (mm->pax_flags & MF_PAX_SEGMEXEC)
-+ pax_task_size = SEGMEXEC_TASK_SIZE;
-+#endif
-+
-+ pax_task_size -= PAGE_SIZE;
-+
-+ if (len > pax_task_size)
-+ return -ENOMEM;
-+
-+ if (flags & MAP_FIXED)
-+ return addr;
-+
-+#ifdef CONFIG_PAX_RANDMMAP
-+ if (!(mm->pax_flags & MF_PAX_RANDMMAP))
-+#endif
-+
-+ if (addr) {
-+ addr = PAGE_ALIGN(addr);
-+ if (pax_task_size - len >= addr) {
-+ vma = find_vma(mm, addr);
-+ if (check_heap_stack_gap(vma, addr, len))
-+ return addr;
-+ }
-+ }
-+ if (len > mm->cached_hole_size) {
-+ start_addr = addr = mm->free_area_cache;
-+ } else {
-+ start_addr = addr = mm->mmap_base;
-+ mm->cached_hole_size = 0;
-+ }
-+
-+#ifdef CONFIG_PAX_PAGEEXEC
-+ if (!(__supported_pte_mask & _PAGE_NX) && (mm->pax_flags & MF_PAX_PAGEEXEC) && (flags & MAP_EXECUTABLE) && start_addr >= mm->mmap_base) {
-+ start_addr = 0x00110000UL;
-+
-+#ifdef CONFIG_PAX_RANDMMAP
-+ if (mm->pax_flags & MF_PAX_RANDMMAP)
-+ start_addr += mm->delta_mmap & 0x03FFF000UL;
-+#endif
-+
-+ if (mm->start_brk <= start_addr && start_addr < mm->mmap_base)
-+ start_addr = addr = mm->mmap_base;
-+ else
-+ addr = start_addr;
-+ }
-+#endif
-+
-+full_search:
-+ for (vma = find_vma(mm, addr); ; vma = vma->vm_next) {
-+ /* At this point: (!vma || addr < vma->vm_end). */
-+ if (pax_task_size - len < addr) {
-+ /*
-+ * Start a new search - just in case we missed
-+ * some holes.
-+ */
-+ if (start_addr != mm->mmap_base) {
-+ start_addr = addr = mm->mmap_base;
-+ mm->cached_hole_size = 0;
-+ goto full_search;
-+ }
-+ return -ENOMEM;
-+ }
-+ if (check_heap_stack_gap(vma, addr, len))
-+ break;
-+ if (addr + mm->cached_hole_size < vma->vm_start)
-+ mm->cached_hole_size = vma->vm_start - addr;
-+ addr = vma->vm_end;
-+ if (mm->start_brk <= addr && addr < mm->mmap_base) {
-+ start_addr = addr = mm->mmap_base;
-+ mm->cached_hole_size = 0;
-+ goto full_search;
-+ }
-+ }
-+
-+ /*
-+ * Remember the place where we stopped the search:
-+ */
-+ mm->free_area_cache = addr + len;
-+ return addr;
-+}
-+
-+unsigned long
-+arch_get_unmapped_area_topdown(struct file *filp, const unsigned long addr0,
-+ const unsigned long len, const unsigned long pgoff,
-+ const unsigned long flags)
-+{
-+ struct vm_area_struct *vma;
-+ struct mm_struct *mm = current->mm;
-+ unsigned long base = mm->mmap_base, addr = addr0, pax_task_size = TASK_SIZE;
-+
-+#ifdef CONFIG_PAX_SEGMEXEC
-+ if (mm->pax_flags & MF_PAX_SEGMEXEC)
-+ pax_task_size = SEGMEXEC_TASK_SIZE;
-+#endif
-+
-+ pax_task_size -= PAGE_SIZE;
-+
-+ /* requested length too big for entire address space */
-+ if (len > pax_task_size)
-+ return -ENOMEM;
-+
-+ if (flags & MAP_FIXED)
-+ return addr;
-+
-+#ifdef CONFIG_PAX_PAGEEXEC
-+ if (!(__supported_pte_mask & _PAGE_NX) && (mm->pax_flags & MF_PAX_PAGEEXEC) && (flags & MAP_EXECUTABLE))
-+ goto bottomup;
-+#endif
-+
-+#ifdef CONFIG_PAX_RANDMMAP
-+ if (!(mm->pax_flags & MF_PAX_RANDMMAP))
-+#endif
-+
-+ /* requesting a specific address */
-+ if (addr) {
-+ addr = PAGE_ALIGN(addr);
-+ if (pax_task_size - len >= addr) {
-+ vma = find_vma(mm, addr);
-+ if (check_heap_stack_gap(vma, addr, len))
-+ return addr;
-+ }
-+ }
-+
-+ /* check if free_area_cache is useful for us */
-+ if (len <= mm->cached_hole_size) {
-+ mm->cached_hole_size = 0;
-+ mm->free_area_cache = mm->mmap_base;
-+ }
-+
-+ /* either no address requested or can't fit in requested address hole */
-+ addr = mm->free_area_cache;
-+
-+ /* make sure it can fit in the remaining address space */
-+ if (addr > len) {
-+ vma = find_vma(mm, addr-len);
-+ if (check_heap_stack_gap(vma, addr - len, len))
-+ /* remember the address as a hint for next time */
-+ return (mm->free_area_cache = addr-len);
-+ }
-+
-+ if (mm->mmap_base < len)
-+ goto bottomup;
-+
-+ addr = mm->mmap_base-len;
-+
-+ do {
-+ /*
-+ * Lookup failure means no vma is above this address,
-+ * else if new region fits below vma->vm_start,
-+ * return with success:
-+ */
-+ vma = find_vma(mm, addr);
-+ if (check_heap_stack_gap(vma, addr, len))
-+ /* remember the address as a hint for next time */
-+ return (mm->free_area_cache = addr);
-+
-+ /* remember the largest hole we saw so far */
-+ if (addr + mm->cached_hole_size < vma->vm_start)
-+ mm->cached_hole_size = vma->vm_start - addr;
-+
-+ /* try just below the current vma->vm_start */
-+ addr = skip_heap_stack_gap(vma, len);
-+ } while (!IS_ERR_VALUE(addr));
-+
-+bottomup:
-+ /*
-+ * A failed mmap() very likely causes application failure,
-+ * so fall back to the bottom-up function here. This scenario
-+ * can happen with large stack limits and large mmap()
-+ * allocations.
-+ */
-+
-+#ifdef CONFIG_PAX_SEGMEXEC
-+ if (mm->pax_flags & MF_PAX_SEGMEXEC)
-+ mm->mmap_base = SEGMEXEC_TASK_UNMAPPED_BASE;
-+ else
-+#endif
-+
-+ mm->mmap_base = TASK_UNMAPPED_BASE;
-+
-+#ifdef CONFIG_PAX_RANDMMAP
-+ if (mm->pax_flags & MF_PAX_RANDMMAP)
-+ mm->mmap_base += mm->delta_mmap;
-+#endif
-+
-+ mm->free_area_cache = mm->mmap_base;
-+ mm->cached_hole_size = ~0UL;
-+ addr = arch_get_unmapped_area(filp, addr0, len, pgoff, flags);
-+ /*
-+ * Restore the topdown base:
-+ */
-+ mm->mmap_base = base;
-+ mm->free_area_cache = base;
-+ mm->cached_hole_size = ~0UL;
-+
-+ return addr;
- }
-diff -urNp linux-2.6.37.4/arch/x86/kernel/sys_x86_64.c linux-2.6.37.4/arch/x86/kernel/sys_x86_64.c
---- linux-2.6.37.4/arch/x86/kernel/sys_x86_64.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/sys_x86_64.c 2011-03-21 17:35:08.000000000 -0400
-@@ -32,8 +32,8 @@ out:
- return error;
- }
-
--static void find_start_end(unsigned long flags, unsigned long *begin,
-- unsigned long *end)
-+static void find_start_end(struct mm_struct *mm, unsigned long flags,
-+ unsigned long *begin, unsigned long *end)
- {
- if (!test_thread_flag(TIF_IA32) && (flags & MAP_32BIT)) {
- unsigned long new_begin;
-@@ -52,7 +52,7 @@ static void find_start_end(unsigned long
- *begin = new_begin;
- }
- } else {
-- *begin = TASK_UNMAPPED_BASE;
-+ *begin = mm->mmap_base;
- *end = TASK_SIZE;
- }
- }
-@@ -69,16 +69,19 @@ arch_get_unmapped_area(struct file *filp
- if (flags & MAP_FIXED)
- return addr;
-
-- find_start_end(flags, &begin, &end);
-+ find_start_end(mm, flags, &begin, &end);
-
- if (len > end)
- return -ENOMEM;
-
-+#ifdef CONFIG_PAX_RANDMMAP
-+ if (!(mm->pax_flags & MF_PAX_RANDMMAP))
-+#endif
-+
- if (addr) {
- addr = PAGE_ALIGN(addr);
- vma = find_vma(mm, addr);
-- if (end - len >= addr &&
-- (!vma || addr + len <= vma->vm_start))
-+ if (end - len >= addr && check_heap_stack_gap(vma, addr, len))
- return addr;
- }
- if (((flags & MAP_32BIT) || test_thread_flag(TIF_IA32))
-@@ -106,7 +109,7 @@ full_search:
- }
- return -ENOMEM;
- }
-- if (!vma || addr + len <= vma->vm_start) {
-+ if (check_heap_stack_gap(vma, addr, len)) {
- /*
- * Remember the place where we stopped the search:
- */
-@@ -128,7 +131,7 @@ arch_get_unmapped_area_topdown(struct fi
- {
- struct vm_area_struct *vma;
- struct mm_struct *mm = current->mm;
-- unsigned long addr = addr0;
-+ unsigned long base = mm->mmap_base, addr = addr0;
-
- /* requested length too big for entire address space */
- if (len > TASK_SIZE)
-@@ -141,13 +144,18 @@ arch_get_unmapped_area_topdown(struct fi
- if (!test_thread_flag(TIF_IA32) && (flags & MAP_32BIT))
- goto bottomup;
-
-+#ifdef CONFIG_PAX_RANDMMAP
-+ if (!(mm->pax_flags & MF_PAX_RANDMMAP))
-+#endif
-+
- /* requesting a specific address */
- if (addr) {
- addr = PAGE_ALIGN(addr);
-- vma = find_vma(mm, addr);
-- if (TASK_SIZE - len >= addr &&
-- (!vma || addr + len <= vma->vm_start))
-- return addr;
-+ if (TASK_SIZE - len >= addr) {
-+ vma = find_vma(mm, addr);
-+ if (check_heap_stack_gap(vma, addr, len))
-+ return addr;
-+ }
- }
-
- /* check if free_area_cache is useful for us */
-@@ -162,7 +170,7 @@ arch_get_unmapped_area_topdown(struct fi
- /* make sure it can fit in the remaining address space */
- if (addr > len) {
- vma = find_vma(mm, addr-len);
-- if (!vma || addr <= vma->vm_start)
-+ if (check_heap_stack_gap(vma, addr - len, len))
- /* remember the address as a hint for next time */
- return mm->free_area_cache = addr-len;
- }
-@@ -179,7 +187,7 @@ arch_get_unmapped_area_topdown(struct fi
- * return with success:
- */
- vma = find_vma(mm, addr);
-- if (!vma || addr+len <= vma->vm_start)
-+ if (check_heap_stack_gap(vma, addr, len))
- /* remember the address as a hint for next time */
- return mm->free_area_cache = addr;
-
-@@ -188,8 +196,8 @@ arch_get_unmapped_area_topdown(struct fi
- mm->cached_hole_size = vma->vm_start - addr;
-
- /* try just below the current vma->vm_start */
-- addr = vma->vm_start-len;
-- } while (len < vma->vm_start);
-+ addr = skip_heap_stack_gap(vma, len);
-+ } while (!IS_ERR_VALUE(addr));
-
- bottomup:
- /*
-@@ -198,13 +206,21 @@ bottomup:
- * can happen with large stack limits and large mmap()
- * allocations.
- */
-+ mm->mmap_base = TASK_UNMAPPED_BASE;
-+
-+#ifdef CONFIG_PAX_RANDMMAP
-+ if (mm->pax_flags & MF_PAX_RANDMMAP)
-+ mm->mmap_base += mm->delta_mmap;
-+#endif
-+
-+ mm->free_area_cache = mm->mmap_base;
- mm->cached_hole_size = ~0UL;
-- mm->free_area_cache = TASK_UNMAPPED_BASE;
- addr = arch_get_unmapped_area(filp, addr0, len, pgoff, flags);
- /*
- * Restore the topdown base:
- */
-- mm->free_area_cache = mm->mmap_base;
-+ mm->mmap_base = base;
-+ mm->free_area_cache = base;
- mm->cached_hole_size = ~0UL;
-
- return addr;
-diff -urNp linux-2.6.37.4/arch/x86/kernel/time.c linux-2.6.37.4/arch/x86/kernel/time.c
---- linux-2.6.37.4/arch/x86/kernel/time.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/time.c 2011-01-17 02:41:01.000000000 -0500
-@@ -26,17 +26,13 @@
- int timer_ack;
- #endif
-
--#ifdef CONFIG_X86_64
--volatile unsigned long __jiffies __section_jiffies = INITIAL_JIFFIES;
--#endif
--
- unsigned long profile_pc(struct pt_regs *regs)
- {
- unsigned long pc = instruction_pointer(regs);
-
-- if (!user_mode_vm(regs) && in_lock_functions(pc)) {
-+ if (!user_mode(regs) && in_lock_functions(pc)) {
- #ifdef CONFIG_FRAME_POINTER
-- return *(unsigned long *)(regs->bp + sizeof(long));
-+ return ktla_ktva(*(unsigned long *)(regs->bp + sizeof(long)));
- #else
- unsigned long *sp =
- (unsigned long *)kernel_stack_pointer(regs);
-@@ -45,11 +41,17 @@ unsigned long profile_pc(struct pt_regs
- * or above a saved flags. Eflags has bits 22-31 zero,
- * kernel addresses don't.
- */
-+
-+#ifdef CONFIG_PAX_KERNEXEC
-+ return ktla_ktva(sp[0]);
-+#else
- if (sp[0] >> 22)
- return sp[0];
- if (sp[1] >> 22)
- return sp[1];
- #endif
-+
-+#endif
- }
- return pc;
- }
-diff -urNp linux-2.6.37.4/arch/x86/kernel/tls.c linux-2.6.37.4/arch/x86/kernel/tls.c
---- linux-2.6.37.4/arch/x86/kernel/tls.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/tls.c 2011-01-17 02:41:01.000000000 -0500
-@@ -85,6 +85,11 @@ int do_set_thread_area(struct task_struc
- if (idx < GDT_ENTRY_TLS_MIN || idx > GDT_ENTRY_TLS_MAX)
- return -EINVAL;
-
-+#ifdef CONFIG_PAX_SEGMEXEC
-+ if ((p->mm->pax_flags & MF_PAX_SEGMEXEC) && (info.contents & MODIFY_LDT_CONTENTS_CODE))
-+ return -EINVAL;
-+#endif
-+
- set_tls_desc(p, idx, &info, 1);
-
- return 0;
-diff -urNp linux-2.6.37.4/arch/x86/kernel/trampoline_32.S linux-2.6.37.4/arch/x86/kernel/trampoline_32.S
---- linux-2.6.37.4/arch/x86/kernel/trampoline_32.S 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/trampoline_32.S 2011-01-17 02:41:01.000000000 -0500
-@@ -32,6 +32,12 @@
- #include <asm/segment.h>
- #include <asm/page_types.h>
-
-+#ifdef CONFIG_PAX_KERNEXEC
-+#define ta(X) (X)
-+#else
-+#define ta(X) ((X) - __PAGE_OFFSET)
-+#endif
-+
- /* We can free up trampoline after bootup if cpu hotplug is not supported. */
- __CPUINITRODATA
- .code16
-@@ -60,7 +66,7 @@ r_base = .
- inc %ax # protected mode (PE) bit
- lmsw %ax # into protected mode
- # flush prefetch and jump to startup_32_smp in arch/i386/kernel/head.S
-- ljmpl $__BOOT_CS, $(startup_32_smp-__PAGE_OFFSET)
-+ ljmpl $__BOOT_CS, $ta(startup_32_smp)
-
- # These need to be in the same 64K segment as the above;
- # hence we don't use the boot_gdt_descr defined in head.S
-diff -urNp linux-2.6.37.4/arch/x86/kernel/trampoline_64.S linux-2.6.37.4/arch/x86/kernel/trampoline_64.S
---- linux-2.6.37.4/arch/x86/kernel/trampoline_64.S 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/trampoline_64.S 2011-01-17 02:41:01.000000000 -0500
-@@ -91,7 +91,7 @@ startup_32:
- movl $__KERNEL_DS, %eax # Initialize the %ds segment register
- movl %eax, %ds
-
-- movl $X86_CR4_PAE, %eax
-+ movl $(X86_CR4_PSE | X86_CR4_PAE | X86_CR4_PGE), %eax
- movl %eax, %cr4 # Enable PAE mode
-
- # Setup trampoline 4 level pagetables
-@@ -138,7 +138,7 @@ tidt:
- # so the kernel can live anywhere
- .balign 4
- tgdt:
-- .short tgdt_end - tgdt # gdt limit
-+ .short tgdt_end - tgdt - 1 # gdt limit
- .long tgdt - r_base
- .short 0
- .quad 0x00cf9b000000ffff # __KERNEL32_CS
-diff -urNp linux-2.6.37.4/arch/x86/kernel/traps.c linux-2.6.37.4/arch/x86/kernel/traps.c
---- linux-2.6.37.4/arch/x86/kernel/traps.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/traps.c 2011-01-17 02:41:01.000000000 -0500
-@@ -70,12 +70,6 @@ asmlinkage int system_call(void);
-
- /* Do we ignore FPU interrupts ? */
- char ignore_fpu_irq;
--
--/*
-- * The IDT has to be page-aligned to simplify the Pentium
-- * F0 0F bug workaround.
-- */
--gate_desc idt_table[NR_VECTORS] __page_aligned_data = { { { { 0, 0 } } }, };
- #endif
-
- DECLARE_BITMAP(used_vectors, NR_VECTORS);
-@@ -110,13 +104,13 @@ static inline void preempt_conditional_c
- }
-
- static void __kprobes
--do_trap(int trapnr, int signr, char *str, struct pt_regs *regs,
-+do_trap(int trapnr, int signr, const char *str, struct pt_regs *regs,
- long error_code, siginfo_t *info)
- {
- struct task_struct *tsk = current;
-
- #ifdef CONFIG_X86_32
-- if (regs->flags & X86_VM_MASK) {
-+ if (v8086_mode(regs)) {
- /*
- * traps 0, 1, 3, 4, and 5 should be forwarded to vm86.
- * On nmi (interrupt 2), do_trap should not be called.
-@@ -127,7 +121,7 @@ do_trap(int trapnr, int signr, char *str
- }
- #endif
-
-- if (!user_mode(regs))
-+ if (!user_mode_novm(regs))
- goto kernel_trap;
-
- #ifdef CONFIG_X86_32
-@@ -150,7 +144,7 @@ trap_signal:
- printk_ratelimit()) {
- printk(KERN_INFO
- "%s[%d] trap %s ip:%lx sp:%lx error:%lx",
-- tsk->comm, tsk->pid, str,
-+ tsk->comm, task_pid_nr(tsk), str,
- regs->ip, regs->sp, error_code);
- print_vma_addr(" in ", regs->ip);
- printk("\n");
-@@ -167,8 +161,20 @@ kernel_trap:
- if (!fixup_exception(regs)) {
- tsk->thread.error_code = error_code;
- tsk->thread.trap_no = trapnr;
-+
-+#if defined(CONFIG_X86_32) && defined(CONFIG_PAX_KERNEXEC)
-+ if (trapnr == 12 && ((regs->cs & 0xFFFF) == __KERNEL_CS || (regs->cs & 0xFFFF) == __KERNEXEC_KERNEL_CS))
-+ str = "PAX: suspicious stack segment fault";
-+#endif
-+
- die(str, regs, error_code);
- }
-+
-+#ifdef CONFIG_PAX_REFCOUNT
-+ if (trapnr == 4)
-+ pax_report_refcount_overflow(regs);
-+#endif
-+
- return;
-
- #ifdef CONFIG_X86_32
-@@ -257,14 +263,30 @@ do_general_protection(struct pt_regs *re
- conditional_sti(regs);
-
- #ifdef CONFIG_X86_32
-- if (regs->flags & X86_VM_MASK)
-+ if (v8086_mode(regs))
- goto gp_in_vm86;
- #endif
-
- tsk = current;
-- if (!user_mode(regs))
-+ if (!user_mode_novm(regs))
- goto gp_in_kernel;
-
-+#if defined(CONFIG_X86_32) && defined(CONFIG_PAX_PAGEEXEC)
-+ if (!(__supported_pte_mask & _PAGE_NX) && tsk->mm && (tsk->mm->pax_flags & MF_PAX_PAGEEXEC)) {
-+ struct mm_struct *mm = tsk->mm;
-+ unsigned long limit;
-+
-+ down_write(&mm->mmap_sem);
-+ limit = mm->context.user_cs_limit;
-+ if (limit < TASK_SIZE) {
-+ track_exec_limit(mm, limit, TASK_SIZE, VM_EXEC);
-+ up_write(&mm->mmap_sem);
-+ return;
-+ }
-+ up_write(&mm->mmap_sem);
-+ }
-+#endif
-+
- tsk->thread.error_code = error_code;
- tsk->thread.trap_no = 13;
-
-@@ -297,6 +319,13 @@ gp_in_kernel:
- if (notify_die(DIE_GPF, "general protection fault", regs,
- error_code, 13, SIGSEGV) == NOTIFY_STOP)
- return;
-+
-+#if defined(CONFIG_X86_32) && defined(CONFIG_PAX_KERNEXEC)
-+ if ((regs->cs & 0xFFFF) == __KERNEL_CS || (regs->cs & 0xFFFF) == __KERNEXEC_KERNEL_CS)
-+ die("PAX: suspicious general protection fault", regs, error_code);
-+ else
-+#endif
-+
- die("general protection fault", regs, error_code);
- }
-
-@@ -572,7 +601,7 @@ dotraplinkage void __kprobes do_debug(st
- /* It's safe to allow irq's after DR6 has been saved */
- preempt_conditional_sti(regs);
-
-- if (regs->flags & X86_VM_MASK) {
-+ if (v8086_mode(regs)) {
- handle_vm86_trap((struct kernel_vm86_regs *) regs,
- error_code, 1);
- preempt_conditional_cli(regs);
-@@ -586,7 +615,7 @@ dotraplinkage void __kprobes do_debug(st
- * We already checked v86 mode above, so we can check for kernel mode
- * by just checking the CPL of CS.
- */
-- if ((dr6 & DR_STEP) && !user_mode(regs)) {
-+ if ((dr6 & DR_STEP) && !user_mode_novm(regs)) {
- tsk->thread.debugreg6 &= ~DR_STEP;
- set_tsk_thread_flag(tsk, TIF_SINGLESTEP);
- regs->flags &= ~X86_EFLAGS_TF;
-@@ -615,7 +644,7 @@ void math_error(struct pt_regs *regs, in
- return;
- conditional_sti(regs);
-
-- if (!user_mode_vm(regs))
-+ if (!user_mode(regs))
- {
- if (!fixup_exception(regs)) {
- task->thread.error_code = error_code;
-diff -urNp linux-2.6.37.4/arch/x86/kernel/tsc.c linux-2.6.37.4/arch/x86/kernel/tsc.c
---- linux-2.6.37.4/arch/x86/kernel/tsc.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/tsc.c 2011-01-17 02:41:01.000000000 -0500
-@@ -837,7 +837,7 @@ static struct dmi_system_id __initdata b
- DMI_MATCH(DMI_BOARD_NAME, "2635FA0"),
- },
- },
-- {}
-+ { NULL, NULL, {{0, {0}}}, NULL}
- };
-
- static void __init check_system_tsc_reliable(void)
-diff -urNp linux-2.6.37.4/arch/x86/kernel/vm86_32.c linux-2.6.37.4/arch/x86/kernel/vm86_32.c
---- linux-2.6.37.4/arch/x86/kernel/vm86_32.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/vm86_32.c 2011-01-17 02:41:01.000000000 -0500
-@@ -41,6 +41,7 @@
- #include <linux/ptrace.h>
- #include <linux/audit.h>
- #include <linux/stddef.h>
-+#include <linux/grsecurity.h>
-
- #include <asm/uaccess.h>
- #include <asm/io.h>
-@@ -148,7 +149,7 @@ struct pt_regs *save_v86_state(struct ke
- do_exit(SIGSEGV);
- }
-
-- tss = &per_cpu(init_tss, get_cpu());
-+ tss = init_tss + get_cpu();
- current->thread.sp0 = current->thread.saved_sp0;
- current->thread.sysenter_cs = __KERNEL_CS;
- load_sp0(tss, &current->thread);
-@@ -207,6 +208,13 @@ int sys_vm86old(struct vm86_struct __use
- struct task_struct *tsk;
- int tmp, ret = -EPERM;
-
-+#ifdef CONFIG_GRKERNSEC_VM86
-+ if (!capable(CAP_SYS_RAWIO)) {
-+ gr_handle_vm86();
-+ goto out;
-+ }
-+#endif
-+
- tsk = current;
- if (tsk->thread.saved_sp0)
- goto out;
-@@ -237,6 +245,14 @@ int sys_vm86(unsigned long cmd, unsigned
- int tmp, ret;
- struct vm86plus_struct __user *v86;
-
-+#ifdef CONFIG_GRKERNSEC_VM86
-+ if (!capable(CAP_SYS_RAWIO)) {
-+ gr_handle_vm86();
-+ ret = -EPERM;
-+ goto out;
-+ }
-+#endif
-+
- tsk = current;
- switch (cmd) {
- case VM86_REQUEST_IRQ:
-@@ -323,7 +339,7 @@ static void do_sys_vm86(struct kernel_vm
- tsk->thread.saved_fs = info->regs32->fs;
- tsk->thread.saved_gs = get_user_gs(info->regs32);
-
-- tss = &per_cpu(init_tss, get_cpu());
-+ tss = init_tss + get_cpu();
- tsk->thread.sp0 = (unsigned long) &info->VM86_TSS_ESP0;
- if (cpu_has_sep)
- tsk->thread.sysenter_cs = 0;
-@@ -528,7 +544,7 @@ static void do_int(struct kernel_vm86_re
- goto cannot_handle;
- if (i == 0x21 && is_revectored(AH(regs), &KVM86->int21_revectored))
- goto cannot_handle;
-- intr_ptr = (unsigned long __user *) (i << 2);
-+ intr_ptr = (__force unsigned long __user *) (i << 2);
- if (get_user(segoffs, intr_ptr))
- goto cannot_handle;
- if ((segoffs >> 16) == BIOSSEG)
-diff -urNp linux-2.6.37.4/arch/x86/kernel/vmlinux.lds.S linux-2.6.37.4/arch/x86/kernel/vmlinux.lds.S
---- linux-2.6.37.4/arch/x86/kernel/vmlinux.lds.S 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/vmlinux.lds.S 2011-01-17 02:41:01.000000000 -0500
-@@ -26,6 +26,13 @@
- #include <asm/page_types.h>
- #include <asm/cache.h>
- #include <asm/boot.h>
-+#include <asm/segment.h>
-+
-+#if defined(CONFIG_X86_32) && defined(CONFIG_PAX_KERNEXEC)
-+#define __KERNEL_TEXT_OFFSET (LOAD_OFFSET + ____LOAD_PHYSICAL_ADDR)
-+#else
-+#define __KERNEL_TEXT_OFFSET 0
-+#endif
-
- #undef i386 /* in case the preprocessor is a 32bit one */
-
-@@ -34,11 +41,9 @@ OUTPUT_FORMAT(CONFIG_OUTPUT_FORMAT, CONF
- #ifdef CONFIG_X86_32
- OUTPUT_ARCH(i386)
- ENTRY(phys_startup_32)
--jiffies = jiffies_64;
- #else
- OUTPUT_ARCH(i386:x86-64)
- ENTRY(phys_startup_64)
--jiffies_64 = jiffies;
- #endif
-
- #if defined(CONFIG_X86_64) && defined(CONFIG_DEBUG_RODATA)
-@@ -69,31 +74,46 @@ jiffies_64 = jiffies;
-
- PHDRS {
- text PT_LOAD FLAGS(5); /* R_E */
-- data PT_LOAD FLAGS(7); /* RWE */
-+#ifdef CONFIG_X86_32
-+ module PT_LOAD FLAGS(5); /* R_E */
-+#endif
-+#ifdef CONFIG_XEN
-+ rodata PT_LOAD FLAGS(5); /* R_E */
-+#else
-+ rodata PT_LOAD FLAGS(4); /* R__ */
-+#endif
-+ data PT_LOAD FLAGS(6); /* RW_ */
- #ifdef CONFIG_X86_64
- user PT_LOAD FLAGS(5); /* R_E */
-+#endif
-+ init.begin PT_LOAD FLAGS(6); /* RW_ */
- #ifdef CONFIG_SMP
- percpu PT_LOAD FLAGS(6); /* RW_ */
- #endif
-+ text.init PT_LOAD FLAGS(5); /* R_E */
-+ text.exit PT_LOAD FLAGS(5); /* R_E */
- init PT_LOAD FLAGS(7); /* RWE */
--#endif
- note PT_NOTE FLAGS(0); /* ___ */
- }
-
- SECTIONS
- {
- #ifdef CONFIG_X86_32
-- . = LOAD_OFFSET + LOAD_PHYSICAL_ADDR;
-- phys_startup_32 = startup_32 - LOAD_OFFSET;
-+ . = LOAD_OFFSET + ____LOAD_PHYSICAL_ADDR;
- #else
-- . = __START_KERNEL;
-- phys_startup_64 = startup_64 - LOAD_OFFSET;
-+ . = __START_KERNEL;
- #endif
-
- /* Text and read-only data */
-- .text : AT(ADDR(.text) - LOAD_OFFSET) {
-- _text = .;
-+ .text (. - __KERNEL_TEXT_OFFSET): AT(ADDR(.text) - LOAD_OFFSET + __KERNEL_TEXT_OFFSET) {
- /* bootstrapping code */
-+#ifdef CONFIG_X86_32
-+ phys_startup_32 = startup_32 - LOAD_OFFSET + __KERNEL_TEXT_OFFSET;
-+#else
-+ phys_startup_64 = startup_64 - LOAD_OFFSET + __KERNEL_TEXT_OFFSET;
-+#endif
-+ __LOAD_PHYSICAL_ADDR = . - LOAD_OFFSET + __KERNEL_TEXT_OFFSET;
-+ _text = .;
- HEAD_TEXT
- #ifdef CONFIG_X86_32
- . = ALIGN(PAGE_SIZE);
-@@ -108,13 +128,47 @@ SECTIONS
- IRQENTRY_TEXT
- *(.fixup)
- *(.gnu.warning)
-- /* End of text section */
-- _etext = .;
- } :text = 0x9090
-
-- NOTES :text :note
-+ . += __KERNEL_TEXT_OFFSET;
-+
-+#ifdef CONFIG_X86_32
-+ . = ALIGN(PAGE_SIZE);
-+ .module.text : AT(ADDR(.module.text) - LOAD_OFFSET) {
-+
-+#if defined(CONFIG_PAX_KERNEXEC) && defined(CONFIG_MODULES)
-+ MODULES_EXEC_VADDR = .;
-+ BYTE(0)
-+ . += (CONFIG_PAX_KERNEXEC_MODULE_TEXT * 1024 * 1024);
-+ . = ALIGN(HPAGE_SIZE);
-+ MODULES_EXEC_END = . - 1;
-+#endif
-+
-+ } :module
-+#endif
-+
-+ .text.end : AT(ADDR(.text.end) - LOAD_OFFSET) {
-+ /* End of text section */
-+ _etext = . - __KERNEL_TEXT_OFFSET;
-+ }
-
-- EXCEPTION_TABLE(16) :text = 0x9090
-+#ifdef CONFIG_X86_32
-+ . = ALIGN(PAGE_SIZE);
-+ .rodata.page_aligned : AT(ADDR(.rodata.page_aligned) - LOAD_OFFSET) {
-+ *(.idt)
-+ . = ALIGN(PAGE_SIZE);
-+ *(.empty_zero_page)
-+ *(.initial_pg_fixmap)
-+ *(.initial_pg_pmd)
-+ *(.initial_page_table)
-+ *(.swapper_pg_dir)
-+ } :rodata
-+#endif
-+
-+ . = ALIGN(PAGE_SIZE);
-+ NOTES :rodata :note
-+
-+ EXCEPTION_TABLE(16) :rodata
-
- X64_ALIGN_DEBUG_RODATA_BEGIN
- RO_DATA(PAGE_SIZE)
-@@ -122,16 +176,20 @@ SECTIONS
-
- /* Data */
- .data : AT(ADDR(.data) - LOAD_OFFSET) {
-+
-+#ifdef CONFIG_PAX_KERNEXEC
-+ . = ALIGN(HPAGE_SIZE);
-+#else
-+ . = ALIGN(PAGE_SIZE);
-+#endif
-+
- /* Start of data section */
- _sdata = .;
-
- /* init_task */
- INIT_TASK_DATA(THREAD_SIZE)
-
--#ifdef CONFIG_X86_32
-- /* 32 bit has nosave before _edata */
- NOSAVE_DATA
--#endif
-
- PAGE_ALIGNED_DATA(PAGE_SIZE)
-
-@@ -140,6 +198,8 @@ SECTIONS
- DATA_DATA
- CONSTRUCTORS
-
-+ jiffies = jiffies_64;
-+
- /* rarely changed data like cpu maps */
- READ_MOSTLY_DATA(INTERNODE_CACHE_BYTES)
-
-@@ -194,12 +254,6 @@ SECTIONS
- }
- vgetcpu_mode = VVIRT(.vgetcpu_mode);
-
-- . = ALIGN(L1_CACHE_BYTES);
-- .jiffies : AT(VLOAD(.jiffies)) {
-- *(.jiffies)
-- }
-- jiffies = VVIRT(.jiffies);
--
- .vsyscall_3 ADDR(.vsyscall_0) + 3072: AT(VLOAD(.vsyscall_3)) {
- *(.vsyscall_3)
- }
-@@ -215,12 +269,19 @@ SECTIONS
- #endif /* CONFIG_X86_64 */
-
- /* Init code and data - will be freed after init */
-- . = ALIGN(PAGE_SIZE);
- .init.begin : AT(ADDR(.init.begin) - LOAD_OFFSET) {
-+ BYTE(0)
-+
-+#ifdef CONFIG_PAX_KERNEXEC
-+ . = ALIGN(HPAGE_SIZE);
-+#else
-+ . = ALIGN(PAGE_SIZE);
-+#endif
-+
- __init_begin = .; /* paired with __init_end */
-- }
-+ } :init.begin
-
--#if defined(CONFIG_X86_64) && defined(CONFIG_SMP)
-+#ifdef CONFIG_SMP
- /*
- * percpu offsets are zero-based on SMP. PERCPU_VADDR() changes the
- * output PHDR, so the next output section - .init.text - should
-@@ -229,12 +290,27 @@ SECTIONS
- PERCPU_VADDR(0, :percpu)
- #endif
-
-- INIT_TEXT_SECTION(PAGE_SIZE)
--#ifdef CONFIG_X86_64
-- :init
--#endif
-+ . = ALIGN(PAGE_SIZE);
-+ init_begin = .;
-+ .init.text (. - __KERNEL_TEXT_OFFSET): AT(init_begin - LOAD_OFFSET) {
-+ VMLINUX_SYMBOL(_sinittext) = .;
-+ INIT_TEXT
-+ VMLINUX_SYMBOL(_einittext) = .;
-+ . = ALIGN(PAGE_SIZE);
-+ } :text.init
-
-- INIT_DATA_SECTION(16)
-+ /*
-+ * .exit.text is discard at runtime, not link time, to deal with
-+ * references from .altinstructions and .eh_frame
-+ */
-+ .exit.text : AT(ADDR(.exit.text) - LOAD_OFFSET + __KERNEL_TEXT_OFFSET) {
-+ EXIT_TEXT
-+ . = ALIGN(16);
-+ } :text.exit
-+ . = init_begin + SIZEOF(.init.text) + SIZEOF(.exit.text);
-+
-+ . = ALIGN(PAGE_SIZE);
-+ INIT_DATA_SECTION(16) :init
-
- .x86_cpu_dev.init : AT(ADDR(.x86_cpu_dev.init) - LOAD_OFFSET) {
- __x86_cpu_dev_start = .;
-@@ -288,19 +364,12 @@ SECTIONS
- __iommu_table_end = .;
- }
- . = ALIGN(8);
-- /*
-- * .exit.text is discard at runtime, not link time, to deal with
-- * references from .altinstructions and .eh_frame
-- */
-- .exit.text : AT(ADDR(.exit.text) - LOAD_OFFSET) {
-- EXIT_TEXT
-- }
-
- .exit.data : AT(ADDR(.exit.data) - LOAD_OFFSET) {
- EXIT_DATA
- }
-
--#if !defined(CONFIG_X86_64) || !defined(CONFIG_SMP)
-+#ifndef CONFIG_SMP
- PERCPU(THREAD_SIZE)
- #endif
-
-@@ -319,16 +388,10 @@ SECTIONS
- .smp_locks : AT(ADDR(.smp_locks) - LOAD_OFFSET) {
- __smp_locks = .;
- *(.smp_locks)
-- . = ALIGN(PAGE_SIZE);
- __smp_locks_end = .;
-+ . = ALIGN(PAGE_SIZE);
- }
-
--#ifdef CONFIG_X86_64
-- .data_nosave : AT(ADDR(.data_nosave) - LOAD_OFFSET) {
-- NOSAVE_DATA
-- }
--#endif
--
- /* BSS */
- . = ALIGN(PAGE_SIZE);
- .bss : AT(ADDR(.bss) - LOAD_OFFSET) {
-@@ -344,6 +407,7 @@ SECTIONS
- __brk_base = .;
- . += 64 * 1024; /* 64k alignment slop space */
- *(.brk_reservation) /* areas brk users have reserved */
-+ . = ALIGN(HPAGE_SIZE);
- __brk_limit = .;
- }
-
-@@ -370,13 +434,12 @@ SECTIONS
- * for the boot processor.
- */
- #define INIT_PER_CPU(x) init_per_cpu__##x = x + __per_cpu_load
--INIT_PER_CPU(gdt_page);
- INIT_PER_CPU(irq_stack_union);
-
- /*
- * Build-time check on the image size:
- */
--. = ASSERT((_end - _text <= KERNEL_IMAGE_SIZE),
-+. = ASSERT((_end - _text - __KERNEL_TEXT_OFFSET <= KERNEL_IMAGE_SIZE),
- "kernel image bigger than KERNEL_IMAGE_SIZE");
-
- #ifdef CONFIG_SMP
-diff -urNp linux-2.6.37.4/arch/x86/kernel/vsyscall_64.c linux-2.6.37.4/arch/x86/kernel/vsyscall_64.c
---- linux-2.6.37.4/arch/x86/kernel/vsyscall_64.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/vsyscall_64.c 2011-01-17 02:41:01.000000000 -0500
-@@ -80,6 +80,7 @@ void update_vsyscall(struct timespec *wa
-
- write_seqlock_irqsave(&vsyscall_gtod_data.lock, flags);
- /* copy vsyscall data */
-+ strlcpy(vsyscall_gtod_data.clock.name, clock->name, sizeof vsyscall_gtod_data.clock.name);
- vsyscall_gtod_data.clock.vread = clock->vread;
- vsyscall_gtod_data.clock.cycle_last = clock->cycle_last;
- vsyscall_gtod_data.clock.mask = clock->mask;
-@@ -208,7 +209,7 @@ vgetcpu(unsigned *cpu, unsigned *node, s
- We do this here because otherwise user space would do it on
- its own in a likely inferior way (no access to jiffies).
- If you don't like it pass NULL. */
-- if (tcache && tcache->blob[0] == (j = __jiffies)) {
-+ if (tcache && tcache->blob[0] == (j = jiffies)) {
- p = tcache->blob[1];
- } else if (__vgetcpu_mode == VGETCPU_RDTSCP) {
- /* Load per CPU data from RDTSCP */
-diff -urNp linux-2.6.37.4/arch/x86/kernel/x8664_ksyms_64.c linux-2.6.37.4/arch/x86/kernel/x8664_ksyms_64.c
---- linux-2.6.37.4/arch/x86/kernel/x8664_ksyms_64.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/x8664_ksyms_64.c 2011-01-17 02:41:01.000000000 -0500
-@@ -29,8 +29,6 @@ EXPORT_SYMBOL(__put_user_8);
- EXPORT_SYMBOL(copy_user_generic_string);
- EXPORT_SYMBOL(copy_user_generic_unrolled);
- EXPORT_SYMBOL(__copy_user_nocache);
--EXPORT_SYMBOL(_copy_from_user);
--EXPORT_SYMBOL(_copy_to_user);
-
- EXPORT_SYMBOL(copy_page);
- EXPORT_SYMBOL(clear_page);
-diff -urNp linux-2.6.37.4/arch/x86/kernel/xsave.c linux-2.6.37.4/arch/x86/kernel/xsave.c
---- linux-2.6.37.4/arch/x86/kernel/xsave.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kernel/xsave.c 2011-01-17 02:41:01.000000000 -0500
-@@ -130,7 +130,7 @@ int check_for_xstate(struct i387_fxsave_
- fx_sw_user->xstate_size > fx_sw_user->extended_size)
- return -EINVAL;
-
-- err = __get_user(magic2, (__u32 *) (((void *)fpstate) +
-+ err = __get_user(magic2, (__u32 __user *) (((void __user *)fpstate) +
- fx_sw_user->extended_size -
- FP_XSTATE_MAGIC2_SIZE));
- if (err)
-@@ -267,7 +267,7 @@ fx_only:
- * the other extended state.
- */
- xrstor_state(init_xstate_buf, pcntxt_mask & ~XSTATE_FPSSE);
-- return fxrstor_checking((__force struct i387_fxsave_struct *)buf);
-+ return fxrstor_checking((struct i387_fxsave_struct __user *)buf);
- }
-
- /*
-@@ -299,7 +299,7 @@ int restore_i387_xstate(void __user *buf
- if (use_xsave())
- err = restore_user_xstate(buf);
- else
-- err = fxrstor_checking((__force struct i387_fxsave_struct *)
-+ err = fxrstor_checking((struct i387_fxsave_struct __user *)
- buf);
- if (unlikely(err)) {
- /*
-diff -urNp linux-2.6.37.4/arch/x86/kvm/emulate.c linux-2.6.37.4/arch/x86/kvm/emulate.c
---- linux-2.6.37.4/arch/x86/kvm/emulate.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kvm/emulate.c 2011-01-17 02:41:01.000000000 -0500
-@@ -96,7 +96,7 @@
- #define Src2ImmByte (2<<29)
- #define Src2One (3<<29)
- #define Src2Imm (4<<29)
--#define Src2Mask (7<<29)
-+#define Src2Mask (7U<<29)
-
- #define X2(x...) x, x
- #define X3(x...) X2(x), x
-@@ -197,6 +197,7 @@ struct group_dual {
-
- #define ____emulate_2op(_op, _src, _dst, _eflags, _x, _y, _suffix, _dsttype) \
- do { \
-+ unsigned long _tmp; \
- __asm__ __volatile__ ( \
- _PRE_EFLAGS("0", "4", "2") \
- _op _suffix " %"_x"3,%1; " \
-@@ -210,8 +211,6 @@ struct group_dual {
- /* Raw emulation: instruction has two explicit operands. */
- #define __emulate_2op_nobyte(_op,_src,_dst,_eflags,_wx,_wy,_lx,_ly,_qx,_qy) \
- do { \
-- unsigned long _tmp; \
-- \
- switch ((_dst).bytes) { \
- case 2: \
- ____emulate_2op(_op,_src,_dst,_eflags,_wx,_wy,"w",u16);\
-@@ -227,7 +226,6 @@ struct group_dual {
-
- #define __emulate_2op(_op,_src,_dst,_eflags,_bx,_by,_wx,_wy,_lx,_ly,_qx,_qy) \
- do { \
-- unsigned long _tmp; \
- switch ((_dst).bytes) { \
- case 1: \
- ____emulate_2op(_op,_src,_dst,_eflags,_bx,_by,"b",u8); \
-diff -urNp linux-2.6.37.4/arch/x86/kvm/lapic.c linux-2.6.37.4/arch/x86/kvm/lapic.c
---- linux-2.6.37.4/arch/x86/kvm/lapic.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kvm/lapic.c 2011-01-17 02:41:01.000000000 -0500
-@@ -53,7 +53,7 @@
- #define APIC_BUS_CYCLE_NS 1
-
- /* #define apic_debug(fmt,arg...) printk(KERN_WARNING fmt,##arg) */
--#define apic_debug(fmt, arg...)
-+#define apic_debug(fmt, arg...) do {} while (0)
-
- #define APIC_LVT_NUM 6
- /* 14 is the version for Xeon and Pentium 8.4.8*/
-diff -urNp linux-2.6.37.4/arch/x86/kvm/svm.c linux-2.6.37.4/arch/x86/kvm/svm.c
---- linux-2.6.37.4/arch/x86/kvm/svm.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kvm/svm.c 2011-01-17 02:41:01.000000000 -0500
-@@ -3023,7 +3023,11 @@ static void reload_tss(struct kvm_vcpu *
- int cpu = raw_smp_processor_id();
-
- struct svm_cpu_data *sd = per_cpu(svm_data, cpu);
-+
-+ pax_open_kernel();
- sd->tss_desc->type = 9; /* available 32/64-bit TSS */
-+ pax_close_kernel();
-+
- load_TR_desc();
- }
-
-@@ -3600,7 +3604,7 @@ static void svm_fpu_deactivate(struct kv
- update_cr0_intercept(svm);
- }
-
--static struct kvm_x86_ops svm_x86_ops = {
-+static const struct kvm_x86_ops svm_x86_ops = {
- .cpu_has_kvm_support = has_svm,
- .disabled_by_bios = is_disabled,
- .hardware_setup = svm_hardware_setup,
-diff -urNp linux-2.6.37.4/arch/x86/kvm/vmx.c linux-2.6.37.4/arch/x86/kvm/vmx.c
---- linux-2.6.37.4/arch/x86/kvm/vmx.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kvm/vmx.c 2011-01-17 02:41:01.000000000 -0500
-@@ -705,7 +705,11 @@ static void reload_tss(void)
- struct desc_struct *descs;
-
- descs = (void *)gdt->address;
-+
-+ pax_open_kernel();
- descs[GDT_ENTRY_TSS].type = 9; /* available TSS */
-+ pax_close_kernel();
-+
- load_TR_desc();
- }
-
-@@ -1589,8 +1593,11 @@ static __init int hardware_setup(void)
- if (!cpu_has_vmx_flexpriority())
- flexpriority_enabled = 0;
-
-- if (!cpu_has_vmx_tpr_shadow())
-- kvm_x86_ops->update_cr8_intercept = NULL;
-+ if (!cpu_has_vmx_tpr_shadow()) {
-+ pax_open_kernel();
-+ *(void **)&kvm_x86_ops->update_cr8_intercept = NULL;
-+ pax_close_kernel();
-+ }
-
- if (enable_ept && !cpu_has_vmx_ept_2m_page())
- kvm_disable_largepages();
-@@ -2575,7 +2582,7 @@ static int vmx_vcpu_setup(struct vcpu_vm
- vmcs_writel(HOST_IDTR_BASE, dt.address); /* 22.2.4 */
-
- asm("mov $.Lkvm_vmx_return, %0" : "=r"(kvm_vmx_return));
-- vmcs_writel(HOST_RIP, kvm_vmx_return); /* 22.2.5 */
-+ vmcs_writel(HOST_RIP, ktla_ktva(kvm_vmx_return)); /* 22.2.5 */
- vmcs_write32(VM_EXIT_MSR_STORE_COUNT, 0);
- vmcs_write32(VM_EXIT_MSR_LOAD_COUNT, 0);
- vmcs_write64(VM_EXIT_MSR_LOAD_ADDR, __pa(vmx->msr_autoload.host));
-@@ -3959,6 +3966,12 @@ static void vmx_vcpu_run(struct kvm_vcpu
- "jmp .Lkvm_vmx_return \n\t"
- ".Llaunched: " __ex(ASM_VMX_VMRESUME) "\n\t"
- ".Lkvm_vmx_return: "
-+
-+#if defined(CONFIG_X86_32) && defined(CONFIG_PAX_KERNEXEC)
-+ "ljmp %[cs],$.Lkvm_vmx_return2\n\t"
-+ ".Lkvm_vmx_return2: "
-+#endif
-+
- /* Save guest registers, load host registers, keep flags */
- "xchg %0, (%%"R"sp) \n\t"
- "mov %%"R"ax, %c[rax](%0) \n\t"
-@@ -4005,6 +4018,11 @@ static void vmx_vcpu_run(struct kvm_vcpu
- [r15]"i"(offsetof(struct vcpu_vmx, vcpu.arch.regs[VCPU_REGS_R15])),
- #endif
- [cr2]"i"(offsetof(struct vcpu_vmx, vcpu.arch.cr2))
-+
-+#if defined(CONFIG_X86_32) && defined(CONFIG_PAX_KERNEXEC)
-+ ,[cs]"i"(__KERNEL_CS)
-+#endif
-+
- : "cc", "memory"
- , R"ax", R"bx", R"di", R"si"
- #ifdef CONFIG_X86_64
-@@ -4018,7 +4036,7 @@ static void vmx_vcpu_run(struct kvm_vcpu
-
- vmx->idt_vectoring_info = vmcs_read32(IDT_VECTORING_INFO_FIELD);
-
-- asm("mov %0, %%ds; mov %0, %%es" : : "r"(__USER_DS));
-+ asm("mov %0, %%ds; mov %0, %%es; mov %0, %%ss" : : "r"(__KERNEL_DS));
- vmx->launched = 1;
-
- vmx->exit_reason = vmcs_read32(VM_EXIT_REASON);
-@@ -4253,7 +4271,7 @@ static void vmx_set_supported_cpuid(u32
- {
- }
-
--static struct kvm_x86_ops vmx_x86_ops = {
-+static const struct kvm_x86_ops vmx_x86_ops = {
- .cpu_has_kvm_support = cpu_has_kvm_support,
- .disabled_by_bios = vmx_disabled_by_bios,
- .hardware_setup = hardware_setup,
-diff -urNp linux-2.6.37.4/arch/x86/kvm/x86.c linux-2.6.37.4/arch/x86/kvm/x86.c
---- linux-2.6.37.4/arch/x86/kvm/x86.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/kvm/x86.c 2011-01-17 02:41:01.000000000 -0500
-@@ -92,7 +92,7 @@ static void update_cr8_intercept(struct
- static int kvm_dev_ioctl_get_supported_cpuid(struct kvm_cpuid2 *cpuid,
- struct kvm_cpuid_entry2 __user *entries);
-
--struct kvm_x86_ops *kvm_x86_ops;
-+const struct kvm_x86_ops *kvm_x86_ops;
- EXPORT_SYMBOL_GPL(kvm_x86_ops);
-
- int ignore_msrs = 0;
-@@ -118,38 +118,38 @@ static struct kvm_shared_msrs_global __r
- static DEFINE_PER_CPU(struct kvm_shared_msrs, shared_msrs);
-
- struct kvm_stats_debugfs_item debugfs_entries[] = {
-- { "pf_fixed", VCPU_STAT(pf_fixed) },
-- { "pf_guest", VCPU_STAT(pf_guest) },
-- { "tlb_flush", VCPU_STAT(tlb_flush) },
-- { "invlpg", VCPU_STAT(invlpg) },
-- { "exits", VCPU_STAT(exits) },
-- { "io_exits", VCPU_STAT(io_exits) },
-- { "mmio_exits", VCPU_STAT(mmio_exits) },
-- { "signal_exits", VCPU_STAT(signal_exits) },
-- { "irq_window", VCPU_STAT(irq_window_exits) },
-- { "nmi_window", VCPU_STAT(nmi_window_exits) },
-- { "halt_exits", VCPU_STAT(halt_exits) },
-- { "halt_wakeup", VCPU_STAT(halt_wakeup) },
-- { "hypercalls", VCPU_STAT(hypercalls) },
-- { "request_irq", VCPU_STAT(request_irq_exits) },
-- { "irq_exits", VCPU_STAT(irq_exits) },
-- { "host_state_reload", VCPU_STAT(host_state_reload) },
-- { "efer_reload", VCPU_STAT(efer_reload) },
-- { "fpu_reload", VCPU_STAT(fpu_reload) },
-- { "insn_emulation", VCPU_STAT(insn_emulation) },
-- { "insn_emulation_fail", VCPU_STAT(insn_emulation_fail) },
-- { "irq_injections", VCPU_STAT(irq_injections) },
-- { "nmi_injections", VCPU_STAT(nmi_injections) },
-- { "mmu_shadow_zapped", VM_STAT(mmu_shadow_zapped) },
-- { "mmu_pte_write", VM_STAT(mmu_pte_write) },
-- { "mmu_pte_updated", VM_STAT(mmu_pte_updated) },
-- { "mmu_pde_zapped", VM_STAT(mmu_pde_zapped) },
-- { "mmu_flooded", VM_STAT(mmu_flooded) },
-- { "mmu_recycled", VM_STAT(mmu_recycled) },
-- { "mmu_cache_miss", VM_STAT(mmu_cache_miss) },
-- { "mmu_unsync", VM_STAT(mmu_unsync) },
-- { "remote_tlb_flush", VM_STAT(remote_tlb_flush) },
-- { "largepages", VM_STAT(lpages) },
-+ { "pf_fixed", VCPU_STAT(pf_fixed), NULL },
-+ { "pf_guest", VCPU_STAT(pf_guest), NULL },
-+ { "tlb_flush", VCPU_STAT(tlb_flush), NULL },
-+ { "invlpg", VCPU_STAT(invlpg), NULL },
-+ { "exits", VCPU_STAT(exits), NULL },
-+ { "io_exits", VCPU_STAT(io_exits), NULL },
-+ { "mmio_exits", VCPU_STAT(mmio_exits), NULL },
-+ { "signal_exits", VCPU_STAT(signal_exits), NULL },
-+ { "irq_window", VCPU_STAT(irq_window_exits), NULL },
-+ { "nmi_window", VCPU_STAT(nmi_window_exits), NULL },
-+ { "halt_exits", VCPU_STAT(halt_exits), NULL },
-+ { "halt_wakeup", VCPU_STAT(halt_wakeup), NULL },
-+ { "hypercalls", VCPU_STAT(hypercalls), NULL },
-+ { "request_irq", VCPU_STAT(request_irq_exits), NULL },
-+ { "irq_exits", VCPU_STAT(irq_exits), NULL },
-+ { "host_state_reload", VCPU_STAT(host_state_reload), NULL },
-+ { "efer_reload", VCPU_STAT(efer_reload), NULL },
-+ { "fpu_reload", VCPU_STAT(fpu_reload), NULL },
-+ { "insn_emulation", VCPU_STAT(insn_emulation), NULL },
-+ { "insn_emulation_fail", VCPU_STAT(insn_emulation_fail), NULL },
-+ { "irq_injections", VCPU_STAT(irq_injections), NULL },
-+ { "nmi_injections", VCPU_STAT(nmi_injections), NULL },
-+ { "mmu_shadow_zapped", VM_STAT(mmu_shadow_zapped), NULL },
-+ { "mmu_pte_write", VM_STAT(mmu_pte_write), NULL },
-+ { "mmu_pte_updated", VM_STAT(mmu_pte_updated), NULL },
-+ { "mmu_pde_zapped", VM_STAT(mmu_pde_zapped), NULL },
-+ { "mmu_flooded", VM_STAT(mmu_flooded), NULL },
-+ { "mmu_recycled", VM_STAT(mmu_recycled), NULL },
-+ { "mmu_cache_miss", VM_STAT(mmu_cache_miss), NULL },
-+ { "mmu_unsync", VM_STAT(mmu_unsync), NULL },
-+ { "remote_tlb_flush", VM_STAT(remote_tlb_flush), NULL },
-+ { "largepages", VM_STAT(lpages), NULL },
- { NULL }
- };
-
-@@ -1979,6 +1979,8 @@ long kvm_arch_dev_ioctl(struct file *fil
- if (n < msr_list.nmsrs)
- goto out;
- r = -EFAULT;
-+ if (num_msrs_to_save > ARRAY_SIZE(msrs_to_save))
-+ goto out;
- if (copy_to_user(user_msr_list->indices, &msrs_to_save,
- num_msrs_to_save * sizeof(u32)))
- goto out;
-@@ -2446,7 +2448,7 @@ static int kvm_vcpu_ioctl_set_lapic(stru
- static int kvm_vcpu_ioctl_interrupt(struct kvm_vcpu *vcpu,
- struct kvm_interrupt *irq)
- {
-- if (irq->irq < 0 || irq->irq >= 256)
-+ if (irq->irq >= 256)
- return -EINVAL;
- if (irqchip_in_kernel(vcpu->kvm))
- return -ENXIO;
-@@ -4623,10 +4625,10 @@ void kvm_after_handle_nmi(struct kvm_vcp
- }
- EXPORT_SYMBOL_GPL(kvm_after_handle_nmi);
-
--int kvm_arch_init(void *opaque)
-+int kvm_arch_init(const void *opaque)
- {
- int r;
-- struct kvm_x86_ops *ops = (struct kvm_x86_ops *)opaque;
-+ const struct kvm_x86_ops *ops = (const struct kvm_x86_ops *)opaque;
-
- if (kvm_x86_ops) {
- printk(KERN_ERR "kvm: already loaded the other module\n");
-diff -urNp linux-2.6.37.4/arch/x86/lib/atomic64_cx8_32.S linux-2.6.37.4/arch/x86/lib/atomic64_cx8_32.S
---- linux-2.6.37.4/arch/x86/lib/atomic64_cx8_32.S 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/lib/atomic64_cx8_32.S 2011-01-17 02:41:01.000000000 -0500
-@@ -86,13 +86,23 @@ ENTRY(atomic64_\func\()_return_cx8)
- movl %edx, %ecx
- \ins\()l %esi, %ebx
- \insc\()l %edi, %ecx
-+
-+#ifdef CONFIG_PAX_REFCOUNT
-+ into
-+2:
-+ _ASM_EXTABLE(2b, 3f)
-+#endif
-+
- LOCK_PREFIX
- cmpxchg8b (%ebp)
- jne 1b
--
--10:
- movl %ebx, %eax
- movl %ecx, %edx
-+
-+#ifdef CONFIG_PAX_REFCOUNT
-+3:
-+#endif
-+
- RESTORE edi
- RESTORE esi
- RESTORE ebx
-@@ -116,13 +126,24 @@ ENTRY(atomic64_\func\()_return_cx8)
- movl %edx, %ecx
- \ins\()l $1, %ebx
- \insc\()l $0, %ecx
-+
-+#ifdef CONFIG_PAX_REFCOUNT
-+ into
-+2:
-+ _ASM_EXTABLE(2b, 3f)
-+#endif
-+
- LOCK_PREFIX
- cmpxchg8b (%esi)
- jne 1b
-
--10:
- movl %ebx, %eax
- movl %ecx, %edx
-+
-+#ifdef CONFIG_PAX_REFCOUNT
-+3:
-+#endif
-+
- RESTORE ebx
- ret
- CFI_ENDPROC
-@@ -176,6 +197,13 @@ ENTRY(atomic64_add_unless_cx8)
- movl %edx, %ecx
- addl %esi, %ebx
- adcl %edi, %ecx
-+
-+#ifdef CONFIG_PAX_REFCOUNT
-+ into
-+1234:
-+ _ASM_EXTABLE(1234b, 1234b)
-+#endif
-+
- LOCK_PREFIX
- cmpxchg8b (%ebp)
- jne 1b
-@@ -208,6 +236,13 @@ ENTRY(atomic64_inc_not_zero_cx8)
- movl %edx, %ecx
- addl $1, %ebx
- adcl $0, %ecx
-+
-+#ifdef CONFIG_PAX_REFCOUNT
-+ into
-+1234:
-+ _ASM_EXTABLE(1234b, 1234b)
-+#endif
-+
- LOCK_PREFIX
- cmpxchg8b (%esi)
- jne 1b
-diff -urNp linux-2.6.37.4/arch/x86/lib/checksum_32.S linux-2.6.37.4/arch/x86/lib/checksum_32.S
---- linux-2.6.37.4/arch/x86/lib/checksum_32.S 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/lib/checksum_32.S 2011-01-17 02:41:01.000000000 -0500
-@@ -28,7 +28,8 @@
- #include <linux/linkage.h>
- #include <asm/dwarf2.h>
- #include <asm/errno.h>
--
-+#include <asm/segment.h>
-+
- /*
- * computes a partial checksum, e.g. for TCP/UDP fragments
- */
-@@ -304,9 +305,28 @@ unsigned int csum_partial_copy_generic (
-
- #define ARGBASE 16
- #define FP 12
--
--ENTRY(csum_partial_copy_generic)
-+
-+ENTRY(csum_partial_copy_generic_to_user)
- CFI_STARTPROC
-+
-+#ifdef CONFIG_PAX_MEMORY_UDEREF
-+ pushl %gs
-+ CFI_ADJUST_CFA_OFFSET 4
-+ popl %es
-+ CFI_ADJUST_CFA_OFFSET -4
-+ jmp csum_partial_copy_generic
-+#endif
-+
-+ENTRY(csum_partial_copy_generic_from_user)
-+
-+#ifdef CONFIG_PAX_MEMORY_UDEREF
-+ pushl %gs
-+ CFI_ADJUST_CFA_OFFSET 4
-+ popl %ds
-+ CFI_ADJUST_CFA_OFFSET -4
-+#endif
-+
-+ENTRY(csum_partial_copy_generic)
- subl $4,%esp
- CFI_ADJUST_CFA_OFFSET 4
- pushl %edi
-@@ -331,7 +351,7 @@ ENTRY(csum_partial_copy_generic)
- jmp 4f
- SRC(1: movw (%esi), %bx )
- addl $2, %esi
--DST( movw %bx, (%edi) )
-+DST( movw %bx, %es:(%edi) )
- addl $2, %edi
- addw %bx, %ax
- adcl $0, %eax
-@@ -343,30 +363,30 @@ DST( movw %bx, (%edi) )
- SRC(1: movl (%esi), %ebx )
- SRC( movl 4(%esi), %edx )
- adcl %ebx, %eax
--DST( movl %ebx, (%edi) )
-+DST( movl %ebx, %es:(%edi) )
- adcl %edx, %eax
--DST( movl %edx, 4(%edi) )
-+DST( movl %edx, %es:4(%edi) )
-
- SRC( movl 8(%esi), %ebx )
- SRC( movl 12(%esi), %edx )
- adcl %ebx, %eax
--DST( movl %ebx, 8(%edi) )
-+DST( movl %ebx, %es:8(%edi) )
- adcl %edx, %eax
--DST( movl %edx, 12(%edi) )
-+DST( movl %edx, %es:12(%edi) )
-
- SRC( movl 16(%esi), %ebx )
- SRC( movl 20(%esi), %edx )
- adcl %ebx, %eax
--DST( movl %ebx, 16(%edi) )
-+DST( movl %ebx, %es:16(%edi) )
- adcl %edx, %eax
--DST( movl %edx, 20(%edi) )
-+DST( movl %edx, %es:20(%edi) )
-
- SRC( movl 24(%esi), %ebx )
- SRC( movl 28(%esi), %edx )
- adcl %ebx, %eax
--DST( movl %ebx, 24(%edi) )
-+DST( movl %ebx, %es:24(%edi) )
- adcl %edx, %eax
--DST( movl %edx, 28(%edi) )
-+DST( movl %edx, %es:28(%edi) )
-
- lea 32(%esi), %esi
- lea 32(%edi), %edi
-@@ -380,7 +400,7 @@ DST( movl %edx, 28(%edi) )
- shrl $2, %edx # This clears CF
- SRC(3: movl (%esi), %ebx )
- adcl %ebx, %eax
--DST( movl %ebx, (%edi) )
-+DST( movl %ebx, %es:(%edi) )
- lea 4(%esi), %esi
- lea 4(%edi), %edi
- dec %edx
-@@ -392,12 +412,12 @@ DST( movl %ebx, (%edi) )
- jb 5f
- SRC( movw (%esi), %cx )
- leal 2(%esi), %esi
--DST( movw %cx, (%edi) )
-+DST( movw %cx, %es:(%edi) )
- leal 2(%edi), %edi
- je 6f
- shll $16,%ecx
- SRC(5: movb (%esi), %cl )
--DST( movb %cl, (%edi) )
-+DST( movb %cl, %es:(%edi) )
- 6: addl %ecx, %eax
- adcl $0, %eax
- 7:
-@@ -408,7 +428,7 @@ DST( movb %cl, (%edi) )
-
- 6001:
- movl ARGBASE+20(%esp), %ebx # src_err_ptr
-- movl $-EFAULT, (%ebx)
-+ movl $-EFAULT, %ss:(%ebx)
-
- # zero the complete destination - computing the rest
- # is too much work
-@@ -421,11 +441,19 @@ DST( movb %cl, (%edi) )
-
- 6002:
- movl ARGBASE+24(%esp), %ebx # dst_err_ptr
-- movl $-EFAULT,(%ebx)
-+ movl $-EFAULT,%ss:(%ebx)
- jmp 5000b
-
- .previous
-
-+ pushl %ss
-+ CFI_ADJUST_CFA_OFFSET 4
-+ popl %ds
-+ CFI_ADJUST_CFA_OFFSET -4
-+ pushl %ss
-+ CFI_ADJUST_CFA_OFFSET 4
-+ popl %es
-+ CFI_ADJUST_CFA_OFFSET -4
- popl %ebx
- CFI_ADJUST_CFA_OFFSET -4
- CFI_RESTORE ebx
-@@ -439,26 +467,47 @@ DST( movb %cl, (%edi) )
- CFI_ADJUST_CFA_OFFSET -4
- ret
- CFI_ENDPROC
--ENDPROC(csum_partial_copy_generic)
-+ENDPROC(csum_partial_copy_generic_to_user)
-
- #else
-
- /* Version for PentiumII/PPro */
-
- #define ROUND1(x) \
-+ nop; nop; nop; \
- SRC(movl x(%esi), %ebx ) ; \
- addl %ebx, %eax ; \
-- DST(movl %ebx, x(%edi) ) ;
-+ DST(movl %ebx, %es:x(%edi)) ;
-
- #define ROUND(x) \
-+ nop; nop; nop; \
- SRC(movl x(%esi), %ebx ) ; \
- adcl %ebx, %eax ; \
-- DST(movl %ebx, x(%edi) ) ;
-+ DST(movl %ebx, %es:x(%edi)) ;
-
- #define ARGBASE 12
--
--ENTRY(csum_partial_copy_generic)
-+
-+ENTRY(csum_partial_copy_generic_to_user)
- CFI_STARTPROC
-+
-+#ifdef CONFIG_PAX_MEMORY_UDEREF
-+ pushl %gs
-+ CFI_ADJUST_CFA_OFFSET 4
-+ popl %es
-+ CFI_ADJUST_CFA_OFFSET -4
-+ jmp csum_partial_copy_generic
-+#endif
-+
-+ENTRY(csum_partial_copy_generic_from_user)
-+
-+#ifdef CONFIG_PAX_MEMORY_UDEREF
-+ pushl %gs
-+ CFI_ADJUST_CFA_OFFSET 4
-+ popl %ds
-+ CFI_ADJUST_CFA_OFFSET -4
-+#endif
-+
-+ENTRY(csum_partial_copy_generic)
- pushl %ebx
- CFI_ADJUST_CFA_OFFSET 4
- CFI_REL_OFFSET ebx, 0
-@@ -482,7 +531,7 @@ ENTRY(csum_partial_copy_generic)
- subl %ebx, %edi
- lea -1(%esi),%edx
- andl $-32,%edx
-- lea 3f(%ebx,%ebx), %ebx
-+ lea 3f(%ebx,%ebx,2), %ebx
- testl %esi, %esi
- jmp *%ebx
- 1: addl $64,%esi
-@@ -503,19 +552,19 @@ ENTRY(csum_partial_copy_generic)
- jb 5f
- SRC( movw (%esi), %dx )
- leal 2(%esi), %esi
--DST( movw %dx, (%edi) )
-+DST( movw %dx, %es:(%edi) )
- leal 2(%edi), %edi
- je 6f
- shll $16,%edx
- 5:
- SRC( movb (%esi), %dl )
--DST( movb %dl, (%edi) )
-+DST( movb %dl, %es:(%edi) )
- 6: addl %edx, %eax
- adcl $0, %eax
- 7:
- .section .fixup, "ax"
- 6001: movl ARGBASE+20(%esp), %ebx # src_err_ptr
-- movl $-EFAULT, (%ebx)
-+ movl $-EFAULT, %ss:(%ebx)
- # zero the complete destination (computing the rest is too much work)
- movl ARGBASE+8(%esp),%edi # dst
- movl ARGBASE+12(%esp),%ecx # len
-@@ -523,10 +572,21 @@ DST( movb %dl, (%edi) )
- rep; stosb
- jmp 7b
- 6002: movl ARGBASE+24(%esp), %ebx # dst_err_ptr
-- movl $-EFAULT, (%ebx)
-+ movl $-EFAULT, %ss:(%ebx)
- jmp 7b
- .previous
-
-+#ifdef CONFIG_PAX_MEMORY_UDEREF
-+ pushl %ss
-+ CFI_ADJUST_CFA_OFFSET 4
-+ popl %ds
-+ CFI_ADJUST_CFA_OFFSET -4
-+ pushl %ss
-+ CFI_ADJUST_CFA_OFFSET 4
-+ popl %es
-+ CFI_ADJUST_CFA_OFFSET -4
-+#endif
-+
- popl %esi
- CFI_ADJUST_CFA_OFFSET -4
- CFI_RESTORE esi
-@@ -538,7 +598,7 @@ DST( movb %dl, (%edi) )
- CFI_RESTORE ebx
- ret
- CFI_ENDPROC
--ENDPROC(csum_partial_copy_generic)
-+ENDPROC(csum_partial_copy_generic_to_user)
-
- #undef ROUND
- #undef ROUND1
-diff -urNp linux-2.6.37.4/arch/x86/lib/clear_page_64.S linux-2.6.37.4/arch/x86/lib/clear_page_64.S
---- linux-2.6.37.4/arch/x86/lib/clear_page_64.S 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/lib/clear_page_64.S 2011-01-17 02:41:01.000000000 -0500
-@@ -43,7 +43,7 @@ ENDPROC(clear_page)
-
- #include <asm/cpufeature.h>
-
-- .section .altinstr_replacement,"ax"
-+ .section .altinstr_replacement,"a"
- 1: .byte 0xeb /* jmp <disp8> */
- .byte (clear_page_c - clear_page) - (2f - 1b) /* offset */
- 2:
-diff -urNp linux-2.6.37.4/arch/x86/lib/copy_page_64.S linux-2.6.37.4/arch/x86/lib/copy_page_64.S
---- linux-2.6.37.4/arch/x86/lib/copy_page_64.S 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/lib/copy_page_64.S 2011-01-17 02:41:01.000000000 -0500
-@@ -104,7 +104,7 @@ ENDPROC(copy_page)
-
- #include <asm/cpufeature.h>
-
-- .section .altinstr_replacement,"ax"
-+ .section .altinstr_replacement,"a"
- 1: .byte 0xeb /* jmp <disp8> */
- .byte (copy_page_c - copy_page) - (2f - 1b) /* offset */
- 2:
-diff -urNp linux-2.6.37.4/arch/x86/lib/copy_user_64.S linux-2.6.37.4/arch/x86/lib/copy_user_64.S
---- linux-2.6.37.4/arch/x86/lib/copy_user_64.S 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/lib/copy_user_64.S 2011-01-17 02:41:01.000000000 -0500
-@@ -15,13 +15,14 @@
- #include <asm/asm-offsets.h>
- #include <asm/thread_info.h>
- #include <asm/cpufeature.h>
-+#include <asm/pgtable.h>
-
- .macro ALTERNATIVE_JUMP feature,orig,alt
- 0:
- .byte 0xe9 /* 32bit jump */
- .long \orig-1f /* by default jump to orig */
- 1:
-- .section .altinstr_replacement,"ax"
-+ .section .altinstr_replacement,"a"
- 2: .byte 0xe9 /* near jump with 32bit immediate */
- .long \alt-1b /* offset */ /* or alternatively to alt */
- .previous
-@@ -64,37 +65,13 @@
- #endif
- .endm
-
--/* Standard copy_to_user with segment limit checking */
--ENTRY(_copy_to_user)
-- CFI_STARTPROC
-- GET_THREAD_INFO(%rax)
-- movq %rdi,%rcx
-- addq %rdx,%rcx
-- jc bad_to_user
-- cmpq TI_addr_limit(%rax),%rcx
-- jae bad_to_user
-- ALTERNATIVE_JUMP X86_FEATURE_REP_GOOD,copy_user_generic_unrolled,copy_user_generic_string
-- CFI_ENDPROC
--ENDPROC(_copy_to_user)
--
--/* Standard copy_from_user with segment limit checking */
--ENTRY(_copy_from_user)
-- CFI_STARTPROC
-- GET_THREAD_INFO(%rax)
-- movq %rsi,%rcx
-- addq %rdx,%rcx
-- jc bad_from_user
-- cmpq TI_addr_limit(%rax),%rcx
-- jae bad_from_user
-- ALTERNATIVE_JUMP X86_FEATURE_REP_GOOD,copy_user_generic_unrolled,copy_user_generic_string
-- CFI_ENDPROC
--ENDPROC(_copy_from_user)
--
- .section .fixup,"ax"
- /* must zero dest */
- ENTRY(bad_from_user)
- bad_from_user:
- CFI_STARTPROC
-+ testl %edx,%edx
-+ js bad_to_user
- movl %edx,%ecx
- xorl %eax,%eax
- rep
-diff -urNp linux-2.6.37.4/arch/x86/lib/copy_user_nocache_64.S linux-2.6.37.4/arch/x86/lib/copy_user_nocache_64.S
---- linux-2.6.37.4/arch/x86/lib/copy_user_nocache_64.S 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/lib/copy_user_nocache_64.S 2011-01-17 02:41:01.000000000 -0500
-@@ -14,6 +14,7 @@
- #include <asm/current.h>
- #include <asm/asm-offsets.h>
- #include <asm/thread_info.h>
-+#include <asm/pgtable.h>
-
- .macro ALIGN_DESTINATION
- #ifdef FIX_ALIGNMENT
-@@ -50,6 +51,15 @@
- */
- ENTRY(__copy_user_nocache)
- CFI_STARTPROC
-+
-+#ifdef CONFIG_PAX_MEMORY_UDEREF
-+ mov $PAX_USER_SHADOW_BASE,%rcx
-+ cmp %rcx,%rsi
-+ jae 1f
-+ add %rcx,%rsi
-+1:
-+#endif
-+
- cmpl $8,%edx
- jb 20f /* less then 8 bytes, go to byte copy loop */
- ALIGN_DESTINATION
-diff -urNp linux-2.6.37.4/arch/x86/lib/csum-wrappers_64.c linux-2.6.37.4/arch/x86/lib/csum-wrappers_64.c
---- linux-2.6.37.4/arch/x86/lib/csum-wrappers_64.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/lib/csum-wrappers_64.c 2011-01-17 02:41:01.000000000 -0500
-@@ -52,6 +52,8 @@ csum_partial_copy_from_user(const void _
- len -= 2;
- }
- }
-+ if ((unsigned long)src < PAX_USER_SHADOW_BASE)
-+ src += PAX_USER_SHADOW_BASE;
- isum = csum_partial_copy_generic((__force const void *)src,
- dst, len, isum, errp, NULL);
- if (unlikely(*errp))
-@@ -105,6 +107,8 @@ csum_partial_copy_to_user(const void *sr
- }
-
- *errp = 0;
-+ if ((unsigned long)dst < PAX_USER_SHADOW_BASE)
-+ dst += PAX_USER_SHADOW_BASE;
- return csum_partial_copy_generic(src, (void __force *)dst,
- len, isum, NULL, errp);
- }
-diff -urNp linux-2.6.37.4/arch/x86/lib/getuser.S linux-2.6.37.4/arch/x86/lib/getuser.S
---- linux-2.6.37.4/arch/x86/lib/getuser.S 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/lib/getuser.S 2011-03-12 22:50:25.000000000 -0500
-@@ -33,14 +33,35 @@
- #include <asm/asm-offsets.h>
- #include <asm/thread_info.h>
- #include <asm/asm.h>
-+#include <asm/segment.h>
-+#include <asm/pgtable.h>
-+
-+#if defined(CONFIG_X86_32) && defined(CONFIG_PAX_MEMORY_UDEREF)
-+#define __copyuser_seg gs;
-+#else
-+#define __copyuser_seg
-+#endif
-
- .text
- ENTRY(__get_user_1)
- CFI_STARTPROC
-+
-+#if !defined(CONFIG_X86_32) || !defined(CONFIG_PAX_MEMORY_UDEREF)
- GET_THREAD_INFO(%_ASM_DX)
- cmp TI_addr_limit(%_ASM_DX),%_ASM_AX
- jae bad_get_user
--1: movzb (%_ASM_AX),%edx
-+
-+#if defined(CONFIG_X86_64) && defined(CONFIG_PAX_MEMORY_UDEREF)
-+ mov $PAX_USER_SHADOW_BASE,%_ASM_DX
-+ cmp %_ASM_DX,%_ASM_AX
-+ jae 1234f
-+ add %_ASM_DX,%_ASM_AX
-+1234:
-+#endif
-+
-+#endif
-+
-+1: __copyuser_seg movzb (%_ASM_AX),%edx
- xor %eax,%eax
- ret
- CFI_ENDPROC
-@@ -49,11 +70,24 @@ ENDPROC(__get_user_1)
- ENTRY(__get_user_2)
- CFI_STARTPROC
- add $1,%_ASM_AX
-+
-+#if !defined(CONFIG_X86_32) || !defined(CONFIG_PAX_MEMORY_UDEREF)
- jc bad_get_user
- GET_THREAD_INFO(%_ASM_DX)
- cmp TI_addr_limit(%_ASM_DX),%_ASM_AX
- jae bad_get_user
--2: movzwl -1(%_ASM_AX),%edx
-+
-+#if defined(CONFIG_X86_64) && defined(CONFIG_PAX_MEMORY_UDEREF)
-+ mov $PAX_USER_SHADOW_BASE,%_ASM_DX
-+ cmp %_ASM_DX,%_ASM_AX
-+ jae 1234f
-+ add %_ASM_DX,%_ASM_AX
-+1234:
-+#endif
-+
-+#endif
-+
-+2: __copyuser_seg movzwl -1(%_ASM_AX),%edx
- xor %eax,%eax
- ret
- CFI_ENDPROC
-@@ -62,11 +96,24 @@ ENDPROC(__get_user_2)
- ENTRY(__get_user_4)
- CFI_STARTPROC
- add $3,%_ASM_AX
-+
-+#if !defined(CONFIG_X86_32) || !defined(CONFIG_PAX_MEMORY_UDEREF)
- jc bad_get_user
- GET_THREAD_INFO(%_ASM_DX)
- cmp TI_addr_limit(%_ASM_DX),%_ASM_AX
- jae bad_get_user
--3: mov -3(%_ASM_AX),%edx
-+
-+#if defined(CONFIG_X86_64) && defined(CONFIG_PAX_MEMORY_UDEREF)
-+ mov $PAX_USER_SHADOW_BASE,%_ASM_DX
-+ cmp %_ASM_DX,%_ASM_AX
-+ jae 1234f
-+ add %_ASM_DX,%_ASM_AX
-+1234:
-+#endif
-+
-+#endif
-+
-+3: __copyuser_seg mov -3(%_ASM_AX),%edx
- xor %eax,%eax
- ret
- CFI_ENDPROC
-@@ -80,6 +127,15 @@ ENTRY(__get_user_8)
- GET_THREAD_INFO(%_ASM_DX)
- cmp TI_addr_limit(%_ASM_DX),%_ASM_AX
- jae bad_get_user
-+
-+#ifdef CONFIG_PAX_MEMORY_UDEREF
-+ mov $PAX_USER_SHADOW_BASE,%_ASM_DX
-+ cmp %_ASM_DX,%_ASM_AX
-+ jae 1234f
-+ add %_ASM_DX,%_ASM_AX
-+1234:
-+#endif
-+
- 4: movq -7(%_ASM_AX),%_ASM_DX
- xor %eax,%eax
- ret
-diff -urNp linux-2.6.37.4/arch/x86/lib/insn.c linux-2.6.37.4/arch/x86/lib/insn.c
---- linux-2.6.37.4/arch/x86/lib/insn.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/lib/insn.c 2011-01-24 18:04:15.000000000 -0500
-@@ -21,6 +21,11 @@
- #include <linux/string.h>
- #include <asm/inat.h>
- #include <asm/insn.h>
-+#ifdef __KERNEL__
-+#include <asm/pgtable_types.h>
-+#else
-+#define ktla_ktva(addr) addr
-+#endif
-
- #define get_next(t, insn) \
- ({t r; r = *(t*)insn->next_byte; insn->next_byte += sizeof(t); r; })
-@@ -40,8 +45,8 @@
- void insn_init(struct insn *insn, const void *kaddr, int x86_64)
- {
- memset(insn, 0, sizeof(*insn));
-- insn->kaddr = kaddr;
-- insn->next_byte = kaddr;
-+ insn->kaddr = ktla_ktva(kaddr);
-+ insn->next_byte = ktla_ktva(kaddr);
- insn->x86_64 = x86_64 ? 1 : 0;
- insn->opnd_bytes = 4;
- if (x86_64)
-diff -urNp linux-2.6.37.4/arch/x86/lib/mmx_32.c linux-2.6.37.4/arch/x86/lib/mmx_32.c
---- linux-2.6.37.4/arch/x86/lib/mmx_32.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/lib/mmx_32.c 2011-01-17 02:41:01.000000000 -0500
-@@ -29,6 +29,7 @@ void *_mmx_memcpy(void *to, const void *
- {
- void *p;
- int i;
-+ unsigned long cr0;
-
- if (unlikely(in_interrupt()))
- return __memcpy(to, from, len);
-@@ -39,44 +40,72 @@ void *_mmx_memcpy(void *to, const void *
- kernel_fpu_begin();
-
- __asm__ __volatile__ (
-- "1: prefetch (%0)\n" /* This set is 28 bytes */
-- " prefetch 64(%0)\n"
-- " prefetch 128(%0)\n"
-- " prefetch 192(%0)\n"
-- " prefetch 256(%0)\n"
-+ "1: prefetch (%1)\n" /* This set is 28 bytes */
-+ " prefetch 64(%1)\n"
-+ " prefetch 128(%1)\n"
-+ " prefetch 192(%1)\n"
-+ " prefetch 256(%1)\n"
- "2: \n"
- ".section .fixup, \"ax\"\n"
-- "3: movw $0x1AEB, 1b\n" /* jmp on 26 bytes */
-+ "3: \n"
-+
-+#ifdef CONFIG_PAX_KERNEXEC
-+ " movl %%cr0, %0\n"
-+ " movl %0, %%eax\n"
-+ " andl $0xFFFEFFFF, %%eax\n"
-+ " movl %%eax, %%cr0\n"
-+#endif
-+
-+ " movw $0x1AEB, 1b\n" /* jmp on 26 bytes */
-+
-+#ifdef CONFIG_PAX_KERNEXEC
-+ " movl %0, %%cr0\n"
-+#endif
-+
- " jmp 2b\n"
- ".previous\n"
- _ASM_EXTABLE(1b, 3b)
-- : : "r" (from));
-+ : "=&r" (cr0) : "r" (from) : "ax");
-
- for ( ; i > 5; i--) {
- __asm__ __volatile__ (
-- "1: prefetch 320(%0)\n"
-- "2: movq (%0), %%mm0\n"
-- " movq 8(%0), %%mm1\n"
-- " movq 16(%0), %%mm2\n"
-- " movq 24(%0), %%mm3\n"
-- " movq %%mm0, (%1)\n"
-- " movq %%mm1, 8(%1)\n"
-- " movq %%mm2, 16(%1)\n"
-- " movq %%mm3, 24(%1)\n"
-- " movq 32(%0), %%mm0\n"
-- " movq 40(%0), %%mm1\n"
-- " movq 48(%0), %%mm2\n"
-- " movq 56(%0), %%mm3\n"
-- " movq %%mm0, 32(%1)\n"
-- " movq %%mm1, 40(%1)\n"
-- " movq %%mm2, 48(%1)\n"
-- " movq %%mm3, 56(%1)\n"
-+ "1: prefetch 320(%1)\n"
-+ "2: movq (%1), %%mm0\n"
-+ " movq 8(%1), %%mm1\n"
-+ " movq 16(%1), %%mm2\n"
-+ " movq 24(%1), %%mm3\n"
-+ " movq %%mm0, (%2)\n"
-+ " movq %%mm1, 8(%2)\n"
-+ " movq %%mm2, 16(%2)\n"
-+ " movq %%mm3, 24(%2)\n"
-+ " movq 32(%1), %%mm0\n"
-+ " movq 40(%1), %%mm1\n"
-+ " movq 48(%1), %%mm2\n"
-+ " movq 56(%1), %%mm3\n"
-+ " movq %%mm0, 32(%2)\n"
-+ " movq %%mm1, 40(%2)\n"
-+ " movq %%mm2, 48(%2)\n"
-+ " movq %%mm3, 56(%2)\n"
- ".section .fixup, \"ax\"\n"
-- "3: movw $0x05EB, 1b\n" /* jmp on 5 bytes */
-+ "3:\n"
-+
-+#ifdef CONFIG_PAX_KERNEXEC
-+ " movl %%cr0, %0\n"
-+ " movl %0, %%eax\n"
-+ " andl $0xFFFEFFFF, %%eax\n"
-+ " movl %%eax, %%cr0\n"
-+#endif
-+
-+ " movw $0x05EB, 1b\n" /* jmp on 5 bytes */
-+
-+#ifdef CONFIG_PAX_KERNEXEC
-+ " movl %0, %%cr0\n"
-+#endif
-+
- " jmp 2b\n"
- ".previous\n"
- _ASM_EXTABLE(1b, 3b)
-- : : "r" (from), "r" (to) : "memory");
-+ : "=&r" (cr0) : "r" (from), "r" (to) : "memory", "ax");
-
- from += 64;
- to += 64;
-@@ -158,6 +187,7 @@ static void fast_clear_page(void *page)
- static void fast_copy_page(void *to, void *from)
- {
- int i;
-+ unsigned long cr0;
-
- kernel_fpu_begin();
-
-@@ -166,42 +196,70 @@ static void fast_copy_page(void *to, voi
- * but that is for later. -AV
- */
- __asm__ __volatile__(
-- "1: prefetch (%0)\n"
-- " prefetch 64(%0)\n"
-- " prefetch 128(%0)\n"
-- " prefetch 192(%0)\n"
-- " prefetch 256(%0)\n"
-+ "1: prefetch (%1)\n"
-+ " prefetch 64(%1)\n"
-+ " prefetch 128(%1)\n"
-+ " prefetch 192(%1)\n"
-+ " prefetch 256(%1)\n"
- "2: \n"
- ".section .fixup, \"ax\"\n"
-- "3: movw $0x1AEB, 1b\n" /* jmp on 26 bytes */
-+ "3: \n"
-+
-+#ifdef CONFIG_PAX_KERNEXEC
-+ " movl %%cr0, %0\n"
-+ " movl %0, %%eax\n"
-+ " andl $0xFFFEFFFF, %%eax\n"
-+ " movl %%eax, %%cr0\n"
-+#endif
-+
-+ " movw $0x1AEB, 1b\n" /* jmp on 26 bytes */
-+
-+#ifdef CONFIG_PAX_KERNEXEC
-+ " movl %0, %%cr0\n"
-+#endif
-+
- " jmp 2b\n"
- ".previous\n"
-- _ASM_EXTABLE(1b, 3b) : : "r" (from));
-+ _ASM_EXTABLE(1b, 3b) : "=&r" (cr0) : "r" (from) : "ax");
-
- for (i = 0; i < (4096-320)/64; i++) {
- __asm__ __volatile__ (
-- "1: prefetch 320(%0)\n"
-- "2: movq (%0), %%mm0\n"
-- " movntq %%mm0, (%1)\n"
-- " movq 8(%0), %%mm1\n"
-- " movntq %%mm1, 8(%1)\n"
-- " movq 16(%0), %%mm2\n"
-- " movntq %%mm2, 16(%1)\n"
-- " movq 24(%0), %%mm3\n"
-- " movntq %%mm3, 24(%1)\n"
-- " movq 32(%0), %%mm4\n"
-- " movntq %%mm4, 32(%1)\n"
-- " movq 40(%0), %%mm5\n"
-- " movntq %%mm5, 40(%1)\n"
-- " movq 48(%0), %%mm6\n"
-- " movntq %%mm6, 48(%1)\n"
-- " movq 56(%0), %%mm7\n"
-- " movntq %%mm7, 56(%1)\n"
-+ "1: prefetch 320(%1)\n"
-+ "2: movq (%1), %%mm0\n"
-+ " movntq %%mm0, (%2)\n"
-+ " movq 8(%1), %%mm1\n"
-+ " movntq %%mm1, 8(%2)\n"
-+ " movq 16(%1), %%mm2\n"
-+ " movntq %%mm2, 16(%2)\n"
-+ " movq 24(%1), %%mm3\n"
-+ " movntq %%mm3, 24(%2)\n"
-+ " movq 32(%1), %%mm4\n"
-+ " movntq %%mm4, 32(%2)\n"
-+ " movq 40(%1), %%mm5\n"
-+ " movntq %%mm5, 40(%2)\n"
-+ " movq 48(%1), %%mm6\n"
-+ " movntq %%mm6, 48(%2)\n"
-+ " movq 56(%1), %%mm7\n"
-+ " movntq %%mm7, 56(%2)\n"
- ".section .fixup, \"ax\"\n"
-- "3: movw $0x05EB, 1b\n" /* jmp on 5 bytes */
-+ "3:\n"
-+
-+#ifdef CONFIG_PAX_KERNEXEC
-+ " movl %%cr0, %0\n"
-+ " movl %0, %%eax\n"
-+ " andl $0xFFFEFFFF, %%eax\n"
-+ " movl %%eax, %%cr0\n"
-+#endif
-+
-+ " movw $0x05EB, 1b\n" /* jmp on 5 bytes */
-+
-+#ifdef CONFIG_PAX_KERNEXEC
-+ " movl %0, %%cr0\n"
-+#endif
-+
- " jmp 2b\n"
- ".previous\n"
-- _ASM_EXTABLE(1b, 3b) : : "r" (from), "r" (to) : "memory");
-+ _ASM_EXTABLE(1b, 3b) : "=&r" (cr0) : "r" (from), "r" (to) : "memory", "ax");
-
- from += 64;
- to += 64;
-@@ -280,47 +338,76 @@ static void fast_clear_page(void *page)
- static void fast_copy_page(void *to, void *from)
- {
- int i;
-+ unsigned long cr0;
-
- kernel_fpu_begin();
-
- __asm__ __volatile__ (
-- "1: prefetch (%0)\n"
-- " prefetch 64(%0)\n"
-- " prefetch 128(%0)\n"
-- " prefetch 192(%0)\n"
-- " prefetch 256(%0)\n"
-+ "1: prefetch (%1)\n"
-+ " prefetch 64(%1)\n"
-+ " prefetch 128(%1)\n"
-+ " prefetch 192(%1)\n"
-+ " prefetch 256(%1)\n"
- "2: \n"
- ".section .fixup, \"ax\"\n"
-- "3: movw $0x1AEB, 1b\n" /* jmp on 26 bytes */
-+ "3: \n"
-+
-+#ifdef CONFIG_PAX_KERNEXEC
-+ " movl %%cr0, %0\n"
-+ " movl %0, %%eax\n"
-+ " andl $0xFFFEFFFF, %%eax\n"
-+ " movl %%eax, %%cr0\n"
-+#endif
-+
-+ " movw $0x1AEB, 1b\n" /* jmp on 26 bytes */
-+
-+#ifdef CONFIG_PAX_KERNEXEC
-+ " movl %0, %%cr0\n"
-+#endif
-+
- " jmp 2b\n"
- ".previous\n"
-- _ASM_EXTABLE(1b, 3b) : : "r" (from));
-+ _ASM_EXTABLE(1b, 3b) : "=&r" (cr0) : "r" (from) : "ax");
-
- for (i = 0; i < 4096/64; i++) {
- __asm__ __volatile__ (
-- "1: prefetch 320(%0)\n"
-- "2: movq (%0), %%mm0\n"
-- " movq 8(%0), %%mm1\n"
-- " movq 16(%0), %%mm2\n"
-- " movq 24(%0), %%mm3\n"
-- " movq %%mm0, (%1)\n"
-- " movq %%mm1, 8(%1)\n"
-- " movq %%mm2, 16(%1)\n"
-- " movq %%mm3, 24(%1)\n"
-- " movq 32(%0), %%mm0\n"
-- " movq 40(%0), %%mm1\n"
-- " movq 48(%0), %%mm2\n"
-- " movq 56(%0), %%mm3\n"
-- " movq %%mm0, 32(%1)\n"
-- " movq %%mm1, 40(%1)\n"
-- " movq %%mm2, 48(%1)\n"
-- " movq %%mm3, 56(%1)\n"
-+ "1: prefetch 320(%1)\n"
-+ "2: movq (%1), %%mm0\n"
-+ " movq 8(%1), %%mm1\n"
-+ " movq 16(%1), %%mm2\n"
-+ " movq 24(%1), %%mm3\n"
-+ " movq %%mm0, (%2)\n"
-+ " movq %%mm1, 8(%2)\n"
-+ " movq %%mm2, 16(%2)\n"
-+ " movq %%mm3, 24(%2)\n"
-+ " movq 32(%1), %%mm0\n"
-+ " movq 40(%1), %%mm1\n"
-+ " movq 48(%1), %%mm2\n"
-+ " movq 56(%1), %%mm3\n"
-+ " movq %%mm0, 32(%2)\n"
-+ " movq %%mm1, 40(%2)\n"
-+ " movq %%mm2, 48(%2)\n"
-+ " movq %%mm3, 56(%2)\n"
- ".section .fixup, \"ax\"\n"
-- "3: movw $0x05EB, 1b\n" /* jmp on 5 bytes */
-+ "3:\n"
-+
-+#ifdef CONFIG_PAX_KERNEXEC
-+ " movl %%cr0, %0\n"
-+ " movl %0, %%eax\n"
-+ " andl $0xFFFEFFFF, %%eax\n"
-+ " movl %%eax, %%cr0\n"
-+#endif
-+
-+ " movw $0x05EB, 1b\n" /* jmp on 5 bytes */
-+
-+#ifdef CONFIG_PAX_KERNEXEC
-+ " movl %0, %%cr0\n"
-+#endif
-+
- " jmp 2b\n"
- ".previous\n"
- _ASM_EXTABLE(1b, 3b)
-- : : "r" (from), "r" (to) : "memory");
-+ : "=&r" (cr0) : "r" (from), "r" (to) : "memory", "ax");
-
- from += 64;
- to += 64;
-diff -urNp linux-2.6.37.4/arch/x86/lib/putuser.S linux-2.6.37.4/arch/x86/lib/putuser.S
---- linux-2.6.37.4/arch/x86/lib/putuser.S 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/lib/putuser.S 2011-03-12 22:50:25.000000000 -0500
-@@ -15,7 +15,8 @@
- #include <asm/thread_info.h>
- #include <asm/errno.h>
- #include <asm/asm.h>
--
-+#include <asm/segment.h>
-+#include <asm/pgtable.h>
-
- /*
- * __put_user_X
-@@ -29,52 +30,119 @@
- * as they get called from within inline assembly.
- */
-
--#define ENTER CFI_STARTPROC ; \
-- GET_THREAD_INFO(%_ASM_BX)
-+#define ENTER CFI_STARTPROC
- #define EXIT ret ; \
- CFI_ENDPROC
-
-+#if defined(CONFIG_X86_64) && defined(CONFIG_PAX_MEMORY_UDEREF)
-+#define _DEST %_ASM_CX,%_ASM_BX
-+#else
-+#define _DEST %_ASM_CX
-+#endif
-+
-+#if defined(CONFIG_X86_32) && defined(CONFIG_PAX_MEMORY_UDEREF)
-+#define __copyuser_seg gs;
-+#else
-+#define __copyuser_seg
-+#endif
-+
- .text
- ENTRY(__put_user_1)
- ENTER
-+
-+#if !defined(CONFIG_X86_32) || !defined(CONFIG_PAX_MEMORY_UDEREF)
-+ GET_THREAD_INFO(%_ASM_BX)
- cmp TI_addr_limit(%_ASM_BX),%_ASM_CX
- jae bad_put_user
--1: movb %al,(%_ASM_CX)
-+
-+#if defined(CONFIG_X86_64) && defined(CONFIG_PAX_MEMORY_UDEREF)
-+ mov $PAX_USER_SHADOW_BASE,%_ASM_BX
-+ cmp %_ASM_BX,%_ASM_CX
-+ jb 1234f
-+ xor %ebx,%ebx
-+1234:
-+#endif
-+
-+#endif
-+
-+1: __copyuser_seg movb %al,(_DEST)
- xor %eax,%eax
- EXIT
- ENDPROC(__put_user_1)
-
- ENTRY(__put_user_2)
- ENTER
-+
-+#if !defined(CONFIG_X86_32) || !defined(CONFIG_PAX_MEMORY_UDEREF)
-+ GET_THREAD_INFO(%_ASM_BX)
- mov TI_addr_limit(%_ASM_BX),%_ASM_BX
- sub $1,%_ASM_BX
- cmp %_ASM_BX,%_ASM_CX
- jae bad_put_user
--2: movw %ax,(%_ASM_CX)
-+
-+#if defined(CONFIG_X86_64) && defined(CONFIG_PAX_MEMORY_UDEREF)
-+ mov $PAX_USER_SHADOW_BASE,%_ASM_BX
-+ cmp %_ASM_BX,%_ASM_CX
-+ jb 1234f
-+ xor %ebx,%ebx
-+1234:
-+#endif
-+
-+#endif
-+
-+2: __copyuser_seg movw %ax,(_DEST)
- xor %eax,%eax
- EXIT
- ENDPROC(__put_user_2)
-
- ENTRY(__put_user_4)
- ENTER
-+
-+#if !defined(CONFIG_X86_32) || !defined(CONFIG_PAX_MEMORY_UDEREF)
-+ GET_THREAD_INFO(%_ASM_BX)
- mov TI_addr_limit(%_ASM_BX),%_ASM_BX
- sub $3,%_ASM_BX
- cmp %_ASM_BX,%_ASM_CX
- jae bad_put_user
--3: movl %eax,(%_ASM_CX)
-+
-+#if defined(CONFIG_X86_64) && defined(CONFIG_PAX_MEMORY_UDEREF)
-+ mov $PAX_USER_SHADOW_BASE,%_ASM_BX
-+ cmp %_ASM_BX,%_ASM_CX
-+ jb 1234f
-+ xor %ebx,%ebx
-+1234:
-+#endif
-+
-+#endif
-+
-+3: __copyuser_seg movl %eax,(_DEST)
- xor %eax,%eax
- EXIT
- ENDPROC(__put_user_4)
-
- ENTRY(__put_user_8)
- ENTER
-+
-+#if !defined(CONFIG_X86_32) || !defined(CONFIG_PAX_MEMORY_UDEREF)
-+ GET_THREAD_INFO(%_ASM_BX)
- mov TI_addr_limit(%_ASM_BX),%_ASM_BX
- sub $7,%_ASM_BX
- cmp %_ASM_BX,%_ASM_CX
- jae bad_put_user
--4: mov %_ASM_AX,(%_ASM_CX)
-+
-+#if defined(CONFIG_X86_64) && defined(CONFIG_PAX_MEMORY_UDEREF)
-+ mov $PAX_USER_SHADOW_BASE,%_ASM_BX
-+ cmp %_ASM_BX,%_ASM_CX
-+ jb 1234f
-+ xor %ebx,%ebx
-+1234:
-+#endif
-+
-+#endif
-+
-+4: __copyuser_seg mov %_ASM_AX,(_DEST)
- #ifdef CONFIG_X86_32
--5: movl %edx,4(%_ASM_CX)
-+5: __copyuser_seg movl %edx,4(_DEST)
- #endif
- xor %eax,%eax
- EXIT
-diff -urNp linux-2.6.37.4/arch/x86/lib/usercopy_32.c linux-2.6.37.4/arch/x86/lib/usercopy_32.c
---- linux-2.6.37.4/arch/x86/lib/usercopy_32.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/lib/usercopy_32.c 2011-03-12 22:50:25.000000000 -0500
-@@ -43,7 +43,7 @@ do { \
- __asm__ __volatile__( \
- " testl %1,%1\n" \
- " jz 2f\n" \
-- "0: lodsb\n" \
-+ "0: "__copyuser_seg"lodsb\n" \
- " stosb\n" \
- " testb %%al,%%al\n" \
- " jz 1f\n" \
-@@ -128,10 +128,12 @@ do { \
- int __d0; \
- might_fault(); \
- __asm__ __volatile__( \
-+ __COPYUSER_SET_ES \
- "0: rep; stosl\n" \
- " movl %2,%0\n" \
- "1: rep; stosb\n" \
- "2:\n" \
-+ __COPYUSER_RESTORE_ES \
- ".section .fixup,\"ax\"\n" \
- "3: lea 0(%2,%0,4),%0\n" \
- " jmp 2b\n" \
-@@ -200,6 +202,7 @@ long strnlen_user(const char __user *s,
- might_fault();
-
- __asm__ __volatile__(
-+ __COPYUSER_SET_ES
- " testl %0, %0\n"
- " jz 3f\n"
- " andl %0,%%ecx\n"
-@@ -208,6 +211,7 @@ long strnlen_user(const char __user *s,
- " subl %%ecx,%0\n"
- " addl %0,%%eax\n"
- "1:\n"
-+ __COPYUSER_RESTORE_ES
- ".section .fixup,\"ax\"\n"
- "2: xorl %%eax,%%eax\n"
- " jmp 1b\n"
-@@ -227,7 +231,7 @@ EXPORT_SYMBOL(strnlen_user);
-
- #ifdef CONFIG_X86_INTEL_USERCOPY
- static unsigned long
--__copy_user_intel(void __user *to, const void *from, unsigned long size)
-+__generic_copy_to_user_intel(void __user *to, const void *from, unsigned long size)
- {
- int d0, d1;
- __asm__ __volatile__(
-@@ -239,36 +243,36 @@ __copy_user_intel(void __user *to, const
- " .align 2,0x90\n"
- "3: movl 0(%4), %%eax\n"
- "4: movl 4(%4), %%edx\n"
-- "5: movl %%eax, 0(%3)\n"
-- "6: movl %%edx, 4(%3)\n"
-+ "5: "__copyuser_seg" movl %%eax, 0(%3)\n"
-+ "6: "__copyuser_seg" movl %%edx, 4(%3)\n"
- "7: movl 8(%4), %%eax\n"
- "8: movl 12(%4),%%edx\n"
-- "9: movl %%eax, 8(%3)\n"
-- "10: movl %%edx, 12(%3)\n"
-+ "9: "__copyuser_seg" movl %%eax, 8(%3)\n"
-+ "10: "__copyuser_seg" movl %%edx, 12(%3)\n"
- "11: movl 16(%4), %%eax\n"
- "12: movl 20(%4), %%edx\n"
-- "13: movl %%eax, 16(%3)\n"
-- "14: movl %%edx, 20(%3)\n"
-+ "13: "__copyuser_seg" movl %%eax, 16(%3)\n"
-+ "14: "__copyuser_seg" movl %%edx, 20(%3)\n"
- "15: movl 24(%4), %%eax\n"
- "16: movl 28(%4), %%edx\n"
-- "17: movl %%eax, 24(%3)\n"
-- "18: movl %%edx, 28(%3)\n"
-+ "17: "__copyuser_seg" movl %%eax, 24(%3)\n"
-+ "18: "__copyuser_seg" movl %%edx, 28(%3)\n"
- "19: movl 32(%4), %%eax\n"
- "20: movl 36(%4), %%edx\n"
-- "21: movl %%eax, 32(%3)\n"
-- "22: movl %%edx, 36(%3)\n"
-+ "21: "__copyuser_seg" movl %%eax, 32(%3)\n"
-+ "22: "__copyuser_seg" movl %%edx, 36(%3)\n"
- "23: movl 40(%4), %%eax\n"
- "24: movl 44(%4), %%edx\n"
-- "25: movl %%eax, 40(%3)\n"
-- "26: movl %%edx, 44(%3)\n"
-+ "25: "__copyuser_seg" movl %%eax, 40(%3)\n"
-+ "26: "__copyuser_seg" movl %%edx, 44(%3)\n"
- "27: movl 48(%4), %%eax\n"
- "28: movl 52(%4), %%edx\n"
-- "29: movl %%eax, 48(%3)\n"
-- "30: movl %%edx, 52(%3)\n"
-+ "29: "__copyuser_seg" movl %%eax, 48(%3)\n"
-+ "30: "__copyuser_seg" movl %%edx, 52(%3)\n"
- "31: movl 56(%4), %%eax\n"
- "32: movl 60(%4), %%edx\n"
-- "33: movl %%eax, 56(%3)\n"
-- "34: movl %%edx, 60(%3)\n"
-+ "33: "__copyuser_seg" movl %%eax, 56(%3)\n"
-+ "34: "__copyuser_seg" movl %%edx, 60(%3)\n"
- " addl $-64, %0\n"
- " addl $64, %4\n"
- " addl $64, %3\n"
-@@ -278,10 +282,119 @@ __copy_user_intel(void __user *to, const
- " shrl $2, %0\n"
- " andl $3, %%eax\n"
- " cld\n"
-+ __COPYUSER_SET_ES
- "99: rep; movsl\n"
- "36: movl %%eax, %0\n"
- "37: rep; movsb\n"
- "100:\n"
-+ __COPYUSER_RESTORE_ES
-+ ".section .fixup,\"ax\"\n"
-+ "101: lea 0(%%eax,%0,4),%0\n"
-+ " jmp 100b\n"
-+ ".previous\n"
-+ ".section __ex_table,\"a\"\n"
-+ " .align 4\n"
-+ " .long 1b,100b\n"
-+ " .long 2b,100b\n"
-+ " .long 3b,100b\n"
-+ " .long 4b,100b\n"
-+ " .long 5b,100b\n"
-+ " .long 6b,100b\n"
-+ " .long 7b,100b\n"
-+ " .long 8b,100b\n"
-+ " .long 9b,100b\n"
-+ " .long 10b,100b\n"
-+ " .long 11b,100b\n"
-+ " .long 12b,100b\n"
-+ " .long 13b,100b\n"
-+ " .long 14b,100b\n"
-+ " .long 15b,100b\n"
-+ " .long 16b,100b\n"
-+ " .long 17b,100b\n"
-+ " .long 18b,100b\n"
-+ " .long 19b,100b\n"
-+ " .long 20b,100b\n"
-+ " .long 21b,100b\n"
-+ " .long 22b,100b\n"
-+ " .long 23b,100b\n"
-+ " .long 24b,100b\n"
-+ " .long 25b,100b\n"
-+ " .long 26b,100b\n"
-+ " .long 27b,100b\n"
-+ " .long 28b,100b\n"
-+ " .long 29b,100b\n"
-+ " .long 30b,100b\n"
-+ " .long 31b,100b\n"
-+ " .long 32b,100b\n"
-+ " .long 33b,100b\n"
-+ " .long 34b,100b\n"
-+ " .long 35b,100b\n"
-+ " .long 36b,100b\n"
-+ " .long 37b,100b\n"
-+ " .long 99b,101b\n"
-+ ".previous"
-+ : "=&c"(size), "=&D" (d0), "=&S" (d1)
-+ : "1"(to), "2"(from), "0"(size)
-+ : "eax", "edx", "memory");
-+ return size;
-+}
-+
-+static unsigned long
-+__generic_copy_from_user_intel(void *to, const void __user *from, unsigned long size)
-+{
-+ int d0, d1;
-+ __asm__ __volatile__(
-+ " .align 2,0x90\n"
-+ "1: "__copyuser_seg" movl 32(%4), %%eax\n"
-+ " cmpl $67, %0\n"
-+ " jbe 3f\n"
-+ "2: "__copyuser_seg" movl 64(%4), %%eax\n"
-+ " .align 2,0x90\n"
-+ "3: "__copyuser_seg" movl 0(%4), %%eax\n"
-+ "4: "__copyuser_seg" movl 4(%4), %%edx\n"
-+ "5: movl %%eax, 0(%3)\n"
-+ "6: movl %%edx, 4(%3)\n"
-+ "7: "__copyuser_seg" movl 8(%4), %%eax\n"
-+ "8: "__copyuser_seg" movl 12(%4),%%edx\n"
-+ "9: movl %%eax, 8(%3)\n"
-+ "10: movl %%edx, 12(%3)\n"
-+ "11: "__copyuser_seg" movl 16(%4), %%eax\n"
-+ "12: "__copyuser_seg" movl 20(%4), %%edx\n"
-+ "13: movl %%eax, 16(%3)\n"
-+ "14: movl %%edx, 20(%3)\n"
-+ "15: "__copyuser_seg" movl 24(%4), %%eax\n"
-+ "16: "__copyuser_seg" movl 28(%4), %%edx\n"
-+ "17: movl %%eax, 24(%3)\n"
-+ "18: movl %%edx, 28(%3)\n"
-+ "19: "__copyuser_seg" movl 32(%4), %%eax\n"
-+ "20: "__copyuser_seg" movl 36(%4), %%edx\n"
-+ "21: movl %%eax, 32(%3)\n"
-+ "22: movl %%edx, 36(%3)\n"
-+ "23: "__copyuser_seg" movl 40(%4), %%eax\n"
-+ "24: "__copyuser_seg" movl 44(%4), %%edx\n"
-+ "25: movl %%eax, 40(%3)\n"
-+ "26: movl %%edx, 44(%3)\n"
-+ "27: "__copyuser_seg" movl 48(%4), %%eax\n"
-+ "28: "__copyuser_seg" movl 52(%4), %%edx\n"
-+ "29: movl %%eax, 48(%3)\n"
-+ "30: movl %%edx, 52(%3)\n"
-+ "31: "__copyuser_seg" movl 56(%4), %%eax\n"
-+ "32: "__copyuser_seg" movl 60(%4), %%edx\n"
-+ "33: movl %%eax, 56(%3)\n"
-+ "34: movl %%edx, 60(%3)\n"
-+ " addl $-64, %0\n"
-+ " addl $64, %4\n"
-+ " addl $64, %3\n"
-+ " cmpl $63, %0\n"
-+ " ja 1b\n"
-+ "35: movl %0, %%eax\n"
-+ " shrl $2, %0\n"
-+ " andl $3, %%eax\n"
-+ " cld\n"
-+ "99: rep; "__copyuser_seg" movsl\n"
-+ "36: movl %%eax, %0\n"
-+ "37: rep; "__copyuser_seg" movsb\n"
-+ "100:\n"
- ".section .fixup,\"ax\"\n"
- "101: lea 0(%%eax,%0,4),%0\n"
- " jmp 100b\n"
-@@ -339,41 +452,41 @@ __copy_user_zeroing_intel(void *to, cons
- int d0, d1;
- __asm__ __volatile__(
- " .align 2,0x90\n"
-- "0: movl 32(%4), %%eax\n"
-+ "0: "__copyuser_seg" movl 32(%4), %%eax\n"
- " cmpl $67, %0\n"
- " jbe 2f\n"
-- "1: movl 64(%4), %%eax\n"
-+ "1: "__copyuser_seg" movl 64(%4), %%eax\n"
- " .align 2,0x90\n"
-- "2: movl 0(%4), %%eax\n"
-- "21: movl 4(%4), %%edx\n"
-+ "2: "__copyuser_seg" movl 0(%4), %%eax\n"
-+ "21: "__copyuser_seg" movl 4(%4), %%edx\n"
- " movl %%eax, 0(%3)\n"
- " movl %%edx, 4(%3)\n"
-- "3: movl 8(%4), %%eax\n"
-- "31: movl 12(%4),%%edx\n"
-+ "3: "__copyuser_seg" movl 8(%4), %%eax\n"
-+ "31: "__copyuser_seg" movl 12(%4),%%edx\n"
- " movl %%eax, 8(%3)\n"
- " movl %%edx, 12(%3)\n"
-- "4: movl 16(%4), %%eax\n"
-- "41: movl 20(%4), %%edx\n"
-+ "4: "__copyuser_seg" movl 16(%4), %%eax\n"
-+ "41: "__copyuser_seg" movl 20(%4), %%edx\n"
- " movl %%eax, 16(%3)\n"
- " movl %%edx, 20(%3)\n"
-- "10: movl 24(%4), %%eax\n"
-- "51: movl 28(%4), %%edx\n"
-+ "10: "__copyuser_seg" movl 24(%4), %%eax\n"
-+ "51: "__copyuser_seg" movl 28(%4), %%edx\n"
- " movl %%eax, 24(%3)\n"
- " movl %%edx, 28(%3)\n"
-- "11: movl 32(%4), %%eax\n"
-- "61: movl 36(%4), %%edx\n"
-+ "11: "__copyuser_seg" movl 32(%4), %%eax\n"
-+ "61: "__copyuser_seg" movl 36(%4), %%edx\n"
- " movl %%eax, 32(%3)\n"
- " movl %%edx, 36(%3)\n"
-- "12: movl 40(%4), %%eax\n"
-- "71: movl 44(%4), %%edx\n"
-+ "12: "__copyuser_seg" movl 40(%4), %%eax\n"
-+ "71: "__copyuser_seg" movl 44(%4), %%edx\n"
- " movl %%eax, 40(%3)\n"
- " movl %%edx, 44(%3)\n"
-- "13: movl 48(%4), %%eax\n"
-- "81: movl 52(%4), %%edx\n"
-+ "13: "__copyuser_seg" movl 48(%4), %%eax\n"
-+ "81: "__copyuser_seg" movl 52(%4), %%edx\n"
- " movl %%eax, 48(%3)\n"
- " movl %%edx, 52(%3)\n"
-- "14: movl 56(%4), %%eax\n"
-- "91: movl 60(%4), %%edx\n"
-+ "14: "__copyuser_seg" movl 56(%4), %%eax\n"
-+ "91: "__copyuser_seg" movl 60(%4), %%edx\n"
- " movl %%eax, 56(%3)\n"
- " movl %%edx, 60(%3)\n"
- " addl $-64, %0\n"
-@@ -385,9 +498,9 @@ __copy_user_zeroing_intel(void *to, cons
- " shrl $2, %0\n"
- " andl $3, %%eax\n"
- " cld\n"
-- "6: rep; movsl\n"
-+ "6: rep; "__copyuser_seg" movsl\n"
- " movl %%eax,%0\n"
-- "7: rep; movsb\n"
-+ "7: rep; "__copyuser_seg" movsb\n"
- "8:\n"
- ".section .fixup,\"ax\"\n"
- "9: lea 0(%%eax,%0,4),%0\n"
-@@ -440,41 +553,41 @@ static unsigned long __copy_user_zeroing
-
- __asm__ __volatile__(
- " .align 2,0x90\n"
-- "0: movl 32(%4), %%eax\n"
-+ "0: "__copyuser_seg" movl 32(%4), %%eax\n"
- " cmpl $67, %0\n"
- " jbe 2f\n"
-- "1: movl 64(%4), %%eax\n"
-+ "1: "__copyuser_seg" movl 64(%4), %%eax\n"
- " .align 2,0x90\n"
-- "2: movl 0(%4), %%eax\n"
-- "21: movl 4(%4), %%edx\n"
-+ "2: "__copyuser_seg" movl 0(%4), %%eax\n"
-+ "21: "__copyuser_seg" movl 4(%4), %%edx\n"
- " movnti %%eax, 0(%3)\n"
- " movnti %%edx, 4(%3)\n"
-- "3: movl 8(%4), %%eax\n"
-- "31: movl 12(%4),%%edx\n"
-+ "3: "__copyuser_seg" movl 8(%4), %%eax\n"
-+ "31: "__copyuser_seg" movl 12(%4),%%edx\n"
- " movnti %%eax, 8(%3)\n"
- " movnti %%edx, 12(%3)\n"
-- "4: movl 16(%4), %%eax\n"
-- "41: movl 20(%4), %%edx\n"
-+ "4: "__copyuser_seg" movl 16(%4), %%eax\n"
-+ "41: "__copyuser_seg" movl 20(%4), %%edx\n"
- " movnti %%eax, 16(%3)\n"
- " movnti %%edx, 20(%3)\n"
-- "10: movl 24(%4), %%eax\n"
-- "51: movl 28(%4), %%edx\n"
-+ "10: "__copyuser_seg" movl 24(%4), %%eax\n"
-+ "51: "__copyuser_seg" movl 28(%4), %%edx\n"
- " movnti %%eax, 24(%3)\n"
- " movnti %%edx, 28(%3)\n"
-- "11: movl 32(%4), %%eax\n"
-- "61: movl 36(%4), %%edx\n"
-+ "11: "__copyuser_seg" movl 32(%4), %%eax\n"
-+ "61: "__copyuser_seg" movl 36(%4), %%edx\n"
- " movnti %%eax, 32(%3)\n"
- " movnti %%edx, 36(%3)\n"
-- "12: movl 40(%4), %%eax\n"
-- "71: movl 44(%4), %%edx\n"
-+ "12: "__copyuser_seg" movl 40(%4), %%eax\n"
-+ "71: "__copyuser_seg" movl 44(%4), %%edx\n"
- " movnti %%eax, 40(%3)\n"
- " movnti %%edx, 44(%3)\n"
-- "13: movl 48(%4), %%eax\n"
-- "81: movl 52(%4), %%edx\n"
-+ "13: "__copyuser_seg" movl 48(%4), %%eax\n"
-+ "81: "__copyuser_seg" movl 52(%4), %%edx\n"
- " movnti %%eax, 48(%3)\n"
- " movnti %%edx, 52(%3)\n"
-- "14: movl 56(%4), %%eax\n"
-- "91: movl 60(%4), %%edx\n"
-+ "14: "__copyuser_seg" movl 56(%4), %%eax\n"
-+ "91: "__copyuser_seg" movl 60(%4), %%edx\n"
- " movnti %%eax, 56(%3)\n"
- " movnti %%edx, 60(%3)\n"
- " addl $-64, %0\n"
-@@ -487,9 +600,9 @@ static unsigned long __copy_user_zeroing
- " shrl $2, %0\n"
- " andl $3, %%eax\n"
- " cld\n"
-- "6: rep; movsl\n"
-+ "6: rep; "__copyuser_seg" movsl\n"
- " movl %%eax,%0\n"
-- "7: rep; movsb\n"
-+ "7: rep; "__copyuser_seg" movsb\n"
- "8:\n"
- ".section .fixup,\"ax\"\n"
- "9: lea 0(%%eax,%0,4),%0\n"
-@@ -537,41 +650,41 @@ static unsigned long __copy_user_intel_n
-
- __asm__ __volatile__(
- " .align 2,0x90\n"
-- "0: movl 32(%4), %%eax\n"
-+ "0: "__copyuser_seg" movl 32(%4), %%eax\n"
- " cmpl $67, %0\n"
- " jbe 2f\n"
-- "1: movl 64(%4), %%eax\n"
-+ "1: "__copyuser_seg" movl 64(%4), %%eax\n"
- " .align 2,0x90\n"
-- "2: movl 0(%4), %%eax\n"
-- "21: movl 4(%4), %%edx\n"
-+ "2: "__copyuser_seg" movl 0(%4), %%eax\n"
-+ "21: "__copyuser_seg" movl 4(%4), %%edx\n"
- " movnti %%eax, 0(%3)\n"
- " movnti %%edx, 4(%3)\n"
-- "3: movl 8(%4), %%eax\n"
-- "31: movl 12(%4),%%edx\n"
-+ "3: "__copyuser_seg" movl 8(%4), %%eax\n"
-+ "31: "__copyuser_seg" movl 12(%4),%%edx\n"
- " movnti %%eax, 8(%3)\n"
- " movnti %%edx, 12(%3)\n"
-- "4: movl 16(%4), %%eax\n"
-- "41: movl 20(%4), %%edx\n"
-+ "4: "__copyuser_seg" movl 16(%4), %%eax\n"
-+ "41: "__copyuser_seg" movl 20(%4), %%edx\n"
- " movnti %%eax, 16(%3)\n"
- " movnti %%edx, 20(%3)\n"
-- "10: movl 24(%4), %%eax\n"
-- "51: movl 28(%4), %%edx\n"
-+ "10: "__copyuser_seg" movl 24(%4), %%eax\n"
-+ "51: "__copyuser_seg" movl 28(%4), %%edx\n"
- " movnti %%eax, 24(%3)\n"
- " movnti %%edx, 28(%3)\n"
-- "11: movl 32(%4), %%eax\n"
-- "61: movl 36(%4), %%edx\n"
-+ "11: "__copyuser_seg" movl 32(%4), %%eax\n"
-+ "61: "__copyuser_seg" movl 36(%4), %%edx\n"
- " movnti %%eax, 32(%3)\n"
- " movnti %%edx, 36(%3)\n"
-- "12: movl 40(%4), %%eax\n"
-- "71: movl 44(%4), %%edx\n"
-+ "12: "__copyuser_seg" movl 40(%4), %%eax\n"
-+ "71: "__copyuser_seg" movl 44(%4), %%edx\n"
- " movnti %%eax, 40(%3)\n"
- " movnti %%edx, 44(%3)\n"
-- "13: movl 48(%4), %%eax\n"
-- "81: movl 52(%4), %%edx\n"
-+ "13: "__copyuser_seg" movl 48(%4), %%eax\n"
-+ "81: "__copyuser_seg" movl 52(%4), %%edx\n"
- " movnti %%eax, 48(%3)\n"
- " movnti %%edx, 52(%3)\n"
-- "14: movl 56(%4), %%eax\n"
-- "91: movl 60(%4), %%edx\n"
-+ "14: "__copyuser_seg" movl 56(%4), %%eax\n"
-+ "91: "__copyuser_seg" movl 60(%4), %%edx\n"
- " movnti %%eax, 56(%3)\n"
- " movnti %%edx, 60(%3)\n"
- " addl $-64, %0\n"
-@@ -584,9 +697,9 @@ static unsigned long __copy_user_intel_n
- " shrl $2, %0\n"
- " andl $3, %%eax\n"
- " cld\n"
-- "6: rep; movsl\n"
-+ "6: rep; "__copyuser_seg" movsl\n"
- " movl %%eax,%0\n"
-- "7: rep; movsb\n"
-+ "7: rep; "__copyuser_seg" movsb\n"
- "8:\n"
- ".section .fixup,\"ax\"\n"
- "9: lea 0(%%eax,%0,4),%0\n"
-@@ -629,32 +742,36 @@ static unsigned long __copy_user_intel_n
- */
- unsigned long __copy_user_zeroing_intel(void *to, const void __user *from,
- unsigned long size);
--unsigned long __copy_user_intel(void __user *to, const void *from,
-+unsigned long __generic_copy_to_user_intel(void __user *to, const void *from,
-+ unsigned long size);
-+unsigned long __generic_copy_from_user_intel(void *to, const void __user *from,
- unsigned long size);
- unsigned long __copy_user_zeroing_intel_nocache(void *to,
- const void __user *from, unsigned long size);
- #endif /* CONFIG_X86_INTEL_USERCOPY */
-
- /* Generic arbitrary sized copy. */
--#define __copy_user(to, from, size) \
-+#define __copy_user(to, from, size, prefix, set, restore) \
- do { \
- int __d0, __d1, __d2; \
- __asm__ __volatile__( \
-+ set \
- " cmp $7,%0\n" \
- " jbe 1f\n" \
- " movl %1,%0\n" \
- " negl %0\n" \
- " andl $7,%0\n" \
- " subl %0,%3\n" \
-- "4: rep; movsb\n" \
-+ "4: rep; "prefix"movsb\n" \
- " movl %3,%0\n" \
- " shrl $2,%0\n" \
- " andl $3,%3\n" \
- " .align 2,0x90\n" \
-- "0: rep; movsl\n" \
-+ "0: rep; "prefix"movsl\n" \
- " movl %3,%0\n" \
-- "1: rep; movsb\n" \
-+ "1: rep; "prefix"movsb\n" \
- "2:\n" \
-+ restore \
- ".section .fixup,\"ax\"\n" \
- "5: addl %3,%0\n" \
- " jmp 2b\n" \
-@@ -682,14 +799,14 @@ do { \
- " negl %0\n" \
- " andl $7,%0\n" \
- " subl %0,%3\n" \
-- "4: rep; movsb\n" \
-+ "4: rep; "__copyuser_seg"movsb\n" \
- " movl %3,%0\n" \
- " shrl $2,%0\n" \
- " andl $3,%3\n" \
- " .align 2,0x90\n" \
-- "0: rep; movsl\n" \
-+ "0: rep; "__copyuser_seg"movsl\n" \
- " movl %3,%0\n" \
-- "1: rep; movsb\n" \
-+ "1: rep; "__copyuser_seg"movsb\n" \
- "2:\n" \
- ".section .fixup,\"ax\"\n" \
- "5: addl %3,%0\n" \
-@@ -775,9 +892,9 @@ survive:
- }
- #endif
- if (movsl_is_ok(to, from, n))
-- __copy_user(to, from, n);
-+ __copy_user(to, from, n, "", __COPYUSER_SET_ES, __COPYUSER_RESTORE_ES);
- else
-- n = __copy_user_intel(to, from, n);
-+ n = __generic_copy_to_user_intel(to, from, n);
- return n;
- }
- EXPORT_SYMBOL(__copy_to_user_ll);
-@@ -797,10 +914,9 @@ unsigned long __copy_from_user_ll_nozero
- unsigned long n)
- {
- if (movsl_is_ok(to, from, n))
-- __copy_user(to, from, n);
-+ __copy_user(to, from, n, __copyuser_seg, "", "");
- else
-- n = __copy_user_intel((void __user *)to,
-- (const void *)from, n);
-+ n = __generic_copy_from_user_intel(to, from, n);
- return n;
- }
- EXPORT_SYMBOL(__copy_from_user_ll_nozero);
-@@ -827,65 +943,49 @@ unsigned long __copy_from_user_ll_nocach
- if (n > 64 && cpu_has_xmm2)
- n = __copy_user_intel_nocache(to, from, n);
- else
-- __copy_user(to, from, n);
-+ __copy_user(to, from, n, __copyuser_seg, "", "");
- #else
-- __copy_user(to, from, n);
-+ __copy_user(to, from, n, __copyuser_seg, "", "");
- #endif
- return n;
- }
- EXPORT_SYMBOL(__copy_from_user_ll_nocache_nozero);
-
--/**
-- * copy_to_user: - Copy a block of data into user space.
-- * @to: Destination address, in user space.
-- * @from: Source address, in kernel space.
-- * @n: Number of bytes to copy.
-- *
-- * Context: User context only. This function may sleep.
-- *
-- * Copy data from kernel space to user space.
-- *
-- * Returns number of bytes that could not be copied.
-- * On success, this will be zero.
-- */
--unsigned long
--copy_to_user(void __user *to, const void *from, unsigned long n)
-+void copy_from_user_overflow(void)
- {
-- if (access_ok(VERIFY_WRITE, to, n))
-- n = __copy_to_user(to, from, n);
-- return n;
-+ WARN(1, "Buffer overflow detected!\n");
- }
--EXPORT_SYMBOL(copy_to_user);
-+EXPORT_SYMBOL(copy_from_user_overflow);
-
--/**
-- * copy_from_user: - Copy a block of data from user space.
-- * @to: Destination address, in kernel space.
-- * @from: Source address, in user space.
-- * @n: Number of bytes to copy.
-- *
-- * Context: User context only. This function may sleep.
-- *
-- * Copy data from user space to kernel space.
-- *
-- * Returns number of bytes that could not be copied.
-- * On success, this will be zero.
-- *
-- * If some data could not be copied, this function will pad the copied
-- * data to the requested size using zero bytes.
-- */
--unsigned long
--_copy_from_user(void *to, const void __user *from, unsigned long n)
-+void copy_to_user_overflow(void)
- {
-- if (access_ok(VERIFY_READ, from, n))
-- n = __copy_from_user(to, from, n);
-- else
-- memset(to, 0, n);
-- return n;
-+ WARN(1, "Buffer overflow detected!\n");
- }
--EXPORT_SYMBOL(_copy_from_user);
-+EXPORT_SYMBOL(copy_to_user_overflow);
-
--void copy_from_user_overflow(void)
-+#ifdef CONFIG_PAX_MEMORY_UDEREF
-+void __set_fs(mm_segment_t x)
- {
-- WARN(1, "Buffer overflow detected!\n");
-+ switch (x.seg) {
-+ case 0:
-+ loadsegment(gs, 0);
-+ break;
-+ case TASK_SIZE_MAX:
-+ loadsegment(gs, __USER_DS);
-+ break;
-+ case -1UL:
-+ loadsegment(gs, __KERNEL_DS);
-+ break;
-+ default:
-+ BUG();
-+ }
-+ return;
- }
--EXPORT_SYMBOL(copy_from_user_overflow);
-+
-+void set_fs(mm_segment_t x)
-+{
-+ current_thread_info()->addr_limit = x;
-+ __set_fs(x);
-+}
-+EXPORT_SYMBOL(set_fs);
-+#endif
-diff -urNp linux-2.6.37.4/arch/x86/lib/usercopy_64.c linux-2.6.37.4/arch/x86/lib/usercopy_64.c
---- linux-2.6.37.4/arch/x86/lib/usercopy_64.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/lib/usercopy_64.c 2011-01-17 02:41:01.000000000 -0500
-@@ -42,6 +42,8 @@ long
- __strncpy_from_user(char *dst, const char __user *src, long count)
- {
- long res;
-+ if ((unsigned long)src < PAX_USER_SHADOW_BASE)
-+ src += PAX_USER_SHADOW_BASE;
- __do_strncpy_from_user(dst, src, count, res);
- return res;
- }
-@@ -65,6 +67,8 @@ unsigned long __clear_user(void __user *
- {
- long __d0;
- might_fault();
-+ if ((unsigned long)addr < PAX_USER_SHADOW_BASE)
-+ addr += PAX_USER_SHADOW_BASE;
- /* no memory constraint because it doesn't change any memory gcc knows
- about */
- asm volatile(
-@@ -151,10 +155,14 @@ EXPORT_SYMBOL(strlen_user);
-
- unsigned long copy_in_user(void __user *to, const void __user *from, unsigned len)
- {
-- if (access_ok(VERIFY_WRITE, to, len) && access_ok(VERIFY_READ, from, len)) {
-+ if (access_ok(VERIFY_WRITE, to, len) && access_ok(VERIFY_READ, from, len)) {
-+ if ((unsigned long)to < PAX_USER_SHADOW_BASE)
-+ to += PAX_USER_SHADOW_BASE;
-+ if ((unsigned long)from < PAX_USER_SHADOW_BASE)
-+ from += PAX_USER_SHADOW_BASE;
- return copy_user_generic((__force void *)to, (__force void *)from, len);
-- }
-- return len;
-+ }
-+ return len;
- }
- EXPORT_SYMBOL(copy_in_user);
-
-diff -urNp linux-2.6.37.4/arch/x86/Makefile linux-2.6.37.4/arch/x86/Makefile
---- linux-2.6.37.4/arch/x86/Makefile 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/Makefile 2011-01-17 02:41:01.000000000 -0500
-@@ -195,3 +195,12 @@ define archhelp
- echo ' FDARGS="..." arguments for the booted kernel'
- echo ' FDINITRD=file initrd for the booted kernel'
- endef
-+
-+define OLD_LD
-+
-+*** ${VERSION}.${PATCHLEVEL} PaX kernels no longer build correctly with old versions of binutils.
-+*** Please upgrade your binutils to 2.18 or newer
-+endef
-+
-+archprepare:
-+ $(if $(LDFLAGS_BUILD_ID),,$(error $(OLD_LD)))
-diff -urNp linux-2.6.37.4/arch/x86/mm/extable.c linux-2.6.37.4/arch/x86/mm/extable.c
---- linux-2.6.37.4/arch/x86/mm/extable.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/mm/extable.c 2011-01-17 02:41:01.000000000 -0500
-@@ -1,14 +1,71 @@
- #include <linux/module.h>
- #include <linux/spinlock.h>
-+#include <linux/sort.h>
- #include <asm/uaccess.h>
-+#include <asm/pgtable.h>
-
-+/*
-+ * The exception table needs to be sorted so that the binary
-+ * search that we use to find entries in it works properly.
-+ * This is used both for the kernel exception table and for
-+ * the exception tables of modules that get loaded.
-+ */
-+static int cmp_ex(const void *a, const void *b)
-+{
-+ const struct exception_table_entry *x = a, *y = b;
-+
-+ /* avoid overflow */
-+ if (x->insn > y->insn)
-+ return 1;
-+ if (x->insn < y->insn)
-+ return -1;
-+ return 0;
-+}
-+
-+static void swap_ex(void *a, void *b, int size)
-+{
-+ struct exception_table_entry t, *x = a, *y = b;
-+
-+ t = *x;
-+
-+ pax_open_kernel();
-+ *x = *y;
-+ *y = t;
-+ pax_close_kernel();
-+}
-+
-+void sort_extable(struct exception_table_entry *start,
-+ struct exception_table_entry *finish)
-+{
-+ sort(start, finish - start, sizeof(struct exception_table_entry),
-+ cmp_ex, swap_ex);
-+}
-+
-+#ifdef CONFIG_MODULES
-+/*
-+ * If the exception table is sorted, any referring to the module init
-+ * will be at the beginning or the end.
-+ */
-+void trim_init_extable(struct module *m)
-+{
-+ /*trim the beginning*/
-+ while (m->num_exentries && within_module_init(m->extable[0].insn, m)) {
-+ m->extable++;
-+ m->num_exentries--;
-+ }
-+ /*trim the end*/
-+ while (m->num_exentries &&
-+ within_module_init(m->extable[m->num_exentries-1].insn, m))
-+ m->num_exentries--;
-+}
-+#endif /* CONFIG_MODULES */
-
- int fixup_exception(struct pt_regs *regs)
- {
- const struct exception_table_entry *fixup;
-
- #ifdef CONFIG_PNPBIOS
-- if (unlikely(SEGMENT_IS_PNP_CODE(regs->cs))) {
-+ if (unlikely(!v8086_mode(regs) && SEGMENT_IS_PNP_CODE(regs->cs))) {
- extern u32 pnp_bios_fault_eip, pnp_bios_fault_esp;
- extern u32 pnp_bios_is_utter_crap;
- pnp_bios_is_utter_crap = 1;
-diff -urNp linux-2.6.37.4/arch/x86/mm/fault.c linux-2.6.37.4/arch/x86/mm/fault.c
---- linux-2.6.37.4/arch/x86/mm/fault.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/mm/fault.c 2011-03-12 22:50:25.000000000 -0500
-@@ -12,10 +12,18 @@
- #include <linux/mmiotrace.h> /* kmmio_handler, ... */
- #include <linux/perf_event.h> /* perf_sw_event */
- #include <linux/hugetlb.h> /* hstate_index_to_shift */
-+#include <linux/unistd.h>
-+#include <linux/compiler.h>
-
- #include <asm/traps.h> /* dotraplinkage, ... */
- #include <asm/pgalloc.h> /* pgd_*(), ... */
- #include <asm/kmemcheck.h> /* kmemcheck_*(), ... */
-+#include <asm/vsyscall.h>
-+#include <asm/tlbflush.h>
-+
-+#if defined(CONFIG_X86_64) && defined(CONFIG_PAX_MEMORY_UDEREF)
-+#include <asm/stacktrace.h>
-+#endif
-
- /*
- * Page fault error code bits:
-@@ -53,7 +61,7 @@ static inline int __kprobes notify_page_
- int ret = 0;
-
- /* kprobe_running() needs smp_processor_id() */
-- if (kprobes_built_in() && !user_mode_vm(regs)) {
-+ if (kprobes_built_in() && !user_mode(regs)) {
- preempt_disable();
- if (kprobe_running() && kprobe_fault_handler(regs, 14))
- ret = 1;
-@@ -114,7 +122,10 @@ check_prefetch_opcode(struct pt_regs *re
- return !instr_lo || (instr_lo>>1) == 1;
- case 0x00:
- /* Prefetch instruction is 0x0F0D or 0x0F18 */
-- if (probe_kernel_address(instr, opcode))
-+ if (user_mode(regs)) {
-+ if (__copy_from_user_inatomic(&opcode, (__force unsigned char __user *)(instr), 1))
-+ return 0;
-+ } else if (probe_kernel_address(instr, opcode))
- return 0;
-
- *prefetch = (instr_lo == 0xF) &&
-@@ -148,7 +159,10 @@ is_prefetch(struct pt_regs *regs, unsign
- while (instr < max_instr) {
- unsigned char opcode;
-
-- if (probe_kernel_address(instr, opcode))
-+ if (user_mode(regs)) {
-+ if (__copy_from_user_inatomic(&opcode, (__force unsigned char __user *)(instr), 1))
-+ break;
-+ } else if (probe_kernel_address(instr, opcode))
- break;
-
- instr++;
-@@ -179,6 +193,30 @@ force_sig_info_fault(int si_signo, int s
- force_sig_info(si_signo, &info, tsk);
- }
-
-+#ifdef CONFIG_PAX_EMUTRAMP
-+static int pax_handle_fetch_fault(struct pt_regs *regs);
-+#endif
-+
-+#ifdef CONFIG_PAX_PAGEEXEC
-+static inline pmd_t * pax_get_pmd(struct mm_struct *mm, unsigned long address)
-+{
-+ pgd_t *pgd;
-+ pud_t *pud;
-+ pmd_t *pmd;
-+
-+ pgd = pgd_offset(mm, address);
-+ if (!pgd_present(*pgd))
-+ return NULL;
-+ pud = pud_offset(pgd, address);
-+ if (!pud_present(*pud))
-+ return NULL;
-+ pmd = pmd_offset(pud, address);
-+ if (!pmd_present(*pmd))
-+ return NULL;
-+ return pmd;
-+}
-+#endif
-+
- DEFINE_SPINLOCK(pgd_lock);
- LIST_HEAD(pgd_list);
-
-@@ -231,18 +269,35 @@ void vmalloc_sync_all(void)
- address += PMD_SIZE) {
-
- unsigned long flags;
-+
-+#ifdef CONFIG_PAX_PER_CPU_PGD
-+ unsigned long cpu;
-+#else
- struct page *page;
-+#endif
-
- spin_lock_irqsave(&pgd_lock, flags);
-+
-+#ifdef CONFIG_PAX_PER_CPU_PGD
-+ for (cpu = 0; cpu < NR_CPUS; ++cpu) {
-+ pgd_t *pgd = get_cpu_pgd(cpu);
-+ pmd_t *ret;
-+#else
- list_for_each_entry(page, &pgd_list, lru) {
-+ pgd_t *pgd = page_address(page);
- spinlock_t *pgt_lock;
- pmd_t *ret;
-
- pgt_lock = &pgd_page_get_mm(page)->page_table_lock;
-
- spin_lock(pgt_lock);
-- ret = vmalloc_sync_one(page_address(page), address);
-+#endif
-+
-+ ret = vmalloc_sync_one(pgd, address);
-+
-+#ifndef CONFIG_PAX_PER_CPU_PGD
- spin_unlock(pgt_lock);
-+#endif
-
- if (!ret)
- break;
-@@ -276,6 +331,11 @@ static noinline __kprobes int vmalloc_fa
- * an interrupt in the middle of a task switch..
- */
- pgd_paddr = read_cr3();
-+
-+#ifdef CONFIG_PAX_PER_CPU_PGD
-+ BUG_ON(__pa(get_cpu_pgd(smp_processor_id())) != (pgd_paddr & PHYSICAL_PAGE_MASK));
-+#endif
-+
- pmd_k = vmalloc_sync_one(__va(pgd_paddr), address);
- if (!pmd_k)
- return -1;
-@@ -371,7 +431,14 @@ static noinline __kprobes int vmalloc_fa
- * happen within a race in page table update. In the later
- * case just flush:
- */
-+
-+#ifdef CONFIG_PAX_PER_CPU_PGD
-+ BUG_ON(__pa(get_cpu_pgd(smp_processor_id())) != (read_cr3() & PHYSICAL_PAGE_MASK));
-+ pgd = pgd_offset_cpu(smp_processor_id(), address);
-+#else
- pgd = pgd_offset(current->active_mm, address);
-+#endif
-+
- pgd_ref = pgd_offset_k(address);
- if (pgd_none(*pgd_ref))
- return -1;
-@@ -533,7 +600,7 @@ static int is_errata93(struct pt_regs *r
- static int is_errata100(struct pt_regs *regs, unsigned long address)
- {
- #ifdef CONFIG_X86_64
-- if ((regs->cs == __USER32_CS || (regs->cs & (1<<2))) && (address >> 32))
-+ if ((regs->cs == __USER32_CS || (regs->cs & SEGMENT_LDT)) && (address >> 32))
- return 1;
- #endif
- return 0;
-@@ -560,7 +627,7 @@ static int is_f00f_bug(struct pt_regs *r
- }
-
- static const char nx_warning[] = KERN_CRIT
--"kernel tried to execute NX-protected page - exploit attempt? (uid: %d)\n";
-+"kernel tried to execute NX-protected page - exploit attempt? (uid: %d, task: %s, pid: %d)\n";
-
- static void
- show_fault_oops(struct pt_regs *regs, unsigned long error_code,
-@@ -569,15 +636,26 @@ show_fault_oops(struct pt_regs *regs, un
- if (!oops_may_print())
- return;
-
-- if (error_code & PF_INSTR) {
-+ if ((__supported_pte_mask & _PAGE_NX) && (error_code & PF_INSTR)) {
- unsigned int level;
-
- pte_t *pte = lookup_address(address, &level);
-
- if (pte && pte_present(*pte) && !pte_exec(*pte))
-- printk(nx_warning, current_uid());
-+ printk(nx_warning, current_uid(), current->comm, task_pid_nr(current));
- }
-
-+#ifdef CONFIG_PAX_KERNEXEC
-+ if (init_mm.start_code <= address && address < init_mm.end_code) {
-+ if (current->signal->curr_ip)
-+ printk(KERN_ERR "PAX: From %pI4: %s:%d, uid/euid: %u/%u, attempted to modify kernel code\n",
-+ &current->signal->curr_ip, current->comm, task_pid_nr(current), current_uid(), current_euid());
-+ else
-+ printk(KERN_ERR "PAX: %s:%d, uid/euid: %u/%u, attempted to modify kernel code\n",
-+ current->comm, task_pid_nr(current), current_uid(), current_euid());
-+ }
-+#endif
-+
- printk(KERN_ALERT "BUG: unable to handle kernel ");
- if (address < PAGE_SIZE)
- printk(KERN_CONT "NULL pointer dereference");
-@@ -702,6 +780,68 @@ __bad_area_nosemaphore(struct pt_regs *r
- unsigned long address, int si_code)
- {
- struct task_struct *tsk = current;
-+ struct mm_struct *mm = tsk->mm;
-+
-+#ifdef CONFIG_X86_64
-+ if (mm && (error_code & PF_INSTR) && mm->context.vdso) {
-+ if (regs->ip == (unsigned long)vgettimeofday) {
-+ regs->ip = (unsigned long)VDSO64_SYMBOL(mm->context.vdso, fallback_gettimeofday);
-+ return;
-+ } else if (regs->ip == (unsigned long)vtime) {
-+ regs->ip = (unsigned long)VDSO64_SYMBOL(mm->context.vdso, fallback_time);
-+ return;
-+ } else if (regs->ip == (unsigned long)vgetcpu) {
-+ regs->ip = (unsigned long)VDSO64_SYMBOL(mm->context.vdso, getcpu);
-+ return;
-+ }
-+ }
-+#endif
-+
-+#if defined(CONFIG_PAX_PAGEEXEC) || defined(CONFIG_PAX_SEGMEXEC)
-+ if (mm && (error_code & PF_USER)) {
-+ unsigned long ip = regs->ip;
-+
-+ if (v8086_mode(regs))
-+ ip = ((regs->cs & 0xffff) << 4) + (ip & 0xffff);
-+
-+ /*
-+ * It's possible to have interrupts off here:
-+ */
-+ local_irq_enable();
-+
-+#ifdef CONFIG_PAX_PAGEEXEC
-+ if ((mm->pax_flags & MF_PAX_PAGEEXEC) &&
-+ (((__supported_pte_mask & _PAGE_NX) && (error_code & PF_INSTR)) || (!(error_code & (PF_PROT | PF_WRITE)) && ip == address))) {
-+
-+#ifdef CONFIG_PAX_EMUTRAMP
-+ switch (pax_handle_fetch_fault(regs)) {
-+ case 2:
-+ return;
-+ }
-+#endif
-+
-+ pax_report_fault(regs, (void *)ip, (void *)regs->sp);
-+ do_group_exit(SIGKILL);
-+ }
-+#endif
-+
-+#ifdef CONFIG_PAX_SEGMEXEC
-+ if ((mm->pax_flags & MF_PAX_SEGMEXEC) && !(error_code & (PF_PROT | PF_WRITE)) && (ip + SEGMEXEC_TASK_SIZE == address)) {
-+
-+#ifdef CONFIG_PAX_EMUTRAMP
-+ switch (pax_handle_fetch_fault(regs)) {
-+ case 2:
-+ return;
-+ }
-+#endif
-+
-+ pax_report_fault(regs, (void *)ip, (void *)regs->sp);
-+ do_group_exit(SIGKILL);
-+ }
-+#endif
-+
-+ }
-+#endif
-
- /* User mode accesses just cause a SIGSEGV */
- if (error_code & PF_USER) {
-@@ -849,6 +989,99 @@ static int spurious_fault_check(unsigned
- return 1;
- }
-
-+#if defined(CONFIG_X86_32) && defined(CONFIG_PAX_PAGEEXEC)
-+static int pax_handle_pageexec_fault(struct pt_regs *regs, struct mm_struct *mm, unsigned long address, unsigned long error_code)
-+{
-+ pte_t *pte;
-+ pmd_t *pmd;
-+ spinlock_t *ptl;
-+ unsigned char pte_mask;
-+
-+ if ((__supported_pte_mask & _PAGE_NX) || (error_code & (PF_PROT|PF_USER)) != (PF_PROT|PF_USER) || v8086_mode(regs) ||
-+ !(mm->pax_flags & MF_PAX_PAGEEXEC))
-+ return 0;
-+
-+ /* PaX: it's our fault, let's handle it if we can */
-+
-+ /* PaX: take a look at read faults before acquiring any locks */
-+ if (unlikely(!(error_code & PF_WRITE) && (regs->ip == address))) {
-+ /* instruction fetch attempt from a protected page in user mode */
-+ up_read(&mm->mmap_sem);
-+
-+#ifdef CONFIG_PAX_EMUTRAMP
-+ switch (pax_handle_fetch_fault(regs)) {
-+ case 2:
-+ return 1;
-+ }
-+#endif
-+
-+ pax_report_fault(regs, (void *)regs->ip, (void *)regs->sp);
-+ do_group_exit(SIGKILL);
-+ }
-+
-+ pmd = pax_get_pmd(mm, address);
-+ if (unlikely(!pmd))
-+ return 0;
-+
-+ pte = pte_offset_map_lock(mm, pmd, address, &ptl);
-+ if (unlikely(!(pte_val(*pte) & _PAGE_PRESENT) || pte_user(*pte))) {
-+ pte_unmap_unlock(pte, ptl);
-+ return 0;
-+ }
-+
-+ if (unlikely((error_code & PF_WRITE) && !pte_write(*pte))) {
-+ /* write attempt to a protected page in user mode */
-+ pte_unmap_unlock(pte, ptl);
-+ return 0;
-+ }
-+
-+#ifdef CONFIG_SMP
-+ if (likely(address > get_limit(regs->cs) && cpu_isset(smp_processor_id(), mm->context.cpu_user_cs_mask)))
-+#else
-+ if (likely(address > get_limit(regs->cs)))
-+#endif
-+ {
-+ set_pte(pte, pte_mkread(*pte));
-+ __flush_tlb_one(address);
-+ pte_unmap_unlock(pte, ptl);
-+ up_read(&mm->mmap_sem);
-+ return 1;
-+ }
-+
-+ pte_mask = _PAGE_ACCESSED | _PAGE_USER | ((error_code & PF_WRITE) << (_PAGE_BIT_DIRTY-1));
-+
-+ /*
-+ * PaX: fill DTLB with user rights and retry
-+ */
-+ __asm__ __volatile__ (
-+ "orb %2,(%1)\n"
-+#if defined(CONFIG_M586) || defined(CONFIG_M586TSC)
-+/*
-+ * PaX: let this uncommented 'invlpg' remind us on the behaviour of Intel's
-+ * (and AMD's) TLBs. namely, they do not cache PTEs that would raise *any*
-+ * page fault when examined during a TLB load attempt. this is true not only
-+ * for PTEs holding a non-present entry but also present entries that will
-+ * raise a page fault (such as those set up by PaX, or the copy-on-write
-+ * mechanism). in effect it means that we do *not* need to flush the TLBs
-+ * for our target pages since their PTEs are simply not in the TLBs at all.
-+
-+ * the best thing in omitting it is that we gain around 15-20% speed in the
-+ * fast path of the page fault handler and can get rid of tracing since we
-+ * can no longer flush unintended entries.
-+ */
-+ "invlpg (%0)\n"
-+#endif
-+ __copyuser_seg"testb $0,(%0)\n"
-+ "xorb %3,(%1)\n"
-+ :
-+ : "r" (address), "r" (pte), "q" (pte_mask), "i" (_PAGE_USER)
-+ : "memory", "cc");
-+ pte_unmap_unlock(pte, ptl);
-+ up_read(&mm->mmap_sem);
-+ return 1;
-+}
-+#endif
-+
- /*
- * Handle a spurious fault caused by a stale TLB entry.
- *
-@@ -921,6 +1154,9 @@ int show_unhandled_signals = 1;
- static inline int
- access_error(unsigned long error_code, struct vm_area_struct *vma)
- {
-+ if ((__supported_pte_mask & _PAGE_NX) && (error_code & PF_INSTR) && !(vma->vm_flags & VM_EXEC))
-+ return 1;
-+
- if (error_code & PF_WRITE) {
- /* write, present and write, not present: */
- if (unlikely(!(vma->vm_flags & VM_WRITE)))
-@@ -954,19 +1190,33 @@ do_page_fault(struct pt_regs *regs, unsi
- {
- struct vm_area_struct *vma;
- struct task_struct *tsk;
-- unsigned long address;
- struct mm_struct *mm;
- int fault;
- int write = error_code & PF_WRITE;
- unsigned int flags = FAULT_FLAG_ALLOW_RETRY |
- (write ? FAULT_FLAG_WRITE : 0);
-
-+ /* Get the faulting address: */
-+ unsigned long address = read_cr2();
-+
-+#if defined(CONFIG_X86_64) && defined(CONFIG_PAX_MEMORY_UDEREF)
-+ if (!user_mode(regs) && address < 2 * PAX_USER_SHADOW_BASE) {
-+ if (!search_exception_tables(regs->ip)) {
-+ bad_area_nosemaphore(regs, error_code, address);
-+ return;
-+ }
-+ if (address < PAX_USER_SHADOW_BASE) {
-+ printk(KERN_ERR "PAX: please report this to pageexec@freemail.hu\n");
-+ printk(KERN_ERR "PAX: faulting IP: %pA\n", (void *)regs->ip);
-+ show_trace_log_lvl(NULL, NULL, (void *)regs->sp, regs->bp, KERN_ERR);
-+ } else
-+ address -= PAX_USER_SHADOW_BASE;
-+ }
-+#endif
-+
- tsk = current;
- mm = tsk->mm;
-
-- /* Get the faulting address: */
-- address = read_cr2();
--
- /*
- * Detect and handle instructions that would cause a page fault for
- * both a tracked kernel page and a userspace page.
-@@ -1026,7 +1276,7 @@ do_page_fault(struct pt_regs *regs, unsi
- * User-mode registers count as a user access even for any
- * potential system fault or CPU buglet:
- */
-- if (user_mode_vm(regs)) {
-+ if (user_mode(regs)) {
- local_irq_enable();
- error_code |= PF_USER;
- } else {
-@@ -1081,6 +1331,11 @@ retry:
- might_sleep();
- }
-
-+#if defined(CONFIG_X86_32) && defined(CONFIG_PAX_PAGEEXEC)
-+ if (pax_handle_pageexec_fault(regs, mm, address, error_code))
-+ return;
-+#endif
-+
- vma = find_vma(mm, address);
- if (unlikely(!vma)) {
- bad_area(regs, error_code, address);
-@@ -1092,18 +1347,24 @@ retry:
- bad_area(regs, error_code, address);
- return;
- }
-- if (error_code & PF_USER) {
-- /*
-- * Accessing the stack below %sp is always a bug.
-- * The large cushion allows instructions like enter
-- * and pusha to work. ("enter $65535, $31" pushes
-- * 32 pointers and then decrements %sp by 65535.)
-- */
-- if (unlikely(address + 65536 + 32 * sizeof(unsigned long) < regs->sp)) {
-- bad_area(regs, error_code, address);
-- return;
-- }
-+ /*
-+ * Accessing the stack below %sp is always a bug.
-+ * The large cushion allows instructions like enter
-+ * and pusha to work. ("enter $65535, $31" pushes
-+ * 32 pointers and then decrements %sp by 65535.)
-+ */
-+ if (unlikely(address + 65536 + 32 * sizeof(unsigned long) < task_pt_regs(tsk)->sp)) {
-+ bad_area(regs, error_code, address);
-+ return;
- }
-+
-+#ifdef CONFIG_PAX_SEGMEXEC
-+ if (unlikely((mm->pax_flags & MF_PAX_SEGMEXEC) && vma->vm_end - SEGMEXEC_TASK_SIZE - 1 < address - SEGMEXEC_TASK_SIZE - 1)) {
-+ bad_area(regs, error_code, address);
-+ return;
-+ }
-+#endif
-+
- if (unlikely(expand_stack(vma, address))) {
- bad_area(regs, error_code, address);
- return;
-@@ -1158,3 +1419,199 @@ good_area:
-
- up_read(&mm->mmap_sem);
- }
-+
-+#ifdef CONFIG_PAX_EMUTRAMP
-+static int pax_handle_fetch_fault_32(struct pt_regs *regs)
-+{
-+ int err;
-+
-+ do { /* PaX: gcc trampoline emulation #1 */
-+ unsigned char mov1, mov2;
-+ unsigned short jmp;
-+ unsigned int addr1, addr2;
-+
-+#ifdef CONFIG_X86_64
-+ if ((regs->ip + 11) >> 32)
-+ break;
-+#endif
-+
-+ err = get_user(mov1, (unsigned char __user *)regs->ip);
-+ err |= get_user(addr1, (unsigned int __user *)(regs->ip + 1));
-+ err |= get_user(mov2, (unsigned char __user *)(regs->ip + 5));
-+ err |= get_user(addr2, (unsigned int __user *)(regs->ip + 6));
-+ err |= get_user(jmp, (unsigned short __user *)(regs->ip + 10));
-+
-+ if (err)
-+ break;
-+
-+ if (mov1 == 0xB9 && mov2 == 0xB8 && jmp == 0xE0FF) {
-+ regs->cx = addr1;
-+ regs->ax = addr2;
-+ regs->ip = addr2;
-+ return 2;
-+ }
-+ } while (0);
-+
-+ do { /* PaX: gcc trampoline emulation #2 */
-+ unsigned char mov, jmp;
-+ unsigned int addr1, addr2;
-+
-+#ifdef CONFIG_X86_64
-+ if ((regs->ip + 9) >> 32)
-+ break;
-+#endif
-+
-+ err = get_user(mov, (unsigned char __user *)regs->ip);
-+ err |= get_user(addr1, (unsigned int __user *)(regs->ip + 1));
-+ err |= get_user(jmp, (unsigned char __user *)(regs->ip + 5));
-+ err |= get_user(addr2, (unsigned int __user *)(regs->ip + 6));
-+
-+ if (err)
-+ break;
-+
-+ if (mov == 0xB9 && jmp == 0xE9) {
-+ regs->cx = addr1;
-+ regs->ip = (unsigned int)(regs->ip + addr2 + 10);
-+ return 2;
-+ }
-+ } while (0);
-+
-+ return 1; /* PaX in action */
-+}
-+
-+#ifdef CONFIG_X86_64
-+static int pax_handle_fetch_fault_64(struct pt_regs *regs)
-+{
-+ int err;
-+
-+ do { /* PaX: gcc trampoline emulation #1 */
-+ unsigned short mov1, mov2, jmp1;
-+ unsigned char jmp2;
-+ unsigned int addr1;
-+ unsigned long addr2;
-+
-+ err = get_user(mov1, (unsigned short __user *)regs->ip);
-+ err |= get_user(addr1, (unsigned int __user *)(regs->ip + 2));
-+ err |= get_user(mov2, (unsigned short __user *)(regs->ip + 6));
-+ err |= get_user(addr2, (unsigned long __user *)(regs->ip + 8));
-+ err |= get_user(jmp1, (unsigned short __user *)(regs->ip + 16));
-+ err |= get_user(jmp2, (unsigned char __user *)(regs->ip + 18));
-+
-+ if (err)
-+ break;
-+
-+ if (mov1 == 0xBB41 && mov2 == 0xBA49 && jmp1 == 0xFF49 && jmp2 == 0xE3) {
-+ regs->r11 = addr1;
-+ regs->r10 = addr2;
-+ regs->ip = addr1;
-+ return 2;
-+ }
-+ } while (0);
-+
-+ do { /* PaX: gcc trampoline emulation #2 */
-+ unsigned short mov1, mov2, jmp1;
-+ unsigned char jmp2;
-+ unsigned long addr1, addr2;
-+
-+ err = get_user(mov1, (unsigned short __user *)regs->ip);
-+ err |= get_user(addr1, (unsigned long __user *)(regs->ip + 2));
-+ err |= get_user(mov2, (unsigned short __user *)(regs->ip + 10));
-+ err |= get_user(addr2, (unsigned long __user *)(regs->ip + 12));
-+ err |= get_user(jmp1, (unsigned short __user *)(regs->ip + 20));
-+ err |= get_user(jmp2, (unsigned char __user *)(regs->ip + 22));
-+
-+ if (err)
-+ break;
-+
-+ if (mov1 == 0xBB49 && mov2 == 0xBA49 && jmp1 == 0xFF49 && jmp2 == 0xE3) {
-+ regs->r11 = addr1;
-+ regs->r10 = addr2;
-+ regs->ip = addr1;
-+ return 2;
-+ }
-+ } while (0);
-+
-+ return 1; /* PaX in action */
-+}
-+#endif
-+
-+/*
-+ * PaX: decide what to do with offenders (regs->ip = fault address)
-+ *
-+ * returns 1 when task should be killed
-+ * 2 when gcc trampoline was detected
-+ */
-+static int pax_handle_fetch_fault(struct pt_regs *regs)
-+{
-+ if (v8086_mode(regs))
-+ return 1;
-+
-+ if (!(current->mm->pax_flags & MF_PAX_EMUTRAMP))
-+ return 1;
-+
-+#ifdef CONFIG_X86_32
-+ return pax_handle_fetch_fault_32(regs);
-+#else
-+ if (regs->cs == __USER32_CS || (regs->cs & SEGMENT_LDT))
-+ return pax_handle_fetch_fault_32(regs);
-+ else
-+ return pax_handle_fetch_fault_64(regs);
-+#endif
-+}
-+#endif
-+
-+#if defined(CONFIG_PAX_PAGEEXEC) || defined(CONFIG_PAX_SEGMEXEC)
-+void pax_report_insns(void *pc, void *sp)
-+{
-+ long i;
-+
-+ printk(KERN_ERR "PAX: bytes at PC: ");
-+ for (i = 0; i < 20; i++) {
-+ unsigned char c;
-+ if (get_user(c, (__force unsigned char __user *)pc+i))
-+ printk(KERN_CONT "?? ");
-+ else
-+ printk(KERN_CONT "%02x ", c);
-+ }
-+ printk("\n");
-+
-+ printk(KERN_ERR "PAX: bytes at SP-%lu: ", (unsigned long)sizeof(long));
-+ for (i = -1; i < 80 / (long)sizeof(long); i++) {
-+ unsigned long c;
-+ if (get_user(c, (__force unsigned long __user *)sp+i))
-+#ifdef CONFIG_X86_32
-+ printk(KERN_CONT "???????? ");
-+#else
-+ printk(KERN_CONT "???????????????? ");
-+#endif
-+ else
-+ printk(KERN_CONT "%0*lx ", 2 * (int)sizeof(long), c);
-+ }
-+ printk("\n");
-+}
-+#endif
-+
-+/**
-+ * probe_kernel_write(): safely attempt to write to a location
-+ * @dst: address to write to
-+ * @src: pointer to the data that shall be written
-+ * @size: size of the data chunk
-+ *
-+ * Safely write to address @dst from the buffer at @src. If a kernel fault
-+ * happens, handle that and return -EFAULT.
-+ */
-+long notrace probe_kernel_write(void *dst, const void *src, size_t size)
-+{
-+ long ret;
-+ mm_segment_t old_fs = get_fs();
-+
-+ set_fs(KERNEL_DS);
-+ pagefault_disable();
-+ pax_open_kernel();
-+ ret = __copy_to_user_inatomic((__force void __user *)dst, src, size);
-+ pax_close_kernel();
-+ pagefault_enable();
-+ set_fs(old_fs);
-+
-+ return ret ? -EFAULT : 0;
-+}
-diff -urNp linux-2.6.37.4/arch/x86/mm/gup.c linux-2.6.37.4/arch/x86/mm/gup.c
---- linux-2.6.37.4/arch/x86/mm/gup.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/mm/gup.c 2011-01-17 02:41:01.000000000 -0500
-@@ -237,7 +237,7 @@ int __get_user_pages_fast(unsigned long
- addr = start;
- len = (unsigned long) nr_pages << PAGE_SHIFT;
- end = start + len;
-- if (unlikely(!access_ok(write ? VERIFY_WRITE : VERIFY_READ,
-+ if (unlikely(!__access_ok(write ? VERIFY_WRITE : VERIFY_READ,
- (void __user *)start, len)))
- return 0;
-
-diff -urNp linux-2.6.37.4/arch/x86/mm/highmem_32.c linux-2.6.37.4/arch/x86/mm/highmem_32.c
---- linux-2.6.37.4/arch/x86/mm/highmem_32.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/mm/highmem_32.c 2011-01-17 02:41:01.000000000 -0500
-@@ -44,7 +44,10 @@ void *kmap_atomic_prot(struct page *page
- idx = type + KM_TYPE_NR*smp_processor_id();
- vaddr = __fix_to_virt(FIX_KMAP_BEGIN + idx);
- BUG_ON(!pte_none(*(kmap_pte-idx)));
-+
-+ pax_open_kernel();
- set_pte(kmap_pte-idx, mk_pte(page, prot));
-+ pax_close_kernel();
-
- return (void *)vaddr;
- }
-diff -urNp linux-2.6.37.4/arch/x86/mm/hugetlbpage.c linux-2.6.37.4/arch/x86/mm/hugetlbpage.c
---- linux-2.6.37.4/arch/x86/mm/hugetlbpage.c 2011-01-04 19:50:19.000000000 -0500
-+++ linux-2.6.37.4/arch/x86/mm/hugetlbpage.c 2011-03-21 17:35:08.000000000 -0400
-@@ -266,13 +266,20 @@ static unsigned long hugetlb_get_unmappe
- struct hstate *h = hstate_file(file);
- struct mm_struct *mm = current->mm;
- struct vm_area_struct *vma;
-- unsigned long start_addr;
-+ unsigned long start_addr, pax_task_size = TASK_SIZE;
-+
-+#ifdef CONFIG_PAX_SEGMEXEC
-+ if (mm->pax_flags & MF_PAX_SEGMEXEC)
-+ pax_task_size = SEGMEXEC_TASK_SIZE;
-+#endif
-+
-+ pax_task_size -= PAGE_SIZE;
-
- if (len > mm->cached_hole_size) {
-- start_addr = mm->free_area_cache;
-+ start_addr = mm->free_area_cache;
- } else {
-- start_addr = TASK_UNMAPPED_BASE;
-- mm->cached_hole_size = 0;
-+ start_addr = mm->mmap_base;
-+ mm->cached_hole_size = 0;
- }
-
- full_search:
-@@ -280,26 +287,27 @@ full_search:
-
- for (vma = find_vma(mm, addr); ; vma = vma->vm_next) {
- /* At this point: (!vma || addr < vma->vm_end). */
-- if (TASK_SIZE - len < addr) {
-+ if (pax_task_size - len < addr) {
- /*
- * Start a new search - just in case we missed
- * some holes.
- */
-- if (start_addr != TASK_UNMAPPED_BASE) {
-- start_addr = TASK_UNMAPPED_BASE;
-+ if (start_addr != mm->mmap_base) {
-+ start_addr = mm->mmap_base;
- mm->cached_hole_size = 0;
- goto full_search;
- }
- return -ENOMEM;
- }
-- if (!vma || addr + len <= vma->vm_start) {
-- mm->free_area_cache = addr + len;
-- return addr;
-- }
-+ if (check_heap_stack_gap(vma, addr, len))
-+ break;
- if (addr + mm->cached_hole_size < vma->vm_start)
- mm->cached_hole_size = vma->vm_start - addr;
- addr = ALIGN(vma->vm_end, huge_page_size(h));
- }
-+
-+ mm->free_area_cache = addr + len;
-+ return addr;
- }
-
- static unsigned long hugetlb_get_unmapped_area_topdown(struct file *file,
-@@ -308,10 +316,9 @@ static unsigned long hugetlb_get_unmappe
- {
- struct hstate *h = hstate_file(file);
- struct mm_struct *mm = current->mm;
-- struct vm_area_struct *vma, *prev_vma;
-- unsigned long base = mm->mmap_base, addr = addr0;
-+ struct vm_area_struct *vma;
-+ unsigned long base = mm->mmap_base, addr;
- unsigned long largest_hole = mm->cached_hole_size;
-- int first_time = 1;
-
- /* don't allow