From: Ingo Molnar <mingo@elte.hu>
Date: Tue, 24 Feb 2009 20:52:45 +0000 (+0100)
Subject: Merge branch 'tj-percpu' of git://git.kernel.org/pub/scm/linux/kernel/git/tj/misc... 
X-Git-Url: http://git.cdn.openwrt.org/?a=commitdiff_plain;h=0edcf8d6926f4038443dbc24e319530177ca0353;p=openwrt%2Fstaging%2Fblogic.git

Merge branch 'tj-percpu' of git://git./linux/kernel/git/tj/misc into core/percpu

Conflicts:
	arch/x86/include/asm/pgtable.h
---

0edcf8d6926f4038443dbc24e319530177ca0353
diff --cc arch/x86/include/asm/pgtable.h
index 1c097a3a6669,46312eb0d682..d0812e155f1d
--- a/arch/x86/include/asm/pgtable.h
+++ b/arch/x86/include/asm/pgtable.h
@@@ -288,195 -388,84 +288,197 @@@ static inline int is_new_memtype_allowe
  	return 1;
  }
  
 -#ifndef __ASSEMBLY__
 -/* Indicate that x86 has its own track and untrack pfn vma functions */
 -#define __HAVE_PFNMAP_TRACKING
 -
 -#define __HAVE_PHYS_MEM_ACCESS_PROT
 -struct file;
 -pgprot_t phys_mem_access_prot(struct file *file, unsigned long pfn,
 -                              unsigned long size, pgprot_t vma_prot);
 -int phys_mem_access_prot_allowed(struct file *file, unsigned long pfn,
 -                              unsigned long size, pgprot_t *vma_prot);
 -#endif
 -
 -/* Install a pte for a particular vaddr in kernel space. */
 -void set_pte_vaddr(unsigned long vaddr, pte_t pte);
+ pmd_t *populate_extra_pmd(unsigned long vaddr);
+ pte_t *populate_extra_pte(unsigned long vaddr);
 +#endif	/* __ASSEMBLY__ */
  
  #ifdef CONFIG_X86_32
 -extern void native_pagetable_setup_start(pgd_t *base);
 -extern void native_pagetable_setup_done(pgd_t *base);
 +# include "pgtable_32.h"
  #else
 -static inline void native_pagetable_setup_start(pgd_t *base) {}
 -static inline void native_pagetable_setup_done(pgd_t *base) {}
 +# include "pgtable_64.h"
  #endif
  
 -struct seq_file;
 -extern void arch_report_meminfo(struct seq_file *m);
 +#ifndef __ASSEMBLY__
 +#include <linux/mm_types.h>
  
 -#ifdef CONFIG_PARAVIRT
 -#include <asm/paravirt.h>
 -#else  /* !CONFIG_PARAVIRT */
 -#define set_pte(ptep, pte)		native_set_pte(ptep, pte)
 -#define set_pte_at(mm, addr, ptep, pte)	native_set_pte_at(mm, addr, ptep, pte)
 +static inline int pte_none(pte_t pte)
 +{
 +	return !pte.pte;
 +}
  
 -#define set_pte_present(mm, addr, ptep, pte)				\
 -	native_set_pte_present(mm, addr, ptep, pte)
 -#define set_pte_atomic(ptep, pte)					\
 -	native_set_pte_atomic(ptep, pte)
 +#define __HAVE_ARCH_PTE_SAME
 +static inline int pte_same(pte_t a, pte_t b)
 +{
 +	return a.pte == b.pte;
 +}
  
 -#define set_pmd(pmdp, pmd)		native_set_pmd(pmdp, pmd)
 +static inline int pte_present(pte_t a)
 +{
 +	return pte_flags(a) & (_PAGE_PRESENT | _PAGE_PROTNONE);
 +}
  
 -#ifndef __PAGETABLE_PUD_FOLDED
 -#define set_pgd(pgdp, pgd)		native_set_pgd(pgdp, pgd)
 -#define pgd_clear(pgd)			native_pgd_clear(pgd)
 -#endif
 +static inline int pmd_present(pmd_t pmd)
 +{
 +	return pmd_flags(pmd) & _PAGE_PRESENT;
 +}
  
 -#ifndef set_pud
 -# define set_pud(pudp, pud)		native_set_pud(pudp, pud)
 -#endif
 +static inline int pmd_none(pmd_t pmd)
 +{
 +	/* Only check low word on 32-bit platforms, since it might be
 +	   out of sync with upper half. */
 +	return (unsigned long)native_pmd_val(pmd) == 0;
 +}
  
 -#ifndef __PAGETABLE_PMD_FOLDED
 -#define pud_clear(pud)			native_pud_clear(pud)
 -#endif
 +static inline unsigned long pmd_page_vaddr(pmd_t pmd)
 +{
 +	return (unsigned long)__va(pmd_val(pmd) & PTE_PFN_MASK);
 +}
  
 -#define pte_clear(mm, addr, ptep)	native_pte_clear(mm, addr, ptep)
 -#define pmd_clear(pmd)			native_pmd_clear(pmd)
 +/*
 + * Currently stuck as a macro due to indirect forward reference to
 + * linux/mmzone.h's __section_mem_map_addr() definition:
 + */
 +#define pmd_page(pmd)	pfn_to_page(pmd_val(pmd) >> PAGE_SHIFT)
  
 -#define pte_update(mm, addr, ptep)              do { } while (0)
 -#define pte_update_defer(mm, addr, ptep)        do { } while (0)
 +/*
 + * the pmd page can be thought of an array like this: pmd_t[PTRS_PER_PMD]
 + *
 + * this macro returns the index of the entry in the pmd page which would
 + * control the given virtual address
 + */
 +static inline unsigned pmd_index(unsigned long address)
 +{
 +	return (address >> PMD_SHIFT) & (PTRS_PER_PMD - 1);
 +}
  
 -static inline void __init paravirt_pagetable_setup_start(pgd_t *base)
 +/*
 + * Conversion functions: convert a page and protection to a page entry,
 + * and a page entry and page directory to the page they refer to.
 + *
 + * (Currently stuck as a macro because of indirect forward reference
 + * to linux/mm.h:page_to_nid())
 + */
 +#define mk_pte(page, pgprot)   pfn_pte(page_to_pfn(page), (pgprot))
 +
 +/*
 + * the pte page can be thought of an array like this: pte_t[PTRS_PER_PTE]
 + *
 + * this function returns the index of the entry in the pte page which would
 + * control the given virtual address
 + */
 +static inline unsigned pte_index(unsigned long address)
  {
 -	native_pagetable_setup_start(base);
 +	return (address >> PAGE_SHIFT) & (PTRS_PER_PTE - 1);
  }
  
 -static inline void __init paravirt_pagetable_setup_done(pgd_t *base)
 +static inline pte_t *pte_offset_kernel(pmd_t *pmd, unsigned long address)
  {
 -	native_pagetable_setup_done(base);
 +	return (pte_t *)pmd_page_vaddr(*pmd) + pte_index(address);
  }
 -#endif	/* CONFIG_PARAVIRT */
  
 -#endif	/* __ASSEMBLY__ */
 +static inline int pmd_bad(pmd_t pmd)
 +{
 +	return (pmd_flags(pmd) & ~_PAGE_USER) != _KERNPG_TABLE;
 +}
  
 -#ifdef CONFIG_X86_32
 -# include "pgtable_32.h"
 +static inline unsigned long pages_to_mb(unsigned long npg)
 +{
 +	return npg >> (20 - PAGE_SHIFT);
 +}
 +
 +#define io_remap_pfn_range(vma, vaddr, pfn, size, prot)	\
 +	remap_pfn_range(vma, vaddr, pfn, size, prot)
 +
 +#if PAGETABLE_LEVELS > 2
 +static inline int pud_none(pud_t pud)
 +{
 +	return native_pud_val(pud) == 0;
 +}
 +
 +static inline int pud_present(pud_t pud)
 +{
 +	return pud_flags(pud) & _PAGE_PRESENT;
 +}
 +
 +static inline unsigned long pud_page_vaddr(pud_t pud)
 +{
 +	return (unsigned long)__va((unsigned long)pud_val(pud) & PTE_PFN_MASK);
 +}
 +
 +/*
 + * Currently stuck as a macro due to indirect forward reference to
 + * linux/mmzone.h's __section_mem_map_addr() definition:
 + */
 +#define pud_page(pud)		pfn_to_page(pud_val(pud) >> PAGE_SHIFT)
 +
 +/* Find an entry in the second-level page table.. */
 +static inline pmd_t *pmd_offset(pud_t *pud, unsigned long address)
 +{
 +	return (pmd_t *)pud_page_vaddr(*pud) + pmd_index(address);
 +}
 +
 +static inline unsigned long pmd_pfn(pmd_t pmd)
 +{
 +	return (pmd_val(pmd) & PTE_PFN_MASK) >> PAGE_SHIFT;
 +}
 +
 +static inline int pud_large(pud_t pud)
 +{
 +	return (pud_val(pud) & (_PAGE_PSE | _PAGE_PRESENT)) ==
 +		(_PAGE_PSE | _PAGE_PRESENT);
 +}
 +
 +static inline int pud_bad(pud_t pud)
 +{
 +	return (pud_flags(pud) & ~(_KERNPG_TABLE | _PAGE_USER)) != 0;
 +}
  #else
 -# include "pgtable_64.h"
 -#endif
 +static inline int pud_large(pud_t pud)
 +{
 +	return 0;
 +}
 +#endif	/* PAGETABLE_LEVELS > 2 */
 +
 +#if PAGETABLE_LEVELS > 3
 +static inline int pgd_present(pgd_t pgd)
 +{
 +	return pgd_flags(pgd) & _PAGE_PRESENT;
 +}
 +
 +static inline unsigned long pgd_page_vaddr(pgd_t pgd)
 +{
 +	return (unsigned long)__va((unsigned long)pgd_val(pgd) & PTE_PFN_MASK);
 +}
 +
 +/*
 + * Currently stuck as a macro due to indirect forward reference to
 + * linux/mmzone.h's __section_mem_map_addr() definition:
 + */
 +#define pgd_page(pgd)		pfn_to_page(pgd_val(pgd) >> PAGE_SHIFT)
 +
 +/* to find an entry in a page-table-directory. */
 +static inline unsigned pud_index(unsigned long address)
 +{
 +	return (address >> PUD_SHIFT) & (PTRS_PER_PUD - 1);
 +}
 +
 +static inline pud_t *pud_offset(pgd_t *pgd, unsigned long address)
 +{
 +	return (pud_t *)pgd_page_vaddr(*pgd) + pud_index(address);
 +}
 +
 +static inline int pgd_bad(pgd_t pgd)
 +{
 +	return (pgd_flags(pgd) & ~_PAGE_USER) != _KERNPG_TABLE;
 +}
 +
 +static inline int pgd_none(pgd_t pgd)
 +{
 +	return !native_pgd_val(pgd);
 +}
 +#endif	/* PAGETABLE_LEVELS > 3 */
 +
 +#endif	/* __ASSEMBLY__ */
  
  /*
   * the pgd page can be thought of an array like this: pgd_t[PTRS_PER_PGD]