forked from luck/tmp_suning_uos_patched
5e5419734c
(with Martin Schwidefsky <schwidefsky@de.ibm.com>) The pgd/pud/pmd/pte page table allocation functions get a mm_struct pointer as first argument. The free functions do not get the mm_struct argument. This is 1) asymmetrical and 2) to do mm related page table allocations the mm argument is needed on the free function as well. [kamalesh@linux.vnet.ibm.com: i386 fix] [akpm@linux-foundation.org: coding-syle fixes] Signed-off-by: Benjamin Herrenschmidt <benh@kernel.crashing.org> Signed-off-by: Martin Schwidefsky <schwidefsky@de.ibm.com> Cc: <linux-arch@vger.kernel.org> Signed-off-by: Kamalesh Babulal <kamalesh@linux.vnet.ibm.com> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
78 lines
1.5 KiB
C
78 lines
1.5 KiB
C
#ifndef _ALPHA_PGALLOC_H
|
|
#define _ALPHA_PGALLOC_H
|
|
|
|
#include <linux/mm.h>
|
|
#include <linux/mmzone.h>
|
|
|
|
/*
|
|
* Allocate and free page tables. The xxx_kernel() versions are
|
|
* used to allocate a kernel page table - this turns on ASN bits
|
|
* if any.
|
|
*/
|
|
|
|
static inline void
|
|
pmd_populate(struct mm_struct *mm, pmd_t *pmd, struct page *pte)
|
|
{
|
|
pmd_set(pmd, (pte_t *)(page_to_pa(pte) + PAGE_OFFSET));
|
|
}
|
|
|
|
static inline void
|
|
pmd_populate_kernel(struct mm_struct *mm, pmd_t *pmd, pte_t *pte)
|
|
{
|
|
pmd_set(pmd, pte);
|
|
}
|
|
|
|
static inline void
|
|
pgd_populate(struct mm_struct *mm, pgd_t *pgd, pmd_t *pmd)
|
|
{
|
|
pgd_set(pgd, pmd);
|
|
}
|
|
|
|
extern pgd_t *pgd_alloc(struct mm_struct *mm);
|
|
|
|
static inline void
|
|
pgd_free(struct mm_struct *mm, pgd_t *pgd)
|
|
{
|
|
free_page((unsigned long)pgd);
|
|
}
|
|
|
|
static inline pmd_t *
|
|
pmd_alloc_one(struct mm_struct *mm, unsigned long address)
|
|
{
|
|
pmd_t *ret = (pmd_t *)__get_free_page(GFP_KERNEL|__GFP_REPEAT|__GFP_ZERO);
|
|
return ret;
|
|
}
|
|
|
|
static inline void
|
|
pmd_free(struct mm_struct *mm, pmd_t *pmd)
|
|
{
|
|
free_page((unsigned long)pmd);
|
|
}
|
|
|
|
extern pte_t *pte_alloc_one_kernel(struct mm_struct *mm, unsigned long addr);
|
|
|
|
static inline void
|
|
pte_free_kernel(struct mm_struct *mm, pte_t *pte)
|
|
{
|
|
free_page((unsigned long)pte);
|
|
}
|
|
|
|
static inline struct page *
|
|
pte_alloc_one(struct mm_struct *mm, unsigned long addr)
|
|
{
|
|
pte_t *pte = pte_alloc_one_kernel(mm, addr);
|
|
if (pte)
|
|
return virt_to_page(pte);
|
|
return NULL;
|
|
}
|
|
|
|
static inline void
|
|
pte_free(struct mm_struct *mm, struct page *page)
|
|
{
|
|
__free_page(page);
|
|
}
|
|
|
|
#define check_pgt_cache() do { } while (0)
|
|
|
|
#endif /* _ALPHA_PGALLOC_H */
|