 74d4799221
			
		
	
	
	74d4799221
	
	
	
		
			
			Add code to support the ColdFire V4e MMU pgalloc functions. Signed-off-by: Greg Ungerer <gerg@uclinux.org> Acked-by: Geert Uytterhoeven <geert@linux-m68k.org> Acked-by: Matt Waddel <mwaddel@yahoo.com> Acked-by: Kurt Mahan <kmahan@xmission.com>
		
			
				
	
	
		
			102 lines
		
	
	
	
		
			2.2 KiB
			
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			102 lines
		
	
	
	
		
			2.2 KiB
			
		
	
	
	
		
			C
		
	
	
	
	
	
| #ifndef M68K_MCF_PGALLOC_H
 | |
| #define M68K_MCF_PGALLOC_H
 | |
| 
 | |
| #include <asm/tlb.h>
 | |
| #include <asm/tlbflush.h>
 | |
| 
 | |
| extern inline void pte_free_kernel(struct mm_struct *mm, pte_t *pte)
 | |
| {
 | |
| 	free_page((unsigned long) pte);
 | |
| }
 | |
| 
 | |
| extern const char bad_pmd_string[];
 | |
| 
 | |
| extern inline pte_t *pte_alloc_one_kernel(struct mm_struct *mm,
 | |
| 	unsigned long address)
 | |
| {
 | |
| 	unsigned long page = __get_free_page(GFP_DMA|__GFP_REPEAT);
 | |
| 
 | |
| 	if (!page)
 | |
| 		return NULL;
 | |
| 
 | |
| 	memset((void *)page, 0, PAGE_SIZE);
 | |
| 	return (pte_t *) (page);
 | |
| }
 | |
| 
 | |
| extern inline pmd_t *pmd_alloc_kernel(pgd_t *pgd, unsigned long address)
 | |
| {
 | |
| 	return (pmd_t *) pgd;
 | |
| }
 | |
| 
 | |
| #define pmd_alloc_one_fast(mm, address) ({ BUG(); ((pmd_t *)1); })
 | |
| #define pmd_alloc_one(mm, address)      ({ BUG(); ((pmd_t *)2); })
 | |
| 
 | |
| #define pte_alloc_one_fast(mm, addr) pte_alloc_one(mm, addr)
 | |
| 
 | |
| #define pmd_populate(mm, pmd, page) (pmd_val(*pmd) = \
 | |
| 	(unsigned long)(page_address(page)))
 | |
| 
 | |
| #define pmd_populate_kernel(mm, pmd, pte) (pmd_val(*pmd) = (unsigned long)(pte))
 | |
| 
 | |
| #define pmd_pgtable(pmd) pmd_page(pmd)
 | |
| 
 | |
| static inline void __pte_free_tlb(struct mmu_gather *tlb, pgtable_t page,
 | |
| 				  unsigned long address)
 | |
| {
 | |
| 	__free_page(page);
 | |
| }
 | |
| 
 | |
| #define __pmd_free_tlb(tlb, pmd, address) do { } while (0)
 | |
| 
 | |
| static inline struct page *pte_alloc_one(struct mm_struct *mm,
 | |
| 	unsigned long address)
 | |
| {
 | |
| 	struct page *page = alloc_pages(GFP_DMA|__GFP_REPEAT, 0);
 | |
| 	pte_t *pte;
 | |
| 
 | |
| 	if (!page)
 | |
| 		return NULL;
 | |
| 
 | |
| 	pte = kmap(page);
 | |
| 	if (pte) {
 | |
| 		clear_page(pte);
 | |
| 		__flush_page_to_ram(pte);
 | |
| 		flush_tlb_kernel_page(pte);
 | |
| 		nocache_page(pte);
 | |
| 	}
 | |
| 	kunmap(page);
 | |
| 
 | |
| 	return page;
 | |
| }
 | |
| 
 | |
| extern inline void pte_free(struct mm_struct *mm, struct page *page)
 | |
| {
 | |
| 	__free_page(page);
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * In our implementation, each pgd entry contains 1 pmd that is never allocated
 | |
|  * or freed.  pgd_present is always 1, so this should never be called. -NL
 | |
|  */
 | |
| #define pmd_free(mm, pmd) BUG()
 | |
| 
 | |
| static inline void pgd_free(struct mm_struct *mm, pgd_t *pgd)
 | |
| {
 | |
| 	free_page((unsigned long) pgd);
 | |
| }
 | |
| 
 | |
| static inline pgd_t *pgd_alloc(struct mm_struct *mm)
 | |
| {
 | |
| 	pgd_t *new_pgd;
 | |
| 
 | |
| 	new_pgd = (pgd_t *)__get_free_page(GFP_DMA | __GFP_NOWARN);
 | |
| 	if (!new_pgd)
 | |
| 		return NULL;
 | |
| 	memcpy(new_pgd, swapper_pg_dir, PAGE_SIZE);
 | |
| 	memset(new_pgd, 0, PAGE_OFFSET >> PGDIR_SHIFT);
 | |
| 	return new_pgd;
 | |
| }
 | |
| 
 | |
| #define pgd_populate(mm, pmd, pte) BUG()
 | |
| 
 | |
| #endif /* M68K_MCF_PGALLOC_H */
 |