 c177c81e09
			
		
	
	
	c177c81e09
	
	
	
		
			
			Currently hugepage migration is available for all archs which support pmd-level hugepage, but testing is done only for x86_64 and there're bugs for other archs. So to avoid breaking such archs, this patch limits the availability strictly to x86_64 until developers of other archs get interested in enabling this feature. Simply disabling hugepage migration on non-x86_64 archs is not enough to fix the reported problem where sys_move_pages() hits the BUG_ON() in follow_page(FOLL_GET), so let's fix this by checking if hugepage migration is supported in vma_migratable(). Signed-off-by: Naoya Horiguchi <n-horiguchi@ah.jp.nec.com> Reported-by: Michael Ellerman <mpe@ellerman.id.au> Tested-by: Michael Ellerman <mpe@ellerman.id.au> Acked-by: Hugh Dickins <hughd@google.com> Cc: Benjamin Herrenschmidt <benh@kernel.crashing.org> Cc: Tony Luck <tony.luck@intel.com> Cc: Russell King <rmk@arm.linux.org.uk> Cc: Martin Schwidefsky <schwidefsky@de.ibm.com> Cc: James Hogan <james.hogan@imgtec.com> Cc: Ralf Baechle <ralf@linux-mips.org> Cc: David Miller <davem@davemloft.net> Cc: <stable@vger.kernel.org> [3.12+] Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
		
			
				
	
	
		
			90 lines
		
	
	
	
		
			1.5 KiB
			
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			90 lines
		
	
	
	
		
			1.5 KiB
			
		
	
	
	
		
			C
		
	
	
	
	
	
| /*
 | |
|  * arch/sh/mm/hugetlbpage.c
 | |
|  *
 | |
|  * SuperH HugeTLB page support.
 | |
|  *
 | |
|  * Cloned from sparc64 by Paul Mundt.
 | |
|  *
 | |
|  * Copyright (C) 2002, 2003 David S. Miller (davem@redhat.com)
 | |
|  */
 | |
| 
 | |
| #include <linux/init.h>
 | |
| #include <linux/fs.h>
 | |
| #include <linux/mm.h>
 | |
| #include <linux/hugetlb.h>
 | |
| #include <linux/pagemap.h>
 | |
| #include <linux/sysctl.h>
 | |
| 
 | |
| #include <asm/mman.h>
 | |
| #include <asm/pgalloc.h>
 | |
| #include <asm/tlb.h>
 | |
| #include <asm/tlbflush.h>
 | |
| #include <asm/cacheflush.h>
 | |
| 
 | |
| pte_t *huge_pte_alloc(struct mm_struct *mm,
 | |
| 			unsigned long addr, unsigned long sz)
 | |
| {
 | |
| 	pgd_t *pgd;
 | |
| 	pud_t *pud;
 | |
| 	pmd_t *pmd;
 | |
| 	pte_t *pte = NULL;
 | |
| 
 | |
| 	pgd = pgd_offset(mm, addr);
 | |
| 	if (pgd) {
 | |
| 		pud = pud_alloc(mm, pgd, addr);
 | |
| 		if (pud) {
 | |
| 			pmd = pmd_alloc(mm, pud, addr);
 | |
| 			if (pmd)
 | |
| 				pte = pte_alloc_map(mm, NULL, pmd, addr);
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	return pte;
 | |
| }
 | |
| 
 | |
| pte_t *huge_pte_offset(struct mm_struct *mm, unsigned long addr)
 | |
| {
 | |
| 	pgd_t *pgd;
 | |
| 	pud_t *pud;
 | |
| 	pmd_t *pmd;
 | |
| 	pte_t *pte = NULL;
 | |
| 
 | |
| 	pgd = pgd_offset(mm, addr);
 | |
| 	if (pgd) {
 | |
| 		pud = pud_offset(pgd, addr);
 | |
| 		if (pud) {
 | |
| 			pmd = pmd_offset(pud, addr);
 | |
| 			if (pmd)
 | |
| 				pte = pte_offset_map(pmd, addr);
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	return pte;
 | |
| }
 | |
| 
 | |
| int huge_pmd_unshare(struct mm_struct *mm, unsigned long *addr, pte_t *ptep)
 | |
| {
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| struct page *follow_huge_addr(struct mm_struct *mm,
 | |
| 			      unsigned long address, int write)
 | |
| {
 | |
| 	return ERR_PTR(-EINVAL);
 | |
| }
 | |
| 
 | |
| int pmd_huge(pmd_t pmd)
 | |
| {
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| int pud_huge(pud_t pud)
 | |
| {
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| struct page *follow_huge_pmd(struct mm_struct *mm, unsigned long address,
 | |
| 			     pmd_t *pmd, int write)
 | |
| {
 | |
| 	return NULL;
 | |
| }
 |