From: William Lee Irwin III This patch introduces remap_pfn_range(), destined to replace remap_page_range(), to which all callers of remap_page_range() are converted in the sequel. Signed-off-by: Andrew Morton --- 25-akpm/include/linux/mm.h | 11 +++++++++-- 25-akpm/mm/memory.c | 19 ++++++++----------- 25-akpm/mm/nommu.c | 2 +- 3 files changed, 18 insertions(+), 14 deletions(-) diff -puN include/linux/mm.h~introduce-remap_pfn_range-to-replace-remap_page_range include/linux/mm.h --- 25/include/linux/mm.h~introduce-remap_pfn_range-to-replace-remap_page_range 2004-10-01 00:25:42.793134936 -0700 +++ 25-akpm/include/linux/mm.h 2004-10-01 00:25:42.800133872 -0700 @@ -766,8 +766,15 @@ extern struct vm_area_struct *find_exten extern struct page * vmalloc_to_page(void *addr); extern struct page * follow_page(struct mm_struct *mm, unsigned long address, int write); -extern int remap_page_range(struct vm_area_struct *vma, unsigned long from, - unsigned long to, unsigned long size, pgprot_t prot); +int remap_pfn_range(struct vm_area_struct *, unsigned long, + unsigned long, unsigned long, pgprot_t); + +static inline __deprecated /* since 25 Sept 2004 -- wli */ +int remap_page_range(struct vm_area_struct *vma, unsigned long uvaddr, + unsigned long paddr, unsigned long size, pgprot_t prot) +{ + return remap_pfn_range(vma, uvaddr, paddr >> PAGE_SHIFT, size, prot); +} #ifdef CONFIG_PROC_FS void __vm_stat_account(struct mm_struct *, unsigned long, struct file *, long); diff -puN mm/memory.c~introduce-remap_pfn_range-to-replace-remap_page_range mm/memory.c --- 25/mm/memory.c~introduce-remap_pfn_range-to-replace-remap_page_range 2004-10-01 00:25:42.795134632 -0700 +++ 25-akpm/mm/memory.c 2004-10-01 00:25:42.802133568 -0700 @@ -945,16 +945,14 @@ int zeromap_page_range(struct vm_area_st * in null mappings (currently treated as "copy-on-access") */ static inline void remap_pte_range(pte_t * pte, unsigned long address, unsigned long size, - unsigned long phys_addr, pgprot_t prot) + unsigned long pfn, pgprot_t prot) { unsigned long end; - unsigned long pfn; address &= ~PMD_MASK; end = address + size; if (end > PMD_SIZE) end = PMD_SIZE; - pfn = phys_addr >> PAGE_SHIFT; do { BUG_ON(!pte_none(*pte)); if (!pfn_valid(pfn) || PageReserved(pfn_to_page(pfn))) @@ -966,7 +964,7 @@ static inline void remap_pte_range(pte_t } static inline int remap_pmd_range(struct mm_struct *mm, pmd_t * pmd, unsigned long address, unsigned long size, - unsigned long phys_addr, pgprot_t prot) + unsigned long pfn, pgprot_t prot) { unsigned long base, end; @@ -975,12 +973,12 @@ static inline int remap_pmd_range(struct end = address + size; if (end > PGDIR_SIZE) end = PGDIR_SIZE; - phys_addr -= address; + pfn -= address >> PAGE_SHIFT; do { pte_t * pte = pte_alloc_map(mm, pmd, base + address); if (!pte) return -ENOMEM; - remap_pte_range(pte, base + address, end - address, address + phys_addr, prot); + remap_pte_range(pte, base + address, end - address, pfn + (address >> PAGE_SHIFT), prot); pte_unmap(pte); address = (address + PMD_SIZE) & PMD_MASK; pmd++; @@ -989,7 +987,7 @@ static inline int remap_pmd_range(struct } /* Note: this is only safe if the mm semaphore is held when called. */ -int remap_page_range(struct vm_area_struct *vma, unsigned long from, unsigned long phys_addr, unsigned long size, pgprot_t prot) +int remap_pfn_range(struct vm_area_struct *vma, unsigned long from, unsigned long pfn, unsigned long size, pgprot_t prot) { int error = 0; pgd_t * dir; @@ -997,7 +995,7 @@ int remap_page_range(struct vm_area_stru unsigned long end = from + size; struct mm_struct *mm = vma->vm_mm; - phys_addr -= from; + pfn -= from >> PAGE_SHIFT; dir = pgd_offset(mm, from); flush_cache_range(vma, beg, end); if (from >= end) @@ -1009,7 +1007,7 @@ int remap_page_range(struct vm_area_stru error = -ENOMEM; if (!pmd) break; - error = remap_pmd_range(mm, pmd, from, end - from, phys_addr + from, prot); + error = remap_pmd_range(mm, pmd, from, end - from, pfn + (from >> PAGE_SHIFT), prot); if (error) break; from = (from + PGDIR_SIZE) & PGDIR_MASK; @@ -1022,8 +1020,7 @@ int remap_page_range(struct vm_area_stru spin_unlock(&mm->page_table_lock); return error; } - -EXPORT_SYMBOL(remap_page_range); +EXPORT_SYMBOL(remap_pfn_range); /* * Do pte_mkwrite, but only if the vma says VM_WRITE. We do this when diff -puN mm/nommu.c~introduce-remap_pfn_range-to-replace-remap_page_range mm/nommu.c --- 25/mm/nommu.c~introduce-remap_pfn_range-to-replace-remap_page_range 2004-10-01 00:25:42.796134480 -0700 +++ 25-akpm/mm/nommu.c 2004-10-01 00:25:42.803133416 -0700 @@ -560,7 +560,7 @@ struct vm_area_struct *find_extend_vma(s return NULL; } -int remap_page_range(struct vm_area_struct *vma, unsigned long from, +int remap_pfn_range(struct vm_area_struct *vma, unsigned long from, unsigned long to, unsigned long size, pgprot_t prot) { return -EPERM; _