mirror of
https://github.com/adulau/aha.git
synced 2024-12-28 03:36:19 +00:00
x86 PAT: return compatible mapping to remap_pfn_range callers
Impact: avoid warning message, potentially solve 3D performance regression Change x86 PAT code to return compatible memtype if the exact memtype that was requested in remap_pfn_rage and friends is not available due to some conflict. This is done by returning the compatible type in pgprot parameter of track_pfn_vma_new(), and the caller uses that memtype for page table. Note that track_pfn_vma_copy() which is basically called during fork gets the prot from existing page table and should not have any conflict. Hence we use strict memtype check there and do not allow compatible memtypes. This patch fixes the bug reported here: http://marc.info/?l=linux-kernel&m=123108883716357&w=2 Specifically the error message: X:5010 map pfn expected mapping type write-back for d0000000-d0101000, got write-combining Should go away. Reported-and-bisected-by: Kevin Winchester <kjwinchester@gmail.com> Signed-off-by: Venkatesh Pallipadi <venkatesh.pallipadi@intel.com> Signed-off-by: Suresh Siddha <suresh.b.siddha@intel.com> Signed-off-by: Ingo Molnar <mingo@elte.hu>
This commit is contained in:
parent
e4b866ed19
commit
cdecff6864
1 changed files with 28 additions and 15 deletions
|
@ -601,12 +601,13 @@ void unmap_devmem(unsigned long pfn, unsigned long size, pgprot_t vma_prot)
|
|||
* Reserved non RAM regions only and after successful reserve_memtype,
|
||||
* this func also keeps identity mapping (if any) in sync with this new prot.
|
||||
*/
|
||||
static int reserve_pfn_range(u64 paddr, unsigned long size, pgprot_t vma_prot)
|
||||
static int reserve_pfn_range(u64 paddr, unsigned long size, pgprot_t *vma_prot,
|
||||
int strict_prot)
|
||||
{
|
||||
int is_ram = 0;
|
||||
int id_sz, ret;
|
||||
unsigned long flags;
|
||||
unsigned long want_flags = (pgprot_val(vma_prot) & _PAGE_CACHE_MASK);
|
||||
unsigned long want_flags = (pgprot_val(*vma_prot) & _PAGE_CACHE_MASK);
|
||||
|
||||
is_ram = pagerange_is_ram(paddr, paddr + size);
|
||||
|
||||
|
@ -625,15 +626,24 @@ static int reserve_pfn_range(u64 paddr, unsigned long size, pgprot_t vma_prot)
|
|||
return ret;
|
||||
|
||||
if (flags != want_flags) {
|
||||
free_memtype(paddr, paddr + size);
|
||||
printk(KERN_ERR
|
||||
"%s:%d map pfn expected mapping type %s for %Lx-%Lx, got %s\n",
|
||||
current->comm, current->pid,
|
||||
cattr_name(want_flags),
|
||||
(unsigned long long)paddr,
|
||||
(unsigned long long)(paddr + size),
|
||||
cattr_name(flags));
|
||||
return -EINVAL;
|
||||
if (strict_prot || !is_new_memtype_allowed(want_flags, flags)) {
|
||||
free_memtype(paddr, paddr + size);
|
||||
printk(KERN_ERR "%s:%d map pfn expected mapping type %s"
|
||||
" for %Lx-%Lx, got %s\n",
|
||||
current->comm, current->pid,
|
||||
cattr_name(want_flags),
|
||||
(unsigned long long)paddr,
|
||||
(unsigned long long)(paddr + size),
|
||||
cattr_name(flags));
|
||||
return -EINVAL;
|
||||
}
|
||||
/*
|
||||
* We allow returning different type than the one requested in
|
||||
* non strict case.
|
||||
*/
|
||||
*vma_prot = __pgprot((pgprot_val(*vma_prot) &
|
||||
(~_PAGE_CACHE_MASK)) |
|
||||
flags);
|
||||
}
|
||||
|
||||
/* Need to keep identity mapping in sync */
|
||||
|
@ -689,6 +699,7 @@ int track_pfn_vma_copy(struct vm_area_struct *vma)
|
|||
unsigned long vma_start = vma->vm_start;
|
||||
unsigned long vma_end = vma->vm_end;
|
||||
unsigned long vma_size = vma_end - vma_start;
|
||||
pgprot_t pgprot;
|
||||
|
||||
if (!pat_enabled)
|
||||
return 0;
|
||||
|
@ -702,7 +713,8 @@ int track_pfn_vma_copy(struct vm_area_struct *vma)
|
|||
WARN_ON_ONCE(1);
|
||||
return -EINVAL;
|
||||
}
|
||||
return reserve_pfn_range(paddr, vma_size, __pgprot(prot));
|
||||
pgprot = __pgprot(prot);
|
||||
return reserve_pfn_range(paddr, vma_size, &pgprot, 1);
|
||||
}
|
||||
|
||||
/* reserve entire vma page by page, using pfn and prot from pte */
|
||||
|
@ -710,7 +722,8 @@ int track_pfn_vma_copy(struct vm_area_struct *vma)
|
|||
if (follow_phys(vma, vma_start + i, 0, &prot, &paddr))
|
||||
continue;
|
||||
|
||||
retval = reserve_pfn_range(paddr, PAGE_SIZE, __pgprot(prot));
|
||||
pgprot = __pgprot(prot);
|
||||
retval = reserve_pfn_range(paddr, PAGE_SIZE, &pgprot, 1);
|
||||
if (retval)
|
||||
goto cleanup_ret;
|
||||
}
|
||||
|
@ -758,14 +771,14 @@ int track_pfn_vma_new(struct vm_area_struct *vma, pgprot_t *prot,
|
|||
if (is_linear_pfn_mapping(vma)) {
|
||||
/* reserve the whole chunk starting from vm_pgoff */
|
||||
paddr = (resource_size_t)vma->vm_pgoff << PAGE_SHIFT;
|
||||
return reserve_pfn_range(paddr, vma_size, *prot);
|
||||
return reserve_pfn_range(paddr, vma_size, prot, 0);
|
||||
}
|
||||
|
||||
/* reserve page by page using pfn and size */
|
||||
base_paddr = (resource_size_t)pfn << PAGE_SHIFT;
|
||||
for (i = 0; i < size; i += PAGE_SIZE) {
|
||||
paddr = base_paddr + i;
|
||||
retval = reserve_pfn_range(paddr, PAGE_SIZE, *prot);
|
||||
retval = reserve_pfn_range(paddr, PAGE_SIZE, prot, 0);
|
||||
if (retval)
|
||||
goto cleanup_ret;
|
||||
}
|
||||
|
|
Loading…
Reference in a new issue