void flush_dcache_page(struct page *page) { struct address_space *mapping = page_mapping(page); if (mapping && !mapping_mapped(mapping)) { if (!test_bit(PG_arch_1, &page->flags)) set_bit(PG_arch_1, &page->flags); return; } else { unsigned long phys = page_to_phys(page); unsigned long temp = page->index << PAGE_SHIFT; unsigned long alias = !(DCACHE_ALIAS_EQ(temp, phys)); unsigned long virt; if (!alias && !mapping) return; __flush_invalidate_dcache_page((long)page_address(page)); virt = TLBTEMP_BASE_1 + (temp & DCACHE_ALIAS_MASK); if (alias) __flush_invalidate_dcache_page_alias(virt, phys); if (mapping) __invalidate_icache_page_alias(virt, phys); } }
void copy_to_user_page(struct vm_area_struct *vma, struct page *page, unsigned long vaddr, void *dst, const void *src, unsigned long len) { unsigned long phys = page_to_phys(page); unsigned long alias = !(DCACHE_ALIAS_EQ(vaddr, phys)); /* Flush and invalidate user page if aliased. */ if (alias) { unsigned long t = TLBTEMP_BASE_1 + (vaddr & DCACHE_ALIAS_MASK); __flush_invalidate_dcache_page_alias(t, phys); } /* Copy data */ memcpy(dst, src, len); /* * Flush and invalidate kernel page if aliased and synchronize * data and instruction caches for executable pages. */ if (alias) { unsigned long t = TLBTEMP_BASE_1 + (vaddr & DCACHE_ALIAS_MASK); __flush_invalidate_dcache_range((unsigned long) dst, len); if ((vma->vm_flags & VM_EXEC) != 0) __invalidate_icache_page_alias(t, phys); } else if ((vma->vm_flags & VM_EXEC) != 0) { __flush_dcache_range((unsigned long)dst,len); __invalidate_icache_range((unsigned long) dst, len); } }
void copy_to_user_page(struct vm_area_struct *vma, struct page *page, unsigned long vaddr, void *dst, const void *src, unsigned long len) { unsigned long phys = page_to_phys(page); unsigned long alias = !(DCACHE_ALIAS_EQ(vaddr, phys)); if (alias) { unsigned long temp = TLBTEMP_BASE_1 + (vaddr & DCACHE_ALIAS_MASK); __flush_invalidate_dcache_page_alias(temp, phys); } memcpy(dst, src, len); if (alias) { unsigned long temp = TLBTEMP_BASE_1 + (vaddr & DCACHE_ALIAS_MASK); __flush_invalidate_dcache_range((unsigned long) dst, len); if ((vma->vm_flags & VM_EXEC) != 0) { __invalidate_icache_page_alias(temp, phys); } } else if ((vma->vm_flags & VM_EXEC) != 0) { __flush_dcache_range((unsigned long)dst,len); __invalidate_icache_range((unsigned long) dst, len); } }
void flush_dcache_page(struct page *page) { struct address_space *mapping = page_mapping(page); /* * If we have a mapping but the page is not mapped to user-space * yet, we simply mark this page dirty and defer flushing the * caches until update_mmu(). */ if (mapping && !mapping_mapped(mapping)) { if (!test_bit(PG_arch_1, &page->flags)) set_bit(PG_arch_1, &page->flags); return; } else { unsigned long phys = page_to_phys(page); unsigned long temp = page->index << PAGE_SHIFT; unsigned long alias = !(DCACHE_ALIAS_EQ(temp, phys)); unsigned long virt; /* * Flush the page in kernel space and user space. * Note that we can omit that step if aliasing is not * an issue, but we do have to synchronize I$ and D$ * if we have a mapping. */ if (!alias && !mapping) return; virt = TLBTEMP_BASE_1 + (phys & DCACHE_ALIAS_MASK); __flush_invalidate_dcache_page_alias(virt, phys); virt = TLBTEMP_BASE_1 + (temp & DCACHE_ALIAS_MASK); if (alias) __flush_invalidate_dcache_page_alias(virt, phys); if (mapping) __invalidate_icache_page_alias(virt, phys); } /* There shouldn't be an entry in the cache for this page anymore. */ }
void local_flush_cache_page(struct vm_area_struct *vma, unsigned long address, unsigned long pfn) { /* Note that we have to use the 'alias' address to avoid multi-hit */ unsigned long phys = page_to_phys(pfn_to_page(pfn)); unsigned long virt = TLBTEMP_BASE_1 + (address & DCACHE_ALIAS_MASK); __flush_invalidate_dcache_page_alias(virt, phys); __invalidate_icache_page_alias(virt, phys); }
void flush_cache_page(struct vm_area_struct* vma, unsigned long address, unsigned long pfn) { unsigned long phys = page_to_phys(pfn_to_page(pfn)); unsigned long virt = TLBTEMP_BASE_1 + (address & DCACHE_ALIAS_MASK); __flush_invalidate_dcache_page_alias(virt, phys); __invalidate_icache_page_alias(virt, phys); }
void update_mmu_cache(struct vm_area_struct * vma, unsigned long addr, pte_t *ptep) { unsigned long pfn = pte_pfn(*ptep); struct page *page; if (!pfn_valid(pfn)) return; page = pfn_to_page(pfn); /* Invalidate old entry in TLBs */ flush_tlb_page(vma, addr); #if (DCACHE_WAY_SIZE > PAGE_SIZE) if (!PageReserved(page) && test_bit(PG_arch_1, &page->flags)) { unsigned long phys = page_to_phys(page); unsigned long tmp; tmp = TLBTEMP_BASE_1 + (phys & DCACHE_ALIAS_MASK); __flush_invalidate_dcache_page_alias(tmp, phys); tmp = TLBTEMP_BASE_1 + (addr & DCACHE_ALIAS_MASK); __flush_invalidate_dcache_page_alias(tmp, phys); __invalidate_icache_page_alias(tmp, phys); clear_bit(PG_arch_1, &page->flags); } #else if (!PageReserved(page) && !test_bit(PG_arch_1, &page->flags) && (vma->vm_flags & VM_EXEC) != 0) { unsigned long paddr = (unsigned long)kmap_atomic(page); __flush_dcache_page(paddr); __invalidate_icache_page(paddr); set_bit(PG_arch_1, &page->flags); kunmap_atomic((void *)paddr); } #endif }
extern void copy_from_user_page(struct vm_area_struct *vma, struct page *page, unsigned long vaddr, void *dst, const void *src, unsigned long len) { unsigned long phys = page_to_phys(page); unsigned long alias = !(DCACHE_ALIAS_EQ(vaddr, phys)); if (alias) { unsigned long temp = TLBTEMP_BASE_1 + (vaddr & DCACHE_ALIAS_MASK); __flush_invalidate_dcache_page_alias(temp, phys); } memcpy(dst, src, len); }
extern void copy_from_user_page(struct vm_area_struct *vma, struct page *page, unsigned long vaddr, void *dst, const void *src, unsigned long len) { unsigned long phys = page_to_phys(page); unsigned long alias = !(DCACHE_ALIAS_EQ(vaddr, phys)); /* * Flush user page if aliased. * (Note: a simply flush would be sufficient) */ if (alias) { unsigned long t = TLBTEMP_BASE_1 + (vaddr & DCACHE_ALIAS_MASK); __flush_invalidate_dcache_page_alias(t, phys); } memcpy(dst, src, len); }
void update_mmu_cache(struct vm_area_struct * vma, unsigned long addr, pte_t *ptep) { unsigned long pfn = pte_pfn(*ptep); struct page *page; if (!pfn_valid(pfn)) return; page = pfn_to_page(pfn); invalidate_itlb_mapping(addr); invalidate_dtlb_mapping(addr); #if (DCACHE_WAY_SIZE > PAGE_SIZE) && XCHAL_DCACHE_IS_WRITEBACK if (!PageReserved(page) && test_bit(PG_arch_1, &page->flags)) { unsigned long vaddr = TLBTEMP_BASE_1 + (addr & DCACHE_ALIAS_MASK); unsigned long paddr = (unsigned long) page_address(page); unsigned long phys = page_to_phys(page); __flush_invalidate_dcache_page(paddr); __flush_invalidate_dcache_page_alias(vaddr, phys); __invalidate_icache_page_alias(vaddr, phys); clear_bit(PG_arch_1, &page->flags); } #else if (!PageReserved(page) && !test_bit(PG_arch_1, &page->flags) && (vma->vm_flags & VM_EXEC) != 0) { unsigned long paddr = (unsigned long) page_address(page); __flush_dcache_page(paddr); __invalidate_icache_page(paddr); set_bit(PG_arch_1, &page->flags); } #endif }