/* SPDX-License-Identifier: GPL-2.0 */ /* clear_page.S: UltraSparc optimized copy page. * * Copyright (C) 1996, 1998, 1999, 2000, 2004 David S. Miller (davem@redhat.com) * Copyright (C) 1997 Jakub Jelinek (jakub@redhat.com) */ #include <linux/export.h> #include <asm/visasm.h> #include <asm/thread_info.h> #include <asm/page.h> #include <linux/pgtable.h> #include <asm/spitfire.h> #include <asm/head.h> /* What we used to do was lock a TLB entry into a specific * TLB slot, clear the page with interrupts disabled, then * restore the original TLB entry. This was great for * disturbing the TLB as little as possible, but it meant * we had to keep interrupts disabled for a long time. * * Now, we simply use the normal TLB loading mechanism, * and this makes the cpu choose a slot all by itself. * Then we do a normal TLB flush on exit. We need only * disable preemption during the clear. */ #define DCACHE_SIZE (PAGE_SIZE * 2) #if (PAGE_SHIFT == 13) #define PAGE_SIZE_REM 0x80 #elif (PAGE_SHIFT == 16) #define PAGE_SIZE_REM 0x100 #else #error Wrong PAGE_SHIFT specified #endif #define TOUCH(reg0, reg1, reg2, reg3, reg4, reg5, reg6, reg7) \ fsrc2 %reg0, %f48; fsrc2 %reg1, %f50; \ fsrc2 %reg2, %f52; fsrc2 %reg3, %f54; \ fsrc2 %reg4, %f56; fsrc2 %reg5, %f58; \ fsrc2 %reg6, %f60; fsrc2 %reg7, %f62; .text .align 32 .globl copy_user_page .type copy_user_page,#function EXPORT_SYMBOL(copy_user_page) copy_user_page: /* %o0=dest, %o1=src, %o2=vaddr */ lduw [%g6 + TI_PRE_COUNT], %o4 sethi %hi(PAGE_OFFSET), %g2 sethi %hi(PAGE_SIZE), %o3 ldx [%g2 + %lo(PAGE_OFFSET)], %g2 sethi %hi(PAGE_KERNEL_LOCKED), %g3 ldx [%g3 + %lo(PAGE_KERNEL_LOCKED)], %g3 sub %o0, %g2, %g1 ! dest paddr sub %o1, %g2, %g2 ! src paddr and %o2, %o3, %o0 ! vaddr D-cache alias bit or %g1, %g3, %g1 ! dest TTE data or %g2, %g3, %g2 ! src TTE data sethi %hi(TLBTEMP_BASE), %o3 sethi %hi(DCACHE_SIZE), %o1 add %o0, %o3, %o0 ! dest TTE vaddr add %o4, 1, %o2 add %o0, %o1, %o1 ! src TTE vaddr /* Disable preemption. */ mov TLB_TAG_ACCESS, %g3 stw %o2, [%g6 + TI_PRE_COUNT] /* Load TLB entries. */ rdpr %pstate, %o2 wrpr %o2, PSTATE_IE, %pstate stxa %o0, [%g3] ASI_DMMU stxa %g1, [%g0] ASI_DTLB_DATA_IN membar #Sync stxa %o1, [%g3] ASI_DMMU stxa %g2, [%g0] ASI_DTLB_DATA_IN membar #Sync wrpr %o2, 0x0, %pstate cheetah_copy_page_insn: ba,pt %xcc, 9f nop 1: VISEntryHalf membar #StoreLoad | #StoreStore | #LoadStore sethi %hi((PAGE_SIZE/64)-2), %o2 mov %o0, %g1 prefetch [%o1 + 0x000], #one_read or %o2, %lo((PAGE_SIZE/64)-2), %o2 prefetch [%o1 + 0x040], #one_read prefetch [%o1 + 0x080], #one_read prefetch [%o1 + 0x0c0], #one_read ldd [%o1 + 0x000], %f0 prefetch [%o1 + 0x100], #one_read ldd [%o1 + 0x008], %f2 prefetch [%o1 + 0x140], #one_read ldd [%o1 + 0x010], %f4 prefetch [%o1 + 0x180], #one_read fsrc2 %f0, %f16 ldd [%o1 + 0x018], %f6 fsrc2 %f2, %f18 ldd [%o1 + 0x020], %f8 fsrc2 %f4, %f20 ldd [%o1 + 0x028], %f10 fsrc2 %f6, %f22 ldd [%o1 + 0x030], %f12 fsrc2 %f8, %f24 ldd [%o1 + 0x038], %f14 fsrc2 %f10, %f26 ldd [%o1 + 0x040], %f0 1: ldd [%o1 + 0x048], %f2 fsrc2 %f12, %f28 ldd [%o1 + 0x050], %f4 fsrc2 %f14, %f30 stda %f16, [%o0] ASI_BLK_P ldd [%o1 + 0x058], %f6 fsrc2 %f0, %f16 ldd [%o1 + 0x060], %f8 fsrc2 %f2, %f18 ldd [%o1 + 0x068], %f10 fsrc2 %f4, %f20 ldd [%o1 + 0x070], %f12 fsrc2 %f6, %f22 ldd [%o1 + 0x078], %f14 fsrc2 %f8, %f24 ldd [%o1 + 0x080], %f0 prefetch [%o1 + 0x180], #one_read fsrc2 %f10, %f26 subcc %o2, 1, %o2 add %o0, 0x40, %o0 bne,pt %xcc, 1b add %o1, 0x40, %o1 ldd [%o1 + 0x048], %f2 fsrc2 %f12, %f28 ldd [%o1 + 0x050], %f4 fsrc2 %f14, %f30 stda %f16, [%o0] ASI_BLK_P ldd [%o1 + 0x058], %f6 fsrc2 %f0, %f16 ldd [%o1 + 0x060], %f8 fsrc2 %f2, %f18 ldd [%o1 + 0x068], %f10 fsrc2 %f4, %f20 ldd [%o1 + 0x070], %f12 fsrc2 %f6, %f22 add %o0, 0x40, %o0 ldd [%o1 + 0x078], %f14 fsrc2 %f8, %f24 fsrc2 %f10, %f26 fsrc2 %f12, %f28 fsrc2 %f14, %f30 stda %f16, [%o0] ASI_BLK_P membar #Sync VISExitHalf ba,pt %xcc, 5f nop 9: VISEntry ldub [%g6 + TI_FAULT_CODE], %g3 mov %o0, %g1 cmp %g3, 0 rd %asi, %g3 be,a,pt %icc, 1f wr %g0, ASI_BLK_P, %asi wr %g0, ASI_BLK_COMMIT_P, %asi 1: ldda [%o1] ASI_BLK_P, %f0 add %o1, 0x40, %o1 ldda [%o1] ASI_BLK_P, %f16 add %o1, 0x40, %o1 sethi %hi(PAGE_SIZE), %o2 1: TOUCH(f0, f2, f4, f6, f8, f10, f12, f14) ldda [%o1] ASI_BLK_P, %f32 stda %f48, [%o0] %asi add %o1, 0x40, %o1 sub %o2, 0x40, %o2 add %o0, 0x40, %o0 TOUCH(f16, f18, f20, f22, f24, f26, f28, f30) ldda [%o1] ASI_BLK_P, %f0 stda %f48, [%o0] %asi add %o1, 0x40, %o1 sub %o2, 0x40, %o2 add %o0, 0x40, %o0 TOUCH(f32, f34, f36, f38, f40, f42, f44, f46) ldda [%o1] ASI_BLK_P, %f16 stda %f48, [%o0] %asi sub %o2, 0x40, %o2 add %o1, 0x40, %o1 cmp %o2, PAGE_SIZE_REM bne,pt %xcc, 1b add %o0, 0x40, %o0 #if (PAGE_SHIFT == 16) TOUCH(f0, f2, f4, f6, f8, f10, f12, f14) ldda [%o1] ASI_BLK_P, %f32 stda %f48, [%o0] %asi add %o1, 0x40, %o1 sub %o2, 0x40, %o2 add %o0, 0x40, %o0 TOUCH(f16, f18, f20, f22, f24, f26, f28, f30) ldda [%o1] ASI_BLK_P, %f0 stda %f48, [%o0] %asi add %o1, 0x40, %o1 sub %o2, 0x40, %o2 add %o0, 0x40, %o0 membar #Sync stda %f32, [%o0] %asi add %o0, 0x40, %o0 stda %f0, [%o0] %asi #else membar #Sync stda %f0, [%o0] %asi add %o0, 0x40, %o0 stda %f16, [%o0] %asi #endif membar #Sync wr %g3, 0x0, %asi VISExit 5: stxa %g0, [%g1] ASI_DMMU_DEMAP membar #Sync sethi %hi(DCACHE_SIZE), %g2 stxa %g0, [%g1 + %g2] ASI_DMMU_DEMAP membar #Sync retl stw %o4, [%g6 + TI_PRE_COUNT] .size copy_user_page, .-copy_user_page .globl cheetah_patch_copy_page cheetah_patch_copy_page: sethi %hi(0x01000000), %o1 ! NOP sethi %hi(cheetah_copy_page_insn), %o0 or %o0, %lo(cheetah_copy_page_insn), %o0 stw %o1, [%o0] membar #StoreStore flush %o0 retl nop