/* SPDX-License-Identifier: GPL-2.0 */ /* * copy_page, __copy_user_page, __copy_user implementation of SuperH * * Copyright (C) 2001 Niibe Yutaka & Kaz Kojima * Copyright (C) 2002 Toshinobu Sugioka * Copyright (C) 2006 Paul Mundt */ #include <linux/linkage.h> #include <asm/page.h> /* * copy_page * @to: P1 address * @from: P1 address * * void copy_page(void *to, void *from) */ /* * r0, r1, r2, r3, r4, r5, r6, r7 --- scratch * r8 --- from + PAGE_SIZE * r9 --- not used * r10 --- to * r11 --- from */ ENTRY(copy_page) mov.l r8,@-r15 mov.l r10,@-r15 mov.l r11,@-r15 mov r4,r10 mov r5,r11 mov r5,r8 mov #(PAGE_SIZE >> 10), r0 shll8 r0 shll2 r0 add r0,r8 ! 1: mov.l @r11+,r0 mov.l @r11+,r1 mov.l @r11+,r2 mov.l @r11+,r3 mov.l @r11+,r4 mov.l @r11+,r5 mov.l @r11+,r6 mov.l @r11+,r7 #if defined(CONFIG_CPU_SH4) movca.l r0,@r10 #else mov.l r0,@r10 #endif add #32,r10 mov.l r7,@-r10 mov.l r6,@-r10 mov.l r5,@-r10 mov.l r4,@-r10 mov.l r3,@-r10 mov.l r2,@-r10 mov.l r1,@-r10 cmp/eq r11,r8 bf/s 1b add #28,r10 ! mov.l @r15+,r11 mov.l @r15+,r10 mov.l @r15+,r8 rts nop /* * __kernel_size_t __copy_user(void *to, const void *from, __kernel_size_t n); * Return the number of bytes NOT copied */ #define EX(...) \ 9999: __VA_ARGS__ ; \ .section __ex_table, "a"; \ .long 9999b, 6000f ; \ .previous #define EX_NO_POP(...) \ 9999: __VA_ARGS__ ; \ .section __ex_table, "a"; \ .long 9999b, 6005f ; \ .previous ENTRY(__copy_user) ! Check if small number of bytes mov #11,r0 mov r4,r3 cmp/gt r0,r6 ! r6 (len) > r0 (11) bf/s .L_cleanup_loop_no_pop add r6,r3 ! last destination address ! Calculate bytes needed to align to src mov.l r11,@-r15 neg r5,r0 mov.l r10,@-r15 add #4,r0 mov.l r9,@-r15 and #3,r0 mov.l r8,@-r15 tst r0,r0 bt 2f 1: ! Copy bytes to long word align src EX( mov.b @r5+,r1 ) dt r0 add #-1,r6 EX( mov.b r1,@r4 ) bf/s 1b add #1,r4 ! Jump to appropriate routine depending on dest 2: mov #3,r1 mov r6, r2 and r4,r1 shlr2 r2 shll2 r1 mova .L_jump_tbl,r0 mov.l @(r0,r1),r1 jmp @r1 nop .align 2 .L_jump_tbl: .long .L_dest00 .long .L_dest01 .long .L_dest10 .long .L_dest11 /* * Come here if there are less than 12 bytes to copy * * Keep the branch target close, so the bf/s callee doesn't overflow * and result in a more expensive branch being inserted. This is the * fast-path for small copies, the jump via the jump table will hit the * default slow-path cleanup. -PFM. */ .L_cleanup_loop_no_pop: tst r6,r6 ! Check explicitly for zero bt 1f 2: EX_NO_POP( mov.b @r5+,r0 ) dt r6 EX_NO_POP( mov.b r0,@r4 ) bf/s 2b add #1,r4 1: mov #0,r0 ! normal return 5000: # Exception handler: .section .fixup, "ax" 6005: mov.l 8000f,r1 mov r3,r0 jmp @r1 sub r4,r0 .align 2 8000: .long 5000b .previous rts nop ! Destination = 00 .L_dest00: ! Skip the large copy for small transfers mov #(32+32-4), r0 cmp/gt r6, r0 ! r0 (60) > r6 (len) bt 1f ! Align dest to a 32 byte boundary neg r4,r0 add #0x20, r0 and #0x1f, r0 tst r0, r0 bt 2f sub r0, r6 shlr2 r0 3: EX( mov.l @r5+,r1 ) dt r0 EX( mov.l r1,@r4 ) bf/s 3b add #4,r4 2: EX( mov.l @r5+,r0 ) EX( mov.l @r5+,r1 ) EX( mov.l @r5+,r2 ) EX( mov.l @r5+,r7 ) EX( mov.l @r5+,r8 ) EX( mov.l @r5+,r9 ) EX( mov.l @r5+,r10 ) EX( mov.l @r5+,r11 ) #ifdef CONFIG_CPU_SH4 EX( movca.l r0,@r4 ) #else EX( mov.l r0,@r4 ) #endif add #-32, r6 EX( mov.l r1,@(4,r4) ) mov #32, r0 EX( mov.l r2,@(8,r4) ) cmp/gt r6, r0 ! r0 (32) > r6 (len) EX( mov.l r7,@(12,r4) ) EX( mov.l r8,@(16,r4) ) EX( mov.l r9,@(20,r4) ) EX( mov.l r10,@(24,r4) ) EX( mov.l r11,@(28,r4) ) bf/s 2b add #32,r4 1: mov r6, r0 shlr2 r0 tst r0, r0 bt .L_cleanup 1: EX( mov.l @r5+,r1 ) dt r0 EX( mov.l r1,@r4 ) bf/s 1b add #4,r4 bra .L_cleanup nop ! Destination = 10 .L_dest10: mov r2,r7 shlr2 r7 shlr r7 tst r7,r7 mov #7,r0 bt/s 1f and r0,r2 2: dt r7 #ifdef CONFIG_CPU_LITTLE_ENDIAN EX( mov.l @r5+,r0 ) EX( mov.l @r5+,r1 ) EX( mov.l @r5+,r8 ) EX( mov.l @r5+,r9 ) EX( mov.l @r5+,r10 ) EX( mov.w r0,@r4 ) add #2,r4 xtrct r1,r0 xtrct r8,r1 xtrct r9,r8 xtrct r10,r9 EX( mov.l r0,@r4 ) EX( mov.l r1,@(4,r4) ) EX( mov.l r8,@(8,r4) ) EX( mov.l r9,@(12,r4) ) EX( mov.l @r5+,r1 ) EX( mov.l @r5+,r8 ) EX( mov.l @r5+,r0 ) xtrct r1,r10 xtrct r8,r1 xtrct r0,r8 shlr16 r0 EX( mov.l r10,@(16,r4) ) EX( mov.l r1,@(20,r4) ) EX( mov.l r8,@(24,r4) ) EX( mov.w r0,@(28,r4) ) bf/s 2b add #30,r4 #else EX( mov.l @(28,r5),r0 ) EX( mov.l @(24,r5),r8 ) EX( mov.l @(20,r5),r9 ) EX( mov.l @(16,r5),r10 ) EX( mov.w r0,@(30,r4) ) add #-2,r4 xtrct r8,r0 xtrct r9,r8 xtrct r10,r9 EX( mov.l r0,@(28,r4) ) EX( mov.l r8,@(24,r4) ) EX( mov.l r9,@(20,r4) ) EX( mov.l @(12,r5),r0 ) EX( mov.l @(8,r5),r8 ) xtrct r0,r10 EX( mov.l @(4,r5),r9 ) mov.l r10,@(16,r4) EX( mov.l @r5,r10 ) xtrct r8,r0 xtrct r9,r8 xtrct r10,r9 EX( mov.l r0,@(12,r4) ) EX( mov.l r8,@(8,r4) ) swap.w r10,r0 EX( mov.l r9,@(4,r4) ) EX( mov.w r0,@(2,r4) ) add #32,r5 bf/s 2b add #34,r4 #endif tst r2,r2 bt .L_cleanup 1: ! Read longword, write two words per iteration EX( mov.l @r5+,r0 ) dt r2 #ifdef CONFIG_CPU_LITTLE_ENDIAN EX( mov.w r0,@r4 ) shlr16 r0 EX( mov.w r0,@(2,r4) ) #else EX( mov.w r0,@(2,r4) ) shlr16 r0 EX( mov.w r0,@r4 ) #endif bf/s 1b add #4,r4 bra .L_cleanup nop ! Destination = 01 or 11 .L_dest01: .L_dest11: ! Read longword, write byte, word, byte per iteration EX( mov.l @r5+,r0 ) dt r2 #ifdef CONFIG_CPU_LITTLE_ENDIAN EX( mov.b r0,@r4 ) shlr8 r0 add #1,r4 EX( mov.w r0,@r4 ) shlr16 r0 EX( mov.b r0,@(2,r4) ) bf/s .L_dest01 add #3,r4 #else EX( mov.b r0,@(3,r4) ) shlr8 r0 swap.w r0,r7 EX( mov.b r7,@r4 ) add #1,r4 EX( mov.w r0,@r4 ) bf/s .L_dest01 add #3,r4 #endif ! Cleanup last few bytes .L_cleanup: mov r6,r0 and #3,r0 tst r0,r0 bt .L_exit mov r0,r6 .L_cleanup_loop: EX( mov.b @r5+,r0 ) dt r6 EX( mov.b r0,@r4 ) bf/s .L_cleanup_loop add #1,r4 .L_exit: mov #0,r0 ! normal return 5000: # Exception handler: .section .fixup, "ax" 6000: mov.l 8000f,r1 mov r3,r0 jmp @r1 sub r4,r0 .align 2 8000: .long 5000b .previous mov.l @r15+,r8 mov.l @r15+,r9 mov.l @r15+,r10 rts mov.l @r15+,r11