/* SPDX-License-Identifier: GPL-2.0-only */ /* * arch/arm/include/asm/xor.h * * Copyright (C) 2001 Russell King */ #include <linux/hardirq.h> #include <asm-generic/xor.h> #include <asm/hwcap.h> #include <asm/neon.h> #define __XOR(a1, a2) a1 ^= a2 #define GET_BLOCK_2(dst) \ __asm__("ldmia %0, {%1, %2}" \ : "=r" (dst), "=r" (a1), "=r" (a2) \ : "0" (dst)) #define GET_BLOCK_4(dst) \ __asm__("ldmia %0, {%1, %2, %3, %4}" \ : "=r" (dst), "=r" (a1), "=r" (a2), "=r" (a3), "=r" (a4) \ : "0" (dst)) #define XOR_BLOCK_2(src) \ __asm__("ldmia %0!, {%1, %2}" \ : "=r" (src), "=r" (b1), "=r" (b2) \ : "0" (src)); \ __XOR(a1, b1); __XOR(a2, b2); #define XOR_BLOCK_4(src) \ __asm__("ldmia %0!, {%1, %2, %3, %4}" \ : "=r" (src), "=r" (b1), "=r" (b2), "=r" (b3), "=r" (b4) \ : "0" (src)); \ __XOR(a1, b1); __XOR(a2, b2); __XOR(a3, b3); __XOR(a4, b4) #define PUT_BLOCK_2(dst) \ __asm__ __volatile__("stmia %0!, {%2, %3}" \ : "=r" (dst) \ : "0" (dst), "r" (a1), "r" (a2)) #define PUT_BLOCK_4(dst) \ __asm__ __volatile__("stmia %0!, {%2, %3, %4, %5}" \ : "=r" (dst) \ : "0" (dst), "r" (a1), "r" (a2), "r" (a3), "r" (a4)) static void xor_arm4regs_2(unsigned long bytes, unsigned long * __restrict p1, const unsigned long * __restrict p2) { unsigned int lines = bytes / sizeof(unsigned long) / 4; register unsigned int a1 __asm__("r4"); register unsigned int a2 __asm__("r5"); register unsigned int a3 __asm__("r6"); register unsigned int a4 __asm__("r10"); register unsigned int b1 __asm__("r8"); register unsigned int b2 __asm__("r9"); register unsigned int b3 __asm__("ip"); register unsigned int b4 __asm__("lr"); do { GET_BLOCK_4(p1); XOR_BLOCK_4(p2); PUT_BLOCK_4(p1); } while (--lines); } static void xor_arm4regs_3(unsigned long bytes, unsigned long * __restrict p1, const unsigned long * __restrict p2, const unsigned long * __restrict p3) { unsigned int lines = bytes / sizeof(unsigned long) / 4; register unsigned int a1 __asm__("r4"); register unsigned int a2 __asm__("r5"); register unsigned int a3 __asm__("r6"); register unsigned int a4 __asm__("r10"); register unsigned int b1 __asm__("r8"); register unsigned int b2 __asm__("r9"); register unsigned int b3 __asm__("ip"); register unsigned int b4 __asm__("lr"); do { GET_BLOCK_4(p1); XOR_BLOCK_4(p2); XOR_BLOCK_4(p3); PUT_BLOCK_4(p1); } while (--lines); } static void xor_arm4regs_4(unsigned long bytes, unsigned long * __restrict p1, const unsigned long * __restrict p2, const unsigned long * __restrict p3, const unsigned long * __restrict p4) { unsigned int lines = bytes / sizeof(unsigned long) / 2; register unsigned int a1 __asm__("r8"); register unsigned int a2 __asm__("r9"); register unsigned int b1 __asm__("ip"); register unsigned int b2 __asm__("lr"); do { GET_BLOCK_2(p1); XOR_BLOCK_2(p2); XOR_BLOCK_2(p3); XOR_BLOCK_2(p4); PUT_BLOCK_2(p1); } while (--lines); } static void xor_arm4regs_5(unsigned long bytes, unsigned long * __restrict p1, const unsigned long * __restrict p2, const unsigned long * __restrict p3, const unsigned long * __restrict p4, const unsigned long * __restrict p5) { unsigned int lines = bytes / sizeof(unsigned long) / 2; register unsigned int a1 __asm__("r8"); register unsigned int a2 __asm__("r9"); register unsigned int b1 __asm__("ip"); register unsigned int b2 __asm__("lr"); do { GET_BLOCK_2(p1); XOR_BLOCK_2(p2); XOR_BLOCK_2(p3); XOR_BLOCK_2(p4); XOR_BLOCK_2(p5); PUT_BLOCK_2(p1); } while (--lines); } static struct xor_block_template xor_block_arm4regs = { .name = "arm4regs", .do_2 = xor_arm4regs_2, .do_3 = xor_arm4regs_3, .do_4 = xor_arm4regs_4, .do_5 = xor_arm4regs_5, }; #undef XOR_TRY_TEMPLATES #define XOR_TRY_TEMPLATES \ do { \ xor_speed(&xor_block_arm4regs); \ xor_speed(&xor_block_8regs); \ xor_speed(&xor_block_32regs); \ NEON_TEMPLATES; \ } while (0) #ifdef CONFIG_KERNEL_MODE_NEON extern struct xor_block_template const xor_block_neon_inner; static void xor_neon_2(unsigned long bytes, unsigned long * __restrict p1, const unsigned long * __restrict p2) { if (in_interrupt()) { xor_arm4regs_2(bytes, p1, p2); } else { kernel_neon_begin(); xor_block_neon_inner.do_2(bytes, p1, p2); kernel_neon_end(); } } static void xor_neon_3(unsigned long bytes, unsigned long * __restrict p1, const unsigned long * __restrict p2, const unsigned long * __restrict p3) { if (in_interrupt()) { xor_arm4regs_3(bytes, p1, p2, p3); } else { kernel_neon_begin(); xor_block_neon_inner.do_3(bytes, p1, p2, p3); kernel_neon_end(); } } static void xor_neon_4(unsigned long bytes, unsigned long * __restrict p1, const unsigned long * __restrict p2, const unsigned long * __restrict p3, const unsigned long * __restrict p4) { if (in_interrupt()) { xor_arm4regs_4(bytes, p1, p2, p3, p4); } else { kernel_neon_begin(); xor_block_neon_inner.do_4(bytes, p1, p2, p3, p4); kernel_neon_end(); } } static void xor_neon_5(unsigned long bytes, unsigned long * __restrict p1, const unsigned long * __restrict p2, const unsigned long * __restrict p3, const unsigned long * __restrict p4, const unsigned long * __restrict p5) { if (in_interrupt()) { xor_arm4regs_5(bytes, p1, p2, p3, p4, p5); } else { kernel_neon_begin(); xor_block_neon_inner.do_5(bytes, p1, p2, p3, p4, p5); kernel_neon_end(); } } static struct xor_block_template xor_block_neon = { .name = "neon", .do_2 = xor_neon_2, .do_3 = xor_neon_3, .do_4 = xor_neon_4, .do_5 = xor_neon_5 }; #define NEON_TEMPLATES \ do { if (cpu_has_neon()) xor_speed(&xor_block_neon); } while (0) #else #define NEON_TEMPLATES #endif