1/* SPDX-License-Identifier: GPL-2.0-or-later */ 2/* 3 * Copyright (C) 2023 WANG Xuerui <git@xen0n.name> 4 */ 5#ifndef _ASM_LOONGARCH_XOR_H 6#define _ASM_LOONGARCH_XOR_H 7 8#include <asm/cpu-features.h> 9#include <asm/xor_simd.h> 10 11#ifdef CONFIG_CPU_HAS_LSX 12static struct xor_block_template xor_block_lsx = { 13 .name = "lsx", 14 .do_2 = xor_lsx_2, 15 .do_3 = xor_lsx_3, 16 .do_4 = xor_lsx_4, 17 .do_5 = xor_lsx_5, 18}; 19 20#define XOR_SPEED_LSX() \ 21 do { \ 22 if (cpu_has_lsx) \ 23 xor_speed(&xor_block_lsx); \ 24 } while (0) 25#else /* CONFIG_CPU_HAS_LSX */ 26#define XOR_SPEED_LSX() 27#endif /* CONFIG_CPU_HAS_LSX */ 28 29#ifdef CONFIG_CPU_HAS_LASX 30static struct xor_block_template xor_block_lasx = { 31 .name = "lasx", 32 .do_2 = xor_lasx_2, 33 .do_3 = xor_lasx_3, 34 .do_4 = xor_lasx_4, 35 .do_5 = xor_lasx_5, 36}; 37 38#define XOR_SPEED_LASX() \ 39 do { \ 40 if (cpu_has_lasx) \ 41 xor_speed(&xor_block_lasx); \ 42 } while (0) 43#else /* CONFIG_CPU_HAS_LASX */ 44#define XOR_SPEED_LASX() 45#endif /* CONFIG_CPU_HAS_LASX */ 46 47/* 48 * For grins, also test the generic routines. 49 * 50 * More importantly: it cannot be ruled out at this point of time, that some 51 * future (maybe reduced) models could run the vector algorithms slower than 52 * the scalar ones, maybe for errata or micro-op reasons. It may be 53 * appropriate to revisit this after one or two more uarch generations. 54 */ 55#include <asm-generic/xor.h> 56 57#undef XOR_TRY_TEMPLATES 58#define XOR_TRY_TEMPLATES \ 59do { \ 60 xor_speed(&xor_block_8regs); \ 61 xor_speed(&xor_block_8regs_p); \ 62 xor_speed(&xor_block_32regs); \ 63 xor_speed(&xor_block_32regs_p); \ 64 XOR_SPEED_LSX(); \ 65 XOR_SPEED_LASX(); \ 66} while (0) 67 68#endif /* _ASM_LOONGARCH_XOR_H */ 69