mirror of
https://mirrors.bfsu.edu.cn/git/linux.git
synced 2024-11-25 21:24:08 +08:00
0f61f6be1f
Now that we have SYM_FUNC_ALIAS() and SYM_FUNC_ALIAS_WEAK(), use those to simplify and more consistently define function aliases across arch/arm64. Aliases are now defined in terms of a canonical function name. For position-independent functions I've made the __pi_<func> name the canonical name, and defined other alises in terms of this. The SYM_FUNC_{START,END}_PI(func) macros obscure the __pi_<func> name, and make this hard to seatch for. The SYM_FUNC_START_WEAK_PI() macro also obscures the fact that the __pi_<func> fymbol is global and the <func> symbol is weak. For clarity, I have removed these macros and used SYM_FUNC_{START,END}() directly with the __pi_<func> name. For example: SYM_FUNC_START_WEAK_PI(func) ... asm insns ... SYM_FUNC_END_PI(func) EXPORT_SYMBOL(func) ... becomes: SYM_FUNC_START(__pi_func) ... asm insns ... SYM_FUNC_END(__pi_func) SYM_FUNC_ALIAS_WEAK(func, __pi_func) EXPORT_SYMBOL(func) For clarity, where there are multiple annotations such as EXPORT_SYMBOL(), I've tried to keep annotations grouped by symbol. For example, where a function has a name and an alias which are both exported, this is organised as: SYM_FUNC_START(func) ... asm insns ... SYM_FUNC_END(func) EXPORT_SYMBOL(func) SYM_FUNC_ALIAS(alias, func) EXPORT_SYMBOL(alias) For consistency with the other string functions, I've defined strrchr as a position-independent function, as it can safely be used as such even though we have no users today. As we no longer use SYM_FUNC_{START,END}_ALIAS(), our local copies are removed. The common versions will be removed by a subsequent patch. There should be no functional change as a result of this patch. Signed-off-by: Mark Rutland <mark.rutland@arm.com> Acked-by: Ard Biesheuvel <ardb@kernel.org> Acked-by: Catalin Marinas <catalin.marinas@arm.com> Acked-by: Josh Poimboeuf <jpoimboe@redhat.com> Acked-by: Mark Brown <broonie@kernel.org> Cc: Joey Gouly <joey.gouly@arm.com> Cc: Will Deacon <will@kernel.org> Acked-by: Peter Zijlstra (Intel) <peterz@infradead.org> Link: https://lore.kernel.org/r/20220216162229.1076788-3-mark.rutland@arm.com Signed-off-by: Will Deacon <will@kernel.org>
140 lines
2.9 KiB
ArmAsm
140 lines
2.9 KiB
ArmAsm
/* SPDX-License-Identifier: GPL-2.0-only */
|
|
/*
|
|
* Copyright (c) 2013-2021, Arm Limited.
|
|
*
|
|
* Adapted from the original at:
|
|
* https://github.com/ARM-software/optimized-routines/blob/e823e3abf5f89ecb/string/aarch64/memcmp.S
|
|
*/
|
|
|
|
#include <linux/linkage.h>
|
|
#include <asm/assembler.h>
|
|
|
|
/* Assumptions:
|
|
*
|
|
* ARMv8-a, AArch64, unaligned accesses.
|
|
*/
|
|
|
|
#define L(label) .L ## label
|
|
|
|
/* Parameters and result. */
|
|
#define src1 x0
|
|
#define src2 x1
|
|
#define limit x2
|
|
#define result w0
|
|
|
|
/* Internal variables. */
|
|
#define data1 x3
|
|
#define data1w w3
|
|
#define data1h x4
|
|
#define data2 x5
|
|
#define data2w w5
|
|
#define data2h x6
|
|
#define tmp1 x7
|
|
#define tmp2 x8
|
|
|
|
SYM_FUNC_START(__pi_memcmp)
|
|
subs limit, limit, 8
|
|
b.lo L(less8)
|
|
|
|
ldr data1, [src1], 8
|
|
ldr data2, [src2], 8
|
|
cmp data1, data2
|
|
b.ne L(return)
|
|
|
|
subs limit, limit, 8
|
|
b.gt L(more16)
|
|
|
|
ldr data1, [src1, limit]
|
|
ldr data2, [src2, limit]
|
|
b L(return)
|
|
|
|
L(more16):
|
|
ldr data1, [src1], 8
|
|
ldr data2, [src2], 8
|
|
cmp data1, data2
|
|
bne L(return)
|
|
|
|
/* Jump directly to comparing the last 16 bytes for 32 byte (or less)
|
|
strings. */
|
|
subs limit, limit, 16
|
|
b.ls L(last_bytes)
|
|
|
|
/* We overlap loads between 0-32 bytes at either side of SRC1 when we
|
|
try to align, so limit it only to strings larger than 128 bytes. */
|
|
cmp limit, 96
|
|
b.ls L(loop16)
|
|
|
|
/* Align src1 and adjust src2 with bytes not yet done. */
|
|
and tmp1, src1, 15
|
|
add limit, limit, tmp1
|
|
sub src1, src1, tmp1
|
|
sub src2, src2, tmp1
|
|
|
|
/* Loop performing 16 bytes per iteration using aligned src1.
|
|
Limit is pre-decremented by 16 and must be larger than zero.
|
|
Exit if <= 16 bytes left to do or if the data is not equal. */
|
|
.p2align 4
|
|
L(loop16):
|
|
ldp data1, data1h, [src1], 16
|
|
ldp data2, data2h, [src2], 16
|
|
subs limit, limit, 16
|
|
ccmp data1, data2, 0, hi
|
|
ccmp data1h, data2h, 0, eq
|
|
b.eq L(loop16)
|
|
|
|
cmp data1, data2
|
|
bne L(return)
|
|
mov data1, data1h
|
|
mov data2, data2h
|
|
cmp data1, data2
|
|
bne L(return)
|
|
|
|
/* Compare last 1-16 bytes using unaligned access. */
|
|
L(last_bytes):
|
|
add src1, src1, limit
|
|
add src2, src2, limit
|
|
ldp data1, data1h, [src1]
|
|
ldp data2, data2h, [src2]
|
|
cmp data1, data2
|
|
bne L(return)
|
|
mov data1, data1h
|
|
mov data2, data2h
|
|
cmp data1, data2
|
|
|
|
/* Compare data bytes and set return value to 0, -1 or 1. */
|
|
L(return):
|
|
#ifndef __AARCH64EB__
|
|
rev data1, data1
|
|
rev data2, data2
|
|
#endif
|
|
cmp data1, data2
|
|
L(ret_eq):
|
|
cset result, ne
|
|
cneg result, result, lo
|
|
ret
|
|
|
|
.p2align 4
|
|
/* Compare up to 8 bytes. Limit is [-8..-1]. */
|
|
L(less8):
|
|
adds limit, limit, 4
|
|
b.lo L(less4)
|
|
ldr data1w, [src1], 4
|
|
ldr data2w, [src2], 4
|
|
cmp data1w, data2w
|
|
b.ne L(return)
|
|
sub limit, limit, 4
|
|
L(less4):
|
|
adds limit, limit, 4
|
|
beq L(ret_eq)
|
|
L(byte_loop):
|
|
ldrb data1w, [src1], 1
|
|
ldrb data2w, [src2], 1
|
|
subs limit, limit, 1
|
|
ccmp data1w, data2w, 0, ne /* NZCV = 0b0000. */
|
|
b.eq L(byte_loop)
|
|
sub result, data1w, data2w
|
|
ret
|
|
SYM_FUNC_END(__pi_memcmp)
|
|
SYM_FUNC_ALIAS_WEAK(memcmp, __pi_memcmp)
|
|
EXPORT_SYMBOL_NOKASAN(memcmp)
|