This adds the Arm Optimized Routines (see https://github.com/ARM-software/optimized-routines) source code under the the LLVM license. The version of the code provided in this patch is v20.02 of the Arm Optimized Routines project. This entire contribution is being committed as is even though it does not currently fit the LLVM libc model and does not follow the LLVM coding style. In the near future, implementations from this patch will be moved over to their right place in the LLVM-libc tree. This will be done over many small patches, all of which will go through the normal LLVM code review process. See this libc-dev post for the plan: http://lists.llvm.org/pipermail/libc-dev/2020-March/000044.html Differential revision of the original upload: https://reviews.llvm.org/D75355
135 lines
2.8 KiB
ArmAsm
135 lines
2.8 KiB
ArmAsm
/* memcmp - compare memory
|
|
*
|
|
* Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
|
|
* See https://llvm.org/LICENSE.txt for license information.
|
|
* SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
|
|
*/
|
|
|
|
/* Assumptions:
|
|
*
|
|
* ARMv8-a, AArch64, unaligned accesses.
|
|
*/
|
|
|
|
#include "../asmdefs.h"
|
|
|
|
/* Parameters and result. */
|
|
#define src1 x0
|
|
#define src2 x1
|
|
#define limit x2
|
|
#define result w0
|
|
|
|
/* Internal variables. */
|
|
#define data1 x3
|
|
#define data1w w3
|
|
#define data1h x4
|
|
#define data2 x5
|
|
#define data2w w5
|
|
#define data2h x6
|
|
#define tmp1 x7
|
|
#define tmp2 x8
|
|
|
|
ENTRY (__memcmp_aarch64)
|
|
subs limit, limit, 8
|
|
b.lo L(less8)
|
|
|
|
ldr data1, [src1], 8
|
|
ldr data2, [src2], 8
|
|
cmp data1, data2
|
|
b.ne L(return)
|
|
|
|
subs limit, limit, 8
|
|
b.gt L(more16)
|
|
|
|
ldr data1, [src1, limit]
|
|
ldr data2, [src2, limit]
|
|
b L(return)
|
|
|
|
L(more16):
|
|
ldr data1, [src1], 8
|
|
ldr data2, [src2], 8
|
|
cmp data1, data2
|
|
bne L(return)
|
|
|
|
/* Jump directly to comparing the last 16 bytes for 32 byte (or less)
|
|
strings. */
|
|
subs limit, limit, 16
|
|
b.ls L(last_bytes)
|
|
|
|
/* We overlap loads between 0-32 bytes at either side of SRC1 when we
|
|
try to align, so limit it only to strings larger than 128 bytes. */
|
|
cmp limit, 96
|
|
b.ls L(loop16)
|
|
|
|
/* Align src1 and adjust src2 with bytes not yet done. */
|
|
and tmp1, src1, 15
|
|
add limit, limit, tmp1
|
|
sub src1, src1, tmp1
|
|
sub src2, src2, tmp1
|
|
|
|
/* Loop performing 16 bytes per iteration using aligned src1.
|
|
Limit is pre-decremented by 16 and must be larger than zero.
|
|
Exit if <= 16 bytes left to do or if the data is not equal. */
|
|
.p2align 4
|
|
L(loop16):
|
|
ldp data1, data1h, [src1], 16
|
|
ldp data2, data2h, [src2], 16
|
|
subs limit, limit, 16
|
|
ccmp data1, data2, 0, hi
|
|
ccmp data1h, data2h, 0, eq
|
|
b.eq L(loop16)
|
|
|
|
cmp data1, data2
|
|
bne L(return)
|
|
mov data1, data1h
|
|
mov data2, data2h
|
|
cmp data1, data2
|
|
bne L(return)
|
|
|
|
/* Compare last 1-16 bytes using unaligned access. */
|
|
L(last_bytes):
|
|
add src1, src1, limit
|
|
add src2, src2, limit
|
|
ldp data1, data1h, [src1]
|
|
ldp data2, data2h, [src2]
|
|
cmp data1, data2
|
|
bne L(return)
|
|
mov data1, data1h
|
|
mov data2, data2h
|
|
cmp data1, data2
|
|
|
|
/* Compare data bytes and set return value to 0, -1 or 1. */
|
|
L(return):
|
|
#ifndef __AARCH64EB__
|
|
rev data1, data1
|
|
rev data2, data2
|
|
#endif
|
|
cmp data1, data2
|
|
L(ret_eq):
|
|
cset result, ne
|
|
cneg result, result, lo
|
|
ret
|
|
|
|
.p2align 4
|
|
/* Compare up to 8 bytes. Limit is [-8..-1]. */
|
|
L(less8):
|
|
adds limit, limit, 4
|
|
b.lo L(less4)
|
|
ldr data1w, [src1], 4
|
|
ldr data2w, [src2], 4
|
|
cmp data1w, data2w
|
|
b.ne L(return)
|
|
sub limit, limit, 4
|
|
L(less4):
|
|
adds limit, limit, 4
|
|
beq L(ret_eq)
|
|
L(byte_loop):
|
|
ldrb data1w, [src1], 1
|
|
ldrb data2w, [src2], 1
|
|
subs limit, limit, 1
|
|
ccmp data1w, data2w, 0, ne /* NZCV = 0b0000. */
|
|
b.eq L(byte_loop)
|
|
sub result, data1w, data2w
|
|
ret
|
|
|
|
END (__memcmp_aarch64)
|