Files
clang-p2996/llvm/test/CodeGen/LoongArch/unaligned-access.ll
Weining Lu 47601815ec [LoongArch] Define ual feature and override allowsMisalignedMemoryAccesses
Some CPUs do not allow memory accesses to be unaligned, e.g. 2k1000la
who uses the la264 core on which misaligned access will trigger an
exception.

In this patch, a backend feature called `ual` is defined to decribe
whether the CPU supports unaligned memroy accesses. And this feature
can be toggled by clang options `-m[no-]unaligned-access` or the
aliases `-m[no-]strict-align`. When this feature is on,
`allowsMisalignedMemoryAccesses` sets the speed number to 1 and returns
true that allows the codegen to generate unaligned memory access insns.

Clang options `-m[no-]unaligned-access` are moved from `m_arm_Features_Group`
to `m_Group` because now more than one targets use them. And a test
is added to show that they remain unused on a target that does not
support them. In addition, to keep compatible with gcc, a new alias
`-mno-strict-align` is added which is equal to `-munaligned-access`.

The feature name `ual` is consistent with linux kernel [1] and the
output of `lscpu` or `/proc/cpuinfo` [2].

There is an `LLT` variant of `allowsMisalignedMemoryAccesses`, but
seems that curently it is only used in GlobalISel which LoongArch
doesn't support yet. So this variant is not implemented in this patch.

[1]: https://github.com/torvalds/linux/blob/master/arch/loongarch/include/asm/cpu.h#L77
[2]: https://github.com/torvalds/linux/blob/master/arch/loongarch/kernel/proc.c#L75

Reviewed By: xen0n

Differential Revision: https://reviews.llvm.org/D149946
2023-06-07 13:40:58 +08:00

73 lines
2.4 KiB
LLVM

; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py UTC_ARGS: --version 2
;; Test the ual feature which is similar to AArch64/arm64-strict-align.ll.
; RUN: llc --mtriple=loongarch32 < %s | FileCheck %s --check-prefix=LA32-ALIGNED
; RUN: llc --mtriple=loongarch32 --mattr=+ual < %s | FileCheck %s --check-prefix=LA32-UNALIGNED
; RUN: llc --mtriple=loongarch32 --mattr=-ual < %s | FileCheck %s --check-prefix=LA32-ALIGNED
; RUN: llc --mtriple=loongarch64 < %s | FileCheck %s --check-prefix=LA64-UNALIGNED
; RUN: llc --mtriple=loongarch64 --mattr=+ual < %s | FileCheck %s --check-prefix=LA64-UNALIGNED
; RUN: llc --mtriple=loongarch64 --mattr=-ual < %s | FileCheck %s --check-prefix=LA64-ALIGNED
define i32 @f0(ptr %p) nounwind {
; LA32-ALIGNED-LABEL: f0:
; LA32-ALIGNED: # %bb.0:
; LA32-ALIGNED-NEXT: ld.hu $a1, $a0, 0
; LA32-ALIGNED-NEXT: ld.hu $a0, $a0, 2
; LA32-ALIGNED-NEXT: slli.w $a0, $a0, 16
; LA32-ALIGNED-NEXT: or $a0, $a0, $a1
; LA32-ALIGNED-NEXT: ret
;
; LA32-UNALIGNED-LABEL: f0:
; LA32-UNALIGNED: # %bb.0:
; LA32-UNALIGNED-NEXT: ld.w $a0, $a0, 0
; LA32-UNALIGNED-NEXT: ret
;
; LA64-UNALIGNED-LABEL: f0:
; LA64-UNALIGNED: # %bb.0:
; LA64-UNALIGNED-NEXT: ld.w $a0, $a0, 0
; LA64-UNALIGNED-NEXT: ret
;
; LA64-ALIGNED-LABEL: f0:
; LA64-ALIGNED: # %bb.0:
; LA64-ALIGNED-NEXT: ld.hu $a1, $a0, 0
; LA64-ALIGNED-NEXT: ld.h $a0, $a0, 2
; LA64-ALIGNED-NEXT: slli.d $a0, $a0, 16
; LA64-ALIGNED-NEXT: or $a0, $a0, $a1
; LA64-ALIGNED-NEXT: ret
%tmp = load i32, ptr %p, align 2
ret i32 %tmp
}
define i64 @f1(ptr %p) nounwind {
; LA32-ALIGNED-LABEL: f1:
; LA32-ALIGNED: # %bb.0:
; LA32-ALIGNED-NEXT: ld.w $a2, $a0, 0
; LA32-ALIGNED-NEXT: ld.w $a1, $a0, 4
; LA32-ALIGNED-NEXT: move $a0, $a2
; LA32-ALIGNED-NEXT: ret
;
; LA32-UNALIGNED-LABEL: f1:
; LA32-UNALIGNED: # %bb.0:
; LA32-UNALIGNED-NEXT: ld.w $a2, $a0, 0
; LA32-UNALIGNED-NEXT: ld.w $a1, $a0, 4
; LA32-UNALIGNED-NEXT: move $a0, $a2
; LA32-UNALIGNED-NEXT: ret
;
; LA64-UNALIGNED-LABEL: f1:
; LA64-UNALIGNED: # %bb.0:
; LA64-UNALIGNED-NEXT: ld.d $a0, $a0, 0
; LA64-UNALIGNED-NEXT: ret
;
; LA64-ALIGNED-LABEL: f1:
; LA64-ALIGNED: # %bb.0:
; LA64-ALIGNED-NEXT: ld.wu $a1, $a0, 0
; LA64-ALIGNED-NEXT: ld.wu $a0, $a0, 4
; LA64-ALIGNED-NEXT: slli.d $a0, $a0, 32
; LA64-ALIGNED-NEXT: or $a0, $a0, $a1
; LA64-ALIGNED-NEXT: ret
%tmp = load i64, ptr %p, align 4
ret i64 %tmp
}