The motivation is that the update script has at least two deviations (`<...>@GOT`/`<...>@PLT`/ and not hiding pointer arithmetics) from what pretty much all the checklines were generated with, and most of the tests are still not updated, so each time one of the non-up-to-date tests is updated to see the effect of the code change, there is a lot of noise. Instead of having to deal with that each time, let's just deal with everything at once. This has been done via: ``` cd llvm-project/llvm/test/CodeGen/X86 grep -rl "; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py" | xargs -L1 <...>/llvm-project/llvm/utils/update_llc_test_checks.py --llc-binary <...>/llvm-project/build/bin/llc ``` Not all tests were regenerated, however.
106 lines
3.9 KiB
LLVM
106 lines
3.9 KiB
LLVM
; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
|
|
; RUN: llc < %s -mtriple=i686-unknown -mattr=+sse2 | FileCheck %s --check-prefix=X86
|
|
; RUN: llc < %s -mtriple=x86_64-unknown -mattr=+sse2 | FileCheck %s --check-prefix=X64
|
|
|
|
; This test makes sure that the compiler does not crash with an
|
|
; assertion failure when trying to fold a vector shift left
|
|
; by immediate count if the type of the input vector is different
|
|
; to the result type.
|
|
;
|
|
; This happens for example when lowering a shift left of a MVT::v16i8 vector.
|
|
; This is custom lowered into the following sequence:
|
|
; count << 5
|
|
; A = VSHLI(MVT::v8i16, r & (char16)15, 4)
|
|
; B = BITCAST MVT::v16i8, A
|
|
; VSELECT(r, B, count);
|
|
; count += count
|
|
; C = VSHLI(MVT::v8i16, r & (char16)63, 2)
|
|
; D = BITCAST MVT::v16i8, C
|
|
; r = VSELECT(r, C, count);
|
|
; count += count
|
|
; VSELECT(r, r+r, count);
|
|
; count = count << 5;
|
|
;
|
|
; Where 'r' is a vector of type MVT::v16i8, and
|
|
; 'count' is the vector shift count.
|
|
|
|
define <16 x i8> @do_not_crash(i8*, i32*, i64*, i32, i64, i8) {
|
|
; X86-LABEL: do_not_crash:
|
|
; X86: # %bb.0: # %entry
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %ecx
|
|
; X86-NEXT: movb %al, (%ecx)
|
|
; X86-NEXT: movd %eax, %xmm0
|
|
; X86-NEXT: psllq $56, %xmm0
|
|
; X86-NEXT: movdqa {{.*#+}} xmm2 = [255,255,255,255,255,255,255,0,255,255,255,255,255,255,255,255]
|
|
; X86-NEXT: movdqa %xmm2, %xmm1
|
|
; X86-NEXT: pandn %xmm0, %xmm1
|
|
; X86-NEXT: por %xmm2, %xmm1
|
|
; X86-NEXT: pcmpeqd %xmm2, %xmm2
|
|
; X86-NEXT: psllw $5, %xmm1
|
|
; X86-NEXT: pxor %xmm3, %xmm3
|
|
; X86-NEXT: pxor %xmm0, %xmm0
|
|
; X86-NEXT: pcmpgtb %xmm1, %xmm0
|
|
; X86-NEXT: pxor %xmm0, %xmm2
|
|
; X86-NEXT: pand {{\.?LCPI[0-9]+_[0-9]+}}, %xmm0
|
|
; X86-NEXT: por %xmm2, %xmm0
|
|
; X86-NEXT: paddb %xmm1, %xmm1
|
|
; X86-NEXT: pxor %xmm2, %xmm2
|
|
; X86-NEXT: pcmpgtb %xmm1, %xmm2
|
|
; X86-NEXT: movdqa %xmm2, %xmm4
|
|
; X86-NEXT: pandn %xmm0, %xmm4
|
|
; X86-NEXT: psllw $2, %xmm0
|
|
; X86-NEXT: pand %xmm2, %xmm0
|
|
; X86-NEXT: pand {{\.?LCPI[0-9]+_[0-9]+}}, %xmm0
|
|
; X86-NEXT: por %xmm4, %xmm0
|
|
; X86-NEXT: paddb %xmm1, %xmm1
|
|
; X86-NEXT: pcmpgtb %xmm1, %xmm3
|
|
; X86-NEXT: movdqa %xmm3, %xmm1
|
|
; X86-NEXT: pandn %xmm0, %xmm1
|
|
; X86-NEXT: paddb %xmm0, %xmm0
|
|
; X86-NEXT: pand %xmm3, %xmm0
|
|
; X86-NEXT: por %xmm1, %xmm0
|
|
; X86-NEXT: retl
|
|
;
|
|
; X64-LABEL: do_not_crash:
|
|
; X64: # %bb.0: # %entry
|
|
; X64-NEXT: movb %r9b, (%rdi)
|
|
; X64-NEXT: movd %r9d, %xmm0
|
|
; X64-NEXT: psllq $56, %xmm0
|
|
; X64-NEXT: movdqa {{.*#+}} xmm2 = [255,255,255,255,255,255,255,0,255,255,255,255,255,255,255,255]
|
|
; X64-NEXT: movdqa %xmm2, %xmm1
|
|
; X64-NEXT: pandn %xmm0, %xmm1
|
|
; X64-NEXT: por %xmm2, %xmm1
|
|
; X64-NEXT: pcmpeqd %xmm2, %xmm2
|
|
; X64-NEXT: psllw $5, %xmm1
|
|
; X64-NEXT: pxor %xmm3, %xmm3
|
|
; X64-NEXT: pxor %xmm0, %xmm0
|
|
; X64-NEXT: pcmpgtb %xmm1, %xmm0
|
|
; X64-NEXT: pxor %xmm0, %xmm2
|
|
; X64-NEXT: pand {{\.?LCPI[0-9]+_[0-9]+}}(%rip), %xmm0
|
|
; X64-NEXT: por %xmm2, %xmm0
|
|
; X64-NEXT: paddb %xmm1, %xmm1
|
|
; X64-NEXT: pxor %xmm2, %xmm2
|
|
; X64-NEXT: pcmpgtb %xmm1, %xmm2
|
|
; X64-NEXT: movdqa %xmm2, %xmm4
|
|
; X64-NEXT: pandn %xmm0, %xmm4
|
|
; X64-NEXT: psllw $2, %xmm0
|
|
; X64-NEXT: pand %xmm2, %xmm0
|
|
; X64-NEXT: pand {{\.?LCPI[0-9]+_[0-9]+}}(%rip), %xmm0
|
|
; X64-NEXT: por %xmm4, %xmm0
|
|
; X64-NEXT: paddb %xmm1, %xmm1
|
|
; X64-NEXT: pcmpgtb %xmm1, %xmm3
|
|
; X64-NEXT: movdqa %xmm3, %xmm1
|
|
; X64-NEXT: pandn %xmm0, %xmm1
|
|
; X64-NEXT: paddb %xmm0, %xmm0
|
|
; X64-NEXT: pand %xmm3, %xmm0
|
|
; X64-NEXT: por %xmm1, %xmm0
|
|
; X64-NEXT: retq
|
|
entry:
|
|
store i8 %5, i8* %0
|
|
%L5 = load i8, i8* %0
|
|
%I8 = insertelement <16 x i8> <i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1>, i8 %L5, i32 7
|
|
%B51 = shl <16 x i8> <i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1>, %I8
|
|
ret <16 x i8> %B51
|
|
}
|