230 lines
6.1 KiB
LLVM
230 lines
6.1 KiB
LLVM
; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
|
|
; RUN: llc -mtriple=riscv32 -verify-machineinstrs < %s | FileCheck -check-prefix=RV32I %s
|
|
; RUN: llc -mtriple=riscv64 -verify-machineinstrs < %s | FileCheck -check-prefix=RV64I %s
|
|
|
|
;
|
|
; fixed avg(x,y) = add(and(x,y),ashr(xor(x,y),1))
|
|
;
|
|
; ext avg(x,y) = trunc(ashr(add(sext(x),sext(y)),1))
|
|
;
|
|
|
|
define i8 @test_fixed_i8(i8 %a0, i8 %a1) nounwind {
|
|
; RV32I-LABEL: test_fixed_i8:
|
|
; RV32I: # %bb.0:
|
|
; RV32I-NEXT: slli a1, a1, 24
|
|
; RV32I-NEXT: srai a1, a1, 24
|
|
; RV32I-NEXT: slli a0, a0, 24
|
|
; RV32I-NEXT: srai a0, a0, 24
|
|
; RV32I-NEXT: add a0, a0, a1
|
|
; RV32I-NEXT: srai a0, a0, 1
|
|
; RV32I-NEXT: ret
|
|
;
|
|
; RV64I-LABEL: test_fixed_i8:
|
|
; RV64I: # %bb.0:
|
|
; RV64I-NEXT: slli a1, a1, 56
|
|
; RV64I-NEXT: srai a1, a1, 56
|
|
; RV64I-NEXT: slli a0, a0, 56
|
|
; RV64I-NEXT: srai a0, a0, 56
|
|
; RV64I-NEXT: add a0, a0, a1
|
|
; RV64I-NEXT: srai a0, a0, 1
|
|
; RV64I-NEXT: ret
|
|
%and = and i8 %a0, %a1
|
|
%xor = xor i8 %a0, %a1
|
|
%shift = ashr i8 %xor, 1
|
|
%res = add i8 %and, %shift
|
|
ret i8 %res
|
|
}
|
|
|
|
define i8 @test_ext_i8(i8 %a0, i8 %a1) nounwind {
|
|
; RV32I-LABEL: test_ext_i8:
|
|
; RV32I: # %bb.0:
|
|
; RV32I-NEXT: slli a1, a1, 24
|
|
; RV32I-NEXT: srai a1, a1, 24
|
|
; RV32I-NEXT: slli a0, a0, 24
|
|
; RV32I-NEXT: srai a0, a0, 24
|
|
; RV32I-NEXT: add a0, a0, a1
|
|
; RV32I-NEXT: srai a0, a0, 1
|
|
; RV32I-NEXT: ret
|
|
;
|
|
; RV64I-LABEL: test_ext_i8:
|
|
; RV64I: # %bb.0:
|
|
; RV64I-NEXT: slli a1, a1, 56
|
|
; RV64I-NEXT: srai a1, a1, 56
|
|
; RV64I-NEXT: slli a0, a0, 56
|
|
; RV64I-NEXT: srai a0, a0, 56
|
|
; RV64I-NEXT: add a0, a0, a1
|
|
; RV64I-NEXT: srai a0, a0, 1
|
|
; RV64I-NEXT: ret
|
|
%x0 = sext i8 %a0 to i16
|
|
%x1 = sext i8 %a1 to i16
|
|
%sum = add i16 %x0, %x1
|
|
%shift = ashr i16 %sum, 1
|
|
%res = trunc i16 %shift to i8
|
|
ret i8 %res
|
|
}
|
|
|
|
define i16 @test_fixed_i16(i16 %a0, i16 %a1) nounwind {
|
|
; RV32I-LABEL: test_fixed_i16:
|
|
; RV32I: # %bb.0:
|
|
; RV32I-NEXT: slli a1, a1, 16
|
|
; RV32I-NEXT: srai a1, a1, 16
|
|
; RV32I-NEXT: slli a0, a0, 16
|
|
; RV32I-NEXT: srai a0, a0, 16
|
|
; RV32I-NEXT: add a0, a0, a1
|
|
; RV32I-NEXT: srai a0, a0, 1
|
|
; RV32I-NEXT: ret
|
|
;
|
|
; RV64I-LABEL: test_fixed_i16:
|
|
; RV64I: # %bb.0:
|
|
; RV64I-NEXT: slli a1, a1, 48
|
|
; RV64I-NEXT: srai a1, a1, 48
|
|
; RV64I-NEXT: slli a0, a0, 48
|
|
; RV64I-NEXT: srai a0, a0, 48
|
|
; RV64I-NEXT: add a0, a0, a1
|
|
; RV64I-NEXT: srai a0, a0, 1
|
|
; RV64I-NEXT: ret
|
|
%and = and i16 %a0, %a1
|
|
%xor = xor i16 %a0, %a1
|
|
%shift = ashr i16 %xor, 1
|
|
%res = add i16 %and, %shift
|
|
ret i16 %res
|
|
}
|
|
|
|
define i16 @test_ext_i16(i16 %a0, i16 %a1) nounwind {
|
|
; RV32I-LABEL: test_ext_i16:
|
|
; RV32I: # %bb.0:
|
|
; RV32I-NEXT: slli a1, a1, 16
|
|
; RV32I-NEXT: srai a1, a1, 16
|
|
; RV32I-NEXT: slli a0, a0, 16
|
|
; RV32I-NEXT: srai a0, a0, 16
|
|
; RV32I-NEXT: add a0, a0, a1
|
|
; RV32I-NEXT: srai a0, a0, 1
|
|
; RV32I-NEXT: ret
|
|
;
|
|
; RV64I-LABEL: test_ext_i16:
|
|
; RV64I: # %bb.0:
|
|
; RV64I-NEXT: slli a1, a1, 48
|
|
; RV64I-NEXT: srai a1, a1, 48
|
|
; RV64I-NEXT: slli a0, a0, 48
|
|
; RV64I-NEXT: srai a0, a0, 48
|
|
; RV64I-NEXT: add a0, a0, a1
|
|
; RV64I-NEXT: srai a0, a0, 1
|
|
; RV64I-NEXT: ret
|
|
%x0 = sext i16 %a0 to i32
|
|
%x1 = sext i16 %a1 to i32
|
|
%sum = add i32 %x0, %x1
|
|
%shift = ashr i32 %sum, 1
|
|
%res = trunc i32 %shift to i16
|
|
ret i16 %res
|
|
}
|
|
|
|
define i32 @test_fixed_i32(i32 %a0, i32 %a1) nounwind {
|
|
; RV32I-LABEL: test_fixed_i32:
|
|
; RV32I: # %bb.0:
|
|
; RV32I-NEXT: and a2, a0, a1
|
|
; RV32I-NEXT: xor a0, a0, a1
|
|
; RV32I-NEXT: srai a0, a0, 1
|
|
; RV32I-NEXT: add a0, a2, a0
|
|
; RV32I-NEXT: ret
|
|
;
|
|
; RV64I-LABEL: test_fixed_i32:
|
|
; RV64I: # %bb.0:
|
|
; RV64I-NEXT: sext.w a1, a1
|
|
; RV64I-NEXT: sext.w a0, a0
|
|
; RV64I-NEXT: add a0, a0, a1
|
|
; RV64I-NEXT: srai a0, a0, 1
|
|
; RV64I-NEXT: ret
|
|
%and = and i32 %a0, %a1
|
|
%xor = xor i32 %a1, %a0
|
|
%shift = ashr i32 %xor, 1
|
|
%res = add i32 %and, %shift
|
|
ret i32 %res
|
|
}
|
|
|
|
define i32 @test_ext_i32(i32 %a0, i32 %a1) nounwind {
|
|
; RV32I-LABEL: test_ext_i32:
|
|
; RV32I: # %bb.0:
|
|
; RV32I-NEXT: and a2, a0, a1
|
|
; RV32I-NEXT: xor a0, a0, a1
|
|
; RV32I-NEXT: srai a0, a0, 1
|
|
; RV32I-NEXT: add a0, a2, a0
|
|
; RV32I-NEXT: ret
|
|
;
|
|
; RV64I-LABEL: test_ext_i32:
|
|
; RV64I: # %bb.0:
|
|
; RV64I-NEXT: sext.w a1, a1
|
|
; RV64I-NEXT: sext.w a0, a0
|
|
; RV64I-NEXT: add a0, a0, a1
|
|
; RV64I-NEXT: srai a0, a0, 1
|
|
; RV64I-NEXT: ret
|
|
%x0 = sext i32 %a0 to i64
|
|
%x1 = sext i32 %a1 to i64
|
|
%sum = add i64 %x0, %x1
|
|
%shift = ashr i64 %sum, 1
|
|
%res = trunc i64 %shift to i32
|
|
ret i32 %res
|
|
}
|
|
|
|
define i64 @test_fixed_i64(i64 %a0, i64 %a1) nounwind {
|
|
; RV32I-LABEL: test_fixed_i64:
|
|
; RV32I: # %bb.0:
|
|
; RV32I-NEXT: and a4, a1, a3
|
|
; RV32I-NEXT: xor a1, a1, a3
|
|
; RV32I-NEXT: srai a3, a1, 1
|
|
; RV32I-NEXT: add a3, a4, a3
|
|
; RV32I-NEXT: slli a1, a1, 31
|
|
; RV32I-NEXT: xor a4, a0, a2
|
|
; RV32I-NEXT: srli a4, a4, 1
|
|
; RV32I-NEXT: or a1, a4, a1
|
|
; RV32I-NEXT: and a2, a0, a2
|
|
; RV32I-NEXT: add a0, a2, a1
|
|
; RV32I-NEXT: sltu a1, a0, a2
|
|
; RV32I-NEXT: add a1, a3, a1
|
|
; RV32I-NEXT: ret
|
|
;
|
|
; RV64I-LABEL: test_fixed_i64:
|
|
; RV64I: # %bb.0:
|
|
; RV64I-NEXT: and a2, a0, a1
|
|
; RV64I-NEXT: xor a0, a0, a1
|
|
; RV64I-NEXT: srai a0, a0, 1
|
|
; RV64I-NEXT: add a0, a2, a0
|
|
; RV64I-NEXT: ret
|
|
%and = and i64 %a0, %a1
|
|
%xor = xor i64 %a1, %a0
|
|
%shift = ashr i64 %xor, 1
|
|
%res = add i64 %and, %shift
|
|
ret i64 %res
|
|
}
|
|
|
|
define i64 @test_ext_i64(i64 %a0, i64 %a1) nounwind {
|
|
; RV32I-LABEL: test_ext_i64:
|
|
; RV32I: # %bb.0:
|
|
; RV32I-NEXT: and a4, a1, a3
|
|
; RV32I-NEXT: xor a1, a1, a3
|
|
; RV32I-NEXT: srai a3, a1, 1
|
|
; RV32I-NEXT: add a3, a4, a3
|
|
; RV32I-NEXT: slli a1, a1, 31
|
|
; RV32I-NEXT: xor a4, a0, a2
|
|
; RV32I-NEXT: srli a4, a4, 1
|
|
; RV32I-NEXT: or a1, a4, a1
|
|
; RV32I-NEXT: and a2, a0, a2
|
|
; RV32I-NEXT: add a0, a2, a1
|
|
; RV32I-NEXT: sltu a1, a0, a2
|
|
; RV32I-NEXT: add a1, a3, a1
|
|
; RV32I-NEXT: ret
|
|
;
|
|
; RV64I-LABEL: test_ext_i64:
|
|
; RV64I: # %bb.0:
|
|
; RV64I-NEXT: and a2, a0, a1
|
|
; RV64I-NEXT: xor a0, a0, a1
|
|
; RV64I-NEXT: srai a0, a0, 1
|
|
; RV64I-NEXT: add a0, a2, a0
|
|
; RV64I-NEXT: ret
|
|
%x0 = sext i64 %a0 to i128
|
|
%x1 = sext i64 %a1 to i128
|
|
%sum = add i128 %x0, %x1
|
|
%shift = ashr i128 %sum, 1
|
|
%res = trunc i128 %shift to i64
|
|
ret i64 %res
|
|
}
|