added a script to update llvm-mc test file (#107246)

Added a script to update the test file generated by llvm-mc binary. The
script accepts .s and .txt for asm and dasm.

For mc test I am targetting there is no function name which can be used
as a key, thus no clear mapping between input and output. The script
assumes the test are always line-by-line and it update the output marker
for each test line-by-line.

---------

Co-authored-by: Alexander Richardson <mail@alexrichardson.me>
This commit is contained in:
Brox Chen
2024-09-23 15:27:41 -04:00
committed by GitHub
parent 6e6d5eae76
commit 2b892b0502
13 changed files with 394 additions and 1 deletions

View File

@@ -52,3 +52,7 @@ if os.path.isfile(llvm_mca_path):
split_file_path = os.path.join(config.llvm_tools_dir, "split-file")
if os.path.isfile(split_file_path):
add_update_script_substition("%update_test_body")
llvm_mc_path = os.path.join(config.llvm_tools_dir, "llvm-mc")
if os.path.isfile(llvm_mc_path):
add_update_script_substition("%update_mc_test_checks")

View File

@@ -0,0 +1,3 @@
// RUN: llvm-mc -triple=amdgcn -show-encoding %s 2>&1 | FileCheck --check-prefixes=CHECK %s
v_bfrev_b32 v5, v1

View File

@@ -0,0 +1,5 @@
; NOTE: Assertions have been autogenerated by utils/update_mc_test_checks.py
// RUN: llvm-mc -triple=amdgcn -show-encoding %s 2>&1 | FileCheck --check-prefixes=CHECK %s
v_bfrev_b32 v5, v1
// CHECK: v_bfrev_b32_e32 v5, v1 ; encoding: [0x01,0x71,0x0a,0x7e]

View File

@@ -0,0 +1,3 @@
// RUN: not llvm-mc -triple=amdgcn -show-encoding %s 2>&1 | FileCheck --check-prefixes=CHECK %s
v_bfrev_b32 v5, v299

View File

@@ -0,0 +1,5 @@
; NOTE: Assertions have been autogenerated by utils/update_mc_test_checks.py
// RUN: not llvm-mc -triple=amdgcn -show-encoding %s 2>&1 | FileCheck --check-prefixes=CHECK %s
v_bfrev_b32 v5, v299
// CHECK: :[[@LINE-1]]:17: error: register index is out of range

View File

@@ -0,0 +1,5 @@
# RUN: llvm-mc -triple=amdgcn -mcpu=gfx1100 -disassemble -show-encoding %s 2>&1 | FileCheck -check-prefixes=CHECK %s
0x00,0x00,0x00,0x7e
0xfd,0xb8,0x0a,0x7f

View File

@@ -0,0 +1,8 @@
; NOTE: Assertions have been autogenerated by utils/update_mc_test_checks.py
# RUN: llvm-mc -triple=amdgcn -mcpu=gfx1100 -disassemble -show-encoding %s 2>&1 | FileCheck -check-prefixes=CHECK %s
0x00,0x00,0x00,0x7e
# CHECK: v_nop ; encoding: [0x00,0x00,0x00,0x7e]
0xfd,0xb8,0x0a,0x7f
# CHECK: :[[@LINE-1]]:1: warning: invalid instruction encoding

View File

@@ -0,0 +1,6 @@
# RUN: llvm-mc -triple=amdgcn -mcpu=tonga -disassemble -show-encoding %s 2>&1 | FileCheck -check-prefixes=CHECK,CHECKA %s
# RUN: llvm-mc -triple=amdgcn -mcpu=gfx1100 -disassemble -show-encoding %s 2>&1 | FileCheck -check-prefixes=CHECK,CHECKB %s
0x00,0x00,0x00,0x7e
0x01,0x71,0x0a,0x7e

View File

@@ -0,0 +1,10 @@
; NOTE: Assertions have been autogenerated by utils/update_mc_test_checks.py
# RUN: llvm-mc -triple=amdgcn -mcpu=tonga -disassemble -show-encoding %s 2>&1 | FileCheck -check-prefixes=CHECK,CHECKA %s
# RUN: llvm-mc -triple=amdgcn -mcpu=gfx1100 -disassemble -show-encoding %s 2>&1 | FileCheck -check-prefixes=CHECK,CHECKB %s
0x00,0x00,0x00,0x7e
# CHECK: v_nop ; encoding: [0x00,0x00,0x00,0x7e]
0x01,0x71,0x0a,0x7e
# CHECKA: v_movrelsd_b32_e32 v5, v1 ; encoding: [0x01,0x71,0x0a,0x7e]
# CHECKB: v_bfrev_b32_e32 v5, v1 ; encoding: [0x01,0x71,0x0a,0x7e]

View File

@@ -0,0 +1,11 @@
# REQUIRES: amdgpu-registered-target
## Check that basic asm/dasm process is correct
# RUN: cp -f %S/Inputs/amdgpu_asm.s %t.s && %update_mc_test_checks %t.s
# RUN: diff -u %S/Inputs/amdgpu_asm.s.expected %t.s
# RUN: cp -f %S/Inputs/amdgpu_asm_err.s %t.s && %update_mc_test_checks %t.s
# RUN: diff -u %S/Inputs/amdgpu_asm_err.s.expected %t.s
# RUN: cp -f %S/Inputs/amdgpu_dasm.txt %t.txt && %update_mc_test_checks %t.txt
# RUN: diff -u %S/Inputs/amdgpu_dasm.txt.expected %t.txt
# RUN: cp -f %S/Inputs/amdgpu_multirun_dasm.txt %t.txt && %update_mc_test_checks %t.txt
# RUN: diff -u %S/Inputs/amdgpu_multirun_dasm.txt.expected %t.txt

View File

@@ -0,0 +1,4 @@
import platform
if platform.system() == "Windows":
config.unsupported = True

View File

@@ -573,7 +573,7 @@ LOOP_PASS_DEBUG_RE = re.compile(
IR_FUNCTION_RE = re.compile(r'^\s*define\s+(?:internal\s+)?[^@]*@"?([\w.$-]+)"?\s*\(')
TRIPLE_IR_RE = re.compile(r'^\s*target\s+triple\s*=\s*"([^"]+)"$')
TRIPLE_ARG_RE = re.compile(r"-mtriple[= ]([^ ]+)")
TRIPLE_ARG_RE = re.compile(r"-m?triple[= ]([^ ]+)")
MARCH_ARG_RE = re.compile(r"-march[= ]([^ ]+)")
DEBUG_ONLY_ARG_RE = re.compile(r"-debug-only[= ]([^ ]+)")

View File

@@ -0,0 +1,329 @@
#!/usr/bin/env python3
"""
A test update script. This script is a utility to update LLVM 'llvm-mc' based test cases with new FileCheck patterns.
"""
from __future__ import print_function
import argparse
import os # Used to advertise this file's name ("autogenerated_note").
from UpdateTestChecks import common
import subprocess
import re
mc_LIKE_TOOLS = [
"llvm-mc",
"not llvm-mc",
]
ERROR_RE = re.compile(r":\d+: (warning|error): .*")
ERROR_CHECK_RE = re.compile(r"# COM: .*")
OUTPUT_SKIPPED_RE = re.compile(r"(.text)")
COMMENT = {"asm": "//", "dasm": "#"}
def invoke_tool(exe, cmd_args, testline, verbose=False):
if isinstance(cmd_args, list):
args = [applySubstitutions(a, substitutions) for a in cmd_args]
else:
args = cmd_args
cmd = 'echo "' + testline + '" | ' + exe + " " + args
if verbose:
print("Command: ", cmd)
out = subprocess.check_output(cmd, shell=True)
# Fix line endings to unix CR style.
return out.decode().replace("\r\n", "\n")
# create tests line-by-line, here we just filter out the check lines and comments
# and treat all others as tests
def isTestLine(input_line, mc_mode):
line = input_line.strip()
# Skip empty and comment lines
if not line or line.startswith(COMMENT[mc_mode]):
return False
# skip any CHECK lines.
elif common.CHECK_RE.match(input_line):
return False
return True
def hasErr(err):
return err and ERROR_RE.search(err) is not None
def getErrString(err):
if not err:
return ""
# take the first match
for line in err.splitlines():
s = ERROR_RE.search(line)
if s:
return s.group(0)
return ""
def getOutputString(out):
if not out:
return ""
output = ""
for line in out.splitlines():
if OUTPUT_SKIPPED_RE.search(line):
continue
if line.strip("\t ") == "":
continue
output += line.lstrip("\t ")
return output
def should_add_line_to_output(input_line, prefix_set, mc_mode):
# special check line
if mc_mode == "dasm" and ERROR_CHECK_RE.search(input_line):
return False
else:
return common.should_add_line_to_output(
input_line, prefix_set, comment_marker=COMMENT[mc_mode]
)
def getStdCheckLine(prefix, output, mc_mode):
o = ""
for line in output.splitlines():
o += COMMENT[mc_mode] + " " + prefix + ": " + line + "\n"
return o
def getErrCheckLine(prefix, output, mc_mode):
return COMMENT[mc_mode] + " " + prefix + ": " + ":[[@LINE-1]]" + output + "\n"
def main():
parser = argparse.ArgumentParser(description=__doc__)
parser.add_argument(
"--llvm-mc-binary",
default=None,
help='The "mc" binary to use to generate the test case',
)
parser.add_argument(
"--tool",
default=None,
help="Treat the given tool name as an mc-like tool for which check lines should be generated",
)
parser.add_argument(
"--default-march",
default=None,
help="Set a default -march for when neither triple nor arch are found in a RUN line",
)
parser.add_argument("tests", nargs="+")
initial_args = common.parse_commandline_args(parser)
script_name = os.path.basename(__file__)
for ti in common.itertests(
initial_args.tests, parser, script_name="utils/" + script_name
):
if ti.path.endswith(".s"):
mc_mode = "asm"
elif ti.path.endswith(".txt"):
mc_mode = "dasm"
else:
common.warn("Expected .s and .txt, Skipping file : ", ti.path)
continue
triple_in_ir = None
for l in ti.input_lines:
m = common.TRIPLE_IR_RE.match(l)
if m:
triple_in_ir = m.groups()[0]
break
run_list = []
for l in ti.run_lines:
if "|" not in l:
common.warn("Skipping unparsable RUN line: " + l)
continue
commands = [cmd.strip() for cmd in l.split("|")]
assert len(commands) >= 2
mc_cmd = " | ".join(commands[:-1])
filecheck_cmd = commands[-1]
mc_tool = mc_cmd.split(" ")[0]
# special handling for negating exit status
if mc_tool == "not":
mc_tool = mc_tool + " " + mc_cmd.split(" ")[1]
triple_in_cmd = None
m = common.TRIPLE_ARG_RE.search(mc_cmd)
if m:
triple_in_cmd = m.groups()[0]
march_in_cmd = ti.args.default_march
m = common.MARCH_ARG_RE.search(mc_cmd)
if m:
march_in_cmd = m.groups()[0]
common.verify_filecheck_prefixes(filecheck_cmd)
mc_like_tools = mc_LIKE_TOOLS[:]
if ti.args.tool:
mc_like_tools.append(ti.args.tool)
if mc_tool not in mc_like_tools:
common.warn("Skipping non-mc RUN line: " + l)
continue
if not filecheck_cmd.startswith("FileCheck "):
common.warn("Skipping non-FileChecked RUN line: " + l)
continue
mc_cmd_args = mc_cmd[len(mc_tool) :].strip()
mc_cmd_args = mc_cmd_args.replace("< %s", "").replace("%s", "").strip()
check_prefixes = common.get_check_prefixes(filecheck_cmd)
run_list.append(
(
check_prefixes,
mc_tool,
mc_cmd_args,
triple_in_cmd,
march_in_cmd,
)
)
# find all test line from input
testlines = [l for l in ti.input_lines if isTestLine(l, mc_mode)]
run_list_size = len(run_list)
testnum = len(testlines)
raw_output = []
raw_prefixes = []
for (
prefixes,
mc_tool,
mc_args,
triple_in_cmd,
march_in_cmd,
) in run_list:
common.debug("Extracted mc cmd:", mc_tool, mc_args)
common.debug("Extracted FileCheck prefixes:", str(prefixes))
common.debug("Extracted triple :", str(triple_in_cmd))
common.debug("Extracted march:", str(march_in_cmd))
triple = triple_in_cmd or triple_in_ir
if not triple:
triple = common.get_triple_from_march(march_in_cmd)
raw_output.append([])
for line in testlines:
# get output for each testline
out = invoke_tool(
ti.args.llvm_mc_binary or mc_tool,
mc_args,
line,
verbose=ti.args.verbose,
)
raw_output[-1].append(out)
common.debug("Collect raw tool lines:", str(len(raw_output[-1])))
raw_prefixes.append(prefixes)
output_lines = []
generated_prefixes = []
used_prefixes = set()
prefix_set = set([prefix for p in run_list for prefix in p[0]])
common.debug("Rewriting FileCheck prefixes:", str(prefix_set))
for test_id in range(testnum):
input_line = testlines[test_id]
# a {prefix : output, [runid] } dict
# insert output to a prefix-key dict, and do a max sorting
# to select the most-used prefix which share the same output string
p_dict = {}
for run_id in range(run_list_size):
out = raw_output[run_id][test_id]
if hasErr(out):
o = getErrString(out)
else:
o = getOutputString(out)
prefixes = raw_prefixes[run_id]
for p in prefixes:
if p not in p_dict:
p_dict[p] = o, [run_id]
else:
if p_dict[p] == (None, []):
continue
prev_o, run_ids = p_dict[p]
if o == prev_o:
run_ids.append(run_id)
p_dict[p] = o, run_ids
else:
# conflict, discard
p_dict[p] = None, []
p_dict_sorted = dict(
sorted(p_dict.items(), key=lambda item: -len(item[1][1]))
)
# prefix is selected and generated with most shared output lines
# each run_id can only be used once
gen_prefix = ""
used_runid = set()
for prefix, tup in p_dict_sorted.items():
o, run_ids = tup
if len(run_ids) == 0:
continue
skip = False
for i in run_ids:
if i in used_runid:
skip = True
else:
used_runid.add(i)
if not skip:
used_prefixes.add(prefix)
if hasErr(o):
gen_prefix += getErrCheckLine(prefix, o, mc_mode)
else:
gen_prefix += getStdCheckLine(prefix, o, mc_mode)
generated_prefixes.append(gen_prefix.rstrip("\n"))
# write output
prefix_id = 0
for input_info in ti.iterlines(output_lines):
input_line = input_info.line
if isTestLine(input_line, mc_mode):
output_lines.append(input_line)
output_lines.append(generated_prefixes[prefix_id])
prefix_id += 1
elif should_add_line_to_output(input_line, prefix_set, mc_mode):
output_lines.append(input_line)
elif input_line in ti.run_lines or input_line == "":
output_lines.append(input_line)
if ti.args.gen_unused_prefix_body:
output_lines.extend(
ti.get_checks_for_unused_prefixes(run_list, used_prefixes)
)
common.debug("Writing %d lines to %s..." % (len(output_lines), ti.path))
with open(ti.path, "wb") as f:
f.writelines(["{}\n".format(l).encode("utf-8") for l in output_lines])
if __name__ == "__main__":
main()