summaryrefslogtreecommitdiff
path: root/0159-Backport-SME-Improve-immediate-expansion-PR106583.patch
diff options
context:
space:
mode:
authorCoprDistGit <infra@openeuler.org>2025-02-28 10:03:49 +0000
committerCoprDistGit <infra@openeuler.org>2025-02-28 10:03:49 +0000
commit73127104a245052cd5cf29cdaaca3e5c32c70348 (patch)
tree8e28b63e478c43c252f18b49836dff7313affe54 /0159-Backport-SME-Improve-immediate-expansion-PR106583.patch
parent49d3feaf4665cdb07576fc1a2382a4d82a612d35 (diff)
automatic import of gccopeneuler24.03_LTS_SP1
Diffstat (limited to '0159-Backport-SME-Improve-immediate-expansion-PR106583.patch')
-rw-r--r--0159-Backport-SME-Improve-immediate-expansion-PR106583.patch631
1 files changed, 631 insertions, 0 deletions
diff --git a/0159-Backport-SME-Improve-immediate-expansion-PR106583.patch b/0159-Backport-SME-Improve-immediate-expansion-PR106583.patch
new file mode 100644
index 0000000..4ecbdd8
--- /dev/null
+++ b/0159-Backport-SME-Improve-immediate-expansion-PR106583.patch
@@ -0,0 +1,631 @@
+From d5293e2a8db54245553e01ad5d791b7492ad6101 Mon Sep 17 00:00:00 2001
+From: Wilco Dijkstra <wdijkstr@arm.com>
+Date: Mon, 24 Oct 2022 15:14:14 +0100
+Subject: [PATCH 060/157] [Backport][SME] Improve immediate expansion
+ [PR106583]
+
+Reference: https://gcc.gnu.org/git/?p=gcc.git;a=commit;h=a096036589d82175a0f729c2dab73c9a527d075d
+
+Improve immediate expansion of immediates which can be created from a
+bitmask immediate and 2 MOVKs. Simplify, refactor and improve efficiency
+of bitmask checks. Move various immediate handling functions together
+to avoid forward declarations.
+
+This reduces the number of 4-instruction immediates in SPECINT/FP by 10-15%.
+
+gcc/
+
+ PR target/106583
+ * config/aarch64/aarch64.cc (aarch64_internal_mov_immediate)
+ Add support for a bitmask immediate with 2 MOVKs.
+ (aarch64_check_bitmask): New function after refactorization.
+ (aarch64_bitmask_imm): Simplify replication of small modes.
+ Split function into 64-bit only version for efficiency.
+ (aarch64_move_imm): Move near other immediate functions.
+ (aarch64_uimm12_shift): Likewise.
+ (aarch64_clamp_to_uimm12_shift): Likewise.
+ (aarch64_movk_shift): Likewise.
+ (aarch64_replicate_bitmask_imm): Likewise.
+ (aarch64_and_split_imm1): Likewise.
+ (aarch64_and_split_imm2): Likewise.
+ (aarch64_and_bitmask_imm): Likewise.
+ (aarch64_movw_imm): Likewise.
+
+gcc/testsuite/
+ PR target/106583
+ * gcc.target/aarch64/pr106583.c: Add new test.
+---
+ gcc/config/aarch64/aarch64.cc | 485 +++++++++++---------
+ gcc/testsuite/gcc.target/aarch64/pr106583.c | 41 ++
+ 2 files changed, 301 insertions(+), 225 deletions(-)
+ create mode 100644 gcc/testsuite/gcc.target/aarch64/pr106583.c
+
+diff --git a/gcc/config/aarch64/aarch64.cc b/gcc/config/aarch64/aarch64.cc
+index b4b646fa0..cf7736994 100644
+--- a/gcc/config/aarch64/aarch64.cc
++++ b/gcc/config/aarch64/aarch64.cc
+@@ -305,7 +305,6 @@ static bool aarch64_builtin_support_vector_misalignment (machine_mode mode,
+ static machine_mode aarch64_simd_container_mode (scalar_mode, poly_int64);
+ static bool aarch64_print_address_internal (FILE*, machine_mode, rtx,
+ aarch64_addr_query_type);
+-static HOST_WIDE_INT aarch64_clamp_to_uimm12_shift (HOST_WIDE_INT val);
+
+ /* The processor for which instructions should be scheduled. */
+ enum aarch64_processor aarch64_tune = cortexa53;
+@@ -5756,6 +5755,143 @@ aarch64_output_sve_vector_inc_dec (const char *operands, rtx x)
+ factor, nelts_per_vq);
+ }
+
++/* Multipliers for repeating bitmasks of width 32, 16, 8, 4, and 2. */
++
++static const unsigned HOST_WIDE_INT bitmask_imm_mul[] =
++ {
++ 0x0000000100000001ull,
++ 0x0001000100010001ull,
++ 0x0101010101010101ull,
++ 0x1111111111111111ull,
++ 0x5555555555555555ull,
++ };
++
++
++
++/* Return true if 64-bit VAL is a valid bitmask immediate. */
++static bool
++aarch64_bitmask_imm (unsigned HOST_WIDE_INT val)
++{
++ unsigned HOST_WIDE_INT tmp, mask, first_one, next_one;
++ int bits;
++
++ /* Check for a single sequence of one bits and return quickly if so.
++ The special cases of all ones and all zeroes returns false. */
++ tmp = val + (val & -val);
++
++ if (tmp == (tmp & -tmp))
++ return (val + 1) > 1;
++
++ /* Invert if the immediate doesn't start with a zero bit - this means we
++ only need to search for sequences of one bits. */
++ if (val & 1)
++ val = ~val;
++
++ /* Find the first set bit and set tmp to val with the first sequence of one
++ bits removed. Return success if there is a single sequence of ones. */
++ first_one = val & -val;
++ tmp = val & (val + first_one);
++
++ if (tmp == 0)
++ return true;
++
++ /* Find the next set bit and compute the difference in bit position. */
++ next_one = tmp & -tmp;
++ bits = clz_hwi (first_one) - clz_hwi (next_one);
++ mask = val ^ tmp;
++
++ /* Check the bit position difference is a power of 2, and that the first
++ sequence of one bits fits within 'bits' bits. */
++ if ((mask >> bits) != 0 || bits != (bits & -bits))
++ return false;
++
++ /* Check the sequence of one bits is repeated 64/bits times. */
++ return val == mask * bitmask_imm_mul[__builtin_clz (bits) - 26];
++}
++
++
++/* Return true if VAL is a valid bitmask immediate for MODE. */
++bool
++aarch64_bitmask_imm (HOST_WIDE_INT val_in, machine_mode mode)
++{
++ if (mode == DImode)
++ return aarch64_bitmask_imm (val_in);
++
++ unsigned HOST_WIDE_INT val = val_in;
++
++ if (mode == SImode)
++ return aarch64_bitmask_imm ((val & 0xffffffff) | (val << 32));
++
++ /* Replicate small immediates to fit 64 bits. */
++ int size = GET_MODE_UNIT_PRECISION (mode);
++ val &= (HOST_WIDE_INT_1U << size) - 1;
++ val *= bitmask_imm_mul[__builtin_clz (size) - 26];
++
++ return aarch64_bitmask_imm (val);
++}
++
++
++/* Return true if the immediate VAL can be a bitfield immediate
++ by changing the given MASK bits in VAL to zeroes, ones or bits
++ from the other half of VAL. Return the new immediate in VAL2. */
++static inline bool
++aarch64_check_bitmask (unsigned HOST_WIDE_INT val,
++ unsigned HOST_WIDE_INT &val2,
++ unsigned HOST_WIDE_INT mask)
++{
++ val2 = val & ~mask;
++ if (val2 != val && aarch64_bitmask_imm (val2))
++ return true;
++ val2 = val | mask;
++ if (val2 != val && aarch64_bitmask_imm (val2))
++ return true;
++ val = val & ~mask;
++ val2 = val | (((val >> 32) | (val << 32)) & mask);
++ if (val2 != val && aarch64_bitmask_imm (val2))
++ return true;
++ val2 = val | (((val >> 16) | (val << 48)) & mask);
++ if (val2 != val && aarch64_bitmask_imm (val2))
++ return true;
++ return false;
++}
++
++
++/* Return true if val is an immediate that can be loaded into a
++ register by a MOVZ instruction. */
++static bool
++aarch64_movw_imm (HOST_WIDE_INT val, scalar_int_mode mode)
++{
++ if (GET_MODE_SIZE (mode) > 4)
++ {
++ if ((val & (((HOST_WIDE_INT) 0xffff) << 32)) == val
++ || (val & (((HOST_WIDE_INT) 0xffff) << 48)) == val)
++ return 1;
++ }
++ else
++ {
++ /* Ignore sign extension. */
++ val &= (HOST_WIDE_INT) 0xffffffff;
++ }
++ return ((val & (((HOST_WIDE_INT) 0xffff) << 0)) == val
++ || (val & (((HOST_WIDE_INT) 0xffff) << 16)) == val);
++}
++
++
++/* Return true if VAL is an immediate that can be loaded into a
++ register in a single instruction. */
++bool
++aarch64_move_imm (HOST_WIDE_INT val, machine_mode mode)
++{
++ scalar_int_mode int_mode;
++ if (!is_a <scalar_int_mode> (mode, &int_mode))
++ return false;
++
++ if (aarch64_movw_imm (val, int_mode) || aarch64_movw_imm (~val, int_mode))
++ return 1;
++ return aarch64_bitmask_imm (val, int_mode);
++}
++
++
+ static int
+ aarch64_internal_mov_immediate (rtx dest, rtx imm, bool generate,
+ scalar_int_mode mode)
+@@ -5786,7 +5922,7 @@ aarch64_internal_mov_immediate (rtx dest, rtx imm, bool generate,
+ emit_insn (gen_rtx_SET (dest, GEN_INT (val2)));
+
+ /* Check if we have to emit a second instruction by checking to see
+- if any of the upper 32 bits of the original DI mode value is set. */
++ if any of the upper 32 bits of the original DI mode value is set. */
+ if (val == val2)
+ return 1;
+
+@@ -5822,36 +5958,43 @@ aarch64_internal_mov_immediate (rtx dest, rtx imm, bool generate,
+ one_match = ((~val & mask) == 0) + ((~val & (mask << 16)) == 0) +
+ ((~val & (mask << 32)) == 0) + ((~val & (mask << 48)) == 0);
+
+- if (zero_match != 2 && one_match != 2)
++ if (zero_match < 2 && one_match < 2)
+ {
+ /* Try emitting a bitmask immediate with a movk replacing 16 bits.
+ For a 64-bit bitmask try whether changing 16 bits to all ones or
+ zeroes creates a valid bitmask. To check any repeated bitmask,
+ try using 16 bits from the other 32-bit half of val. */
+
+- for (i = 0; i < 64; i += 16, mask <<= 16)
+- {
+- val2 = val & ~mask;
+- if (val2 != val && aarch64_bitmask_imm (val2, mode))
+- break;
+- val2 = val | mask;
+- if (val2 != val && aarch64_bitmask_imm (val2, mode))
+- break;
+- val2 = val2 & ~mask;
+- val2 = val2 | (((val2 >> 32) | (val2 << 32)) & mask);
+- if (val2 != val && aarch64_bitmask_imm (val2, mode))
+- break;
+- }
+- if (i != 64)
+- {
+- if (generate)
++ for (i = 0; i < 64; i += 16)
++ if (aarch64_check_bitmask (val, val2, mask << i))
++ {
++ if (generate)
++ {
++ emit_insn (gen_rtx_SET (dest, GEN_INT (val2)));
++ emit_insn (gen_insv_immdi (dest, GEN_INT (i),
++ GEN_INT ((val >> i) & 0xffff)));
++ }
++ return 2;
++ }
++ }
++
++ /* Try a bitmask plus 2 movk to generate the immediate in 3 instructions. */
++ if (zero_match + one_match == 0)
++ {
++ for (i = 0; i < 48; i += 16)
++ for (int j = i + 16; j < 64; j += 16)
++ if (aarch64_check_bitmask (val, val2, (mask << i) | (mask << j)))
+ {
+- emit_insn (gen_rtx_SET (dest, GEN_INT (val2)));
+- emit_insn (gen_insv_immdi (dest, GEN_INT (i),
+- GEN_INT ((val >> i) & 0xffff)));
++ if (generate)
++ {
++ emit_insn (gen_rtx_SET (dest, GEN_INT (val2)));
++ emit_insn (gen_insv_immdi (dest, GEN_INT (i),
++ GEN_INT ((val >> i) & 0xffff)));
++ emit_insn (gen_insv_immdi (dest, GEN_INT (j),
++ GEN_INT ((val >> j) & 0xffff)));
++ }
++ return 3;
+ }
+- return 2;
+- }
+ }
+
+ /* Generate 2-4 instructions, skipping 16 bits of all zeroes or ones which
+@@ -5898,6 +6041,99 @@ aarch64_mov128_immediate (rtx imm)
+ }
+
+
++/* Return true if val can be encoded as a 12-bit unsigned immediate with
++ a left shift of 0 or 12 bits. */
++bool
++aarch64_uimm12_shift (HOST_WIDE_INT val)
++{
++ return ((val & (((HOST_WIDE_INT) 0xfff) << 0)) == val
++ || (val & (((HOST_WIDE_INT) 0xfff) << 12)) == val
++ );
++}
++
++/* Returns the nearest value to VAL that will fit as a 12-bit unsigned immediate
++ that can be created with a left shift of 0 or 12. */
++static HOST_WIDE_INT
++aarch64_clamp_to_uimm12_shift (HOST_WIDE_INT val)
++{
++ /* Check to see if the value fits in 24 bits, as that is the maximum we can
++ handle correctly. */
++ gcc_assert ((val & 0xffffff) == val);
++
++ if (((val & 0xfff) << 0) == val)
++ return val;
++
++ return val & (0xfff << 12);
++}
++
++
++/* Test whether:
++
++ X = (X & AND_VAL) | IOR_VAL;
++
++ can be implemented using:
++
++ MOVK X, #(IOR_VAL >> shift), LSL #shift
++
++ Return the shift if so, otherwise return -1. */
++int
++aarch64_movk_shift (const wide_int_ref &and_val,
++ const wide_int_ref &ior_val)
++{
++ unsigned int precision = and_val.get_precision ();
++ unsigned HOST_WIDE_INT mask = 0xffff;
++ for (unsigned int shift = 0; shift < precision; shift += 16)
++ {
++ if (and_val == ~mask && (ior_val & mask) == ior_val)
++ return shift;
++ mask <<= 16;
++ }
++ return -1;
++}
++
++/* Create mask of ones, covering the lowest to highest bits set in VAL_IN.
++ Assumed precondition: VAL_IN Is not zero. */
++
++unsigned HOST_WIDE_INT
++aarch64_and_split_imm1 (HOST_WIDE_INT val_in)
++{
++ int lowest_bit_set = ctz_hwi (val_in);
++ int highest_bit_set = floor_log2 (val_in);
++ gcc_assert (val_in != 0);
++
++ return ((HOST_WIDE_INT_UC (2) << highest_bit_set) -
++ (HOST_WIDE_INT_1U << lowest_bit_set));
++}
++
++/* Create constant where bits outside of lowest bit set to highest bit set
++ are set to 1. */
++
++unsigned HOST_WIDE_INT
++aarch64_and_split_imm2 (HOST_WIDE_INT val_in)
++{
++ return val_in | ~aarch64_and_split_imm1 (val_in);
++}
++
++/* Return true if VAL_IN is a valid 'and' bitmask immediate. */
++
++bool
++aarch64_and_bitmask_imm (unsigned HOST_WIDE_INT val_in, machine_mode mode)
++{
++ scalar_int_mode int_mode;
++ if (!is_a <scalar_int_mode> (mode, &int_mode))
++ return false;
++
++ if (aarch64_bitmask_imm (val_in, int_mode))
++ return false;
++
++ if (aarch64_move_imm (val_in, int_mode))
++ return false;
++
++ unsigned HOST_WIDE_INT imm2 = aarch64_and_split_imm2 (val_in);
++
++ return aarch64_bitmask_imm (imm2, int_mode);
++}
++
+ /* Return the number of temporary registers that aarch64_add_offset_1
+ would need to add OFFSET to a register. */
+
+@@ -10379,207 +10615,6 @@ aarch64_tls_referenced_p (rtx x)
+ }
+
+
+-/* Return true if val can be encoded as a 12-bit unsigned immediate with
+- a left shift of 0 or 12 bits. */
+-bool
+-aarch64_uimm12_shift (HOST_WIDE_INT val)
+-{
+- return ((val & (((HOST_WIDE_INT) 0xfff) << 0)) == val
+- || (val & (((HOST_WIDE_INT) 0xfff) << 12)) == val
+- );
+-}
+-
+-/* Returns the nearest value to VAL that will fit as a 12-bit unsigned immediate
+- that can be created with a left shift of 0 or 12. */
+-static HOST_WIDE_INT
+-aarch64_clamp_to_uimm12_shift (HOST_WIDE_INT val)
+-{
+- /* Check to see if the value fits in 24 bits, as that is the maximum we can
+- handle correctly. */
+- gcc_assert ((val & 0xffffff) == val);
+-
+- if (((val & 0xfff) << 0) == val)
+- return val;
+-
+- return val & (0xfff << 12);
+-}
+-
+-/* Return true if val is an immediate that can be loaded into a
+- register by a MOVZ instruction. */
+-static bool
+-aarch64_movw_imm (HOST_WIDE_INT val, scalar_int_mode mode)
+-{
+- if (GET_MODE_SIZE (mode) > 4)
+- {
+- if ((val & (((HOST_WIDE_INT) 0xffff) << 32)) == val
+- || (val & (((HOST_WIDE_INT) 0xffff) << 48)) == val)
+- return 1;
+- }
+- else
+- {
+- /* Ignore sign extension. */
+- val &= (HOST_WIDE_INT) 0xffffffff;
+- }
+- return ((val & (((HOST_WIDE_INT) 0xffff) << 0)) == val
+- || (val & (((HOST_WIDE_INT) 0xffff) << 16)) == val);
+-}
+-
+-/* Test whether:
+-
+- X = (X & AND_VAL) | IOR_VAL;
+-
+- can be implemented using:
+-
+- MOVK X, #(IOR_VAL >> shift), LSL #shift
+-
+- Return the shift if so, otherwise return -1. */
+-int
+-aarch64_movk_shift (const wide_int_ref &and_val,
+- const wide_int_ref &ior_val)
+-{
+- unsigned int precision = and_val.get_precision ();
+- unsigned HOST_WIDE_INT mask = 0xffff;
+- for (unsigned int shift = 0; shift < precision; shift += 16)
+- {
+- if (and_val == ~mask && (ior_val & mask) == ior_val)
+- return shift;
+- mask <<= 16;
+- }
+- return -1;
+-}
+-
+-/* VAL is a value with the inner mode of MODE. Replicate it to fill a
+- 64-bit (DImode) integer. */
+-
+-static unsigned HOST_WIDE_INT
+-aarch64_replicate_bitmask_imm (unsigned HOST_WIDE_INT val, machine_mode mode)
+-{
+- unsigned int size = GET_MODE_UNIT_PRECISION (mode);
+- while (size < 64)
+- {
+- val &= (HOST_WIDE_INT_1U << size) - 1;
+- val |= val << size;
+- size *= 2;
+- }
+- return val;
+-}
+-
+-/* Multipliers for repeating bitmasks of width 32, 16, 8, 4, and 2. */
+-
+-static const unsigned HOST_WIDE_INT bitmask_imm_mul[] =
+- {
+- 0x0000000100000001ull,
+- 0x0001000100010001ull,
+- 0x0101010101010101ull,
+- 0x1111111111111111ull,
+- 0x5555555555555555ull,
+- };
+-
+-
+-/* Return true if val is a valid bitmask immediate. */
+-
+-bool
+-aarch64_bitmask_imm (HOST_WIDE_INT val_in, machine_mode mode)
+-{
+- unsigned HOST_WIDE_INT val, tmp, mask, first_one, next_one;
+- int bits;
+-
+- /* Check for a single sequence of one bits and return quickly if so.
+- The special cases of all ones and all zeroes returns false. */
+- val = aarch64_replicate_bitmask_imm (val_in, mode);
+- tmp = val + (val & -val);
+-
+- if (tmp == (tmp & -tmp))
+- return (val + 1) > 1;
+-
+- /* Replicate 32-bit immediates so we can treat them as 64-bit. */
+- if (mode == SImode)
+- val = (val << 32) | (val & 0xffffffff);
+-
+- /* Invert if the immediate doesn't start with a zero bit - this means we
+- only need to search for sequences of one bits. */
+- if (val & 1)
+- val = ~val;
+-
+- /* Find the first set bit and set tmp to val with the first sequence of one
+- bits removed. Return success if there is a single sequence of ones. */
+- first_one = val & -val;
+- tmp = val & (val + first_one);
+-
+- if (tmp == 0)
+- return true;
+-
+- /* Find the next set bit and compute the difference in bit position. */
+- next_one = tmp & -tmp;
+- bits = clz_hwi (first_one) - clz_hwi (next_one);
+- mask = val ^ tmp;
+-
+- /* Check the bit position difference is a power of 2, and that the first
+- sequence of one bits fits within 'bits' bits. */
+- if ((mask >> bits) != 0 || bits != (bits & -bits))
+- return false;
+-
+- /* Check the sequence of one bits is repeated 64/bits times. */
+- return val == mask * bitmask_imm_mul[__builtin_clz (bits) - 26];
+-}
+-
+-/* Create mask of ones, covering the lowest to highest bits set in VAL_IN.
+- Assumed precondition: VAL_IN Is not zero. */
+-
+-unsigned HOST_WIDE_INT
+-aarch64_and_split_imm1 (HOST_WIDE_INT val_in)
+-{
+- int lowest_bit_set = ctz_hwi (val_in);
+- int highest_bit_set = floor_log2 (val_in);
+- gcc_assert (val_in != 0);
+-
+- return ((HOST_WIDE_INT_UC (2) << highest_bit_set) -
+- (HOST_WIDE_INT_1U << lowest_bit_set));
+-}
+-
+-/* Create constant where bits outside of lowest bit set to highest bit set
+- are set to 1. */
+-
+-unsigned HOST_WIDE_INT
+-aarch64_and_split_imm2 (HOST_WIDE_INT val_in)
+-{
+- return val_in | ~aarch64_and_split_imm1 (val_in);
+-}
+-
+-/* Return true if VAL_IN is a valid 'and' bitmask immediate. */
+-
+-bool
+-aarch64_and_bitmask_imm (unsigned HOST_WIDE_INT val_in, machine_mode mode)
+-{
+- scalar_int_mode int_mode;
+- if (!is_a <scalar_int_mode> (mode, &int_mode))
+- return false;
+-
+- if (aarch64_bitmask_imm (val_in, int_mode))
+- return false;
+-
+- if (aarch64_move_imm (val_in, int_mode))
+- return false;
+-
+- unsigned HOST_WIDE_INT imm2 = aarch64_and_split_imm2 (val_in);
+-
+- return aarch64_bitmask_imm (imm2, int_mode);
+-}
+-
+-/* Return true if val is an immediate that can be loaded into a
+- register in a single instruction. */
+-bool
+-aarch64_move_imm (HOST_WIDE_INT val, machine_mode mode)
+-{
+- scalar_int_mode int_mode;
+- if (!is_a <scalar_int_mode> (mode, &int_mode))
+- return false;
+-
+- if (aarch64_movw_imm (val, int_mode) || aarch64_movw_imm (~val, int_mode))
+- return 1;
+- return aarch64_bitmask_imm (val, int_mode);
+-}
+-
+ static bool
+ aarch64_cannot_force_const_mem (machine_mode mode ATTRIBUTE_UNUSED, rtx x)
+ {
+diff --git a/gcc/testsuite/gcc.target/aarch64/pr106583.c b/gcc/testsuite/gcc.target/aarch64/pr106583.c
+new file mode 100644
+index 000000000..0f9315808
+--- /dev/null
++++ b/gcc/testsuite/gcc.target/aarch64/pr106583.c
+@@ -0,0 +1,41 @@
++/* { dg-do assemble } */
++/* { dg-options "-O2 --save-temps" } */
++
++long f1 (void)
++{
++ return 0x7efefefefefefeff;
++}
++
++long f2 (void)
++{
++ return 0x12345678aaaaaaaa;
++}
++
++long f3 (void)
++{
++ return 0x1234cccccccc5678;
++}
++
++long f4 (void)
++{
++ return 0x7777123456787777;
++}
++
++long f5 (void)
++{
++ return 0x5555555512345678;
++}
++
++long f6 (void)
++{
++ return 0x1234bbbb5678bbbb;
++}
++
++long f7 (void)
++{
++ return 0x4444123444445678;
++}
++
++
++/* { dg-final { scan-assembler-times {\tmovk\t} 14 } } */
++/* { dg-final { scan-assembler-times {\tmov\t} 7 } } */
+--
+2.33.0
+