From: Christian Bruel Date: Wed, 30 Oct 2013 12:35:27 +0000 (+0100) Subject: sh-mem.cc (sh_expand_cmpnstr): New function. X-Git-Url: https://git.libre-soc.org/?a=commitdiff_plain;h=ca494b8d783b3a0d6e259789ccd09714adaf3d46;p=gcc.git sh-mem.cc (sh_expand_cmpnstr): New function. 2013-10-30 Christian Bruel * gcc/config/sh/sh-mem.cc (sh_expand_cmpnstr): New function. (sh_expand_cmpstr): Handle known align and schedule improvements. * gcc/config/sh/sh-protos.h (sh_expand_cmpstrn): Declare. * gcc/config/sh/sh.md (cmpstrnsi): New pattern. From-SVN: r204206 --- diff --git a/gcc/ChangeLog b/gcc/ChangeLog index fff372650f1..8666f23616e 100644 --- a/gcc/ChangeLog +++ b/gcc/ChangeLog @@ -1,3 +1,10 @@ +2013-10-30 Christian Bruel + + * gcc/config/sh/sh-mem.cc (sh_expand_cmpnstr): New function. + (sh_expand_cmpstr): Handle known align and schedule improvements. + * gcc/config/sh/sh-protos.h (sh_expand_cmpstrn): Declare. + * gcc/config/sh/sh.md (cmpstrnsi): New pattern. + 2013-10-30 Martin Jambor PR rtl-optimization/10474 diff --git a/gcc/config/sh/sh-mem.cc b/gcc/config/sh/sh-mem.cc index e6f0843ca30..eabf6871173 100644 --- a/gcc/config/sh/sh-mem.cc +++ b/gcc/config/sh/sh-mem.cc @@ -200,22 +200,25 @@ sh_expand_cmpstr (rtx *operands) rtx L_return = gen_label_rtx (); rtx L_loop_byte = gen_label_rtx (); rtx L_end_loop_byte = gen_label_rtx (); - rtx L_loop_long = gen_label_rtx (); - rtx L_end_loop_long = gen_label_rtx (); rtx jump, addr1, addr2; int prob_unlikely = REG_BR_PROB_BASE / 10; int prob_likely = REG_BR_PROB_BASE / 4; - emit_insn (gen_iorsi3 (tmp1, s1_addr, s2_addr)); - emit_move_insn (tmp0, GEN_INT (3)); + rtx L_loop_long = gen_label_rtx (); + rtx L_end_loop_long = gen_label_rtx (); - emit_insn (gen_tstsi_t (tmp0, tmp1)); + int align = INTVAL (operands[3]); emit_move_insn (tmp0, const0_rtx); - jump = emit_jump_insn (gen_branch_false (L_loop_byte)); - add_int_reg_note (jump, REG_BR_PROB, prob_likely); + if (align < 4) + { + emit_insn (gen_iorsi3 (tmp1, s1_addr, s2_addr)); + emit_insn (gen_tstsi_t (GEN_INT (3), tmp1)); + jump = emit_jump_insn (gen_branch_false (L_loop_byte)); + add_int_reg_note (jump, REG_BR_PROB, prob_likely); + } addr1 = adjust_automodify_address (s1, SImode, s1_addr, 0); addr2 = adjust_automodify_address (s2, SImode, s2_addr, 0); @@ -250,7 +253,7 @@ sh_expand_cmpstr (rtx *operands) add_int_reg_note (jump, REG_BR_PROB, prob_likely); /* end loop. */ - /* Fallthu, check if one of the word is greater. */ + /* Fallthu, substract words. */ if (TARGET_LITTLE_ENDIAN) { rtx low_1 = gen_lowpart (HImode, tmp1); @@ -267,15 +270,15 @@ sh_expand_cmpstr (rtx *operands) jump = emit_jump_insn (gen_jump_compact (L_return)); emit_barrier_after (jump); - /* start byte loop. */ - addr1 = adjust_automodify_address (s1, QImode, s1_addr, 0); - addr2 = adjust_automodify_address (s2, QImode, s2_addr, 0); - emit_label (L_end_loop_long); emit_move_insn (s1_addr, plus_constant (Pmode, s1_addr, -4)); emit_move_insn (s2_addr, plus_constant (Pmode, s2_addr, -4)); + /* start byte loop. */ + addr1 = adjust_automodify_address (s1, QImode, s1_addr, 0); + addr2 = adjust_automodify_address (s2, QImode, s2_addr, 0); + emit_label (L_loop_byte); emit_insn (gen_extendqisi2 (tmp2, addr2)); @@ -289,13 +292,16 @@ sh_expand_cmpstr (rtx *operands) add_int_reg_note (jump, REG_BR_PROB, prob_unlikely); emit_insn (gen_cmpeqsi_t (tmp1, tmp2)); - emit_jump_insn (gen_branch_true (L_loop_byte)); + if (flag_delayed_branch) + emit_insn (gen_zero_extendqisi2 (tmp2, gen_lowpart (QImode, tmp2))); + jump = emit_jump_insn (gen_branch_true (L_loop_byte)); add_int_reg_note (jump, REG_BR_PROB, prob_likely); /* end loop. */ emit_label (L_end_loop_byte); - emit_insn (gen_zero_extendqisi2 (tmp2, gen_lowpart (QImode, tmp2))); + if (! flag_delayed_branch) + emit_insn (gen_zero_extendqisi2 (tmp2, gen_lowpart (QImode, tmp2))); emit_insn (gen_zero_extendqisi2 (tmp1, gen_lowpart (QImode, tmp1))); emit_label (L_return); @@ -305,3 +311,166 @@ sh_expand_cmpstr (rtx *operands) return true; } +/* Emit code to perform a strncmp. + + OPERANDS[0] is the destination. + OPERANDS[1] is the first string. + OPERANDS[2] is the second string. + OPERANDS[3] is the length. + OPERANDS[4] is the align. */ +bool +sh_expand_cmpnstr (rtx *operands) +{ + rtx s1 = copy_rtx (operands[1]); + rtx s2 = copy_rtx (operands[2]); + + rtx s1_addr = copy_addr_to_reg (XEXP (s1, 0)); + rtx s2_addr = copy_addr_to_reg (XEXP (s2, 0)); + rtx tmp0 = gen_reg_rtx (SImode); + rtx tmp1 = gen_reg_rtx (SImode); + rtx tmp2 = gen_reg_rtx (SImode); + + rtx L_return = gen_label_rtx (); + rtx L_loop_byte = gen_label_rtx (); + rtx L_end_loop_byte = gen_label_rtx (); + + rtx jump, addr1, addr2; + int prob_unlikely = REG_BR_PROB_BASE / 10; + int prob_likely = REG_BR_PROB_BASE / 4; + + rtx len = force_reg (SImode, operands[3]); + int constp = (CONST_INT_P (operands[3])); + int bytes = (constp ? INTVAL (operands[3]) : 0); + int witers = bytes / 4; + + /* We could still loop on a register count. Not found very + convincing to optimize yet. */ + if (! constp) + return false; + + if (witers > 1) + { + rtx L_loop_long = gen_label_rtx (); + rtx L_end_loop_long = gen_label_rtx (); + rtx tmp3 = gen_reg_rtx (SImode); + rtx lenw = gen_reg_rtx (SImode); + int align = INTVAL (operands[4]); + + emit_move_insn (tmp0, const0_rtx); + + if (align < 4) + { + emit_insn (gen_iorsi3 (tmp1, s1_addr, s2_addr)); + emit_insn (gen_tstsi_t (GEN_INT (3), tmp1)); + jump = emit_jump_insn (gen_branch_false (L_loop_byte)); + add_int_reg_note (jump, REG_BR_PROB, prob_likely); + } + + addr1 = adjust_automodify_address (s1, SImode, s1_addr, 0); + addr2 = adjust_automodify_address (s2, SImode, s2_addr, 0); + + /* words count. */ + emit_insn (gen_lshrsi3 (lenw, len, GEN_INT (2))); + + /*start long loop. */ + emit_label (L_loop_long); + + /* tmp2 is aligned, OK to load. */ + emit_move_insn (tmp2, addr2); + emit_move_insn (s2_addr, plus_constant (Pmode, s2_addr, 4)); + + /* tmp1 is aligned, OK to load. */ + emit_move_insn (tmp1, addr1); + emit_move_insn (s1_addr, plus_constant (Pmode, s1_addr, 4)); + + /* Is there a 0 byte ? */ + emit_insn (gen_andsi3 (tmp3, tmp2, tmp1)); + + emit_insn (gen_cmpstr_t (tmp0, tmp3)); + jump = emit_jump_insn (gen_branch_true (L_end_loop_long)); + add_int_reg_note (jump, REG_BR_PROB, prob_unlikely); + + emit_insn (gen_cmpeqsi_t (tmp1, tmp2)); + jump = emit_jump_insn (gen_branch_false (L_end_loop_long)); + add_int_reg_note (jump, REG_BR_PROB, prob_unlikely); + + if (TARGET_SH2) + emit_insn (gen_dect (lenw, lenw)); + else + { + emit_insn (gen_addsi3 (lenw, lenw, GEN_INT (-1))); + emit_insn (gen_tstsi_t (lenw, lenw)); + } + jump = emit_jump_insn (gen_branch_false (L_loop_long)); + add_int_reg_note (jump, REG_BR_PROB, prob_likely); + + /* end loop. Reached max iterations. */ + if (bytes % 4 == 0) + { + /* Done. */ + jump = emit_jump_insn (gen_jump_compact (L_return)); + emit_barrier_after (jump); + } + else + { + /* Remaining bytes to read. */ + emit_move_insn (len, GEN_INT (bytes % 4)); + jump = emit_jump_insn (gen_jump_compact (L_loop_byte)); + emit_barrier_after (jump); + } + + emit_label (L_end_loop_long); + + /* Remaining bytes to read. */ + emit_move_insn (len, GEN_INT (4)); + + /* Found last word. Restart it byte per byte. */ + emit_move_insn (s1_addr, plus_constant (Pmode, s1_addr, -4)); + emit_move_insn (s2_addr, plus_constant (Pmode, s2_addr, -4)); + } + + addr1 = adjust_automodify_address (s1, QImode, s1_addr, 0); + addr2 = adjust_automodify_address (s2, QImode, s2_addr, 0); + + emit_label (L_loop_byte); + + emit_insn (gen_extendqisi2 (tmp2, addr2)); + emit_move_insn (s2_addr, plus_constant (Pmode, s2_addr, 1)); + + emit_insn (gen_extendqisi2 (tmp1, addr1)); + emit_move_insn (s1_addr, plus_constant (Pmode, s1_addr, 1)); + + emit_insn (gen_cmpeqsi_t (tmp2, const0_rtx)); + jump = emit_jump_insn (gen_branch_true (L_end_loop_byte)); + add_int_reg_note (jump, REG_BR_PROB, prob_unlikely); + + emit_insn (gen_cmpeqsi_t (tmp1, tmp2)); + if (flag_delayed_branch) + emit_insn (gen_zero_extendqisi2 (tmp2, gen_lowpart (QImode, tmp2))); + jump = emit_jump_insn (gen_branch_false (L_end_loop_byte)); + add_int_reg_note (jump, REG_BR_PROB, prob_unlikely); + + if (TARGET_SH2) + emit_insn (gen_dect (len, len)); + else + { + emit_insn (gen_addsi3 (len, len, GEN_INT (-1))); + emit_insn (gen_tstsi_t (len, len)); + } + + jump = emit_jump_insn (gen_branch_false (L_loop_byte)); + add_int_reg_note (jump, REG_BR_PROB, prob_likely); + /* end byte loop. */ + + emit_label (L_end_loop_byte); + + if (! flag_delayed_branch) + emit_insn (gen_zero_extendqisi2 (tmp2, gen_lowpart (QImode, tmp2))); + emit_insn (gen_zero_extendqisi2 (tmp1, gen_lowpart (QImode, tmp1))); + + emit_label (L_return); + + emit_insn (gen_subsi3 (operands[0], tmp1, tmp2)); + + return true; +} diff --git a/gcc/config/sh/sh-protos.h b/gcc/config/sh/sh-protos.h index 3484e4a2fdf..e7dfce366ab 100644 --- a/gcc/config/sh/sh-protos.h +++ b/gcc/config/sh/sh-protos.h @@ -117,6 +117,7 @@ extern void output_pic_addr_const (FILE *, rtx); extern bool expand_block_move (rtx *); extern void prepare_move_operands (rtx[], enum machine_mode mode); extern bool sh_expand_cmpstr (rtx *); +extern bool sh_expand_cmpnstr (rtx *); extern enum rtx_code prepare_cbranch_operands (rtx *, enum machine_mode mode, enum rtx_code comparison); extern void expand_cbranchsi4 (rtx *operands, enum rtx_code comparison, int); diff --git a/gcc/config/sh/sh.md b/gcc/config/sh/sh.md index 0b952991039..b8109e62c44 100644 --- a/gcc/config/sh/sh.md +++ b/gcc/config/sh/sh.md @@ -12067,6 +12067,20 @@ label: FAIL; }) +(define_expand "cmpstrnsi" + [(set (match_operand:SI 0 "register_operand") + (compare:SI (match_operand:BLK 1 "memory_operand") + (match_operand:BLK 2 "memory_operand"))) + (use (match_operand:SI 3 "immediate_operand")) + (use (match_operand:SI 4 "immediate_operand"))] + "TARGET_SH1" +{ + if (! optimize_insn_for_size_p () && sh_expand_cmpnstr (operands)) + DONE; + else + FAIL; +}) + ;; ------------------------------------------------------------------------- ;; Floating point instructions. diff --git a/gcc/testsuite/ChangeLog b/gcc/testsuite/ChangeLog index f938522ec46..ac28ad16867 100644 --- a/gcc/testsuite/ChangeLog +++ b/gcc/testsuite/ChangeLog @@ -1,3 +1,9 @@ +2013-10-30 Christian Bruel + + * gcc.c-torture/execute/builtins/strncmp-2.c: Enable for SH. + * gcc.target/sh/cmpstr.c: New test. + * gcc.target/sh/cmpstrn.c: New test. + 2013-10-30 Martin Jambor PR rtl-optimization/10474 diff --git a/gcc/testsuite/gcc.c-torture/execute/builtins/strncmp-2.c b/gcc/testsuite/gcc.c-torture/execute/builtins/strncmp-2.c index fe3462724ea..508b2d3309a 100644 --- a/gcc/testsuite/gcc.c-torture/execute/builtins/strncmp-2.c +++ b/gcc/testsuite/gcc.c-torture/execute/builtins/strncmp-2.c @@ -12,7 +12,7 @@ extern int strncmp (const char *, const char *, size_t); void main_test (void) { -#if !defined(__OPTIMIZE__) || ((defined(__i386__) || defined (__x86_64__)) && !defined(__OPTIMIZE_SIZE__)) +#if !defined(__OPTIMIZE__) || ((defined(__sh__) || defined(__i386__) || defined (__x86_64__)) && !defined(__OPTIMIZE_SIZE__)) /* These tests work on platforms which support cmpstrsi. We test it at -O0 on all platforms to ensure the strncmp logic is correct. */ const char *const s1 = "hello world"; diff --git a/gcc/testsuite/gcc.target/sh/cmpstr.c b/gcc/testsuite/gcc.target/sh/cmpstr.c new file mode 100644 index 00000000000..3e75e4a6434 --- /dev/null +++ b/gcc/testsuite/gcc.target/sh/cmpstr.c @@ -0,0 +1,27 @@ +/* Check that the __builtin_strcmp function is inlined with cmp/str + when optimizing for speed. */ +/* { dg-do compile { target "sh*-*-*" } } */ +/* { dg-options "-O2" } */ +/* { dg-skip-if "" { "sh*-*-*" } { "-m5*" } { "" } } */ +/* { dg-final { scan-assembler-not "jmp" } } */ +/* { dg-final { scan-assembler-times "cmp/str" 3 } } */ +/* { dg-final { scan-assembler-times "tst\t#3" 2 } } */ + +test00 (const char *s1, const char *s2) +{ + return __builtin_strcmp (s1, s2); +} + +/* NB: This might change as further optimisation might detect the + max length and fallback to cmpstrn. */ +test01(const char *s2) +{ + return __builtin_strcmp ("abc", s2); +} + +/* Check that no test for alignment is needed. */ +test03(const char *s1, const char *s2) +{ + return __builtin_strcmp (__builtin_assume_aligned (s1, 4), + __builtin_assume_aligned (s2, 4)); +} diff --git a/gcc/testsuite/gcc.target/sh/cmpstrn.c b/gcc/testsuite/gcc.target/sh/cmpstrn.c new file mode 100644 index 00000000000..b2260f92110 --- /dev/null +++ b/gcc/testsuite/gcc.target/sh/cmpstrn.c @@ -0,0 +1,21 @@ +/* Check that the __builtin_strncmp function is inlined + when optimizing for speed. */ +/* { dg-do compile { target "sh*-*-*" } } */ +/* { dg-options "-O2" } */ +/* { dg-skip-if "" { "sh*-*-*" } { "-m5*" } { "" } } */ +/* { dg-final { scan-assembler-not "jmp" } } */ +/* { dg-final { scan-assembler-times "cmp/str" 1 } } */ + +/* Test that the cmp/str loop is optimized out. */ +test01(const char *s1, const char *s2, int n) +{ + return __builtin_strncmp (s1, "abcde", 3); +} + +/* Test that the cmp/str loop is used. */ +test02(const char *s1, const char *s2, int n) +{ + return __builtin_strncmp (s1, "abcdefghi", 8); +} + +