-/* Translation of ISL AST to Gimple.
- Copyright (C) 2014-2015 Free Software Foundation, Inc.
+/* Translation of isl AST to Gimple.
+ Copyright (C) 2014-2019 Free Software Foundation, Inc.
Contributed by Roman Gareev <gareevroman@gmail.com>.
This file is part of GCC.
along with GCC; see the file COPYING3. If not see
<http://www.gnu.org/licenses/>. */
+#define USES_ISL
+
#include "config.h"
#ifdef HAVE_isl
-/* Workaround for GMP 5.1.3 bug, see PR56019. */
-#include <stddef.h>
-
-#include <isl/constraint.h>
-#include <isl/set.h>
-#include <isl/union_set.h>
-#include <isl/map.h>
-#include <isl/union_map.h>
-#include <isl/ast_build.h>
-
-/* Since ISL-0.13, the extern is in val_gmp.h. */
-#if !defined(HAVE_ISL_SCHED_CONSTRAINTS_COMPUTE_SCHEDULE) && defined(__cplusplus)
-extern "C" {
-#endif
-#include <isl/val_gmp.h>
-#if !defined(HAVE_ISL_SCHED_CONSTRAINTS_COMPUTE_SCHEDULE) && defined(__cplusplus)
-}
-#endif
+#define INCLUDE_MAP
#include "system.h"
#include "coretypes.h"
#include "backend.h"
#include "cfghooks.h"
#include "tree.h"
#include "gimple.h"
-#include "params.h"
+#include "ssa.h"
#include "fold-const.h"
+#include "gimple-fold.h"
#include "gimple-iterator.h"
+#include "gimplify.h"
+#include "gimplify-me.h"
+#include "tree-eh.h"
#include "tree-ssa-loop.h"
+#include "tree-ssa-operands.h"
+#include "tree-ssa-propagate.h"
#include "tree-pass.h"
#include "cfgloop.h"
#include "tree-data-ref.h"
-#include "graphite-poly.h"
#include "tree-ssa-loop-manip.h"
#include "tree-scalar-evolution.h"
#include "gimple-ssa.h"
#include "tree-phinodes.h"
#include "tree-into-ssa.h"
#include "ssa-iterators.h"
-#include <map>
-#include "graphite-isl-ast-to-gimple.h"
-
-/* This flag is set when an error occurred during the translation of
- ISL AST to Gimple. */
-
-static bool graphite_regenerate_error;
-
-/* We always try to use signed 128 bit types, but fall back to smaller types
- in case a platform does not provide types of these sizes. In the future we
- should use isl to derive the optimal type for each subexpression. */
-
-static int max_mode_int_precision =
- GET_MODE_PRECISION (mode_for_size (MAX_FIXED_MODE_SIZE, MODE_INT, 0));
-static int graphite_expression_type_precision = 128 <= max_mode_int_precision ?
- 128 : max_mode_int_precision;
+#include "tree-cfg.h"
+#include "gimple-pretty-print.h"
+#include "cfganal.h"
+#include "value-prof.h"
+#include "tree-ssa.h"
+#include "tree-vectorizer.h"
+#include "graphite.h"
struct ast_build_info
{
ast_build_info()
: is_parallelizable(false)
- { };
+ { }
bool is_parallelizable;
};
-/* Converts a GMP constant VAL to a tree and returns it. */
-
-static tree
-gmp_cst_to_tree (tree type, mpz_t val)
-{
- tree t = type ? type : integer_type_node;
- mpz_t tmp;
-
- mpz_init (tmp);
- mpz_set (tmp, val);
- wide_int wi = wi::from_mpz (t, tmp, true);
- mpz_clear (tmp);
-
- return wide_int_to_tree (t, wi);
-}
-
-/* Verifies properties that GRAPHITE should maintain during translation. */
-
-static inline void
-graphite_verify (void)
-{
-#ifdef ENABLE_CHECKING
- verify_loop_structure ();
- verify_loop_closed_ssa (true);
-#endif
-}
-
-/* IVS_PARAMS maps ISL's scattering and parameter identifiers
+/* IVS_PARAMS maps isl's scattering and parameter identifiers
to corresponding trees. */
typedef std::map<isl_id *, tree> ivs_params;
-/* Free all memory allocated for ISL's identifiers. */
+/* Free all memory allocated for isl's identifiers. */
-void ivs_params_clear (ivs_params &ip)
+static void ivs_params_clear (ivs_params &ip)
{
std::map<isl_id *, tree>::iterator it;
for (it = ip.begin ();
}
}
+/* Set the "separate" option for the schedule node. */
+
+static isl_schedule_node *
+set_separate_option (__isl_take isl_schedule_node *node, void *user)
+{
+ if (user)
+ return node;
+
+ if (isl_schedule_node_get_type (node) != isl_schedule_node_band)
+ return node;
+
+ /* Set the "separate" option unless it is set earlier to another option. */
+ if (isl_schedule_node_band_member_get_ast_loop_type (node, 0)
+ == isl_ast_loop_default)
+ return isl_schedule_node_band_member_set_ast_loop_type
+ (node, 0, isl_ast_loop_separate);
+
+ return node;
+}
+
+/* Print SCHEDULE under an AST form on file F. */
+
+void
+print_schedule_ast (FILE *f, __isl_keep isl_schedule *schedule, scop_p scop)
+{
+ isl_set *set = isl_set_params (isl_set_copy (scop->param_context));
+ isl_ast_build *context = isl_ast_build_from_context (set);
+ isl_ast_node *ast
+ = isl_ast_build_node_from_schedule (context, isl_schedule_copy (schedule));
+ isl_ast_build_free (context);
+ print_isl_ast (f, ast);
+ isl_ast_node_free (ast);
+}
+
+DEBUG_FUNCTION void
+debug_schedule_ast (__isl_keep isl_schedule *s, scop_p scop)
+{
+ print_schedule_ast (stderr, s, scop);
+}
+
+enum phi_node_kind
+{
+ unknown_phi,
+ loop_phi,
+ close_phi,
+ cond_phi
+};
+
class translate_isl_ast_to_gimple
{
public:
- translate_isl_ast_to_gimple (sese r)
- : region (r)
- { }
-
- /* Translates an ISL AST node NODE to GCC representation in the
- context of a SESE. */
+ translate_isl_ast_to_gimple (sese_info_p r);
edge translate_isl_ast (loop_p context_loop, __isl_keep isl_ast_node *node,
edge next_e, ivs_params &ip);
-
- /* Translates an isl_ast_node_for to Gimple. */
edge translate_isl_ast_node_for (loop_p context_loop,
__isl_keep isl_ast_node *node,
edge next_e, ivs_params &ip);
-
- /* Create the loop for a isl_ast_node_for.
-
- - NEXT_E is the edge where new generated code should be attached. */
edge translate_isl_ast_for_loop (loop_p context_loop,
__isl_keep isl_ast_node *node_for,
edge next_e,
tree type, tree lb, tree ub,
ivs_params &ip);
-
- /* Translates an isl_ast_node_if to Gimple. */
edge translate_isl_ast_node_if (loop_p context_loop,
__isl_keep isl_ast_node *node,
edge next_e, ivs_params &ip);
-
- /* Translates an isl_ast_node_user to Gimple.
-
- FIXME: We should remove iv_map.create (loop->num + 1), if it is
- possible. */
edge translate_isl_ast_node_user (__isl_keep isl_ast_node *node,
edge next_e, ivs_params &ip);
-
- /* Translates an isl_ast_node_block to Gimple. */
edge translate_isl_ast_node_block (loop_p context_loop,
__isl_keep isl_ast_node *node,
edge next_e, ivs_params &ip);
-
- /* Converts a unary isl_ast_expr_op expression E to a GCC expression tree of
- type TYPE. */
tree unary_op_to_tree (tree type, __isl_take isl_ast_expr *expr,
ivs_params &ip);
-
- /* Converts a binary isl_ast_expr_op expression E to a GCC expression tree of
- type TYPE. */
tree binary_op_to_tree (tree type, __isl_take isl_ast_expr *expr,
ivs_params &ip);
-
- /* Converts a ternary isl_ast_expr_op expression E to a GCC expression tree of
- type TYPE. */
tree ternary_op_to_tree (tree type, __isl_take isl_ast_expr *expr,
ivs_params &ip);
-
- /* Converts an isl_ast_expr_op expression E with unknown number of arguments
- to a GCC expression tree of type TYPE. */
tree nary_op_to_tree (tree type, __isl_take isl_ast_expr *expr,
ivs_params &ip);
-
- /* Converts an ISL AST expression E back to a GCC expression tree of
- type TYPE. */
tree gcc_expression_from_isl_expression (tree type,
__isl_take isl_ast_expr *,
ivs_params &ip);
-
- /* Return the tree variable that corresponds to the given isl ast identifier
- expression (an isl_ast_expr of type isl_ast_expr_id).
-
- FIXME: We should replace blind conversation of id's type with derivation
- of the optimal type when we get the corresponding isl support. Blindly
- converting type sizes may be problematic when we switch to smaller
- types. */
tree gcc_expression_from_isl_ast_expr_id (tree type,
__isl_keep isl_ast_expr *expr_id,
ivs_params &ip);
-
- /* Converts an isl_ast_expr_int expression E to a GCC expression tree of
- type TYPE. */
+ widest_int widest_int_from_isl_expr_int (__isl_keep isl_ast_expr *expr);
tree gcc_expression_from_isl_expr_int (tree type,
__isl_take isl_ast_expr *expr);
-
- /* Converts an isl_ast_expr_op expression E to a GCC expression tree of
- type TYPE. */
tree gcc_expression_from_isl_expr_op (tree type,
__isl_take isl_ast_expr *expr,
ivs_params &ip);
-
- /* Creates a new LOOP corresponding to isl_ast_node_for. Inserts an
- induction variable for the new LOOP. New LOOP is attached to CFG
- starting at ENTRY_EDGE. LOOP is inserted into the loop tree and
- becomes the child loop of the OUTER_LOOP. NEWIVS_INDEX binds
- ISL's scattering name to the induction variable created for the
- loop of STMT. The new induction variable is inserted in the NEWIVS
- vector and is of type TYPE. */
struct loop *graphite_create_new_loop (edge entry_edge,
__isl_keep isl_ast_node *node_for,
loop_p outer, tree type,
tree lb, tree ub, ivs_params &ip);
-
- /* All loops generated by create_empty_loop_on_edge have the form of
- a post-test loop:
-
- do
-
- {
- body of the loop;
- } while (lower bound < upper bound);
-
- We create a new if region protecting the loop to be executed, if
- the execution count is zero (lower bound > upper bound). */
- edge graphite_create_new_loop_guard (edge entry_edge,
- __isl_keep isl_ast_node *node_for,
- tree *type,
- tree *lb, tree *ub, ivs_params &ip);
-
- /* Creates a new if region corresponding to ISL's cond. */
edge graphite_create_new_guard (edge entry_edge,
__isl_take isl_ast_expr *if_cond,
ivs_params &ip);
-
- /* Inserts in iv_map a tuple (OLD_LOOP->num, NEW_NAME) for the induction
- variables of the loops around GBB in SESE.
-
- FIXME: Instead of using a vec<tree> that maps each loop id to a possible
- chrec, we could consider using a map<int, tree> that maps loop ids to the
- corresponding tree expressions. */
void build_iv_mapping (vec<tree> iv_map, gimple_poly_bb_p gbb,
__isl_keep isl_ast_expr *user_expr, ivs_params &ip,
- sese region);
+ sese_l ®ion);
+ void add_parameters_to_ivs_params (scop_p scop, ivs_params &ip);
+ __isl_give isl_ast_build *generate_isl_context (scop_p scop);
+
+ __isl_give isl_ast_node * scop_to_isl_ast (scop_p scop);
+
+ tree get_rename_from_scev (tree old_name, gimple_seq *stmts, loop_p loop,
+ vec<tree> iv_map);
+ void graphite_copy_stmts_from_block (basic_block bb, basic_block new_bb,
+ vec<tree> iv_map);
+ edge copy_bb_and_scalar_dependences (basic_block bb, edge next_e,
+ vec<tree> iv_map);
+ void set_rename (tree old_name, tree expr);
+ void gsi_insert_earliest (gimple_seq seq);
+ bool codegen_error_p () const { return codegen_error; }
+
+ void set_codegen_error ()
+ {
+ codegen_error = true;
+ gcc_assert (! flag_checking
+ || param_graphite_allow_codegen_errors);
+ }
+
+ bool is_constant (tree op) const
+ {
+ return TREE_CODE (op) == INTEGER_CST
+ || TREE_CODE (op) == REAL_CST
+ || TREE_CODE (op) == COMPLEX_CST
+ || TREE_CODE (op) == VECTOR_CST;
+ }
+
private:
- sese region;
+ /* The region to be translated. */
+ sese_info_p region;
+
+ /* This flag is set when an error occurred during the translation of isl AST
+ to Gimple. */
+ bool codegen_error;
+
+ /* A vector of all the edges at if_condition merge points. */
+ auto_vec<edge, 2> merge_points;
+
+ tree graphite_expr_type;
};
+translate_isl_ast_to_gimple::translate_isl_ast_to_gimple (sese_info_p r)
+ : region (r), codegen_error (false)
+{
+ /* We always try to use signed 128 bit types, but fall back to smaller types
+ in case a platform does not provide types of these sizes. In the future we
+ should use isl to derive the optimal type for each subexpression. */
+ int max_mode_int_precision
+ = GET_MODE_PRECISION (int_mode_for_size (MAX_FIXED_MODE_SIZE, 0).require ());
+ int graphite_expr_type_precision
+ = 128 <= max_mode_int_precision ? 128 : max_mode_int_precision;
+ graphite_expr_type
+ = build_nonstandard_integer_type (graphite_expr_type_precision, 0);
+}
+
/* Return the tree variable that corresponds to the given isl ast identifier
expression (an isl_ast_expr of type isl_ast_expr_id).
- FIXME: We should replace blind conversation of id's type with derivation
- of the optimal type when we get the corresponding isl support. Blindly
+ FIXME: We should replace blind conversion of id's type with derivation
+ of the optimal type when we get the corresponding isl support. Blindly
converting type sizes may be problematic when we switch to smaller
types. */
-tree
-translate_isl_ast_to_gimple::
+tree translate_isl_ast_to_gimple::
gcc_expression_from_isl_ast_expr_id (tree type,
- __isl_keep isl_ast_expr *expr_id,
+ __isl_take isl_ast_expr *expr_id,
ivs_params &ip)
{
gcc_assert (isl_ast_expr_get_type (expr_id) == isl_ast_expr_id);
"Could not map isl_id to tree expression");
isl_ast_expr_free (expr_id);
tree t = res->second;
- tree *val = region->parameter_rename_map->get(t);
-
- if (!val)
- val = &t;
- return fold_convert (type, *val);
+ if (useless_type_conversion_p (type, TREE_TYPE (t)))
+ return t;
+ return fold_convert (type, t);
}
-/* Converts an isl_ast_expr_int expression E to a GCC expression tree of
- type TYPE. */
+/* Converts an isl_ast_expr_int expression E to a widest_int.
+ Raises a code generation error when the constant doesn't fit. */
-tree
-translate_isl_ast_to_gimple::
-gcc_expression_from_isl_expr_int (tree type, __isl_take isl_ast_expr *expr)
+widest_int translate_isl_ast_to_gimple::
+widest_int_from_isl_expr_int (__isl_keep isl_ast_expr *expr)
{
gcc_assert (isl_ast_expr_get_type (expr) == isl_ast_expr_int);
isl_val *val = isl_ast_expr_get_val (expr);
- mpz_t val_mpz_t;
- mpz_init (val_mpz_t);
- tree res;
- if (isl_val_get_num_gmp (val, val_mpz_t) == -1)
- res = NULL_TREE;
- else
- res = gmp_cst_to_tree (type, val_mpz_t);
+ size_t n = isl_val_n_abs_num_chunks (val, sizeof (HOST_WIDE_INT));
+ HOST_WIDE_INT *chunks = XALLOCAVEC (HOST_WIDE_INT, n);
+ if (n > WIDE_INT_MAX_ELTS
+ || isl_val_get_abs_num_chunks (val, sizeof (HOST_WIDE_INT), chunks) == -1)
+ {
+ isl_val_free (val);
+ set_codegen_error ();
+ return 0;
+ }
+ widest_int wi = widest_int::from_array (chunks, n, true);
+ if (isl_val_is_neg (val))
+ wi = -wi;
isl_val_free (val);
+ return wi;
+}
+
+/* Converts an isl_ast_expr_int expression E to a GCC expression tree of
+ type TYPE. Raises a code generation error when the constant doesn't fit. */
+
+tree translate_isl_ast_to_gimple::
+gcc_expression_from_isl_expr_int (tree type, __isl_take isl_ast_expr *expr)
+{
+ widest_int wi = widest_int_from_isl_expr_int (expr);
isl_ast_expr_free (expr);
- mpz_clear (val_mpz_t);
- return res;
+ if (codegen_error_p ())
+ return NULL_TREE;
+ if (wi::min_precision (wi, TYPE_SIGN (type)) > TYPE_PRECISION (type))
+ {
+ set_codegen_error ();
+ return NULL_TREE;
+ }
+ return wide_int_to_tree (type, wi);
}
/* Converts a binary isl_ast_expr_op expression E to a GCC expression tree of
type TYPE. */
-tree
-translate_isl_ast_to_gimple::
+tree translate_isl_ast_to_gimple::
binary_op_to_tree (tree type, __isl_take isl_ast_expr *expr, ivs_params &ip)
{
+ enum isl_ast_op_type expr_type = isl_ast_expr_get_op_type (expr);
isl_ast_expr *arg_expr = isl_ast_expr_get_op_arg (expr, 0);
tree tree_lhs_expr = gcc_expression_from_isl_expression (type, arg_expr, ip);
arg_expr = isl_ast_expr_get_op_arg (expr, 1);
- tree tree_rhs_expr = gcc_expression_from_isl_expression (type, arg_expr, ip);
- enum isl_ast_op_type expr_type = isl_ast_expr_get_op_type (expr);
isl_ast_expr_free (expr);
+
+ /* From our constraint generation we may get modulo operations that
+ we cannot represent explicitely but that are no-ops for TYPE.
+ Elide those. */
+ if ((expr_type == isl_ast_op_pdiv_r
+ || expr_type == isl_ast_op_zdiv_r
+ || expr_type == isl_ast_op_add)
+ && isl_ast_expr_get_type (arg_expr) == isl_ast_expr_int
+ && (wi::exact_log2 (widest_int_from_isl_expr_int (arg_expr))
+ >= TYPE_PRECISION (type)))
+ {
+ isl_ast_expr_free (arg_expr);
+ return tree_lhs_expr;
+ }
+
+ tree tree_rhs_expr = gcc_expression_from_isl_expression (type, arg_expr, ip);
+ if (codegen_error_p ())
+ return NULL_TREE;
+
switch (expr_type)
{
case isl_ast_op_add:
case isl_ast_op_pdiv_q:
return fold_build2 (TRUNC_DIV_EXPR, type, tree_lhs_expr, tree_rhs_expr);
+ case isl_ast_op_zdiv_r:
case isl_ast_op_pdiv_r:
return fold_build2 (TRUNC_MOD_EXPR, type, tree_lhs_expr, tree_rhs_expr);
/* Converts a ternary isl_ast_expr_op expression E to a GCC expression tree of
type TYPE. */
-tree
-translate_isl_ast_to_gimple::
+tree translate_isl_ast_to_gimple::
ternary_op_to_tree (tree type, __isl_take isl_ast_expr *expr, ivs_params &ip)
{
- gcc_assert (isl_ast_expr_get_op_type (expr) == isl_ast_op_minus);
+ enum isl_ast_op_type t = isl_ast_expr_get_op_type (expr);
+ gcc_assert (t == isl_ast_op_cond || t == isl_ast_op_select);
isl_ast_expr *arg_expr = isl_ast_expr_get_op_arg (expr, 0);
- tree tree_first_expr
- = gcc_expression_from_isl_expression (type, arg_expr, ip);
+ tree a = gcc_expression_from_isl_expression (type, arg_expr, ip);
arg_expr = isl_ast_expr_get_op_arg (expr, 1);
- tree tree_second_expr
- = gcc_expression_from_isl_expression (type, arg_expr, ip);
+ tree b = gcc_expression_from_isl_expression (type, arg_expr, ip);
arg_expr = isl_ast_expr_get_op_arg (expr, 2);
- tree tree_third_expr
- = gcc_expression_from_isl_expression (type, arg_expr, ip);
+ tree c = gcc_expression_from_isl_expression (type, arg_expr, ip);
isl_ast_expr_free (expr);
- return fold_build3 (COND_EXPR, type, tree_first_expr,
- tree_second_expr, tree_third_expr);
+
+ if (codegen_error_p ())
+ return NULL_TREE;
+
+ return fold_build3 (COND_EXPR, type, a,
+ rewrite_to_non_trapping_overflow (b),
+ rewrite_to_non_trapping_overflow (c));
}
/* Converts a unary isl_ast_expr_op expression E to a GCC expression tree of
type TYPE. */
-tree
-translate_isl_ast_to_gimple::
+tree translate_isl_ast_to_gimple::
unary_op_to_tree (tree type, __isl_take isl_ast_expr *expr, ivs_params &ip)
{
gcc_assert (isl_ast_expr_get_op_type (expr) == isl_ast_op_minus);
isl_ast_expr *arg_expr = isl_ast_expr_get_op_arg (expr, 0);
tree tree_expr = gcc_expression_from_isl_expression (type, arg_expr, ip);
isl_ast_expr_free (expr);
- return fold_build1 (NEGATE_EXPR, type, tree_expr);
+ return codegen_error_p () ? NULL_TREE
+ : fold_build1 (NEGATE_EXPR, type, tree_expr);
}
/* Converts an isl_ast_expr_op expression E with unknown number of arguments
to a GCC expression tree of type TYPE. */
-tree
-translate_isl_ast_to_gimple::
+tree translate_isl_ast_to_gimple::
nary_op_to_tree (tree type, __isl_take isl_ast_expr *expr, ivs_params &ip)
{
enum tree_code op_code;
}
isl_ast_expr *arg_expr = isl_ast_expr_get_op_arg (expr, 0);
tree res = gcc_expression_from_isl_expression (type, arg_expr, ip);
+
+ if (codegen_error_p ())
+ {
+ isl_ast_expr_free (expr);
+ return NULL_TREE;
+ }
+
int i;
for (i = 1; i < isl_ast_expr_get_op_n_arg (expr); i++)
{
arg_expr = isl_ast_expr_get_op_arg (expr, i);
tree t = gcc_expression_from_isl_expression (type, arg_expr, ip);
+
+ if (codegen_error_p ())
+ {
+ isl_ast_expr_free (expr);
+ return NULL_TREE;
+ }
+
res = fold_build2 (op_code, type, res, t);
}
isl_ast_expr_free (expr);
/* Converts an isl_ast_expr_op expression E to a GCC expression tree of
type TYPE. */
-tree
-translate_isl_ast_to_gimple::
+tree translate_isl_ast_to_gimple::
gcc_expression_from_isl_expr_op (tree type, __isl_take isl_ast_expr *expr,
ivs_params &ip)
{
+ if (codegen_error_p ())
+ {
+ isl_ast_expr_free (expr);
+ return NULL_TREE;
+ }
+
gcc_assert (isl_ast_expr_get_type (expr) == isl_ast_expr_op);
switch (isl_ast_expr_get_op_type (expr))
{
case isl_ast_op_call:
case isl_ast_op_and_then:
case isl_ast_op_or_else:
- case isl_ast_op_select:
gcc_unreachable ();
case isl_ast_op_max:
case isl_ast_op_pdiv_q:
case isl_ast_op_pdiv_r:
case isl_ast_op_fdiv_q:
+ case isl_ast_op_zdiv_r:
case isl_ast_op_and:
case isl_ast_op_or:
case isl_ast_op_eq:
return unary_op_to_tree (type, expr, ip);
case isl_ast_op_cond:
+ case isl_ast_op_select:
return ternary_op_to_tree (type, expr, ip);
default:
return NULL_TREE;
}
-/* Converts an ISL AST expression E back to a GCC expression tree of
+/* Converts an isl AST expression E back to a GCC expression tree of
type TYPE. */
-tree
-translate_isl_ast_to_gimple::
+tree translate_isl_ast_to_gimple::
gcc_expression_from_isl_expression (tree type, __isl_take isl_ast_expr *expr,
ivs_params &ip)
{
+ if (codegen_error_p ())
+ {
+ isl_ast_expr_free (expr);
+ return NULL_TREE;
+ }
+
switch (isl_ast_expr_get_type (expr))
{
case isl_ast_expr_id:
induction variable for the new LOOP. New LOOP is attached to CFG
starting at ENTRY_EDGE. LOOP is inserted into the loop tree and
becomes the child loop of the OUTER_LOOP. NEWIVS_INDEX binds
- ISL's scattering name to the induction variable created for the
+ isl's scattering name to the induction variable created for the
loop of STMT. The new induction variable is inserted in the NEWIVS
vector and is of type TYPE. */
-struct loop *
-translate_isl_ast_to_gimple::
+struct loop *translate_isl_ast_to_gimple::
graphite_create_new_loop (edge entry_edge, __isl_keep isl_ast_node *node_for,
loop_p outer, tree type, tree lb, tree ub,
ivs_params &ip)
{
isl_ast_expr *for_inc = isl_ast_node_for_get_inc (node_for);
tree stride = gcc_expression_from_isl_expression (type, for_inc, ip);
+
+ /* To fail code generation, we generate wrong code until we discard it. */
+ if (codegen_error_p ())
+ stride = integer_zero_node;
+
tree ivvar = create_tmp_var (type, "graphite_IV");
tree iv, iv_after_increment;
loop_p loop = create_empty_loop_on_edge
- NEXT_E is the edge where new generated code should be attached. */
-edge
-translate_isl_ast_to_gimple::
+edge translate_isl_ast_to_gimple::
translate_isl_ast_for_loop (loop_p context_loop,
__isl_keep isl_ast_node *node_for, edge next_e,
tree type, tree lb, tree ub,
edge to_body = single_succ_edge (loop->header);
basic_block after = to_body->dest;
- /* Create a basic block for loop close phi nodes. */
- last_e = single_succ_edge (split_edge (last_e));
-
/* Translate the body of the loop. */
isl_ast_node *for_body = isl_ast_node_for_get_body (node_for);
next_e = translate_isl_ast (loop, for_body, to_body, ip);
isl_ast_node_free (for_body);
- redirect_edge_succ_nodup (next_e, after);
+
+ /* Early return if we failed to translate loop body. */
+ if (!next_e || codegen_error_p ())
+ return NULL;
+
+ if (next_e->dest != after)
+ redirect_edge_succ_nodup (next_e, after);
set_immediate_dominator (CDI_DOMINATORS, next_e->dest, next_e->src);
if (flag_loop_parallelize_all)
- {
- isl_id *id = isl_ast_node_get_annotation (node_for);
- gcc_assert (id);
- ast_build_info *for_info = (ast_build_info *) isl_id_get_user (id);
- loop->can_be_parallel = for_info->is_parallelizable;
- free (for_info);
- isl_id_free (id);
- }
+ {
+ isl_id *id = isl_ast_node_get_annotation (node_for);
+ gcc_assert (id);
+ ast_build_info *for_info = (ast_build_info *) isl_id_get_user (id);
+ loop->can_be_parallel = for_info->is_parallelizable;
+ free (for_info);
+ isl_id_free (id);
+ }
return last_e;
}
{
- ...
+ ...
}
case isl_ast_op_lt:
{
- // (iterator < ub) => (iterator <= ub - 1)
+ /* (iterator < ub) => (iterator <= ub - 1). */
isl_val *one =
isl_val_int_from_si (isl_ast_expr_get_ctx (for_cond), 1);
isl_ast_expr *ub = isl_ast_expr_get_op_arg (for_cond, 1);
return res;
}
-/* All loops generated by create_empty_loop_on_edge have the form of
- a post-test loop:
-
- do
-
- {
- body of the loop;
- } while (lower bound < upper bound);
-
- We create a new if region protecting the loop to be executed, if
- the execution count is zero (lower bound > upper bound). */
+/* Translates an isl_ast_node_for to Gimple. */
-edge
-translate_isl_ast_to_gimple::
-graphite_create_new_loop_guard (edge entry_edge,
- __isl_keep isl_ast_node *node_for, tree *type,
- tree *lb, tree *ub, ivs_params &ip)
+edge translate_isl_ast_to_gimple::
+translate_isl_ast_node_for (loop_p context_loop, __isl_keep isl_ast_node *node,
+ edge next_e, ivs_params &ip)
{
- gcc_assert (isl_ast_node_get_type (node_for) == isl_ast_node_for);
- tree cond_expr;
- edge exit_edge;
-
- *type =
- build_nonstandard_integer_type (graphite_expression_type_precision, 0);
- isl_ast_expr *for_init = isl_ast_node_for_get_init (node_for);
- *lb = gcc_expression_from_isl_expression (*type, for_init, ip);
- isl_ast_expr *upper_bound = get_upper_bound (node_for);
- *ub = gcc_expression_from_isl_expression (*type, upper_bound, ip);
-
- /* When ub is simply a constant or a parameter, use lb <= ub. */
- if (TREE_CODE (*ub) == INTEGER_CST || TREE_CODE (*ub) == SSA_NAME)
- cond_expr = fold_build2 (LE_EXPR, boolean_type_node, *lb, *ub);
+ gcc_assert (isl_ast_node_get_type (node) == isl_ast_node_for);
+ tree type = graphite_expr_type;
+
+ isl_ast_expr *for_init = isl_ast_node_for_get_init (node);
+ tree lb = gcc_expression_from_isl_expression (type, for_init, ip);
+ /* To fail code generation, we generate wrong code until we discard it. */
+ if (codegen_error_p ())
+ lb = integer_zero_node;
+
+ isl_ast_expr *upper_bound = get_upper_bound (node);
+ tree ub = gcc_expression_from_isl_expression (type, upper_bound, ip);
+ /* To fail code generation, we generate wrong code until we discard it. */
+ if (codegen_error_p ())
+ ub = integer_zero_node;
+
+ edge last_e = single_succ_edge (split_edge (next_e));
+
+ /* Compensate for the fact that we emit a do { } while loop from
+ a for ISL AST.
+ ??? We often miss constraints on niter because the SESE region
+ doesn't cover loop header copies. Ideally we'd add constraints
+ for all relevant dominating conditions. */
+ if (TREE_CODE (lb) == INTEGER_CST && TREE_CODE (ub) == INTEGER_CST
+ && tree_int_cst_compare (lb, ub) <= 0)
+ ;
else
{
- tree one = (POINTER_TYPE_P (*type)
- ? convert_to_ptrofftype (integer_one_node)
- : fold_convert (*type, integer_one_node));
+ tree one = build_one_cst (POINTER_TYPE_P (type) ? sizetype : type);
/* Adding +1 and using LT_EXPR helps with loop latches that have a
loop iteration count of "PARAMETER - 1". For PARAMETER == 0 this
becomes 2^k-1 due to integer overflow, and the condition lb <= ub
is true, even if we do not want this. However lb < ub + 1 is false,
as expected. */
- tree ub_one = fold_build2 (POINTER_TYPE_P (*type) ? POINTER_PLUS_EXPR
- : PLUS_EXPR, *type, *ub, one);
-
- cond_expr = fold_build2 (LT_EXPR, boolean_type_node, *lb, ub_one);
+ tree ub_one = fold_build2 (POINTER_TYPE_P (type)
+ ? POINTER_PLUS_EXPR : PLUS_EXPR,
+ type, unshare_expr (ub), one);
+ create_empty_if_region_on_edge (next_e,
+ fold_build2 (LT_EXPR, boolean_type_node,
+ unshare_expr (lb), ub_one));
+ next_e = get_true_edge_from_guard_bb (next_e->dest);
}
- exit_edge = create_empty_if_region_on_edge (entry_edge, cond_expr);
-
- return exit_edge;
-}
-
-/* Translates an isl_ast_node_for to Gimple. */
-
-edge
-translate_isl_ast_to_gimple::
-translate_isl_ast_node_for (loop_p context_loop, __isl_keep isl_ast_node *node,
- edge next_e, ivs_params &ip)
-{
- gcc_assert (isl_ast_node_get_type (node) == isl_ast_node_for);
- tree type, lb, ub;
- edge last_e = graphite_create_new_loop_guard (next_e, node, &type,
- &lb, &ub, ip);
- edge true_e = get_true_edge_from_guard_bb (next_e->dest);
-
- translate_isl_ast_for_loop (context_loop, node, true_e,
+ translate_isl_ast_for_loop (context_loop, node, next_e,
type, lb, ub, ip);
return last_e;
}
chrec, we could consider using a map<int, tree> that maps loop ids to the
corresponding tree expressions. */
-void
-translate_isl_ast_to_gimple::
+void translate_isl_ast_to_gimple::
build_iv_mapping (vec<tree> iv_map, gimple_poly_bb_p gbb,
__isl_keep isl_ast_expr *user_expr, ivs_params &ip,
- sese region)
+ sese_l ®ion)
{
gcc_assert (isl_ast_expr_get_type (user_expr) == isl_ast_expr_op &&
isl_ast_expr_get_op_type (user_expr) == isl_ast_op_call);
for (i = 1; i < isl_ast_expr_get_op_n_arg (user_expr); i++)
{
arg_expr = isl_ast_expr_get_op_arg (user_expr, i);
- tree type =
- build_nonstandard_integer_type (graphite_expression_type_precision, 0);
+ tree type = graphite_expr_type;
tree t = gcc_expression_from_isl_expression (type, arg_expr, ip);
+
+ /* To fail code generation, we generate wrong code until we discard it. */
+ if (codegen_error_p ())
+ t = integer_zero_node;
+
loop_p old_loop = gbb_loop_at_index (gbb, region, i - 1);
iv_map[old_loop->num] = t;
}
-
}
/* Translates an isl_ast_node_user to Gimple.
FIXME: We should remove iv_map.create (loop->num + 1), if it is possible. */
-edge
-translate_isl_ast_to_gimple::
+edge translate_isl_ast_to_gimple::
translate_isl_ast_node_user (__isl_keep isl_ast_node *node,
edge next_e, ivs_params &ip)
{
gcc_assert (isl_ast_node_get_type (node) == isl_ast_node_user);
+
isl_ast_expr *user_expr = isl_ast_node_user_get_expr (node);
isl_ast_expr *name_expr = isl_ast_expr_get_op_arg (user_expr, 0);
gcc_assert (isl_ast_expr_get_type (name_expr) == isl_ast_expr_id);
+
isl_id *name_id = isl_ast_expr_get_id (name_expr);
poly_bb_p pbb = (poly_bb_p) isl_id_get_user (name_id);
gcc_assert (pbb);
+
gimple_poly_bb_p gbb = PBB_BLACK_BOX (pbb);
- vec<tree> iv_map;
+
isl_ast_expr_free (name_expr);
isl_id_free (name_id);
gcc_assert (GBB_BB (gbb) != ENTRY_BLOCK_PTR_FOR_FN (cfun) &&
"The entry block should not even appear within a scop");
- int nb_loops = number_of_loops (cfun);
+ const int nb_loops = number_of_loops (cfun);
+ vec<tree> iv_map;
iv_map.create (nb_loops);
iv_map.safe_grow_cleared (nb_loops);
- build_iv_mapping (iv_map, gbb, user_expr, ip, SCOP_REGION (pbb->scop));
+ build_iv_mapping (iv_map, gbb, user_expr, ip, pbb->scop->scop_info->region);
isl_ast_expr_free (user_expr);
- next_e = copy_bb_and_scalar_dependences (GBB_BB (gbb),
- SCOP_REGION (pbb->scop), next_e,
- iv_map,
- &graphite_regenerate_error);
+
+ basic_block old_bb = GBB_BB (gbb);
+ if (dump_file && (dump_flags & TDF_DETAILS))
+ {
+ fprintf (dump_file,
+ "[codegen] copying from bb_%d on edge (bb_%d, bb_%d)\n",
+ old_bb->index, next_e->src->index, next_e->dest->index);
+ print_loops_bb (dump_file, GBB_BB (gbb), 0, 3);
+ }
+
+ next_e = copy_bb_and_scalar_dependences (old_bb, next_e, iv_map);
+
iv_map.release ();
- mark_virtual_operands_for_renaming (cfun);
- update_ssa (TODO_update_ssa);
+
+ if (codegen_error_p ())
+ return NULL;
+
+ if (dump_file && (dump_flags & TDF_DETAILS))
+ {
+ fprintf (dump_file, "[codegen] (after copy) new basic block\n");
+ print_loops_bb (dump_file, next_e->src, 0, 3);
+ }
+
return next_e;
}
/* Translates an isl_ast_node_block to Gimple. */
-edge
-translate_isl_ast_to_gimple::
+edge translate_isl_ast_to_gimple::
translate_isl_ast_node_block (loop_p context_loop,
__isl_keep isl_ast_node *node,
edge next_e, ivs_params &ip)
return next_e;
}
-/* Creates a new if region corresponding to ISL's cond. */
+/* Creates a new if region corresponding to isl's cond. */
-edge
-translate_isl_ast_to_gimple::
+edge translate_isl_ast_to_gimple::
graphite_create_new_guard (edge entry_edge, __isl_take isl_ast_expr *if_cond,
ivs_params &ip)
{
- tree type =
- build_nonstandard_integer_type (graphite_expression_type_precision, 0);
+ tree type = graphite_expr_type;
tree cond_expr = gcc_expression_from_isl_expression (type, if_cond, ip);
+
+ /* To fail code generation, we generate wrong code until we discard it. */
+ if (codegen_error_p ())
+ cond_expr = integer_zero_node;
+
edge exit_edge = create_empty_if_region_on_edge (entry_edge, cond_expr);
return exit_edge;
}
/* Translates an isl_ast_node_if to Gimple. */
-edge
-translate_isl_ast_to_gimple::
+edge translate_isl_ast_to_gimple::
translate_isl_ast_node_if (loop_p context_loop,
__isl_keep isl_ast_node *node,
edge next_e, ivs_params &ip)
gcc_assert (isl_ast_node_get_type (node) == isl_ast_node_if);
isl_ast_expr *if_cond = isl_ast_node_if_get_cond (node);
edge last_e = graphite_create_new_guard (next_e, if_cond, ip);
-
edge true_e = get_true_edge_from_guard_bb (next_e->dest);
+ merge_points.safe_push (last_e);
+
isl_ast_node *then_node = isl_ast_node_if_get_then (node);
translate_isl_ast (context_loop, then_node, true_e, ip);
isl_ast_node_free (then_node);
isl_ast_node *else_node = isl_ast_node_if_get_else (node);
if (isl_ast_node_get_type (else_node) != isl_ast_node_error)
translate_isl_ast (context_loop, else_node, false_e, ip);
+
isl_ast_node_free (else_node);
return last_e;
}
-/* Translates an ISL AST node NODE to GCC representation in the
+/* Translates an isl AST node NODE to GCC representation in the
context of a SESE. */
-edge
-translate_isl_ast_to_gimple::translate_isl_ast (loop_p context_loop,
- __isl_keep isl_ast_node *node,
- edge next_e, ivs_params &ip)
+edge translate_isl_ast_to_gimple::
+translate_isl_ast (loop_p context_loop, __isl_keep isl_ast_node *node,
+ edge next_e, ivs_params &ip)
{
+ if (codegen_error_p ())
+ return NULL;
+
switch (isl_ast_node_get_type (node))
{
case isl_ast_node_error:
return translate_isl_ast_node_block (context_loop, node,
next_e, ip);
+ case isl_ast_node_mark:
+ {
+ isl_ast_node *n = isl_ast_node_mark_get_node (node);
+ edge e = translate_isl_ast (context_loop, n, next_e, ip);
+ isl_ast_node_free (n);
+ return e;
+ }
+
default:
gcc_unreachable ();
}
}
-/* Prints NODE to FILE. */
+/* Register in RENAME_MAP the rename tuple (OLD_NAME, EXPR).
+ When OLD_NAME and EXPR are the same we assert. */
-void
-print_isl_ast_node (FILE *file, __isl_keep isl_ast_node *node,
- __isl_keep isl_ctx *ctx)
+void translate_isl_ast_to_gimple::
+set_rename (tree old_name, tree expr)
{
- isl_printer *prn = isl_printer_to_file (ctx, file);
- prn = isl_printer_set_output_format (prn, ISL_FORMAT_C);
- prn = isl_printer_print_ast_node (prn, node);
- prn = isl_printer_print_str (prn, "\n");
- isl_printer_free (prn);
+ if (dump_file)
+ {
+ fprintf (dump_file, "[codegen] setting rename: old_name = ");
+ print_generic_expr (dump_file, old_name);
+ fprintf (dump_file, ", new decl = ");
+ print_generic_expr (dump_file, expr);
+ fprintf (dump_file, "\n");
+ }
+ bool res = region->rename_map->put (old_name, expr);
+ gcc_assert (! res);
}
-/* Add ISL's parameter identifiers and corresponding.trees to ivs_params */
+/* Return an iterator to the instructions comes last in the execution order.
+ Either GSI1 and GSI2 should belong to the same basic block or one of their
+ respective basic blocks should dominate the other. */
-static void
-add_parameters_to_ivs_params (scop_p scop, ivs_params &ip)
+gimple_stmt_iterator
+later_of_the_two (gimple_stmt_iterator gsi1, gimple_stmt_iterator gsi2)
{
- sese region = SCOP_REGION (scop);
- unsigned nb_parameters = isl_set_dim (scop->param_context, isl_dim_param);
- gcc_assert (nb_parameters == SESE_PARAMS (region).length ());
- unsigned i;
- for (i = 0; i < nb_parameters; i++)
+ basic_block bb1 = gsi_bb (gsi1);
+ basic_block bb2 = gsi_bb (gsi2);
+
+ /* Find the iterator which is the latest. */
+ if (bb1 == bb2)
{
- isl_id *tmp_id = isl_set_get_dim_id (scop->param_context,
- isl_dim_param, i);
- ip[tmp_id] = SESE_PARAMS (region)[i];
+ gimple *stmt1 = gsi_stmt (gsi1);
+ gimple *stmt2 = gsi_stmt (gsi2);
+
+ if (stmt1 != NULL && stmt2 != NULL)
+ {
+ bool is_phi1 = gimple_code (stmt1) == GIMPLE_PHI;
+ bool is_phi2 = gimple_code (stmt2) == GIMPLE_PHI;
+
+ if (is_phi1 != is_phi2)
+ return is_phi1 ? gsi2 : gsi1;
+ }
+
+ /* For empty basic blocks gsis point to the end of the sequence. Since
+ there is no operator== defined for gimple_stmt_iterator and for gsis
+ not pointing to a valid statement gsi_next would assert. */
+ gimple_stmt_iterator gsi = gsi1;
+ do {
+ if (gsi_stmt (gsi) == gsi_stmt (gsi2))
+ return gsi2;
+ gsi_next (&gsi);
+ } while (!gsi_end_p (gsi));
+
+ return gsi1;
}
-}
+ /* Find the basic block closest to the basic block which defines stmt. */
+ if (dominated_by_p (CDI_DOMINATORS, bb1, bb2))
+ return gsi1;
-/* Generates a build, which specifies the constraints on the parameters. */
+ gcc_assert (dominated_by_p (CDI_DOMINATORS, bb2, bb1));
+ return gsi2;
+}
-static __isl_give isl_ast_build *
-generate_isl_context (scop_p scop)
+/* Insert each statement from SEQ at its earliest insertion p. */
+
+void translate_isl_ast_to_gimple::
+gsi_insert_earliest (gimple_seq seq)
{
- isl_set *context_isl = isl_set_params (isl_set_copy (scop->param_context));
- return isl_ast_build_from_context (context_isl);
+ update_modified_stmts (seq);
+ sese_l &codegen_region = region->if_region->true_region->region;
+ basic_block begin_bb = get_entry_bb (codegen_region);
+
+ /* Inserting the gimple statements in a vector because gimple_seq behave
+ in strage ways when inserting the stmts from it into different basic
+ blocks one at a time. */
+ auto_vec<gimple *, 3> stmts;
+ for (gimple_stmt_iterator gsi = gsi_start (seq); !gsi_end_p (gsi);
+ gsi_next (&gsi))
+ stmts.safe_push (gsi_stmt (gsi));
+
+ int i;
+ gimple *use_stmt;
+ FOR_EACH_VEC_ELT (stmts, i, use_stmt)
+ {
+ gcc_assert (gimple_code (use_stmt) != GIMPLE_PHI);
+ gimple_stmt_iterator gsi_def_stmt = gsi_start_nondebug_bb (begin_bb);
+
+ use_operand_p use_p;
+ ssa_op_iter op_iter;
+ FOR_EACH_SSA_USE_OPERAND (use_p, use_stmt, op_iter, SSA_OP_USE)
+ {
+ /* Iterator to the current def of use_p. For function parameters or
+ anything where def is not found, insert at the beginning of the
+ generated region. */
+ gimple_stmt_iterator gsi_stmt = gsi_def_stmt;
+
+ tree op = USE_FROM_PTR (use_p);
+ gimple *stmt = SSA_NAME_DEF_STMT (op);
+ if (stmt && (gimple_code (stmt) != GIMPLE_NOP))
+ gsi_stmt = gsi_for_stmt (stmt);
+
+ /* For region parameters, insert at the beginning of the generated
+ region. */
+ if (!bb_in_sese_p (gsi_bb (gsi_stmt), codegen_region))
+ gsi_stmt = gsi_def_stmt;
+
+ gsi_def_stmt = later_of_the_two (gsi_stmt, gsi_def_stmt);
+ }
+
+ if (!gsi_stmt (gsi_def_stmt))
+ {
+ gimple_stmt_iterator gsi = gsi_after_labels (gsi_bb (gsi_def_stmt));
+ gsi_insert_before (&gsi, use_stmt, GSI_NEW_STMT);
+ }
+ else if (gimple_code (gsi_stmt (gsi_def_stmt)) == GIMPLE_PHI)
+ {
+ gimple_stmt_iterator bsi
+ = gsi_start_nondebug_bb (gsi_bb (gsi_def_stmt));
+ /* Insert right after the PHI statements. */
+ gsi_insert_before (&bsi, use_stmt, GSI_NEW_STMT);
+ }
+ else
+ gsi_insert_after (&gsi_def_stmt, use_stmt, GSI_NEW_STMT);
+
+ if (dump_file)
+ {
+ fprintf (dump_file, "[codegen] inserting statement in BB %d: ",
+ gimple_bb (use_stmt)->index);
+ print_gimple_stmt (dump_file, use_stmt, 0, TDF_VOPS | TDF_MEMSYMS);
+ }
+ }
}
-/* Get the maximal number of schedule dimensions in the scop SCOP. */
+/* For ops which are scev_analyzeable, we can regenerate a new name from its
+ scalar evolution around LOOP. */
-static
-int get_max_schedule_dimensions (scop_p scop)
+tree translate_isl_ast_to_gimple::
+get_rename_from_scev (tree old_name, gimple_seq *stmts, loop_p loop,
+ vec<tree> iv_map)
{
- int i;
- poly_bb_p pbb;
- int schedule_dims = 0;
+ tree scev = cached_scalar_evolution_in_region (region->region,
+ loop, old_name);
+
+ /* At this point we should know the exact scev for each
+ scalar SSA_NAME used in the scop: all the other scalar
+ SSA_NAMEs should have been translated out of SSA using
+ arrays with one element. */
+ tree new_expr;
+ if (chrec_contains_undetermined (scev))
+ {
+ set_codegen_error ();
+ return build_zero_cst (TREE_TYPE (old_name));
+ }
- FOR_EACH_VEC_ELT (SCOP_BBS (scop), i, pbb)
+ new_expr = chrec_apply_map (scev, iv_map);
+
+ /* The apply should produce an expression tree containing
+ the uses of the new induction variables. We should be
+ able to use new_expr instead of the old_name in the newly
+ generated loop nest. */
+ if (chrec_contains_undetermined (new_expr)
+ || tree_contains_chrecs (new_expr, NULL))
{
- int pbb_schedule_dims = isl_map_dim (pbb->transformed, isl_dim_out);
- if (pbb_schedule_dims > schedule_dims)
- schedule_dims = pbb_schedule_dims;
+ set_codegen_error ();
+ return build_zero_cst (TREE_TYPE (old_name));
}
- return schedule_dims;
+ /* Replace the old_name with the new_expr. */
+ return force_gimple_operand (unshare_expr (new_expr), stmts,
+ true, NULL_TREE);
}
-/* Extend the schedule to NB_SCHEDULE_DIMS schedule dimensions.
- For schedules with different dimensionality, the isl AST generator can not
- define an order and will just randomly choose an order. The solution to this
- problem is to extend all schedules to the maximal number of schedule
- dimensions (using '0's for the remaining values). */
+/* Return true if STMT should be copied from region to the new code-generated
+ region. LABELs, CONDITIONS, induction-variables and region parameters need
+ not be copied. */
-static __isl_give isl_map *
-extend_schedule (__isl_take isl_map *schedule, int nb_schedule_dims)
+static bool
+should_copy_to_new_region (gimple *stmt, sese_info_p region)
{
- int tmp_dims = isl_map_dim (schedule, isl_dim_out);
- schedule =
- isl_map_add_dims (schedule, isl_dim_out, nb_schedule_dims - tmp_dims);
- isl_val *zero =
- isl_val_int_from_si (isl_map_get_ctx (schedule), 0);
- int i;
- for (i = tmp_dims; i < nb_schedule_dims; i++)
+ /* Do not copy labels or conditions. */
+ if (gimple_code (stmt) == GIMPLE_LABEL
+ || gimple_code (stmt) == GIMPLE_COND)
+ return false;
+
+ tree lhs;
+ /* Do not copy induction variables. */
+ if (is_gimple_assign (stmt)
+ && (lhs = gimple_assign_lhs (stmt))
+ && TREE_CODE (lhs) == SSA_NAME
+ && scev_analyzable_p (lhs, region->region)
+ /* But to code-generate liveouts - liveout PHI generation is
+ in generic sese.c code that cannot do code generation. */
+ && ! bitmap_bit_p (region->liveout, SSA_NAME_VERSION (lhs)))
+ return false;
+
+ return true;
+}
+
+/* Duplicates the statements of basic block BB into basic block NEW_BB
+ and compute the new induction variables according to the IV_MAP. */
+
+void translate_isl_ast_to_gimple::
+graphite_copy_stmts_from_block (basic_block bb, basic_block new_bb,
+ vec<tree> iv_map)
+{
+ /* Iterator poining to the place where new statement (s) will be inserted. */
+ gimple_stmt_iterator gsi_tgt = gsi_last_bb (new_bb);
+
+ for (gimple_stmt_iterator gsi = gsi_start_bb (bb); !gsi_end_p (gsi);
+ gsi_next (&gsi))
{
- schedule =
- isl_map_fix_val (schedule, isl_dim_out, i, isl_val_copy (zero));
+ gimple *stmt = gsi_stmt (gsi);
+ if (!should_copy_to_new_region (stmt, region))
+ continue;
+
+ /* Create a new copy of STMT and duplicate STMT's virtual
+ operands. */
+ gimple *copy = gimple_copy (stmt);
+
+ /* Rather than not copying debug stmts we reset them.
+ ??? Where we can rewrite uses without inserting new
+ stmts we could simply do that. */
+ if (is_gimple_debug (copy))
+ {
+ if (gimple_debug_bind_p (copy))
+ gimple_debug_bind_reset_value (copy);
+ else if (gimple_debug_source_bind_p (copy)
+ || gimple_debug_nonbind_marker_p (copy))
+ ;
+ else
+ gcc_unreachable ();
+ }
+
+ maybe_duplicate_eh_stmt (copy, stmt);
+ gimple_duplicate_stmt_histograms (cfun, copy, cfun, stmt);
+
+ /* Crete new names for each def in the copied stmt. */
+ def_operand_p def_p;
+ ssa_op_iter op_iter;
+ FOR_EACH_SSA_DEF_OPERAND (def_p, copy, op_iter, SSA_OP_ALL_DEFS)
+ {
+ tree old_name = DEF_FROM_PTR (def_p);
+ create_new_def_for (old_name, copy, def_p);
+ }
+
+ gsi_insert_after (&gsi_tgt, copy, GSI_NEW_STMT);
+ if (dump_file)
+ {
+ fprintf (dump_file, "[codegen] inserting statement: ");
+ print_gimple_stmt (dump_file, copy, 0);
+ }
+
+ /* For each SCEV analyzable SSA_NAME, rename their usage. */
+ ssa_op_iter iter;
+ use_operand_p use_p;
+ if (!is_gimple_debug (copy))
+ {
+ bool changed = false;
+ FOR_EACH_SSA_USE_OPERAND (use_p, copy, iter, SSA_OP_USE)
+ {
+ tree old_name = USE_FROM_PTR (use_p);
+
+ if (TREE_CODE (old_name) != SSA_NAME
+ || SSA_NAME_IS_DEFAULT_DEF (old_name)
+ || ! scev_analyzable_p (old_name, region->region))
+ continue;
+
+ gimple_seq stmts = NULL;
+ tree new_name = get_rename_from_scev (old_name, &stmts,
+ bb->loop_father, iv_map);
+ if (! codegen_error_p ())
+ gsi_insert_earliest (stmts);
+ replace_exp (use_p, new_name);
+ changed = true;
+ }
+ if (changed)
+ fold_stmt_inplace (&gsi_tgt);
+ }
+
+ update_stmt (copy);
}
- isl_val_free (zero);
- return schedule;
}
-/* Generates a schedule, which specifies an order used to
- visit elements in a domain. */
-static __isl_give isl_union_map *
-generate_isl_schedule (scop_p scop)
-{
- int nb_schedule_dims = get_max_schedule_dimensions (scop);
- int i;
- poly_bb_p pbb;
- isl_union_map *schedule_isl =
- isl_union_map_empty (isl_set_get_space (scop->param_context));
+/* Copies BB and includes in the copied BB all the statements that can
+ be reached following the use-def chains from the memory accesses,
+ and returns the next edge following this new block. */
- FOR_EACH_VEC_ELT (SCOP_BBS (scop), i, pbb)
+edge translate_isl_ast_to_gimple::
+copy_bb_and_scalar_dependences (basic_block bb, edge next_e, vec<tree> iv_map)
+{
+ basic_block new_bb = split_edge (next_e);
+ gimple_stmt_iterator gsi_tgt = gsi_last_bb (new_bb);
+ for (gphi_iterator psi = gsi_start_phis (bb); !gsi_end_p (psi);
+ gsi_next (&psi))
{
- /* Dead code elimination: when the domain of a PBB is empty,
- don't generate code for the PBB. */
- if (isl_set_is_empty (pbb->domain))
+ gphi *phi = psi.phi ();
+ tree res = gimple_phi_result (phi);
+ if (virtual_operand_p (res)
+ || scev_analyzable_p (res, region->region))
continue;
- isl_map *bb_schedule = isl_map_copy (pbb->transformed);
- bb_schedule = isl_map_intersect_domain (bb_schedule,
- isl_set_copy (pbb->domain));
- bb_schedule = extend_schedule (bb_schedule, nb_schedule_dims);
- schedule_isl =
- isl_union_map_union (schedule_isl,
- isl_union_map_from_map (bb_schedule));
+ tree new_phi_def;
+ tree *rename = region->rename_map->get (res);
+ if (! rename)
+ {
+ new_phi_def = create_tmp_reg (TREE_TYPE (res));
+ set_rename (res, new_phi_def);
+ }
+ else
+ new_phi_def = *rename;
+
+ gassign *ass = gimple_build_assign (NULL_TREE, new_phi_def);
+ create_new_def_for (res, ass, NULL);
+ gsi_insert_after (&gsi_tgt, ass, GSI_NEW_STMT);
}
- return schedule_isl;
+
+ graphite_copy_stmts_from_block (bb, new_bb, iv_map);
+
+ /* Insert out-of SSA copies on the original BB outgoing edges. */
+ gsi_tgt = gsi_last_bb (new_bb);
+ basic_block bb_for_succs = bb;
+ if (bb_for_succs == bb_for_succs->loop_father->latch
+ && bb_in_sese_p (bb_for_succs, region->region)
+ && sese_trivially_empty_bb_p (bb_for_succs))
+ bb_for_succs = NULL;
+ while (bb_for_succs)
+ {
+ basic_block latch = NULL;
+ edge_iterator ei;
+ edge e;
+ FOR_EACH_EDGE (e, ei, bb_for_succs->succs)
+ {
+ for (gphi_iterator psi = gsi_start_phis (e->dest); !gsi_end_p (psi);
+ gsi_next (&psi))
+ {
+ gphi *phi = psi.phi ();
+ tree res = gimple_phi_result (phi);
+ if (virtual_operand_p (res)
+ || scev_analyzable_p (res, region->region))
+ continue;
+
+ tree new_phi_def;
+ tree *rename = region->rename_map->get (res);
+ if (! rename)
+ {
+ new_phi_def = create_tmp_reg (TREE_TYPE (res));
+ set_rename (res, new_phi_def);
+ }
+ else
+ new_phi_def = *rename;
+
+ tree arg = PHI_ARG_DEF_FROM_EDGE (phi, e);
+ if (TREE_CODE (arg) == SSA_NAME
+ && scev_analyzable_p (arg, region->region))
+ {
+ gimple_seq stmts = NULL;
+ tree new_name = get_rename_from_scev (arg, &stmts,
+ bb->loop_father,
+ iv_map);
+ if (! codegen_error_p ())
+ gsi_insert_earliest (stmts);
+ arg = new_name;
+ }
+ gassign *ass = gimple_build_assign (new_phi_def, arg);
+ gsi_insert_after (&gsi_tgt, ass, GSI_NEW_STMT);
+ }
+ if (e->dest == bb_for_succs->loop_father->latch
+ && bb_in_sese_p (e->dest, region->region)
+ && sese_trivially_empty_bb_p (e->dest))
+ latch = e->dest;
+ }
+ bb_for_succs = latch;
+ }
+
+ return single_succ_edge (new_bb);
+}
+
+/* Add isl's parameter identifiers and corresponding trees to ivs_params. */
+
+void translate_isl_ast_to_gimple::
+add_parameters_to_ivs_params (scop_p scop, ivs_params &ip)
+{
+ sese_info_p region = scop->scop_info;
+ unsigned nb_parameters = isl_set_dim (scop->param_context, isl_dim_param);
+ gcc_assert (nb_parameters == sese_nb_params (region));
+ unsigned i;
+ tree param;
+ FOR_EACH_VEC_ELT (region->params, i, param)
+ {
+ isl_id *tmp_id = isl_set_get_dim_id (scop->param_context,
+ isl_dim_param, i);
+ ip[tmp_id] = param;
+ }
+}
+
+
+/* Generates a build, which specifies the constraints on the parameters. */
+
+__isl_give isl_ast_build *translate_isl_ast_to_gimple::
+generate_isl_context (scop_p scop)
+{
+ isl_set *context_isl = isl_set_params (isl_set_copy (scop->param_context));
+ return isl_ast_build_from_context (context_isl);
}
/* This method is executed before the construction of a for node. */
-static __isl_give isl_id *
+__isl_give isl_id *
ast_build_before_for (__isl_keep isl_ast_build *build, void *user)
{
isl_union_map *dependences = (isl_union_map *) user;
return id;
}
-/* Set the separate option for all dimensions.
- This helps to reduce control overhead. */
-
-static __isl_give isl_ast_build *
-set_options (__isl_take isl_ast_build *control,
- __isl_keep isl_union_map *schedule)
-{
- isl_ctx *ctx = isl_union_map_get_ctx (schedule);
- isl_space *range_space = isl_space_set_alloc (ctx, 0, 1);
- range_space =
- isl_space_set_tuple_name (range_space, isl_dim_set, "separate");
- isl_union_set *range =
- isl_union_set_from_set (isl_set_universe (range_space));
- isl_union_set *domain = isl_union_map_range (isl_union_map_copy (schedule));
- domain = isl_union_set_universe (domain);
- isl_union_map *options = isl_union_map_from_domain_and_range (domain, range);
- return isl_ast_build_set_options (control, options);
-}
+/* Generate isl AST from schedule of SCOP. */
-static __isl_give isl_ast_node *
-scop_to_isl_ast (scop_p scop, ivs_params &ip)
+__isl_give isl_ast_node *translate_isl_ast_to_gimple::
+scop_to_isl_ast (scop_p scop)
{
- /* Generate loop upper bounds that consist of the current loop iterator,
- an operator (< or <=) and an expression not involving the iterator.
- If this option is not set, then the current loop iterator may appear several
- times in the upper bound. See the isl manual for more details. */
- isl_options_set_ast_build_atomic_upper_bound (scop->isl_context, true);
-
- add_parameters_to_ivs_params (scop, ip);
- isl_union_map *schedule_isl = generate_isl_schedule (scop);
+ int old_err = isl_options_get_on_error (scop->isl_context);
+ int old_max_operations = isl_ctx_get_max_operations (scop->isl_context);
+ int max_operations = param_max_isl_operations;
+ if (max_operations)
+ isl_ctx_set_max_operations (scop->isl_context, max_operations);
+ isl_options_set_on_error (scop->isl_context, ISL_ON_ERROR_CONTINUE);
+
+ gcc_assert (scop->transformed_schedule);
+
+ /* Set the separate option to reduce control flow overhead. */
+ isl_schedule *schedule = isl_schedule_map_schedule_node_bottom_up
+ (isl_schedule_copy (scop->transformed_schedule), set_separate_option, NULL);
isl_ast_build *context_isl = generate_isl_context (scop);
- context_isl = set_options (context_isl, schedule_isl);
- isl_union_map *dependences = NULL;
- if (flag_loop_parallelize_all)
- {
- dependences = scop_get_dependences (scop);
- context_isl =
- isl_ast_build_set_before_each_for (context_isl, ast_build_before_for,
- dependences);
- }
- isl_ast_node *ast_isl = isl_ast_build_ast_from_schedule (context_isl,
- schedule_isl);
- if(dependences)
- isl_union_map_free (dependences);
- isl_ast_build_free (context_isl);
- return ast_isl;
-}
-
-/* Copy def from sese REGION to the newly created TO_REGION. TR is defined by
- DEF_STMT. GSI points to entry basic block of the TO_REGION. */
-static void
-copy_def(tree tr, gimple *def_stmt, sese region, sese to_region, gimple_stmt_iterator *gsi)
-{
- if (!defined_in_sese_p (tr, region))
- return;
- ssa_op_iter iter;
- use_operand_p use_p;
-
- FOR_EACH_SSA_USE_OPERAND (use_p, def_stmt, iter, SSA_OP_USE)
+ if (flag_loop_parallelize_all)
{
- tree use_tr = USE_FROM_PTR (use_p);
-
- /* Do not copy parameters that have been generated in the header of the
- scop. */
- if (region->parameter_rename_map->get(use_tr))
- continue;
-
- gimple *def_of_use = SSA_NAME_DEF_STMT (use_tr);
- if (!def_of_use)
- continue;
-
- copy_def (use_tr, def_of_use, region, to_region, gsi);
+ scop_get_dependences (scop);
+ context_isl =
+ isl_ast_build_set_before_each_for (context_isl, ast_build_before_for,
+ scop->dependence);
}
- gimple *copy = gimple_copy (def_stmt);
- gsi_insert_after (gsi, copy, GSI_NEW_STMT);
+ isl_ast_node *ast_isl = isl_ast_build_node_from_schedule
+ (context_isl, schedule);
+ isl_ast_build_free (context_isl);
- /* Create new names for all the definitions created by COPY and
- add replacement mappings for each new name. */
- def_operand_p def_p;
- ssa_op_iter op_iter;
- FOR_EACH_SSA_DEF_OPERAND (def_p, copy, op_iter, SSA_OP_ALL_DEFS)
+ isl_options_set_on_error (scop->isl_context, old_err);
+ isl_ctx_reset_operations (scop->isl_context);
+ isl_ctx_set_max_operations (scop->isl_context, old_max_operations);
+ if (isl_ctx_last_error (scop->isl_context) != isl_error_none)
{
- tree old_name = DEF_FROM_PTR (def_p);
- tree new_name = create_new_def_for (old_name, copy, def_p);
- region->parameter_rename_map->put(old_name, new_name);
+ if (dump_enabled_p ())
+ {
+ dump_user_location_t loc = find_loop_location
+ (scop->scop_info->region.entry->dest->loop_father);
+ if (isl_ctx_last_error (scop->isl_context) == isl_error_quota)
+ dump_printf_loc (MSG_MISSED_OPTIMIZATION, loc,
+ "loop nest not optimized, AST generation timed out "
+ "after %d operations [--param max-isl-operations]\n",
+ max_operations);
+ else
+ dump_printf_loc (MSG_MISSED_OPTIMIZATION, loc,
+ "loop nest not optimized, ISL AST generation "
+ "signalled an error\n");
+ }
+ isl_ast_node_free (ast_isl);
+ return NULL;
}
- update_stmt (copy);
+ return ast_isl;
}
+/* Generate out-of-SSA copies for the entry edge FALSE_ENTRY/TRUE_ENTRY
+ in REGION. */
+
static void
-copy_internal_parameters(sese region, sese to_region)
+generate_entry_out_of_ssa_copies (edge false_entry,
+ edge true_entry,
+ sese_info_p region)
{
- /* For all the parameters which definitino is in the if_region->false_region,
- insert code on true_region (if_region->true_region->entry). */
-
- int i;
- tree tr;
- gimple_stmt_iterator gsi = gsi_start_bb(to_region->entry->dest);
-
- FOR_EACH_VEC_ELT (region->params, i, tr)
+ gimple_stmt_iterator gsi_tgt = gsi_start_bb (true_entry->dest);
+ for (gphi_iterator psi = gsi_start_phis (false_entry->dest);
+ !gsi_end_p (psi); gsi_next (&psi))
{
- // If def is not in region.
- gimple *def_stmt = SSA_NAME_DEF_STMT (tr);
- if (def_stmt)
- copy_def (tr, def_stmt, region, to_region, &gsi);
+ gphi *phi = psi.phi ();
+ tree res = gimple_phi_result (phi);
+ if (virtual_operand_p (res))
+ continue;
+ /* When there's no out-of-SSA var registered do not bother
+ to create one. */
+ tree *rename = region->rename_map->get (res);
+ if (! rename)
+ continue;
+ tree new_phi_def = *rename;
+ gassign *ass = gimple_build_assign (new_phi_def,
+ PHI_ARG_DEF_FROM_EDGE (phi,
+ false_entry));
+ gsi_insert_after (&gsi_tgt, ass, GSI_NEW_STMT);
}
}
-/* GIMPLE Loop Generator: generates loops from STMT in GIMPLE form for
- the given SCOP. Return true if code generation succeeded.
-
- FIXME: This is not yet a full implementation of the code generator
- with ISL ASTs. Generation of GIMPLE code has to be completed. */
+/* GIMPLE Loop Generator: generates loops in GIMPLE form for the given SCOP.
+ Return true if code generation succeeded. */
bool
graphite_regenerate_ast_isl (scop_p scop)
{
- loop_p context_loop;
- sese region = SCOP_REGION (scop);
+ sese_info_p region = scop->scop_info;
+ translate_isl_ast_to_gimple t (region);
+
ifsese if_region = NULL;
isl_ast_node *root_node;
ivs_params ip;
timevar_push (TV_GRAPHITE_CODE_GEN);
- graphite_regenerate_error = false;
- root_node = scop_to_isl_ast (scop, ip);
+ t.add_parameters_to_ivs_params (scop, ip);
+ root_node = t.scop_to_isl_ast (scop);
+ if (! root_node)
+ {
+ ivs_params_clear (ip);
+ timevar_pop (TV_GRAPHITE_CODE_GEN);
+ return false;
+ }
if (dump_file && (dump_flags & TDF_DETAILS))
{
- fprintf (dump_file, "\nISL AST generated by ISL: \n");
- print_isl_ast_node (dump_file, root_node, scop->isl_context);
- fprintf (dump_file, "\n");
+ fprintf (dump_file, "[scheduler] original schedule:\n");
+ print_isl_schedule (dump_file, scop->original_schedule);
+ fprintf (dump_file, "[scheduler] isl transformed schedule:\n");
+ print_isl_schedule (dump_file, scop->transformed_schedule);
+
+ fprintf (dump_file, "[scheduler] original ast:\n");
+ print_schedule_ast (dump_file, scop->original_schedule, scop);
+ fprintf (dump_file, "[scheduler] AST generated by isl:\n");
+ print_isl_ast (dump_file, root_node);
}
- recompute_all_dominators ();
- graphite_verify ();
-
if_region = move_sese_in_condition (region);
- sese_insert_phis_for_liveouts (region,
- if_region->region->exit->src,
- if_region->false_region->exit,
- if_region->true_region->exit);
- recompute_all_dominators ();
- graphite_verify ();
+ region->if_region = if_region;
- context_loop = SESE_ENTRY (region)->src->loop_father;
+ loop_p context_loop = region->region.entry->src->loop_father;
+ edge e = single_succ_edge (if_region->true_region->region.entry->dest);
+ basic_block bb = split_edge (e);
- /* Copy all the parameters which are defined in the region. */
- copy_internal_parameters(if_region->false_region, if_region->true_region);
+ /* Update the true_region exit edge. */
+ region->if_region->true_region->region.exit = single_succ_edge (bb);
- translate_isl_ast_to_gimple t(region);
- edge e = single_succ_edge (if_region->true_region->entry->dest);
- split_edge (e);
t.translate_isl_ast (context_loop, root_node, e, ip);
+ if (! t.codegen_error_p ())
+ {
+ generate_entry_out_of_ssa_copies (if_region->false_region->region.entry,
+ if_region->true_region->region.entry,
+ region);
+ sese_insert_phis_for_liveouts (region,
+ if_region->region->region.exit->src,
+ if_region->false_region->region.exit,
+ if_region->true_region->region.exit);
+ if (dump_file)
+ fprintf (dump_file, "[codegen] isl AST to Gimple succeeded.\n");
+ }
- mark_virtual_operands_for_renaming (cfun);
- update_ssa (TODO_update_ssa);
-
- graphite_verify ();
- scev_reset ();
- recompute_all_dominators ();
- graphite_verify ();
+ if (t.codegen_error_p ())
+ {
+ if (dump_enabled_p ())
+ {
+ dump_user_location_t loc = find_loop_location
+ (scop->scop_info->region.entry->dest->loop_father);
+ dump_printf_loc (MSG_MISSED_OPTIMIZATION, loc,
+ "loop nest not optimized, code generation error\n");
+ }
+
+ /* Remove the unreachable region. */
+ remove_edge_and_dominated_blocks (if_region->true_region->region.entry);
+ basic_block ifb = if_region->false_region->region.entry->src;
+ gimple_stmt_iterator gsi = gsi_last_bb (ifb);
+ gsi_remove (&gsi, true);
+ if_region->false_region->region.entry->flags &= ~EDGE_FALSE_VALUE;
+ if_region->false_region->region.entry->flags |= EDGE_FALLTHRU;
+ /* remove_edge_and_dominated_blocks marks loops for removal but
+ doesn't actually remove them (fix that...). */
+ loop_p loop;
+ FOR_EACH_LOOP (loop, LI_FROM_INNERMOST)
+ if (! loop->header)
+ delete_loop (loop);
+ }
- if (graphite_regenerate_error)
- set_ifsese_condition (if_region, integer_zero_node);
+ /* We are delaying SSA update to after code-generating all SCOPs.
+ This is because we analyzed DRs and parameters on the unmodified
+ IL and thus rely on SSA update to pick up new dominating definitions
+ from for example SESE liveout PHIs. This is also for efficiency
+ as SSA update does work depending on the size of the function. */
free (if_region->true_region);
free (if_region->region);
isl_ast_node_free (root_node);
timevar_pop (TV_GRAPHITE_CODE_GEN);
- if (dump_file && (dump_flags & TDF_DETAILS))
- {
- loop_p loop;
- int num_no_dependency = 0;
-
- FOR_EACH_LOOP (loop, 0)
- if (loop->can_be_parallel)
- num_no_dependency++;
-
- fprintf (dump_file, "\n%d loops carried no dependency.\n",
- num_no_dependency);
- }
-
- return !graphite_regenerate_error;
+ return !t.codegen_error_p ();
}
+
#endif /* HAVE_isl */