2012-03-27 23:13:14 +00:00
|
|
|
|
/* Conditional Dead Call Elimination pass for the GNU compiler.
|
2022-10-27 18:55:19 +00:00
|
|
|
|
Copyright (C) 2008-2022 Free Software Foundation, Inc.
|
2012-03-27 23:13:14 +00:00
|
|
|
|
Contributed by Xinliang David Li <davidxl@google.com>
|
|
|
|
|
|
|
|
|
|
This file is part of GCC.
|
|
|
|
|
|
|
|
|
|
GCC is free software; you can redistribute it and/or modify it
|
|
|
|
|
under the terms of the GNU General Public License as published by the
|
|
|
|
|
Free Software Foundation; either version 3, or (at your option) any
|
|
|
|
|
later version.
|
|
|
|
|
|
|
|
|
|
GCC is distributed in the hope that it will be useful, but WITHOUT
|
|
|
|
|
ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
|
|
|
|
FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
|
|
|
|
|
for more details.
|
|
|
|
|
|
|
|
|
|
You should have received a copy of the GNU General Public License
|
|
|
|
|
along with GCC; see the file COPYING3. If not see
|
|
|
|
|
<http://www.gnu.org/licenses/>. */
|
|
|
|
|
|
|
|
|
|
#include "config.h"
|
|
|
|
|
#include "system.h"
|
|
|
|
|
#include "coretypes.h"
|
2017-04-10 11:32:00 +00:00
|
|
|
|
#include "backend.h"
|
2012-03-27 23:13:14 +00:00
|
|
|
|
#include "tree.h"
|
2017-04-10 11:32:00 +00:00
|
|
|
|
#include "gimple.h"
|
|
|
|
|
#include "cfghooks.h"
|
|
|
|
|
#include "tree-pass.h"
|
|
|
|
|
#include "ssa.h"
|
|
|
|
|
#include "gimple-pretty-print.h"
|
2015-08-28 15:33:40 +00:00
|
|
|
|
#include "fold-const.h"
|
2014-09-21 17:33:12 +00:00
|
|
|
|
#include "stor-layout.h"
|
|
|
|
|
#include "gimple-iterator.h"
|
|
|
|
|
#include "tree-cfg.h"
|
|
|
|
|
#include "tree-into-ssa.h"
|
2017-04-10 11:32:00 +00:00
|
|
|
|
#include "builtins.h"
|
|
|
|
|
#include "internal-fn.h"
|
2017-10-07 00:16:47 +00:00
|
|
|
|
#include "tree-dfa.h"
|
2012-03-27 23:13:14 +00:00
|
|
|
|
|
|
|
|
|
|
2017-04-10 11:32:00 +00:00
|
|
|
|
/* This pass serves two closely-related purposes:
|
|
|
|
|
|
|
|
|
|
1. It conditionally executes calls that set errno if (a) the result of
|
|
|
|
|
the call is unused and (b) a simple range check on the arguments can
|
|
|
|
|
detect most cases where errno does not need to be set.
|
|
|
|
|
|
|
|
|
|
This is the "conditional dead-code elimination" that gave the pass
|
|
|
|
|
its original name, since the call is dead for most argument values.
|
|
|
|
|
The calls for which it helps are usually part of the C++ abstraction
|
|
|
|
|
penalty exposed after inlining.
|
|
|
|
|
|
|
|
|
|
2. It looks for calls to built-in functions that set errno and whose
|
|
|
|
|
result is used. It checks whether there is an associated internal
|
|
|
|
|
function that doesn't set errno and whether the target supports
|
|
|
|
|
that internal function. If so, the pass uses the internal function
|
|
|
|
|
to compute the result of the built-in function but still arranges
|
|
|
|
|
for errno to be set when necessary. There are two ways of setting
|
|
|
|
|
errno:
|
|
|
|
|
|
|
|
|
|
a. by protecting the original call with the same argument checks as (1)
|
|
|
|
|
|
|
|
|
|
b. by protecting the original call with a check that the result
|
|
|
|
|
of the internal function is not equal to itself (i.e. is NaN).
|
|
|
|
|
|
|
|
|
|
(b) requires that NaNs are the only erroneous results. It is not
|
|
|
|
|
appropriate for functions like log, which returns ERANGE for zero
|
|
|
|
|
arguments. (b) is also likely to perform worse than (a) because it
|
|
|
|
|
requires the result to be calculated first. The pass therefore uses
|
|
|
|
|
(a) when it can and uses (b) as a fallback.
|
2012-03-27 23:13:14 +00:00
|
|
|
|
|
2017-04-10 11:32:00 +00:00
|
|
|
|
For (b) the pass can replace the original call with a call to
|
|
|
|
|
IFN_SET_EDOM, if the target supports direct assignments to errno.
|
2012-03-27 23:13:14 +00:00
|
|
|
|
|
2017-04-10 11:32:00 +00:00
|
|
|
|
In both cases, arguments that require errno to be set should occur
|
|
|
|
|
rarely in practice. Checks of the errno result should also be rare,
|
|
|
|
|
but the compiler would need powerful interprocedural analysis to
|
|
|
|
|
prove that errno is not checked. It's much easier to add argument
|
|
|
|
|
checks or result checks instead.
|
2012-03-27 23:13:14 +00:00
|
|
|
|
|
2017-04-10 11:32:00 +00:00
|
|
|
|
An example of (1) is:
|
|
|
|
|
|
|
|
|
|
log (x); // Mostly dead call
|
2012-03-27 23:13:14 +00:00
|
|
|
|
==>
|
2017-04-10 11:32:00 +00:00
|
|
|
|
if (__builtin_islessequal (x, 0))
|
|
|
|
|
log (x);
|
|
|
|
|
|
2012-03-27 23:13:14 +00:00
|
|
|
|
With this change, call to log (x) is effectively eliminated, as
|
2017-04-10 11:32:00 +00:00
|
|
|
|
in the majority of the cases, log won't be called with x out of
|
2012-03-27 23:13:14 +00:00
|
|
|
|
range. The branch is totally predictable, so the branch cost
|
|
|
|
|
is low.
|
|
|
|
|
|
2017-04-10 11:32:00 +00:00
|
|
|
|
An example of (2) is:
|
|
|
|
|
|
|
|
|
|
y = sqrt (x);
|
|
|
|
|
==>
|
|
|
|
|
if (__builtin_isless (x, 0))
|
2022-10-27 18:55:19 +00:00
|
|
|
|
y = sqrt (x);
|
|
|
|
|
else
|
|
|
|
|
y = IFN_SQRT (x);
|
2017-04-10 11:32:00 +00:00
|
|
|
|
In the vast majority of cases we should then never need to call sqrt.
|
|
|
|
|
|
2012-03-27 23:13:14 +00:00
|
|
|
|
Note that library functions are not supposed to clear errno to zero without
|
|
|
|
|
error. See IEEE Std 1003.1, section 2.3 Error Numbers, and section 7.5:3 of
|
|
|
|
|
ISO/IEC 9899 (C99).
|
|
|
|
|
|
|
|
|
|
The condition wrapping the builtin call is conservatively set to avoid too
|
2017-04-10 11:32:00 +00:00
|
|
|
|
aggressive (wrong) shrink wrapping. */
|
2012-03-27 23:13:14 +00:00
|
|
|
|
|
|
|
|
|
|
|
|
|
|
/* A structure for representing input domain of
|
|
|
|
|
a function argument in integer. If the lower
|
|
|
|
|
bound is -inf, has_lb is set to false. If the
|
|
|
|
|
upper bound is +inf, has_ub is false.
|
|
|
|
|
is_lb_inclusive and is_ub_inclusive are flags
|
|
|
|
|
to indicate if lb and ub value are inclusive
|
|
|
|
|
respectively. */
|
|
|
|
|
|
2017-04-10 11:32:00 +00:00
|
|
|
|
struct inp_domain
|
2012-03-27 23:13:14 +00:00
|
|
|
|
{
|
|
|
|
|
int lb;
|
|
|
|
|
int ub;
|
|
|
|
|
bool has_lb;
|
|
|
|
|
bool has_ub;
|
|
|
|
|
bool is_lb_inclusive;
|
|
|
|
|
bool is_ub_inclusive;
|
2017-04-10 11:32:00 +00:00
|
|
|
|
};
|
2012-03-27 23:13:14 +00:00
|
|
|
|
|
|
|
|
|
/* A helper function to construct and return an input
|
|
|
|
|
domain object. LB is the lower bound, HAS_LB is
|
|
|
|
|
a boolean flag indicating if the lower bound exists,
|
|
|
|
|
and LB_INCLUSIVE is a boolean flag indicating if the
|
|
|
|
|
lower bound is inclusive or not. UB, HAS_UB, and
|
|
|
|
|
UB_INCLUSIVE have the same meaning, but for upper
|
|
|
|
|
bound of the domain. */
|
|
|
|
|
|
|
|
|
|
static inp_domain
|
|
|
|
|
get_domain (int lb, bool has_lb, bool lb_inclusive,
|
|
|
|
|
int ub, bool has_ub, bool ub_inclusive)
|
|
|
|
|
{
|
|
|
|
|
inp_domain domain;
|
|
|
|
|
domain.lb = lb;
|
|
|
|
|
domain.has_lb = has_lb;
|
|
|
|
|
domain.is_lb_inclusive = lb_inclusive;
|
|
|
|
|
domain.ub = ub;
|
|
|
|
|
domain.has_ub = has_ub;
|
|
|
|
|
domain.is_ub_inclusive = ub_inclusive;
|
|
|
|
|
return domain;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* A helper function to check the target format for the
|
|
|
|
|
argument type. In this implementation, only IEEE formats
|
|
|
|
|
are supported. ARG is the call argument to be checked.
|
|
|
|
|
Returns true if the format is supported. To support other
|
|
|
|
|
target formats, function get_no_error_domain needs to be
|
|
|
|
|
enhanced to have range bounds properly computed. Since
|
|
|
|
|
the check is cheap (very small number of candidates
|
|
|
|
|
to be checked), the result is not cached for each float type. */
|
|
|
|
|
|
|
|
|
|
static bool
|
|
|
|
|
check_target_format (tree arg)
|
|
|
|
|
{
|
|
|
|
|
tree type;
|
2015-08-28 15:33:40 +00:00
|
|
|
|
machine_mode mode;
|
2012-03-27 23:13:14 +00:00
|
|
|
|
const struct real_format *rfmt;
|
|
|
|
|
|
|
|
|
|
type = TREE_TYPE (arg);
|
|
|
|
|
mode = TYPE_MODE (type);
|
|
|
|
|
rfmt = REAL_MODE_FORMAT (mode);
|
|
|
|
|
if ((mode == SFmode
|
|
|
|
|
&& (rfmt == &ieee_single_format || rfmt == &mips_single_format
|
|
|
|
|
|| rfmt == &motorola_single_format))
|
|
|
|
|
|| (mode == DFmode
|
|
|
|
|
&& (rfmt == &ieee_double_format || rfmt == &mips_double_format
|
|
|
|
|
|| rfmt == &motorola_double_format))
|
2019-06-02 15:48:37 +00:00
|
|
|
|
/* For long double, we cannot really check XFmode
|
2012-03-27 23:13:14 +00:00
|
|
|
|
which is only defined on intel platforms.
|
|
|
|
|
Candidate pre-selection using builtin function
|
|
|
|
|
code guarantees that we are checking formats
|
|
|
|
|
for long double modes: double, quad, and extended. */
|
|
|
|
|
|| (mode != SFmode && mode != DFmode
|
|
|
|
|
&& (rfmt == &ieee_quad_format
|
|
|
|
|
|| rfmt == &mips_quad_format
|
|
|
|
|
|| rfmt == &ieee_extended_motorola_format
|
|
|
|
|
|| rfmt == &ieee_extended_intel_96_format
|
|
|
|
|
|| rfmt == &ieee_extended_intel_128_format
|
|
|
|
|
|| rfmt == &ieee_extended_intel_96_round_53_format)))
|
|
|
|
|
return true;
|
|
|
|
|
|
|
|
|
|
return false;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
/* A helper function to help select calls to pow that are suitable for
|
|
|
|
|
conditional DCE transformation. It looks for pow calls that can be
|
|
|
|
|
guided with simple conditions. Such calls either have constant base
|
|
|
|
|
values or base values converted from integers. Returns true if
|
|
|
|
|
the pow call POW_CALL is a candidate. */
|
|
|
|
|
|
|
|
|
|
/* The maximum integer bit size for base argument of a pow call
|
|
|
|
|
that is suitable for shrink-wrapping transformation. */
|
|
|
|
|
#define MAX_BASE_INT_BIT_SIZE 32
|
|
|
|
|
|
|
|
|
|
static bool
|
2015-08-28 15:33:40 +00:00
|
|
|
|
check_pow (gcall *pow_call)
|
2012-03-27 23:13:14 +00:00
|
|
|
|
{
|
|
|
|
|
tree base, expn;
|
|
|
|
|
enum tree_code bc, ec;
|
|
|
|
|
|
|
|
|
|
if (gimple_call_num_args (pow_call) != 2)
|
|
|
|
|
return false;
|
|
|
|
|
|
|
|
|
|
base = gimple_call_arg (pow_call, 0);
|
|
|
|
|
expn = gimple_call_arg (pow_call, 1);
|
|
|
|
|
|
|
|
|
|
if (!check_target_format (expn))
|
|
|
|
|
return false;
|
|
|
|
|
|
|
|
|
|
bc = TREE_CODE (base);
|
|
|
|
|
ec = TREE_CODE (expn);
|
|
|
|
|
|
|
|
|
|
/* Folding candidates are not interesting.
|
|
|
|
|
Can actually assert that it is already folded. */
|
|
|
|
|
if (ec == REAL_CST && bc == REAL_CST)
|
|
|
|
|
return false;
|
|
|
|
|
|
|
|
|
|
if (bc == REAL_CST)
|
|
|
|
|
{
|
|
|
|
|
/* Only handle a fixed range of constant. */
|
|
|
|
|
REAL_VALUE_TYPE mv;
|
|
|
|
|
REAL_VALUE_TYPE bcv = TREE_REAL_CST (base);
|
2017-04-10 11:32:00 +00:00
|
|
|
|
if (real_equal (&bcv, &dconst1))
|
2012-03-27 23:13:14 +00:00
|
|
|
|
return false;
|
2017-04-10 11:32:00 +00:00
|
|
|
|
if (real_less (&bcv, &dconst1))
|
2012-03-27 23:13:14 +00:00
|
|
|
|
return false;
|
2015-08-28 15:33:40 +00:00
|
|
|
|
real_from_integer (&mv, TYPE_MODE (TREE_TYPE (base)), 256, UNSIGNED);
|
2017-04-10 11:32:00 +00:00
|
|
|
|
if (real_less (&mv, &bcv))
|
2012-03-27 23:13:14 +00:00
|
|
|
|
return false;
|
|
|
|
|
return true;
|
|
|
|
|
}
|
|
|
|
|
else if (bc == SSA_NAME)
|
|
|
|
|
{
|
2014-09-21 17:33:12 +00:00
|
|
|
|
tree base_val0, type;
|
2017-04-10 11:32:00 +00:00
|
|
|
|
gimple *base_def;
|
2012-03-27 23:13:14 +00:00
|
|
|
|
int bit_sz;
|
|
|
|
|
|
|
|
|
|
/* Only handles cases where base value is converted
|
|
|
|
|
from integer values. */
|
|
|
|
|
base_def = SSA_NAME_DEF_STMT (base);
|
|
|
|
|
if (gimple_code (base_def) != GIMPLE_ASSIGN)
|
|
|
|
|
return false;
|
|
|
|
|
|
|
|
|
|
if (gimple_assign_rhs_code (base_def) != FLOAT_EXPR)
|
|
|
|
|
return false;
|
|
|
|
|
base_val0 = gimple_assign_rhs1 (base_def);
|
|
|
|
|
|
2014-09-21 17:33:12 +00:00
|
|
|
|
type = TREE_TYPE (base_val0);
|
2012-03-27 23:13:14 +00:00
|
|
|
|
if (TREE_CODE (type) != INTEGER_TYPE)
|
|
|
|
|
return false;
|
|
|
|
|
bit_sz = TYPE_PRECISION (type);
|
|
|
|
|
/* If the type of the base is too wide,
|
|
|
|
|
the resulting shrink wrapping condition
|
|
|
|
|
will be too conservative. */
|
|
|
|
|
if (bit_sz > MAX_BASE_INT_BIT_SIZE)
|
|
|
|
|
return false;
|
|
|
|
|
|
|
|
|
|
return true;
|
|
|
|
|
}
|
|
|
|
|
else
|
|
|
|
|
return false;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* A helper function to help select candidate function calls that are
|
|
|
|
|
suitable for conditional DCE. Candidate functions must have single
|
|
|
|
|
valid input domain in this implementation except for pow (see check_pow).
|
|
|
|
|
Returns true if the function call is a candidate. */
|
|
|
|
|
|
|
|
|
|
static bool
|
2015-08-28 15:33:40 +00:00
|
|
|
|
check_builtin_call (gcall *bcall)
|
2012-03-27 23:13:14 +00:00
|
|
|
|
{
|
|
|
|
|
tree arg;
|
|
|
|
|
|
|
|
|
|
arg = gimple_call_arg (bcall, 0);
|
|
|
|
|
return check_target_format (arg);
|
|
|
|
|
}
|
|
|
|
|
|
2017-04-10 11:32:00 +00:00
|
|
|
|
/* Return true if built-in function call CALL calls a math function
|
|
|
|
|
and if we know how to test the range of its arguments to detect _most_
|
|
|
|
|
situations in which errno is not set. The test must err on the side
|
|
|
|
|
of treating non-erroneous values as potentially erroneous. */
|
2012-03-27 23:13:14 +00:00
|
|
|
|
|
|
|
|
|
static bool
|
2017-04-10 11:32:00 +00:00
|
|
|
|
can_test_argument_range (gcall *call)
|
2012-03-27 23:13:14 +00:00
|
|
|
|
{
|
2017-04-10 11:32:00 +00:00
|
|
|
|
switch (DECL_FUNCTION_CODE (gimple_call_fndecl (call)))
|
2012-03-27 23:13:14 +00:00
|
|
|
|
{
|
|
|
|
|
/* Trig functions. */
|
|
|
|
|
CASE_FLT_FN (BUILT_IN_ACOS):
|
|
|
|
|
CASE_FLT_FN (BUILT_IN_ASIN):
|
|
|
|
|
/* Hyperbolic functions. */
|
|
|
|
|
CASE_FLT_FN (BUILT_IN_ACOSH):
|
|
|
|
|
CASE_FLT_FN (BUILT_IN_ATANH):
|
|
|
|
|
CASE_FLT_FN (BUILT_IN_COSH):
|
|
|
|
|
CASE_FLT_FN (BUILT_IN_SINH):
|
|
|
|
|
/* Log functions. */
|
|
|
|
|
CASE_FLT_FN (BUILT_IN_LOG):
|
|
|
|
|
CASE_FLT_FN (BUILT_IN_LOG2):
|
|
|
|
|
CASE_FLT_FN (BUILT_IN_LOG10):
|
|
|
|
|
CASE_FLT_FN (BUILT_IN_LOG1P):
|
|
|
|
|
/* Exp functions. */
|
|
|
|
|
CASE_FLT_FN (BUILT_IN_EXP):
|
|
|
|
|
CASE_FLT_FN (BUILT_IN_EXP2):
|
|
|
|
|
CASE_FLT_FN (BUILT_IN_EXP10):
|
|
|
|
|
CASE_FLT_FN (BUILT_IN_EXPM1):
|
|
|
|
|
CASE_FLT_FN (BUILT_IN_POW10):
|
|
|
|
|
/* Sqrt. */
|
|
|
|
|
CASE_FLT_FN (BUILT_IN_SQRT):
|
2018-12-28 15:30:48 +00:00
|
|
|
|
CASE_FLT_FN_FLOATN_NX (BUILT_IN_SQRT):
|
2012-03-27 23:13:14 +00:00
|
|
|
|
return check_builtin_call (call);
|
|
|
|
|
/* Special one: two argument pow. */
|
|
|
|
|
case BUILT_IN_POW:
|
|
|
|
|
return check_pow (call);
|
|
|
|
|
default:
|
|
|
|
|
break;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
return false;
|
|
|
|
|
}
|
|
|
|
|
|
2017-04-10 11:32:00 +00:00
|
|
|
|
/* Return true if CALL can produce a domain error (EDOM) but can never
|
|
|
|
|
produce a pole, range overflow or range underflow error (all ERANGE).
|
|
|
|
|
This means that we can tell whether a function would have set errno
|
|
|
|
|
by testing whether the result is a NaN. */
|
|
|
|
|
|
|
|
|
|
static bool
|
|
|
|
|
edom_only_function (gcall *call)
|
|
|
|
|
{
|
|
|
|
|
switch (DECL_FUNCTION_CODE (gimple_call_fndecl (call)))
|
|
|
|
|
{
|
|
|
|
|
CASE_FLT_FN (BUILT_IN_ACOS):
|
|
|
|
|
CASE_FLT_FN (BUILT_IN_ASIN):
|
|
|
|
|
CASE_FLT_FN (BUILT_IN_ATAN):
|
|
|
|
|
CASE_FLT_FN (BUILT_IN_COS):
|
|
|
|
|
CASE_FLT_FN (BUILT_IN_SIGNIFICAND):
|
|
|
|
|
CASE_FLT_FN (BUILT_IN_SIN):
|
|
|
|
|
CASE_FLT_FN (BUILT_IN_SQRT):
|
2018-12-28 15:30:48 +00:00
|
|
|
|
CASE_FLT_FN_FLOATN_NX (BUILT_IN_SQRT):
|
2017-04-10 11:32:00 +00:00
|
|
|
|
CASE_FLT_FN (BUILT_IN_FMOD):
|
|
|
|
|
CASE_FLT_FN (BUILT_IN_REMAINDER):
|
|
|
|
|
return true;
|
|
|
|
|
|
|
|
|
|
default:
|
|
|
|
|
return false;
|
|
|
|
|
}
|
|
|
|
|
}
|
2017-10-07 00:16:47 +00:00
|
|
|
|
|
|
|
|
|
/* Return true if it is structurally possible to guard CALL. */
|
|
|
|
|
|
|
|
|
|
static bool
|
|
|
|
|
can_guard_call_p (gimple *call)
|
|
|
|
|
{
|
|
|
|
|
return (!stmt_ends_bb_p (call)
|
|
|
|
|
|| find_fallthru_edge (gimple_bb (call)->succs));
|
|
|
|
|
}
|
2012-03-27 23:13:14 +00:00
|
|
|
|
|
2019-06-02 15:48:37 +00:00
|
|
|
|
/* For a comparison code return the comparison code we should use if we don't
|
|
|
|
|
HONOR_NANS. */
|
|
|
|
|
|
|
|
|
|
static enum tree_code
|
|
|
|
|
comparison_code_if_no_nans (tree_code code)
|
|
|
|
|
{
|
|
|
|
|
switch (code)
|
|
|
|
|
{
|
|
|
|
|
case UNLT_EXPR:
|
|
|
|
|
return LT_EXPR;
|
|
|
|
|
case UNGT_EXPR:
|
|
|
|
|
return GT_EXPR;
|
|
|
|
|
case UNLE_EXPR:
|
|
|
|
|
return LE_EXPR;
|
|
|
|
|
case UNGE_EXPR:
|
|
|
|
|
return GE_EXPR;
|
|
|
|
|
case UNEQ_EXPR:
|
|
|
|
|
return EQ_EXPR;
|
|
|
|
|
case LTGT_EXPR:
|
|
|
|
|
return NE_EXPR;
|
|
|
|
|
|
|
|
|
|
case LT_EXPR:
|
|
|
|
|
case GT_EXPR:
|
|
|
|
|
case LE_EXPR:
|
|
|
|
|
case GE_EXPR:
|
|
|
|
|
case EQ_EXPR:
|
|
|
|
|
case NE_EXPR:
|
|
|
|
|
return code;
|
|
|
|
|
|
|
|
|
|
default:
|
|
|
|
|
gcc_unreachable ();
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
2017-04-10 11:32:00 +00:00
|
|
|
|
/* A helper function to generate gimple statements for one bound
|
|
|
|
|
comparison, so that the built-in function is called whenever
|
|
|
|
|
TCODE <ARG, LBUB> is *false*. TEMP_NAME1/TEMP_NAME2 are names
|
|
|
|
|
of the temporaries, CONDS is a vector holding the produced GIMPLE
|
|
|
|
|
statements, and NCONDS points to the variable holding the number of
|
|
|
|
|
logical comparisons. CONDS is either empty or a list ended with a
|
|
|
|
|
null tree. */
|
2012-03-27 23:13:14 +00:00
|
|
|
|
|
|
|
|
|
static void
|
|
|
|
|
gen_one_condition (tree arg, int lbub,
|
|
|
|
|
enum tree_code tcode,
|
|
|
|
|
const char *temp_name1,
|
|
|
|
|
const char *temp_name2,
|
2017-04-10 11:32:00 +00:00
|
|
|
|
vec<gimple *> conds,
|
2012-03-27 23:13:14 +00:00
|
|
|
|
unsigned *nconds)
|
|
|
|
|
{
|
2019-06-02 15:48:37 +00:00
|
|
|
|
if (!HONOR_NANS (arg))
|
|
|
|
|
tcode = comparison_code_if_no_nans (tcode);
|
|
|
|
|
|
2012-03-27 23:13:14 +00:00
|
|
|
|
tree lbub_real_cst, lbub_cst, float_type;
|
|
|
|
|
tree temp, tempn, tempc, tempcn;
|
2015-08-28 15:33:40 +00:00
|
|
|
|
gassign *stmt1;
|
|
|
|
|
gassign *stmt2;
|
|
|
|
|
gcond *stmt3;
|
2012-03-27 23:13:14 +00:00
|
|
|
|
|
|
|
|
|
float_type = TREE_TYPE (arg);
|
|
|
|
|
lbub_cst = build_int_cst (integer_type_node, lbub);
|
|
|
|
|
lbub_real_cst = build_real_from_int_cst (float_type, lbub_cst);
|
|
|
|
|
|
|
|
|
|
temp = create_tmp_var (float_type, temp_name1);
|
|
|
|
|
stmt1 = gimple_build_assign (temp, arg);
|
|
|
|
|
tempn = make_ssa_name (temp, stmt1);
|
|
|
|
|
gimple_assign_set_lhs (stmt1, tempn);
|
|
|
|
|
|
|
|
|
|
tempc = create_tmp_var (boolean_type_node, temp_name2);
|
|
|
|
|
stmt2 = gimple_build_assign (tempc,
|
|
|
|
|
fold_build2 (tcode,
|
|
|
|
|
boolean_type_node,
|
|
|
|
|
tempn, lbub_real_cst));
|
|
|
|
|
tempcn = make_ssa_name (tempc, stmt2);
|
|
|
|
|
gimple_assign_set_lhs (stmt2, tempcn);
|
|
|
|
|
|
|
|
|
|
stmt3 = gimple_build_cond_from_tree (tempcn, NULL_TREE, NULL_TREE);
|
2014-09-21 17:33:12 +00:00
|
|
|
|
conds.quick_push (stmt1);
|
|
|
|
|
conds.quick_push (stmt2);
|
|
|
|
|
conds.quick_push (stmt3);
|
2012-03-27 23:13:14 +00:00
|
|
|
|
(*nconds)++;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* A helper function to generate GIMPLE statements for
|
|
|
|
|
out of input domain check. ARG is the call argument
|
|
|
|
|
to be runtime checked, DOMAIN holds the valid domain
|
|
|
|
|
for the given function, CONDS points to the vector
|
|
|
|
|
holding the result GIMPLE statements. *NCONDS is
|
|
|
|
|
the number of logical comparisons. This function
|
|
|
|
|
produces no more than two logical comparisons, one
|
|
|
|
|
for lower bound check, one for upper bound check. */
|
|
|
|
|
|
|
|
|
|
static void
|
|
|
|
|
gen_conditions_for_domain (tree arg, inp_domain domain,
|
2017-04-10 11:32:00 +00:00
|
|
|
|
vec<gimple *> conds,
|
2012-03-27 23:13:14 +00:00
|
|
|
|
unsigned *nconds)
|
|
|
|
|
{
|
|
|
|
|
if (domain.has_lb)
|
|
|
|
|
gen_one_condition (arg, domain.lb,
|
|
|
|
|
(domain.is_lb_inclusive
|
2017-04-10 11:32:00 +00:00
|
|
|
|
? UNGE_EXPR : UNGT_EXPR),
|
2012-03-27 23:13:14 +00:00
|
|
|
|
"DCE_COND_LB", "DCE_COND_LB_TEST",
|
|
|
|
|
conds, nconds);
|
|
|
|
|
|
|
|
|
|
if (domain.has_ub)
|
|
|
|
|
{
|
|
|
|
|
/* Now push a separator. */
|
|
|
|
|
if (domain.has_lb)
|
2014-09-21 17:33:12 +00:00
|
|
|
|
conds.quick_push (NULL);
|
2012-03-27 23:13:14 +00:00
|
|
|
|
|
|
|
|
|
gen_one_condition (arg, domain.ub,
|
|
|
|
|
(domain.is_ub_inclusive
|
2017-04-10 11:32:00 +00:00
|
|
|
|
? UNLE_EXPR : UNLT_EXPR),
|
2012-03-27 23:13:14 +00:00
|
|
|
|
"DCE_COND_UB", "DCE_COND_UB_TEST",
|
|
|
|
|
conds, nconds);
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
/* A helper function to generate condition
|
|
|
|
|
code for the y argument in call pow (some_const, y).
|
|
|
|
|
See candidate selection in check_pow. Since the
|
|
|
|
|
candidates' base values have a limited range,
|
|
|
|
|
the guarded code generated for y are simple:
|
2017-04-10 11:32:00 +00:00
|
|
|
|
if (__builtin_isgreater (y, max_y))
|
2012-03-27 23:13:14 +00:00
|
|
|
|
pow (const, y);
|
|
|
|
|
Note max_y can be computed separately for each
|
|
|
|
|
const base, but in this implementation, we
|
|
|
|
|
choose to compute it using the max base
|
|
|
|
|
in the allowed range for the purpose of
|
|
|
|
|
simplicity. BASE is the constant base value,
|
|
|
|
|
EXPN is the expression for the exponent argument,
|
|
|
|
|
*CONDS is the vector to hold resulting statements,
|
|
|
|
|
and *NCONDS is the number of logical conditions. */
|
|
|
|
|
|
|
|
|
|
static void
|
|
|
|
|
gen_conditions_for_pow_cst_base (tree base, tree expn,
|
2017-04-10 11:32:00 +00:00
|
|
|
|
vec<gimple *> conds,
|
2012-03-27 23:13:14 +00:00
|
|
|
|
unsigned *nconds)
|
|
|
|
|
{
|
|
|
|
|
inp_domain exp_domain;
|
|
|
|
|
/* Validate the range of the base constant to make
|
|
|
|
|
sure it is consistent with check_pow. */
|
|
|
|
|
REAL_VALUE_TYPE mv;
|
|
|
|
|
REAL_VALUE_TYPE bcv = TREE_REAL_CST (base);
|
2017-04-10 11:32:00 +00:00
|
|
|
|
gcc_assert (!real_equal (&bcv, &dconst1)
|
|
|
|
|
&& !real_less (&bcv, &dconst1));
|
2015-08-28 15:33:40 +00:00
|
|
|
|
real_from_integer (&mv, TYPE_MODE (TREE_TYPE (base)), 256, UNSIGNED);
|
2017-04-10 11:32:00 +00:00
|
|
|
|
gcc_assert (!real_less (&mv, &bcv));
|
2012-03-27 23:13:14 +00:00
|
|
|
|
|
|
|
|
|
exp_domain = get_domain (0, false, false,
|
|
|
|
|
127, true, false);
|
|
|
|
|
|
|
|
|
|
gen_conditions_for_domain (expn, exp_domain,
|
|
|
|
|
conds, nconds);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Generate error condition code for pow calls with
|
|
|
|
|
non constant base values. The candidates selected
|
|
|
|
|
have their base argument value converted from
|
|
|
|
|
integer (see check_pow) value (1, 2, 4 bytes), and
|
|
|
|
|
the max exp value is computed based on the size
|
|
|
|
|
of the integer type (i.e. max possible base value).
|
|
|
|
|
The resulting input domain for exp argument is thus
|
|
|
|
|
conservative (smaller than the max value allowed by
|
|
|
|
|
the runtime value of the base). BASE is the integer
|
|
|
|
|
base value, EXPN is the expression for the exponent
|
|
|
|
|
argument, *CONDS is the vector to hold resulting
|
|
|
|
|
statements, and *NCONDS is the number of logical
|
|
|
|
|
conditions. */
|
|
|
|
|
|
|
|
|
|
static void
|
|
|
|
|
gen_conditions_for_pow_int_base (tree base, tree expn,
|
2017-04-10 11:32:00 +00:00
|
|
|
|
vec<gimple *> conds,
|
2012-03-27 23:13:14 +00:00
|
|
|
|
unsigned *nconds)
|
|
|
|
|
{
|
2017-04-10 11:32:00 +00:00
|
|
|
|
gimple *base_def;
|
2012-03-27 23:13:14 +00:00
|
|
|
|
tree base_val0;
|
2014-09-21 17:33:12 +00:00
|
|
|
|
tree int_type;
|
2012-03-27 23:13:14 +00:00
|
|
|
|
tree temp, tempn;
|
|
|
|
|
tree cst0;
|
2017-04-10 11:32:00 +00:00
|
|
|
|
gimple *stmt1, *stmt2;
|
2012-03-27 23:13:14 +00:00
|
|
|
|
int bit_sz, max_exp;
|
|
|
|
|
inp_domain exp_domain;
|
|
|
|
|
|
|
|
|
|
base_def = SSA_NAME_DEF_STMT (base);
|
|
|
|
|
base_val0 = gimple_assign_rhs1 (base_def);
|
2014-09-21 17:33:12 +00:00
|
|
|
|
int_type = TREE_TYPE (base_val0);
|
2012-03-27 23:13:14 +00:00
|
|
|
|
bit_sz = TYPE_PRECISION (int_type);
|
|
|
|
|
gcc_assert (bit_sz > 0
|
|
|
|
|
&& bit_sz <= MAX_BASE_INT_BIT_SIZE);
|
|
|
|
|
|
|
|
|
|
/* Determine the max exp argument value according to
|
|
|
|
|
the size of the base integer. The max exp value
|
|
|
|
|
is conservatively estimated assuming IEEE754 double
|
|
|
|
|
precision format. */
|
|
|
|
|
if (bit_sz == 8)
|
|
|
|
|
max_exp = 128;
|
|
|
|
|
else if (bit_sz == 16)
|
|
|
|
|
max_exp = 64;
|
|
|
|
|
else
|
|
|
|
|
{
|
|
|
|
|
gcc_assert (bit_sz == MAX_BASE_INT_BIT_SIZE);
|
|
|
|
|
max_exp = 32;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* For pow ((double)x, y), generate the following conditions:
|
|
|
|
|
cond 1:
|
|
|
|
|
temp1 = x;
|
2017-04-10 11:32:00 +00:00
|
|
|
|
if (__builtin_islessequal (temp1, 0))
|
2012-03-27 23:13:14 +00:00
|
|
|
|
|
|
|
|
|
cond 2:
|
|
|
|
|
temp2 = y;
|
2017-04-10 11:32:00 +00:00
|
|
|
|
if (__builtin_isgreater (temp2, max_exp_real_cst)) */
|
2012-03-27 23:13:14 +00:00
|
|
|
|
|
|
|
|
|
/* Generate condition in reverse order -- first
|
|
|
|
|
the condition for the exp argument. */
|
|
|
|
|
|
|
|
|
|
exp_domain = get_domain (0, false, false,
|
|
|
|
|
max_exp, true, true);
|
|
|
|
|
|
|
|
|
|
gen_conditions_for_domain (expn, exp_domain,
|
|
|
|
|
conds, nconds);
|
|
|
|
|
|
|
|
|
|
/* Now generate condition for the base argument.
|
|
|
|
|
Note it does not use the helper function
|
|
|
|
|
gen_conditions_for_domain because the base
|
|
|
|
|
type is integer. */
|
|
|
|
|
|
|
|
|
|
/* Push a separator. */
|
2014-09-21 17:33:12 +00:00
|
|
|
|
conds.quick_push (NULL);
|
2012-03-27 23:13:14 +00:00
|
|
|
|
|
|
|
|
|
temp = create_tmp_var (int_type, "DCE_COND1");
|
|
|
|
|
cst0 = build_int_cst (int_type, 0);
|
|
|
|
|
stmt1 = gimple_build_assign (temp, base_val0);
|
|
|
|
|
tempn = make_ssa_name (temp, stmt1);
|
|
|
|
|
gimple_assign_set_lhs (stmt1, tempn);
|
2017-04-10 11:32:00 +00:00
|
|
|
|
stmt2 = gimple_build_cond (GT_EXPR, tempn, cst0, NULL_TREE, NULL_TREE);
|
2012-03-27 23:13:14 +00:00
|
|
|
|
|
2014-09-21 17:33:12 +00:00
|
|
|
|
conds.quick_push (stmt1);
|
|
|
|
|
conds.quick_push (stmt2);
|
2012-03-27 23:13:14 +00:00
|
|
|
|
(*nconds)++;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Method to generate conditional statements for guarding conditionally
|
|
|
|
|
dead calls to pow. One or more statements can be generated for
|
|
|
|
|
each logical condition. Statement groups of different conditions
|
2014-09-21 17:33:12 +00:00
|
|
|
|
are separated by a NULL tree and they are stored in the vec
|
2012-03-27 23:13:14 +00:00
|
|
|
|
conds. The number of logical conditions are stored in *nconds.
|
|
|
|
|
|
|
|
|
|
See C99 standard, 7.12.7.4:2, for description of pow (x, y).
|
|
|
|
|
The precise condition for domain errors are complex. In this
|
|
|
|
|
implementation, a simplified (but conservative) valid domain
|
|
|
|
|
for x and y are used: x is positive to avoid dom errors, while
|
|
|
|
|
y is smaller than a upper bound (depending on x) to avoid range
|
|
|
|
|
errors. Runtime code is generated to check x (if not constant)
|
|
|
|
|
and y against the valid domain. If it is out, jump to the call,
|
|
|
|
|
otherwise the call is bypassed. POW_CALL is the call statement,
|
|
|
|
|
*CONDS is a vector holding the resulting condition statements,
|
|
|
|
|
and *NCONDS is the number of logical conditions. */
|
|
|
|
|
|
|
|
|
|
static void
|
2017-04-10 11:32:00 +00:00
|
|
|
|
gen_conditions_for_pow (gcall *pow_call, vec<gimple *> conds,
|
2012-03-27 23:13:14 +00:00
|
|
|
|
unsigned *nconds)
|
|
|
|
|
{
|
|
|
|
|
tree base, expn;
|
|
|
|
|
enum tree_code bc;
|
|
|
|
|
|
|
|
|
|
gcc_checking_assert (check_pow (pow_call));
|
|
|
|
|
|
|
|
|
|
*nconds = 0;
|
|
|
|
|
|
|
|
|
|
base = gimple_call_arg (pow_call, 0);
|
|
|
|
|
expn = gimple_call_arg (pow_call, 1);
|
|
|
|
|
|
|
|
|
|
bc = TREE_CODE (base);
|
|
|
|
|
|
|
|
|
|
if (bc == REAL_CST)
|
|
|
|
|
gen_conditions_for_pow_cst_base (base, expn, conds, nconds);
|
|
|
|
|
else if (bc == SSA_NAME)
|
|
|
|
|
gen_conditions_for_pow_int_base (base, expn, conds, nconds);
|
|
|
|
|
else
|
|
|
|
|
gcc_unreachable ();
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* A helper routine to help computing the valid input domain
|
|
|
|
|
for a builtin function. See C99 7.12.7 for details. In this
|
|
|
|
|
implementation, we only handle single region domain. The
|
|
|
|
|
resulting region can be conservative (smaller) than the actual
|
|
|
|
|
one and rounded to integers. Some of the bounds are documented
|
|
|
|
|
in the standard, while other limit constants are computed
|
|
|
|
|
assuming IEEE floating point format (for SF and DF modes).
|
|
|
|
|
Since IEEE only sets minimum requirements for long double format,
|
|
|
|
|
different long double formats exist under different implementations
|
|
|
|
|
(e.g, 64 bit double precision (DF), 80 bit double-extended
|
|
|
|
|
precision (XF), and 128 bit quad precision (QF) ). For simplicity,
|
|
|
|
|
in this implementation, the computed bounds for long double assume
|
|
|
|
|
64 bit format (DF), and are therefore conservative. Another
|
|
|
|
|
assumption is that single precision float type is always SF mode,
|
|
|
|
|
and double type is DF mode. This function is quite
|
|
|
|
|
implementation specific, so it may not be suitable to be part of
|
2022-10-27 18:55:19 +00:00
|
|
|
|
builtins.cc. This needs to be revisited later to see if it can
|
2012-03-27 23:13:14 +00:00
|
|
|
|
be leveraged in x87 assembly expansion. */
|
|
|
|
|
|
|
|
|
|
static inp_domain
|
|
|
|
|
get_no_error_domain (enum built_in_function fnc)
|
|
|
|
|
{
|
|
|
|
|
switch (fnc)
|
|
|
|
|
{
|
|
|
|
|
/* Trig functions: return [-1, +1] */
|
|
|
|
|
CASE_FLT_FN (BUILT_IN_ACOS):
|
|
|
|
|
CASE_FLT_FN (BUILT_IN_ASIN):
|
|
|
|
|
return get_domain (-1, true, true,
|
|
|
|
|
1, true, true);
|
|
|
|
|
/* Hyperbolic functions. */
|
|
|
|
|
CASE_FLT_FN (BUILT_IN_ACOSH):
|
|
|
|
|
/* acosh: [1, +inf) */
|
|
|
|
|
return get_domain (1, true, true,
|
|
|
|
|
1, false, false);
|
|
|
|
|
CASE_FLT_FN (BUILT_IN_ATANH):
|
|
|
|
|
/* atanh: (-1, +1) */
|
|
|
|
|
return get_domain (-1, true, false,
|
|
|
|
|
1, true, false);
|
|
|
|
|
case BUILT_IN_COSHF:
|
|
|
|
|
case BUILT_IN_SINHF:
|
|
|
|
|
/* coshf: (-89, +89) */
|
|
|
|
|
return get_domain (-89, true, false,
|
|
|
|
|
89, true, false);
|
|
|
|
|
case BUILT_IN_COSH:
|
|
|
|
|
case BUILT_IN_SINH:
|
|
|
|
|
case BUILT_IN_COSHL:
|
|
|
|
|
case BUILT_IN_SINHL:
|
|
|
|
|
/* cosh: (-710, +710) */
|
|
|
|
|
return get_domain (-710, true, false,
|
|
|
|
|
710, true, false);
|
|
|
|
|
/* Log functions: (0, +inf) */
|
|
|
|
|
CASE_FLT_FN (BUILT_IN_LOG):
|
|
|
|
|
CASE_FLT_FN (BUILT_IN_LOG2):
|
|
|
|
|
CASE_FLT_FN (BUILT_IN_LOG10):
|
|
|
|
|
return get_domain (0, true, false,
|
|
|
|
|
0, false, false);
|
|
|
|
|
CASE_FLT_FN (BUILT_IN_LOG1P):
|
|
|
|
|
return get_domain (-1, true, false,
|
|
|
|
|
0, false, false);
|
|
|
|
|
/* Exp functions. */
|
|
|
|
|
case BUILT_IN_EXPF:
|
|
|
|
|
case BUILT_IN_EXPM1F:
|
|
|
|
|
/* expf: (-inf, 88) */
|
|
|
|
|
return get_domain (-1, false, false,
|
|
|
|
|
88, true, false);
|
|
|
|
|
case BUILT_IN_EXP:
|
|
|
|
|
case BUILT_IN_EXPM1:
|
|
|
|
|
case BUILT_IN_EXPL:
|
|
|
|
|
case BUILT_IN_EXPM1L:
|
|
|
|
|
/* exp: (-inf, 709) */
|
|
|
|
|
return get_domain (-1, false, false,
|
|
|
|
|
709, true, false);
|
|
|
|
|
case BUILT_IN_EXP2F:
|
|
|
|
|
/* exp2f: (-inf, 128) */
|
|
|
|
|
return get_domain (-1, false, false,
|
|
|
|
|
128, true, false);
|
|
|
|
|
case BUILT_IN_EXP2:
|
|
|
|
|
case BUILT_IN_EXP2L:
|
|
|
|
|
/* exp2: (-inf, 1024) */
|
|
|
|
|
return get_domain (-1, false, false,
|
|
|
|
|
1024, true, false);
|
|
|
|
|
case BUILT_IN_EXP10F:
|
|
|
|
|
case BUILT_IN_POW10F:
|
|
|
|
|
/* exp10f: (-inf, 38) */
|
|
|
|
|
return get_domain (-1, false, false,
|
|
|
|
|
38, true, false);
|
|
|
|
|
case BUILT_IN_EXP10:
|
|
|
|
|
case BUILT_IN_POW10:
|
|
|
|
|
case BUILT_IN_EXP10L:
|
|
|
|
|
case BUILT_IN_POW10L:
|
|
|
|
|
/* exp10: (-inf, 308) */
|
|
|
|
|
return get_domain (-1, false, false,
|
|
|
|
|
308, true, false);
|
|
|
|
|
/* sqrt: [0, +inf) */
|
|
|
|
|
CASE_FLT_FN (BUILT_IN_SQRT):
|
2018-12-28 15:30:48 +00:00
|
|
|
|
CASE_FLT_FN_FLOATN_NX (BUILT_IN_SQRT):
|
2012-03-27 23:13:14 +00:00
|
|
|
|
return get_domain (0, true, true,
|
|
|
|
|
0, false, false);
|
|
|
|
|
default:
|
|
|
|
|
gcc_unreachable ();
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
gcc_unreachable ();
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* The function to generate shrink wrap conditions for a partially
|
|
|
|
|
dead builtin call whose return value is not used anywhere,
|
|
|
|
|
but has to be kept live due to potential error condition.
|
|
|
|
|
BI_CALL is the builtin call, CONDS is the vector of statements
|
|
|
|
|
for condition code, NCODES is the pointer to the number of
|
|
|
|
|
logical conditions. Statements belonging to different logical
|
|
|
|
|
condition are separated by NULL tree in the vector. */
|
|
|
|
|
|
|
|
|
|
static void
|
2022-10-27 18:55:19 +00:00
|
|
|
|
gen_shrink_wrap_conditions (gcall *bi_call, const vec<gimple *> &conds,
|
2012-03-27 23:13:14 +00:00
|
|
|
|
unsigned int *nconds)
|
|
|
|
|
{
|
2015-08-28 15:33:40 +00:00
|
|
|
|
gcall *call;
|
2012-03-27 23:13:14 +00:00
|
|
|
|
tree fn;
|
|
|
|
|
enum built_in_function fnc;
|
|
|
|
|
|
2014-09-21 17:33:12 +00:00
|
|
|
|
gcc_assert (nconds && conds.exists ());
|
|
|
|
|
gcc_assert (conds.length () == 0);
|
2012-03-27 23:13:14 +00:00
|
|
|
|
gcc_assert (is_gimple_call (bi_call));
|
|
|
|
|
|
|
|
|
|
call = bi_call;
|
|
|
|
|
fn = gimple_call_fndecl (call);
|
2019-06-02 15:48:37 +00:00
|
|
|
|
gcc_assert (fn && fndecl_built_in_p (fn));
|
2012-03-27 23:13:14 +00:00
|
|
|
|
fnc = DECL_FUNCTION_CODE (fn);
|
|
|
|
|
*nconds = 0;
|
|
|
|
|
|
|
|
|
|
if (fnc == BUILT_IN_POW)
|
|
|
|
|
gen_conditions_for_pow (call, conds, nconds);
|
|
|
|
|
else
|
|
|
|
|
{
|
|
|
|
|
tree arg;
|
|
|
|
|
inp_domain domain = get_no_error_domain (fnc);
|
|
|
|
|
*nconds = 0;
|
|
|
|
|
arg = gimple_call_arg (bi_call, 0);
|
|
|
|
|
gen_conditions_for_domain (arg, domain, conds, nconds);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
return;
|
|
|
|
|
}
|
|
|
|
|
|
2017-04-10 11:32:00 +00:00
|
|
|
|
/* Shrink-wrap BI_CALL so that it is only called when one of the NCONDS
|
2022-10-27 18:55:19 +00:00
|
|
|
|
conditions in CONDS is false. Also move BI_NEWCALL to a new basic block
|
|
|
|
|
when it is non-null, it is called while all of the CONDS are true. */
|
2012-03-27 23:13:14 +00:00
|
|
|
|
|
2017-10-07 00:16:47 +00:00
|
|
|
|
static void
|
2022-10-27 18:55:19 +00:00
|
|
|
|
shrink_wrap_one_built_in_call_with_conds (gcall *bi_call,
|
|
|
|
|
const vec <gimple *> &conds,
|
|
|
|
|
unsigned int nconds,
|
|
|
|
|
gcall *bi_newcall = NULL)
|
2012-03-27 23:13:14 +00:00
|
|
|
|
{
|
|
|
|
|
gimple_stmt_iterator bi_call_bsi;
|
2022-10-27 18:55:19 +00:00
|
|
|
|
basic_block bi_call_bb, bi_newcall_bb, join_tgt_bb, guard_bb;
|
2012-03-27 23:13:14 +00:00
|
|
|
|
edge join_tgt_in_edge_from_call, join_tgt_in_edge_fall_thru;
|
|
|
|
|
edge bi_call_in_edge0, guard_bb_in_edge;
|
2017-04-10 11:32:00 +00:00
|
|
|
|
unsigned tn_cond_stmts;
|
2012-03-27 23:13:14 +00:00
|
|
|
|
unsigned ci;
|
2017-04-10 11:32:00 +00:00
|
|
|
|
gimple *cond_expr = NULL;
|
|
|
|
|
gimple *cond_expr_start;
|
|
|
|
|
|
|
|
|
|
/* The cfg we want to create looks like this:
|
2022-10-27 18:55:19 +00:00
|
|
|
|
[guard n-1] <- guard_bb (old block)
|
|
|
|
|
| \
|
|
|
|
|
| [guard n-2] }
|
|
|
|
|
| / \ }
|
|
|
|
|
| / ... } new blocks
|
|
|
|
|
| / [guard 0] }
|
|
|
|
|
| / / | }
|
|
|
|
|
[call] | <- bi_call_bb }
|
|
|
|
|
\ [newcall] <-bi_newcall_bb}
|
|
|
|
|
\ |
|
|
|
|
|
[join] <- join_tgt_bb (old iff call must end bb)
|
2017-04-10 11:32:00 +00:00
|
|
|
|
possible EH edges (only if [join] is old)
|
|
|
|
|
|
|
|
|
|
When [join] is new, the immediate dominators for these blocks are:
|
|
|
|
|
|
|
|
|
|
1. [guard n-1]: unchanged
|
|
|
|
|
2. [call]: [guard n-1]
|
2022-10-27 18:55:19 +00:00
|
|
|
|
3. [newcall]: [guard 0]
|
|
|
|
|
4. [guard m]: [guard m+1] for 0 <= m <= n-2
|
|
|
|
|
5. [join]: [guard n-1]
|
2017-04-10 11:32:00 +00:00
|
|
|
|
|
2022-10-27 18:55:19 +00:00
|
|
|
|
We punt for the more complex case of [join] being old and
|
2017-04-10 11:32:00 +00:00
|
|
|
|
simply free the dominance info. We also punt on postdominators,
|
|
|
|
|
which aren't expected to be available at this point anyway. */
|
2012-03-27 23:13:14 +00:00
|
|
|
|
bi_call_bb = gimple_bb (bi_call);
|
|
|
|
|
|
2014-09-21 17:33:12 +00:00
|
|
|
|
/* Now find the join target bb -- split bi_call_bb if needed. */
|
|
|
|
|
if (stmt_ends_bb_p (bi_call))
|
|
|
|
|
{
|
2017-10-07 00:16:47 +00:00
|
|
|
|
/* We checked that there was a fallthrough edge in
|
|
|
|
|
can_guard_call_p. */
|
2014-09-21 17:33:12 +00:00
|
|
|
|
join_tgt_in_edge_from_call = find_fallthru_edge (bi_call_bb->succs);
|
2017-10-07 00:16:47 +00:00
|
|
|
|
gcc_assert (join_tgt_in_edge_from_call);
|
2017-04-10 11:32:00 +00:00
|
|
|
|
/* We don't want to handle PHIs. */
|
|
|
|
|
if (EDGE_COUNT (join_tgt_in_edge_from_call->dest->preds) > 1)
|
|
|
|
|
join_tgt_bb = split_edge (join_tgt_in_edge_from_call);
|
|
|
|
|
else
|
2017-10-07 00:16:47 +00:00
|
|
|
|
{
|
|
|
|
|
join_tgt_bb = join_tgt_in_edge_from_call->dest;
|
|
|
|
|
/* We may have degenerate PHIs in the destination. Propagate
|
|
|
|
|
those out. */
|
|
|
|
|
for (gphi_iterator i = gsi_start_phis (join_tgt_bb); !gsi_end_p (i);)
|
|
|
|
|
{
|
|
|
|
|
gphi *phi = i.phi ();
|
|
|
|
|
replace_uses_by (gimple_phi_result (phi),
|
|
|
|
|
gimple_phi_arg_def (phi, 0));
|
|
|
|
|
remove_phi_node (&i, true);
|
|
|
|
|
}
|
|
|
|
|
}
|
2014-09-21 17:33:12 +00:00
|
|
|
|
}
|
|
|
|
|
else
|
2017-04-10 11:32:00 +00:00
|
|
|
|
{
|
|
|
|
|
join_tgt_in_edge_from_call = split_block (bi_call_bb, bi_call);
|
|
|
|
|
join_tgt_bb = join_tgt_in_edge_from_call->dest;
|
|
|
|
|
}
|
2012-03-27 23:13:14 +00:00
|
|
|
|
|
|
|
|
|
bi_call_bsi = gsi_for_stmt (bi_call);
|
|
|
|
|
|
|
|
|
|
/* Now it is time to insert the first conditional expression
|
|
|
|
|
into bi_call_bb and split this bb so that bi_call is
|
|
|
|
|
shrink-wrapped. */
|
2014-09-21 17:33:12 +00:00
|
|
|
|
tn_cond_stmts = conds.length ();
|
2012-03-27 23:13:14 +00:00
|
|
|
|
cond_expr = NULL;
|
2014-09-21 17:33:12 +00:00
|
|
|
|
cond_expr_start = conds[0];
|
2012-03-27 23:13:14 +00:00
|
|
|
|
for (ci = 0; ci < tn_cond_stmts; ci++)
|
|
|
|
|
{
|
2017-04-10 11:32:00 +00:00
|
|
|
|
gimple *c = conds[ci];
|
2012-03-27 23:13:14 +00:00
|
|
|
|
gcc_assert (c || ci != 0);
|
|
|
|
|
if (!c)
|
|
|
|
|
break;
|
|
|
|
|
gsi_insert_before (&bi_call_bsi, c, GSI_SAME_STMT);
|
|
|
|
|
cond_expr = c;
|
|
|
|
|
}
|
|
|
|
|
ci++;
|
|
|
|
|
gcc_assert (cond_expr && gimple_code (cond_expr) == GIMPLE_COND);
|
|
|
|
|
|
2017-10-07 00:16:47 +00:00
|
|
|
|
typedef std::pair<edge, edge> edge_pair;
|
|
|
|
|
auto_vec<edge_pair, 8> edges;
|
|
|
|
|
|
2012-03-27 23:13:14 +00:00
|
|
|
|
bi_call_in_edge0 = split_block (bi_call_bb, cond_expr);
|
|
|
|
|
bi_call_in_edge0->flags &= ~EDGE_FALLTHRU;
|
2017-04-10 11:32:00 +00:00
|
|
|
|
bi_call_in_edge0->flags |= EDGE_FALSE_VALUE;
|
|
|
|
|
guard_bb = bi_call_bb;
|
2012-03-27 23:13:14 +00:00
|
|
|
|
bi_call_bb = bi_call_in_edge0->dest;
|
2017-04-10 11:32:00 +00:00
|
|
|
|
join_tgt_in_edge_fall_thru = make_edge (guard_bb, join_tgt_bb,
|
|
|
|
|
EDGE_TRUE_VALUE);
|
2012-03-27 23:13:14 +00:00
|
|
|
|
|
2017-10-07 00:16:47 +00:00
|
|
|
|
edges.reserve (nconds);
|
|
|
|
|
edges.quick_push (edge_pair (bi_call_in_edge0, join_tgt_in_edge_fall_thru));
|
2012-03-27 23:13:14 +00:00
|
|
|
|
|
|
|
|
|
/* Code generation for the rest of the conditions */
|
2017-10-07 00:16:47 +00:00
|
|
|
|
for (unsigned int i = 1; i < nconds; ++i)
|
2012-03-27 23:13:14 +00:00
|
|
|
|
{
|
|
|
|
|
unsigned ci0;
|
|
|
|
|
edge bi_call_in_edge;
|
|
|
|
|
gimple_stmt_iterator guard_bsi = gsi_for_stmt (cond_expr_start);
|
|
|
|
|
ci0 = ci;
|
2014-09-21 17:33:12 +00:00
|
|
|
|
cond_expr_start = conds[ci0];
|
2012-03-27 23:13:14 +00:00
|
|
|
|
for (; ci < tn_cond_stmts; ci++)
|
|
|
|
|
{
|
2017-04-10 11:32:00 +00:00
|
|
|
|
gimple *c = conds[ci];
|
2012-03-27 23:13:14 +00:00
|
|
|
|
gcc_assert (c || ci != ci0);
|
|
|
|
|
if (!c)
|
|
|
|
|
break;
|
|
|
|
|
gsi_insert_before (&guard_bsi, c, GSI_SAME_STMT);
|
|
|
|
|
cond_expr = c;
|
|
|
|
|
}
|
|
|
|
|
ci++;
|
|
|
|
|
gcc_assert (cond_expr && gimple_code (cond_expr) == GIMPLE_COND);
|
|
|
|
|
guard_bb_in_edge = split_block (guard_bb, cond_expr);
|
|
|
|
|
guard_bb_in_edge->flags &= ~EDGE_FALLTHRU;
|
2017-04-10 11:32:00 +00:00
|
|
|
|
guard_bb_in_edge->flags |= EDGE_TRUE_VALUE;
|
2012-03-27 23:13:14 +00:00
|
|
|
|
|
2017-04-10 11:32:00 +00:00
|
|
|
|
bi_call_in_edge = make_edge (guard_bb, bi_call_bb, EDGE_FALSE_VALUE);
|
2017-10-07 00:16:47 +00:00
|
|
|
|
edges.quick_push (edge_pair (bi_call_in_edge, guard_bb_in_edge));
|
|
|
|
|
}
|
2012-03-27 23:13:14 +00:00
|
|
|
|
|
2022-10-27 18:55:19 +00:00
|
|
|
|
/* Move BI_NEWCALL to new basic block when it is non-null. */
|
|
|
|
|
if (bi_newcall)
|
|
|
|
|
{
|
|
|
|
|
/* Get bi_newcall_bb by split join_tgt_in_edge_fall_thru edge,
|
|
|
|
|
and move BI_NEWCALL to bi_newcall_bb. */
|
|
|
|
|
bi_newcall_bb = split_edge (join_tgt_in_edge_fall_thru);
|
|
|
|
|
gimple_stmt_iterator to_gsi = gsi_start_bb (bi_newcall_bb);
|
|
|
|
|
gimple_stmt_iterator from_gsi = gsi_for_stmt (bi_newcall);
|
|
|
|
|
gsi_move_before (&from_gsi, &to_gsi);
|
|
|
|
|
join_tgt_in_edge_fall_thru = EDGE_SUCC (bi_newcall_bb, 0);
|
|
|
|
|
join_tgt_bb = join_tgt_in_edge_fall_thru->dest;
|
|
|
|
|
|
|
|
|
|
tree bi_newcall_lhs = gimple_call_lhs (bi_newcall);
|
|
|
|
|
tree bi_call_lhs = gimple_call_lhs (bi_call);
|
|
|
|
|
if (!bi_call_lhs)
|
|
|
|
|
{
|
|
|
|
|
bi_call_lhs = copy_ssa_name (bi_newcall_lhs);
|
|
|
|
|
gimple_call_set_lhs (bi_call, bi_call_lhs);
|
|
|
|
|
SSA_NAME_DEF_STMT (bi_call_lhs) = bi_call;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Create phi node for lhs of BI_CALL and BI_NEWCALL. */
|
|
|
|
|
gphi *new_phi = create_phi_node (copy_ssa_name (bi_newcall_lhs),
|
|
|
|
|
join_tgt_bb);
|
|
|
|
|
SSA_NAME_OCCURS_IN_ABNORMAL_PHI (PHI_RESULT (new_phi))
|
|
|
|
|
= SSA_NAME_OCCURS_IN_ABNORMAL_PHI (bi_newcall_lhs);
|
|
|
|
|
add_phi_arg (new_phi, bi_call_lhs, join_tgt_in_edge_from_call,
|
|
|
|
|
gimple_location (bi_call));
|
|
|
|
|
add_phi_arg (new_phi, bi_newcall_lhs, join_tgt_in_edge_fall_thru,
|
|
|
|
|
gimple_location (bi_newcall));
|
|
|
|
|
|
|
|
|
|
/* Replace all use of original return value with result of phi node. */
|
|
|
|
|
use_operand_p use_p;
|
|
|
|
|
gimple *use_stmt;
|
|
|
|
|
imm_use_iterator iterator;
|
|
|
|
|
FOR_EACH_IMM_USE_STMT (use_stmt, iterator, bi_newcall_lhs)
|
|
|
|
|
if (use_stmt != new_phi)
|
|
|
|
|
FOR_EACH_IMM_USE_ON_STMT (use_p, iterator)
|
|
|
|
|
SET_USE (use_p, PHI_RESULT (new_phi));
|
|
|
|
|
}
|
|
|
|
|
|
2017-10-07 00:16:47 +00:00
|
|
|
|
/* Now update the probability and profile information, processing the
|
|
|
|
|
guards in order of execution.
|
|
|
|
|
|
|
|
|
|
There are two approaches we could take here. On the one hand we
|
|
|
|
|
could assign a probability of X to the call block and distribute
|
|
|
|
|
that probability among its incoming edges. On the other hand we
|
|
|
|
|
could assign a probability of X to each individual call edge.
|
|
|
|
|
|
|
|
|
|
The choice only affects calls that have more than one condition.
|
|
|
|
|
In those cases, the second approach would give the call block
|
|
|
|
|
a greater probability than the first. However, the difference
|
|
|
|
|
is only small, and our chosen X is a pure guess anyway.
|
|
|
|
|
|
|
|
|
|
Here we take the second approach because it's slightly simpler
|
|
|
|
|
and because it's easy to see that it doesn't lose profile counts. */
|
2018-12-28 15:30:48 +00:00
|
|
|
|
bi_call_bb->count = profile_count::zero ();
|
2017-10-07 00:16:47 +00:00
|
|
|
|
while (!edges.is_empty ())
|
|
|
|
|
{
|
|
|
|
|
edge_pair e = edges.pop ();
|
|
|
|
|
edge call_edge = e.first;
|
|
|
|
|
edge nocall_edge = e.second;
|
|
|
|
|
basic_block src_bb = call_edge->src;
|
|
|
|
|
gcc_assert (src_bb == nocall_edge->src);
|
|
|
|
|
|
2018-12-28 15:30:48 +00:00
|
|
|
|
call_edge->probability = profile_probability::very_unlikely ();
|
|
|
|
|
nocall_edge->probability = profile_probability::always ()
|
|
|
|
|
- call_edge->probability;
|
2017-10-07 00:16:47 +00:00
|
|
|
|
|
2018-12-28 15:30:48 +00:00
|
|
|
|
bi_call_bb->count += call_edge->count ();
|
2017-10-07 00:16:47 +00:00
|
|
|
|
|
|
|
|
|
if (nocall_edge->dest != join_tgt_bb)
|
2018-12-28 15:30:48 +00:00
|
|
|
|
nocall_edge->dest->count = src_bb->count - bi_call_bb->count;
|
2012-03-27 23:13:14 +00:00
|
|
|
|
}
|
|
|
|
|
|
2017-04-10 11:32:00 +00:00
|
|
|
|
if (dom_info_available_p (CDI_DOMINATORS))
|
|
|
|
|
{
|
|
|
|
|
/* The split_blocks leave [guard 0] as the immediate dominator
|
|
|
|
|
of [call] and [call] as the immediate dominator of [join].
|
|
|
|
|
Fix them up. */
|
|
|
|
|
set_immediate_dominator (CDI_DOMINATORS, bi_call_bb, guard_bb);
|
|
|
|
|
set_immediate_dominator (CDI_DOMINATORS, join_tgt_bb, guard_bb);
|
|
|
|
|
}
|
|
|
|
|
|
2012-03-27 23:13:14 +00:00
|
|
|
|
if (dump_file && (dump_flags & TDF_DETAILS))
|
|
|
|
|
{
|
|
|
|
|
location_t loc;
|
|
|
|
|
loc = gimple_location (bi_call);
|
|
|
|
|
fprintf (dump_file,
|
|
|
|
|
"%s:%d: note: function call is shrink-wrapped"
|
|
|
|
|
" into error conditions.\n",
|
|
|
|
|
LOCATION_FILE (loc), LOCATION_LINE (loc));
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
2017-04-10 11:32:00 +00:00
|
|
|
|
/* Shrink-wrap BI_CALL so that it is only called when it might set errno
|
2017-10-07 00:16:47 +00:00
|
|
|
|
(but is always called if it would set errno). */
|
2017-04-10 11:32:00 +00:00
|
|
|
|
|
2017-10-07 00:16:47 +00:00
|
|
|
|
static void
|
2017-04-10 11:32:00 +00:00
|
|
|
|
shrink_wrap_one_built_in_call (gcall *bi_call)
|
|
|
|
|
{
|
|
|
|
|
unsigned nconds = 0;
|
|
|
|
|
auto_vec<gimple *, 12> conds;
|
|
|
|
|
gen_shrink_wrap_conditions (bi_call, conds, &nconds);
|
2017-10-07 00:16:47 +00:00
|
|
|
|
gcc_assert (nconds != 0);
|
|
|
|
|
shrink_wrap_one_built_in_call_with_conds (bi_call, conds, nconds);
|
2017-04-10 11:32:00 +00:00
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Return true if built-in function call CALL could be implemented using
|
|
|
|
|
a combination of an internal function to compute the result and a
|
|
|
|
|
separate call to set errno. */
|
|
|
|
|
|
|
|
|
|
static bool
|
|
|
|
|
can_use_internal_fn (gcall *call)
|
|
|
|
|
{
|
|
|
|
|
/* Only replace calls that set errno. */
|
|
|
|
|
if (!gimple_vdef (call))
|
|
|
|
|
return false;
|
|
|
|
|
|
|
|
|
|
/* See whether there is an internal function for this built-in. */
|
|
|
|
|
if (replacement_internal_fn (call) == IFN_LAST)
|
|
|
|
|
return false;
|
|
|
|
|
|
|
|
|
|
/* See whether we can catch all cases where errno would be set,
|
|
|
|
|
while still avoiding the call in most cases. */
|
|
|
|
|
if (!can_test_argument_range (call)
|
|
|
|
|
&& !edom_only_function (call))
|
|
|
|
|
return false;
|
|
|
|
|
|
|
|
|
|
return true;
|
|
|
|
|
}
|
|
|
|
|
|
2017-10-07 00:16:47 +00:00
|
|
|
|
/* Implement built-in function call CALL using an internal function. */
|
2017-04-10 11:32:00 +00:00
|
|
|
|
|
2017-10-07 00:16:47 +00:00
|
|
|
|
static void
|
2017-04-10 11:32:00 +00:00
|
|
|
|
use_internal_fn (gcall *call)
|
|
|
|
|
{
|
2017-10-07 00:16:47 +00:00
|
|
|
|
/* We'll be inserting another call with the same arguments after the
|
|
|
|
|
lhs has been set, so prevent any possible coalescing failure from
|
|
|
|
|
having both values live at once. See PR 71020. */
|
|
|
|
|
replace_abnormal_ssa_names (call);
|
|
|
|
|
|
2017-04-10 11:32:00 +00:00
|
|
|
|
unsigned nconds = 0;
|
|
|
|
|
auto_vec<gimple *, 12> conds;
|
2022-10-27 18:55:19 +00:00
|
|
|
|
bool is_arg_conds = false;
|
2017-04-10 11:32:00 +00:00
|
|
|
|
if (can_test_argument_range (call))
|
2017-10-07 00:16:47 +00:00
|
|
|
|
{
|
|
|
|
|
gen_shrink_wrap_conditions (call, conds, &nconds);
|
2022-10-27 18:55:19 +00:00
|
|
|
|
is_arg_conds = true;
|
2017-10-07 00:16:47 +00:00
|
|
|
|
gcc_assert (nconds != 0);
|
|
|
|
|
}
|
|
|
|
|
else
|
|
|
|
|
gcc_assert (edom_only_function (call));
|
2017-04-10 11:32:00 +00:00
|
|
|
|
|
|
|
|
|
internal_fn ifn = replacement_internal_fn (call);
|
|
|
|
|
gcc_assert (ifn != IFN_LAST);
|
|
|
|
|
|
|
|
|
|
/* Construct the new call, with the same arguments as the original one. */
|
|
|
|
|
auto_vec <tree, 16> args;
|
|
|
|
|
unsigned int nargs = gimple_call_num_args (call);
|
|
|
|
|
for (unsigned int i = 0; i < nargs; ++i)
|
|
|
|
|
args.safe_push (gimple_call_arg (call, i));
|
|
|
|
|
gcall *new_call = gimple_build_call_internal_vec (ifn, args);
|
|
|
|
|
gimple_set_location (new_call, gimple_location (call));
|
2018-12-28 15:30:48 +00:00
|
|
|
|
gimple_call_set_nothrow (new_call, gimple_call_nothrow_p (call));
|
2017-04-10 11:32:00 +00:00
|
|
|
|
|
|
|
|
|
/* Transfer the LHS to the new call. */
|
|
|
|
|
tree lhs = gimple_call_lhs (call);
|
|
|
|
|
gimple_call_set_lhs (new_call, lhs);
|
|
|
|
|
gimple_call_set_lhs (call, NULL_TREE);
|
|
|
|
|
SSA_NAME_DEF_STMT (lhs) = new_call;
|
|
|
|
|
|
|
|
|
|
/* Insert the new call. */
|
|
|
|
|
gimple_stmt_iterator gsi = gsi_for_stmt (call);
|
|
|
|
|
gsi_insert_before (&gsi, new_call, GSI_SAME_STMT);
|
|
|
|
|
|
|
|
|
|
if (nconds == 0)
|
|
|
|
|
{
|
|
|
|
|
/* Skip the call if LHS == LHS. If we reach here, EDOM is the only
|
|
|
|
|
valid errno value and it is used iff the result is NaN. */
|
|
|
|
|
conds.quick_push (gimple_build_cond (EQ_EXPR, lhs, lhs,
|
|
|
|
|
NULL_TREE, NULL_TREE));
|
|
|
|
|
nconds++;
|
|
|
|
|
|
|
|
|
|
/* Try replacing the original call with a direct assignment to
|
|
|
|
|
errno, via an internal function. */
|
|
|
|
|
if (set_edom_supported_p () && !stmt_ends_bb_p (call))
|
|
|
|
|
{
|
|
|
|
|
gimple_stmt_iterator gsi = gsi_for_stmt (call);
|
|
|
|
|
gcall *new_call = gimple_build_call_internal (IFN_SET_EDOM, 0);
|
2022-10-27 18:55:19 +00:00
|
|
|
|
gimple_move_vops (new_call, call);
|
2017-04-10 11:32:00 +00:00
|
|
|
|
gimple_set_location (new_call, gimple_location (call));
|
|
|
|
|
gsi_replace (&gsi, new_call, false);
|
|
|
|
|
call = new_call;
|
|
|
|
|
}
|
|
|
|
|
}
|
2022-10-27 18:55:19 +00:00
|
|
|
|
shrink_wrap_one_built_in_call_with_conds (call, conds, nconds,
|
|
|
|
|
is_arg_conds ? new_call : NULL);
|
2017-04-10 11:32:00 +00:00
|
|
|
|
}
|
|
|
|
|
|
2012-03-27 23:13:14 +00:00
|
|
|
|
/* The top level function for conditional dead code shrink
|
|
|
|
|
wrapping transformation. */
|
|
|
|
|
|
2017-10-07 00:16:47 +00:00
|
|
|
|
static void
|
2022-10-27 18:55:19 +00:00
|
|
|
|
shrink_wrap_conditional_dead_built_in_calls (const vec<gcall *> &calls)
|
2012-03-27 23:13:14 +00:00
|
|
|
|
{
|
|
|
|
|
unsigned i = 0;
|
|
|
|
|
|
2014-09-21 17:33:12 +00:00
|
|
|
|
unsigned n = calls.length ();
|
2012-03-27 23:13:14 +00:00
|
|
|
|
for (; i < n ; i++)
|
|
|
|
|
{
|
2015-08-28 15:33:40 +00:00
|
|
|
|
gcall *bi_call = calls[i];
|
2017-04-10 11:32:00 +00:00
|
|
|
|
if (gimple_call_lhs (bi_call))
|
2017-10-07 00:16:47 +00:00
|
|
|
|
use_internal_fn (bi_call);
|
2017-04-10 11:32:00 +00:00
|
|
|
|
else
|
2017-10-07 00:16:47 +00:00
|
|
|
|
shrink_wrap_one_built_in_call (bi_call);
|
2012-03-27 23:13:14 +00:00
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
2015-08-28 15:33:40 +00:00
|
|
|
|
namespace {
|
|
|
|
|
|
|
|
|
|
const pass_data pass_data_call_cdce =
|
|
|
|
|
{
|
|
|
|
|
GIMPLE_PASS, /* type */
|
|
|
|
|
"cdce", /* name */
|
|
|
|
|
OPTGROUP_NONE, /* optinfo_flags */
|
|
|
|
|
TV_TREE_CALL_CDCE, /* tv_id */
|
|
|
|
|
( PROP_cfg | PROP_ssa ), /* properties_required */
|
|
|
|
|
0, /* properties_provided */
|
|
|
|
|
0, /* properties_destroyed */
|
|
|
|
|
0, /* todo_flags_start */
|
|
|
|
|
0, /* todo_flags_finish */
|
|
|
|
|
};
|
|
|
|
|
|
|
|
|
|
class pass_call_cdce : public gimple_opt_pass
|
|
|
|
|
{
|
|
|
|
|
public:
|
|
|
|
|
pass_call_cdce (gcc::context *ctxt)
|
|
|
|
|
: gimple_opt_pass (pass_data_call_cdce, ctxt)
|
|
|
|
|
{}
|
|
|
|
|
|
|
|
|
|
/* opt_pass methods: */
|
2017-04-10 11:32:00 +00:00
|
|
|
|
virtual bool gate (function *)
|
2015-08-28 15:33:40 +00:00
|
|
|
|
{
|
|
|
|
|
/* The limit constants used in the implementation
|
|
|
|
|
assume IEEE floating point format. Other formats
|
|
|
|
|
can be supported in the future if needed. */
|
2017-04-10 11:32:00 +00:00
|
|
|
|
return flag_tree_builtin_call_dce != 0;
|
2015-08-28 15:33:40 +00:00
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
virtual unsigned int execute (function *);
|
2012-03-27 23:13:14 +00:00
|
|
|
|
|
2015-08-28 15:33:40 +00:00
|
|
|
|
}; // class pass_call_cdce
|
|
|
|
|
|
|
|
|
|
unsigned int
|
|
|
|
|
pass_call_cdce::execute (function *fun)
|
2012-03-27 23:13:14 +00:00
|
|
|
|
{
|
|
|
|
|
basic_block bb;
|
|
|
|
|
gimple_stmt_iterator i;
|
2015-08-28 15:33:40 +00:00
|
|
|
|
auto_vec<gcall *> cond_dead_built_in_calls;
|
|
|
|
|
FOR_EACH_BB_FN (bb, fun)
|
2012-03-27 23:13:14 +00:00
|
|
|
|
{
|
2017-04-10 11:32:00 +00:00
|
|
|
|
/* Skip blocks that are being optimized for size, since our
|
|
|
|
|
transformation always increases code size. */
|
|
|
|
|
if (optimize_bb_for_size_p (bb))
|
|
|
|
|
continue;
|
|
|
|
|
|
2012-03-27 23:13:14 +00:00
|
|
|
|
/* Collect dead call candidates. */
|
|
|
|
|
for (i = gsi_start_bb (bb); !gsi_end_p (i); gsi_next (&i))
|
|
|
|
|
{
|
2015-08-28 15:33:40 +00:00
|
|
|
|
gcall *stmt = dyn_cast <gcall *> (gsi_stmt (i));
|
2017-04-10 11:32:00 +00:00
|
|
|
|
if (stmt
|
|
|
|
|
&& gimple_call_builtin_p (stmt, BUILT_IN_NORMAL)
|
|
|
|
|
&& (gimple_call_lhs (stmt)
|
|
|
|
|
? can_use_internal_fn (stmt)
|
2017-10-07 00:16:47 +00:00
|
|
|
|
: can_test_argument_range (stmt))
|
|
|
|
|
&& can_guard_call_p (stmt))
|
2012-03-27 23:13:14 +00:00
|
|
|
|
{
|
|
|
|
|
if (dump_file && (dump_flags & TDF_DETAILS))
|
|
|
|
|
{
|
|
|
|
|
fprintf (dump_file, "Found conditional dead call: ");
|
|
|
|
|
print_gimple_stmt (dump_file, stmt, 0, TDF_SLIM);
|
|
|
|
|
fprintf (dump_file, "\n");
|
|
|
|
|
}
|
2014-09-21 17:33:12 +00:00
|
|
|
|
if (!cond_dead_built_in_calls.exists ())
|
|
|
|
|
cond_dead_built_in_calls.create (64);
|
|
|
|
|
cond_dead_built_in_calls.safe_push (stmt);
|
2012-03-27 23:13:14 +00:00
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
2014-09-21 17:33:12 +00:00
|
|
|
|
if (!cond_dead_built_in_calls.exists ())
|
2012-03-27 23:13:14 +00:00
|
|
|
|
return 0;
|
|
|
|
|
|
2017-10-07 00:16:47 +00:00
|
|
|
|
shrink_wrap_conditional_dead_built_in_calls (cond_dead_built_in_calls);
|
|
|
|
|
free_dominance_info (CDI_POST_DOMINATORS);
|
|
|
|
|
/* As we introduced new control-flow we need to insert PHI-nodes
|
|
|
|
|
for the call-clobbers of the remaining call. */
|
|
|
|
|
mark_virtual_operands_for_renaming (fun);
|
|
|
|
|
return TODO_update_ssa;
|
2012-03-27 23:13:14 +00:00
|
|
|
|
}
|
|
|
|
|
|
2014-09-21 17:33:12 +00:00
|
|
|
|
} // anon namespace
|
|
|
|
|
|
|
|
|
|
gimple_opt_pass *
|
|
|
|
|
make_pass_call_cdce (gcc::context *ctxt)
|
|
|
|
|
{
|
|
|
|
|
return new pass_call_cdce (ctxt);
|
|
|
|
|
}
|