+2004-09-14 Kazu Hirata <kazu@cs.umass.edu>
+
+ * cgraph.h, cgraphunit.c, fold-const.c, lambda-mat.c,
+ lambda.h, profile.c, tree-optimize.c, tree-sra.c, vec.h: Fix
+ comment typos. Follow spelling conventions.
+
2004-09-14 Kazu Hirata <kazu@cs.umass.edu>
* config/darwin-c.c, config/alpha/alpha.c, config/i386/i386.c,
rest of the compilation continues. When this reorgination is done,
it will no longer be necessary to have the _decl_uid versions of
local_static_vars_info and global_static_vars_info structures.
- Having both structures is now requirred because the _ann_uid values
+ Having both structures is now required because the _ann_uid values
for static variables are reset as each function is compiled.
Currently, the analysis is done using the _decl_uid versions and
converted to the _var_ann versions on demand.
/* Additionally this file gathers information about how local statics
- are used. This is done in cgraph_charactize_statics. After the
+ are used. This is done in cgraph_characterize_statics. After the
call graph has been built, each function is analyzed to determine
which local static variables are either read or written or have
their address taken. Any local static that has its address taken
}
/* Check to see if T is a read or address of operation on a static var
- we are interrested in analyzing. FN is passed in to get access to
+ we are interested in analyzing. FN is passed in to get access to
its bit vectors. */
static void
}
}
-/* Check to see if T is an assignement to a static var we are
+/* Check to see if T is an assignment to a static var we are
interrested in analyzing. FN is passed in to get access to its bit
vectors.
*/
/* FIXME -- PROFILE-RESTRUCTURE: Change all *_decl_uid to *_ann_uid. */
/* Or in all of the bits from every callee into X, the caller's, bit
- vector. There are several cases to check to avoid the sparce
+ vector. There are several cases to check to avoid the sparse
bitmap oring. */
static void
cgraph_propagate_bits (struct cgraph_node *x)
indicates whether constant overflow has already occurred. We force
T's value to be within range of T's type (by setting to 0 or 1 all
the bits outside the type's range). We set TREE_OVERFLOWED if,
- OVERFLOWED is non-zero,
+ OVERFLOWED is nonzero,
or OVERFLOWABLE is >0 and signed overflow occurs
or OVERFLOWABLE is <0 and any overflow occurs
We set TREE_CONSTANT_OVERFLOWED if,
- CONST_OVERFLOWED is non-zero
+ CONST_OVERFLOWED is nonzero
or we set TREE_OVERFLOWED.
We return either the original T, or a copy. */
}
}
- /* Stop when only the diagonal element is non-zero. */
+ /* Stop when only the diagonal element is nonzero. */
while (lambda_vector_first_nz (row, n, j + 1) < n)
{
minimum_col = lambda_vector_min_nz (row, n, j);
}
}
-/* When it exists, return the first non-zero row in MAT after row
+/* When it exists, return the first nonzero row in MAT after row
STARTROW. Otherwise return rowsize. */
int
return true;
}
-/* Return the minimum non-zero element in vector VEC1 between START and N.
+/* Return the minimum nonzero element in vector VEC1 between START and N.
We must have START <= N. */
static inline int
#define BB_TO_GCOV_INDEX(bb) ((bb)->index + 1)
/* When passed NULL as file_name, initialize.
- When passed something else, output the neccesary commands to change
+ When passed something else, output the necessary commands to change
line to LINE and offset to FILE_NAME. */
static void
output_location (char const *file_name, int line,
&offset, bb);
}
- /* Notice GOTO expressions we elliminated while constructing the
+ /* Notice GOTO expressions we eliminated while constructing the
CFG. */
if (bb->succ && !bb->succ->succ_next && bb->succ->goto_locus)
{
next_pass_1 (struct tree_opt_pass **list, struct tree_opt_pass *pass)
{
- /* A non-zero static_pass_number indicates that the
+ /* A nonzero static_pass_number indicates that the
pass is already in the list. */
if (pass->static_pass_number)
{
if (bitmap_bit_p (sra_type_decomp_cache, cache+1))
return false;
- /* The type must have a definite non-zero size. */
+ /* The type must have a definite nonzero size. */
if (TYPE_SIZE (type) == NULL || integer_zerop (TYPE_SIZE (type)))
goto fail;
int VEC_T_space (VEC(T) *v,int reserve)
- If V has space for RESERVE additional entries, return non-zero. If
+ If V has space for RESERVE additional entries, return nonzero. If
RESERVE is < 0, ensure there is at least one space slot. You
usually only need to use this if you are doing your own vector
reallocation, for instance on an embedded vector. This returns
- non-zero in exactly the same circumstances that VEC_T_reserve
+ nonzero in exactly the same circumstances that VEC_T_reserve
will. */
#define VEC_space(TDEF,V,R) (VEC_OP(TDEF,space)(V,R))
slot. These differ in their reallocation behaviour, the first will
not create additional headroom, but the second mechanism will
perform the usual exponential headroom increase. Note this can
- cause V to be reallocated. Returns non-zero iff reallocation
+ cause V to be reallocated. Returns nonzero iff reallocation
actually occurred. */
#define VEC_reserve(TDEF,V,R) (VEC_OP(TDEF,reserve)(&(V),R MEM_STAT_INFO))