From: Pierre Muller Date: Fri, 2 Nov 2012 18:19:36 +0000 (+0000) Subject: Re-added with correct permissions X-Git-Url: https://git.libre-soc.org/?a=commitdiff_plain;h=b786c521573fce15b05c4b879e05f9cc592e1e4b;p=binutils-gdb.git Re-added with correct permissions --- diff --git a/gdb/contrib/ari/create-web-ari-in-src.sh b/gdb/contrib/ari/create-web-ari-in-src.sh new file mode 100644 index 00000000000..fe9155b005b --- /dev/null +++ b/gdb/contrib/ari/create-web-ari-in-src.sh @@ -0,0 +1,77 @@ +#! /bin/sh + +# GDB script to create web ARI page directly from within gdb/ari directory. +# +# Copyright (C) 2012 Free Software Foundation, Inc. +# +# This file is part of GDB. +# +# This program is free software; you can redistribute it and/or modify +# it under the terms of the GNU General Public License as published by +# the Free Software Foundation; either version 3 of the License, or +# (at your option) any later version. +# +# This program is distributed in the hope that it will be useful, +# but WITHOUT ANY WARRANTY; without even the implied warranty of +# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the +# GNU General Public License for more details. +# +# You should have received a copy of the GNU General Public License +# along with this program. If not, see . + +# Determine directory of current script. +scriptpath=`dirname $0` +# If "scriptpath" is a relative path, then convert it to absolute. +if [ "`echo ${scriptpath} | cut -b1`" != '/' ] ; then + scriptpath="`pwd`/${scriptpath}" +fi + +# update-web-ari.sh script wants four parameters +# 1: directory of checkout src or gdb-RELEASE for release sources. +# 2: a temp directory. +# 3: a directory for generated web page. +# 4: The name of the current package, must be gdb here. +# Here we provide default values for these 4 parameters + +# srcdir parameter +if [ -z "${srcdir}" ] ; then + srcdir=${scriptpath}/../../.. +fi + +# Determine location of a temporary directory to be used by +# update-web-ari.sh script. +if [ -z "${tempdir}" ] ; then + if [ ! -z "$TMP" ] ; then + tempdir=$TMP/create-ari + elif [ ! -z "$TEMP" ] ; then + tempdir=$TEMP/create-ari + else + tempdir=/tmp/create-ari + fi +fi + +# Default location of generate index.hmtl web page. +if [ -z "${webdir}" ] ; then +# Use 'branch' subdir name if Tag contains branch + if [ -f "${srcdir}/gdb/CVS/Tag" ] ; then + tagname=`cat "${srcdir}/gdb/CVS/Tag"` + else + tagname=trunk + fi + if [ "${tagname#branch}" != "${tagname}" ] ; then + subdir=branch + else + subdir=trunk + fi + webdir=`pwd`/${subdir}/ari +fi + +# Launch update-web-ari.sh in same directory as current script. +${scriptpath}/update-web-ari.sh ${srcdir} ${tempdir} ${webdir} gdb + +if [ -f "${webdir}/index.html" ] ; then + echo "ARI output can be viewed in file \"${webdir}/index.html\"" +else + echo "ARI script failed to generate file \"${webdir}/index.html\"" +fi + diff --git a/gdb/contrib/ari/gdb_ari.sh b/gdb/contrib/ari/gdb_ari.sh new file mode 100644 index 00000000000..21ad58eb918 --- /dev/null +++ b/gdb/contrib/ari/gdb_ari.sh @@ -0,0 +1,1351 @@ +#!/bin/sh + +# GDB script to list of problems using awk. +# +# Copyright (C) 2002-2012 Free Software Foundation, Inc. +# +# This file is part of GDB. +# +# This program is free software; you can redistribute it and/or modify +# it under the terms of the GNU General Public License as published by +# the Free Software Foundation; either version 3 of the License, or +# (at your option) any later version. +# +# This program is distributed in the hope that it will be useful, +# but WITHOUT ANY WARRANTY; without even the implied warranty of +# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the +# GNU General Public License for more details. +# +# You should have received a copy of the GNU General Public License +# along with this program. If not, see . + +# Make certain that the script is not running in an internationalized +# environment. + +LANG=c ; export LANG +LC_ALL=c ; export LC_ALL + +# Permanent checks take the form: + +# Do not use XXXX, ISO C 90 implies YYYY +# Do not use XXXX, instead use YYYY''. + +# and should never be removed. + +# Temporary checks take the form: + +# Replace XXXX with YYYY + +# and once they reach zero, can be eliminated. + +# FIXME: It should be able to override this on the command line +error="regression" +warning="regression" +ari="regression eol code comment deprecated legacy obsolete gettext" +all="regression eol code comment deprecated legacy obsolete gettext deprecate internal gdbarch macro" +print_doc=0 +print_idx=0 + +usage () +{ + cat <&2 +Error: $1 + +Usage: + $0 --print-doc --print-idx -Wall -Werror -W ... +Options: + --print-doc Print a list of all potential problems, then exit. + --print-idx Include the problems IDX (index or key) in every message. + --src=file Write source lines to file. + -Werror Treat all problems as errors. + -Wall Report all problems. + -Wari Report problems that should be fixed in new code. + -W Report problems in the specifed category. Vaid categories + are: ${all} +EOF + exit 1 +} + + +# Parse the various options +Woptions= +srclines="" +while test $# -gt 0 +do + case "$1" in + -Wall ) Woptions="${all}" ;; + -Wari ) Woptions="${ari}" ;; + -Werror ) Werror=1 ;; + -W* ) Woptions="${Woptions} `echo x$1 | sed -e 's/x-W//'`" ;; + --print-doc ) print_doc=1 ;; + --print-idx ) print_idx=1 ;; + --src=* ) srclines="`echo $1 | sed -e 's/--src=/srclines=\"/'`\"" ;; + -- ) shift ; break ;; + - ) break ;; + -* ) usage "$1: unknown option" ;; + * ) break ;; + esac + shift +done +if test -n "$Woptions" ; then + warning="$Woptions" + error= +fi + + +# -Werror implies treating all warnings as errors. +if test -n "${Werror}" ; then + error="${error} ${warning}" +fi + + +# Validate all errors and warnings. +for w in ${warning} ${error} +do + case " ${all} " in + *" ${w} "* ) ;; + * ) usage "Unknown option -W${w}" ;; + esac +done + + +# make certain that there is at least one file. +if test $# -eq 0 -a ${print_doc} = 0 +then + usage "Missing file." +fi + + +# Convert the errors/warnings into corresponding array entries. +for a in ${all} +do + aris="${aris} ari_${a} = \"${a}\";" +done +for w in ${warning} +do + warnings="${warnings} warning[ari_${w}] = 1;" +done +for e in ${error} +do + errors="${errors} error[ari_${e}] = 1;" +done + +if [ "$AWK" == "" ] ; then + AWK=awk +fi + +${AWK} -- ' +BEGIN { + # NOTE, for a per-file begin use "FNR == 1". + '"${aris}"' + '"${errors}"' + '"${warnings}"' + '"${srclines}"' + print_doc = '$print_doc' + print_idx = '$print_idx' + PWD = "'`pwd`'" +} + +# Print the error message for BUG. Append SUPLEMENT if non-empty. +function print_bug(file,line,prefix,category,bug,doc,supplement, suffix,idx) { + if (print_idx) { + idx = bug ": " + } else { + idx = "" + } + if (supplement) { + suffix = " (" supplement ")" + } else { + suffix = "" + } + # ari.*.bug: :: : : + print file ":" line ": " prefix category ": " idx doc suffix + if (srclines != "") { + print file ":" line ":" $0 >> srclines + } +} + +function fix(bug,file,count) { + skip[bug, file] = count + skipped[bug, file] = 0 +} + +function fail(bug,supplement) { + if (doc[bug] == "") { + print_bug("", 0, "internal: ", "internal", "internal", "Missing doc for bug " bug) + exit + } + if (category[bug] == "") { + print_bug("", 0, "internal: ", "internal", "internal", "Missing category for bug " bug) + exit + } + + if (ARI_OK == bug) { + return + } + # Trim the filename down to just DIRECTORY/FILE so that it can be + # robustly used by the FIX code. + + if (FILENAME ~ /^\//) { + canonicalname = FILENAME + } else { + canonicalname = PWD "/" FILENAME + } + shortname = gensub (/^.*\/([^\\]*\/[^\\]*)$/, "\\1", 1, canonicalname) + + skipped[bug, shortname]++ + if (skip[bug, shortname] >= skipped[bug, shortname]) { + # print FILENAME, FNR, skip[bug, FILENAME], skipped[bug, FILENAME], bug + # Do nothing + } else if (error[category[bug]]) { + # ari.*.bug: :: : : + print_bug(FILENAME, FNR, "", category[bug], bug, doc[bug], supplement) + } else if (warning[category[bug]]) { + # ari.*.bug: :: : : + print_bug(FILENAME, FNR, "warning: ", category[bug], bug, doc[bug], supplement) + } +} + +FNR == 1 { + seen[FILENAME] = 1 + if (match(FILENAME, "\\.[ly]$")) { + # FILENAME is a lex or yacc source + is_yacc_or_lex = 1 + } + else { + is_yacc_or_lex = 0 + } +} +END { + if (print_idx) { + idx = bug ": " + } else { + idx = "" + } + # Did we do only a partial skip? + for (bug_n_file in skip) { + split (bug_n_file, a, SUBSEP) + bug = a[1] + file = a[2] + if (seen[file] && (skipped[bug_n_file] < skip[bug_n_file])) { + # ari.*.bug: :: : : + b = file " missing " bug + print_bug(file, 0, "", "internal", file " missing " bug, "Expecting " skip[bug_n_file] " occurances of bug " bug " in file " file ", only found " skipped[bug_n_file]) + } + } +} + + +# Skip OBSOLETE lines +/(^|[^_[:alnum:]])OBSOLETE([^_[:alnum:]]|$)/ { next; } + +# Skip ARI lines + +BEGIN { + ARI_OK = "" +} + +/\/\* ARI:[[:space:]]*(.*)[[:space:]]*\*\// { + ARI_OK = gensub(/^.*\/\* ARI:[[:space:]]*(.*[^[:space:]])[[:space:]]*\*\/.*$/, "\\1", 1, $0) + # print "ARI line found \"" $0 "\"" + # print "ARI_OK \"" ARI_OK "\"" +} +! /\/\* ARI:[[:space:]]*(.*)[[:space:]]*\*\// { + ARI_OK = "" +} + + +# Things in comments + +BEGIN { doc["GNU/Linux"] = "\ +Do not use `Linux'\'', instead use `Linux kernel'\'' or `GNU/Linux system'\'';\ + comments should clearly differentiate between the two (this test assumes that\ + word `Linux'\'' appears on the same line as the word `GNU'\'' or `kernel'\''\ + or a kernel version" + category["GNU/Linux"] = ari_comment +} +/(^|[^_[:alnum:]])Linux([^_[:alnum:]]|$)/ \ +&& !/(^|[^_[:alnum:]])Linux\[sic\]([^_[:alnum:]]|$)/ \ +&& !/(^|[^_[:alnum:]])GNU\/Linux([^_[:alnum:]]|$)/ \ +&& !/(^|[^_[:alnum:]])Linux kernel([^_[:alnum:]]|$)/ \ +&& !/(^|[^_[:alnum:]])Linux [:digit:]\.[:digit:]+)/ { + fail("GNU/Linux") +} + +BEGIN { doc["ARGSUSED"] = "\ +Do not use ARGSUSED, unnecessary" + category["ARGSUSED"] = ari_regression +} +/(^|[^_[:alnum:]])ARGSUSED([^_[:alnum:]]|$)/ { + fail("ARGSUSED") +} + + +# SNIP - Strip out comments - SNIP + +FNR == 1 { + comment_p = 0 +} +comment_p && /\*\// { gsub (/^([^\*]|\*+[^\/\*])*\*+\//, " "); comment_p = 0; } +comment_p { next; } +!comment_p { gsub (/\/\*([^\*]|\*+[^\/\*])*\*+\//, " "); } +!comment_p && /(^|[^"])\/\*/ { gsub (/\/\*.*$/, " "); comment_p = 1; } + + +BEGIN { doc["_ markup"] = "\ +All messages should be marked up with _." + category["_ markup"] = ari_gettext +} +/^[^"]*[[:space:]](warning|error|error_no_arg|query|perror_with_name)[[:space:]]*\([^_\(a-z]/ { + if (! /\("%s"/) { + fail("_ markup") + } +} + +BEGIN { doc["trailing new line"] = "\ +A message should not have a trailing new line" + category["trailing new line"] = ari_gettext +} +/(^|[^_[:alnum:]])(warning|error)[[:space:]]*\(_\(".*\\n"\)[\),]/ { + fail("trailing new line") +} + +# Include files for which GDB has a custom version. + +BEGIN { doc["assert.h"] = "\ +Do not include assert.h, instead include \"gdb_assert.h\""; + category["assert.h"] = ari_regression + fix("assert.h", "gdb/gdb_assert.h", 0) # it does not use it +} +/^#[[:space:]]*include[[:space:]]+.assert\.h./ { + fail("assert.h") +} + +BEGIN { doc["dirent.h"] = "\ +Do not include dirent.h, instead include gdb_dirent.h" + category["dirent.h"] = ari_regression +} +/^#[[:space:]]*include[[:space:]]*.dirent\.h./ { + fail("dirent.h") +} + +BEGIN { doc["regex.h"] = "\ +Do not include regex.h, instead include gdb_regex.h" + category["regex.h"] = ari_regression + fix("regex.h", "gdb/gdb_regex.h", 1) +} +/^#[[:space:]]*include[[:space:]]*.regex\.h./ { + fail("regex.h") +} + +BEGIN { doc["xregex.h"] = "\ +Do not include xregex.h, instead include gdb_regex.h" + category["xregex.h"] = ari_regression + fix("xregex.h", "gdb/gdb_regex.h", 1) +} +/^#[[:space:]]*include[[:space:]]*.xregex\.h./ { + fail("xregex.h") +} + +BEGIN { doc["gnu-regex.h"] = "\ +Do not include gnu-regex.h, instead include gdb_regex.h" + category["gnu-regex.h"] = ari_regression +} +/^#[[:space:]]*include[[:space:]]*.gnu-regex\.h./ { + fail("gnu regex.h") +} + +BEGIN { doc["stat.h"] = "\ +Do not include stat.h or sys/stat.h, instead include gdb_stat.h" + category["stat.h"] = ari_regression + fix("stat.h", "gdb/gdb_stat.h", 1) +} +/^#[[:space:]]*include[[:space:]]*.stat\.h./ \ +|| /^#[[:space:]]*include[[:space:]]*.sys\/stat\.h./ { + fail("stat.h") +} + +BEGIN { doc["wait.h"] = "\ +Do not include wait.h or sys/wait.h, instead include gdb_wait.h" + fix("wait.h", "gdb/gdb_wait.h", 2); + category["wait.h"] = ari_regression +} +/^#[[:space:]]*include[[:space:]]*.wait\.h./ \ +|| /^#[[:space:]]*include[[:space:]]*.sys\/wait\.h./ { + fail("wait.h") +} + +BEGIN { doc["vfork.h"] = "\ +Do not include vfork.h, instead include gdb_vfork.h" + fix("vfork.h", "gdb/gdb_vfork.h", 1); + category["vfork.h"] = ari_regression +} +/^#[[:space:]]*include[[:space:]]*.vfork\.h./ { + fail("vfork.h") +} + +BEGIN { doc["error not internal-warning"] = "\ +Do not use error(\"internal-warning\"), instead use internal_warning" + category["error not internal-warning"] = ari_regression +} +/error.*\"[Ii]nternal.warning/ { + fail("error not internal-warning") +} + +BEGIN { doc["%p"] = "\ +Do not use printf(\"%p\"), instead use printf(\"%s\",paddr()) to dump a \ +target address, or host_address_to_string() for a host address" + category["%p"] = ari_code +} +/%p/ && !/%prec/ { + fail("%p") +} + +BEGIN { doc["%ll"] = "\ +Do not use printf(\"%ll\"), instead use printf(\"%s\",phex()) to dump a \ +`long long'\'' value" + category["%ll"] = ari_code +} +# Allow %ll in scanf +/%[0-9]*ll/ && !/scanf \(.*%[0-9]*ll/ { + fail("%ll") +} + + +# SNIP - Strip out strings - SNIP + +# Test on top.c, scm-valprint.c, remote-rdi.c, ada-lang.c +FNR == 1 { + string_p = 0 + trace_string = 0 +} +# Strip escaped characters. +{ gsub(/\\./, "."); } +# Strip quoted quotes. +{ gsub(/'\''.'\''/, "'\''.'\''"); } +# End of multi-line string +string_p && /\"/ { + if (trace_string) print "EOS:" FNR, $0; + gsub (/^[^\"]*\"/, "'\''"); + string_p = 0; +} +# Middle of multi-line string, discard line. +string_p { + if (trace_string) print "MOS:" FNR, $0; + $0 = "" +} +# Strip complete strings from the middle of the line +!string_p && /\"[^\"]*\"/ { + if (trace_string) print "COS:" FNR, $0; + gsub (/\"[^\"]*\"/, "'\''"); +} +# Start of multi-line string +BEGIN { doc["multi-line string"] = "\ +Multi-line string must have the newline escaped" + category["multi-line string"] = ari_regression +} +!string_p && /\"/ { + if (trace_string) print "SOS:" FNR, $0; + if (/[^\\]$/) { + fail("multi-line string") + } + gsub (/\"[^\"]*$/, "'\''"); + string_p = 1; +} +# { print } + +# Multi-line string +string_p && + +# Accumulate continuation lines +FNR == 1 { + cont_p = 0 +} +!cont_p { full_line = ""; } +/[^\\]\\$/ { gsub (/\\$/, ""); full_line = full_line $0; cont_p = 1; next; } +cont_p { $0 = full_line $0; cont_p = 0; full_line = ""; } + + +# GDB uses ISO C 90. Check for any non pure ISO C 90 code + +BEGIN { doc["PARAMS"] = "\ +Do not use PARAMS(), ISO C 90 implies prototypes" + category["PARAMS"] = ari_regression +} +/(^|[^_[:alnum:]])PARAMS([^_[:alnum:]]|$)/ { + fail("PARAMS") +} + +BEGIN { doc["__func__"] = "\ +Do not use __func__, ISO C 90 does not support this macro" + category["__func__"] = ari_regression + fix("__func__", "gdb/gdb_assert.h", 1) +} +/(^|[^_[:alnum:]])__func__([^_[:alnum:]]|$)/ { + fail("__func__") +} + +BEGIN { doc["__FUNCTION__"] = "\ +Do not use __FUNCTION__, ISO C 90 does not support this macro" + category["__FUNCTION__"] = ari_regression +} +/(^|[^_[:alnum:]])__FUNCTION__([^_[:alnum:]]|$)/ { + fail("__FUNCTION__") +} + +BEGIN { doc["__CYGWIN32__"] = "\ +Do not use __CYGWIN32__, instead use __CYGWIN__ or, better, an explicit \ +autoconf tests" + category["__CYGWIN32__"] = ari_regression +} +/(^|[^_[:alnum:]])__CYGWIN32__([^_[:alnum:]]|$)/ { + fail("__CYGWIN32__") +} + +BEGIN { doc["PTR"] = "\ +Do not use PTR, ISO C 90 implies `void *'\''" + category["PTR"] = ari_regression + #fix("PTR", "gdb/utils.c", 6) +} +/(^|[^_[:alnum:]])PTR([^_[:alnum:]]|$)/ { + fail("PTR") +} + +BEGIN { doc["UCASE function"] = "\ +Function name is uppercase." + category["UCASE function"] = ari_code + possible_UCASE = 0 + UCASE_full_line = "" +} +(possible_UCASE) { + if (ARI_OK == "UCASE function") { + possible_UCASE = 0 + } + # Closing brace found? + else if (UCASE_full_line ~ \ + /^[A-Z][[:alnum:]_]*[[:space:]]*\([^()]*\).*$/) { + if ((UCASE_full_line ~ \ + /^[A-Z][[:alnum:]_]*[[:space:]]*\([^()]*\)[[:space:]]*$/) \ + && ($0 ~ /^\{/) && (is_yacc_or_lex == 0)) { + store_FNR = FNR + FNR = possible_FNR + store_0 = $0; + $0 = UCASE_full_line; + fail("UCASE function") + FNR = store_FNR + $0 = store_0; + } + possible_UCASE = 0 + UCASE_full_line = "" + } else { + UCASE_full_line = UCASE_full_line $0; + } +} +/^[A-Z][[:alnum:]_]*[[:space:]]*\([^()]*(|\))[[:space:]]*$/ { + possible_UCASE = 1 + if (ARI_OK == "UCASE function") { + possible_UCASE = 0 + } + possible_FNR = FNR + UCASE_full_line = $0 +} + + +BEGIN { doc["editCase function"] = "\ +Function name starts lower case but has uppercased letters." + category["editCase function"] = ari_code + possible_editCase = 0 + editCase_full_line = "" +} +(possible_editCase) { + if (ARI_OK == "ediCase function") { + possible_editCase = 0 + } + # Closing brace found? + else if (editCase_full_line ~ \ +/^[a-z][a-z0-9_]*[A-Z][a-z0-9A-Z_]*[[:space:]]*\([^()]*\).*$/) { + if ((editCase_full_line ~ \ +/^[a-z][a-z0-9_]*[A-Z][a-z0-9A-Z_]*[[:space:]]*\([^()]*\)[[:space:]]*$/) \ + && ($0 ~ /^\{/) && (is_yacc_or_lex == 0)) { + store_FNR = FNR + FNR = possible_FNR + store_0 = $0; + $0 = editCase_full_line; + fail("editCase function") + FNR = store_FNR + $0 = store_0; + } + possible_editCase = 0 + editCase_full_line = "" + } else { + editCase_full_line = editCase_full_line $0; + } +} +/^[a-z][a-z0-9_]*[A-Z][a-z0-9A-Z_]*[[:space:]]*\([^()]*(|\))[[:space:]]*$/ { + possible_editCase = 1 + if (ARI_OK == "editCase function") { + possible_editCase = 0 + } + possible_FNR = FNR + editCase_full_line = $0 +} + +# Only function implementation should be on first column +BEGIN { doc["function call in first column"] = "\ +Function name in first column should be restricted to function implementation" + category["function call in first column"] = ari_code +} +/^[a-z][a-z0-9_]*[[:space:]]*\((|[^*][^()]*)\)[[:space:]]*[^ \t]+/ { + fail("function call in first column") +} + + +# Functions without any parameter should have (void) +# after their name not simply (). +BEGIN { doc["no parameter function"] = "\ +Function having no parameter should be declared with funcname (void)." + category["no parameter function"] = ari_code +} +/^[a-zA-Z][a-z0-9A-Z_]*[[:space:]]*\(\)/ { + fail("no parameter function") +} + +BEGIN { doc["hash"] = "\ +Do not use ` #...'\'', instead use `#...'\''(some compilers only correctly \ +parse a C preprocessor directive when `#'\'' is the first character on \ +the line)" + category["hash"] = ari_regression +} +/^[[:space:]]+#/ { + fail("hash") +} + +BEGIN { doc["OP eol"] = "\ +Do not use &&, or || at the end of a line" + category["OP eol"] = ari_code +} +/(\|\||\&\&|==|!=)[[:space:]]*$/ { + fail("OP eol") +} + +BEGIN { doc["strerror"] = "\ +Do not use strerror(), instead use safe_strerror()" + category["strerror"] = ari_regression + fix("strerror", "gdb/gdb_string.h", 1) + fix("strerror", "gdb/mingw-hdep.c", 1) + fix("strerror", "gdb/posix-hdep.c", 1) +} +/(^|[^_[:alnum:]])strerror[[:space:]]*\(/ { + fail("strerror") +} + +BEGIN { doc["long long"] = "\ +Do not use `long long'\'', instead use LONGEST" + category["long long"] = ari_code + # defs.h needs two such patterns for LONGEST and ULONGEST definitions + fix("long long", "gdb/defs.h", 2) +} +/(^|[^_[:alnum:]])long[[:space:]]+long([^_[:alnum:]]|$)/ { + fail("long long") +} + +BEGIN { doc["ATTRIBUTE_UNUSED"] = "\ +Do not use ATTRIBUTE_UNUSED, do not bother (GDB is compiled with -Werror and, \ +consequently, is not able to tolerate false warnings. Since -Wunused-param \ +produces such warnings, neither that warning flag nor ATTRIBUTE_UNUSED \ +are used by GDB" + category["ATTRIBUTE_UNUSED"] = ari_regression +} +/(^|[^_[:alnum:]])ATTRIBUTE_UNUSED([^_[:alnum:]]|$)/ { + fail("ATTRIBUTE_UNUSED") +} + +BEGIN { doc["ATTR_FORMAT"] = "\ +Do not use ATTR_FORMAT, use ATTRIBUTE_PRINTF instead" + category["ATTR_FORMAT"] = ari_regression +} +/(^|[^_[:alnum:]])ATTR_FORMAT([^_[:alnum:]]|$)/ { + fail("ATTR_FORMAT") +} + +BEGIN { doc["ATTR_NORETURN"] = "\ +Do not use ATTR_NORETURN, use ATTRIBUTE_NORETURN instead" + category["ATTR_NORETURN"] = ari_regression +} +/(^|[^_[:alnum:]])ATTR_NORETURN([^_[:alnum:]]|$)/ { + fail("ATTR_NORETURN") +} + +BEGIN { doc["NORETURN"] = "\ +Do not use NORETURN, use ATTRIBUTE_NORETURN instead" + category["NORETURN"] = ari_regression +} +/(^|[^_[:alnum:]])NORETURN([^_[:alnum:]]|$)/ { + fail("NORETURN") +} + + +# General problems + +BEGIN { doc["multiple messages"] = "\ +Do not use multiple calls to warning or error, instead use a single call" + category["multiple messages"] = ari_gettext +} +FNR == 1 { + warning_fnr = -1 +} +/(^|[^_[:alnum:]])(warning|error)[[:space:]]*\(/ { + if (FNR == warning_fnr + 1) { + fail("multiple messages") + } else { + warning_fnr = FNR + } +} + +# Commented out, but left inside sources, just in case. +# BEGIN { doc["inline"] = "\ +# Do not use the inline attribute; \ +# since the compiler generally ignores this, better algorithm selection \ +# is needed to improved performance" +# category["inline"] = ari_code +# } +# /(^|[^_[:alnum:]])inline([^_[:alnum:]]|$)/ { +# fail("inline") +# } + +# This test is obsolete as this type +# has been deprecated and finally suppressed from GDB sources +#BEGIN { doc["obj_private"] = "\ +#Replace obj_private with objfile_data" +# category["obj_private"] = ari_obsolete +#} +#/(^|[^_[:alnum:]])obj_private([^_[:alnum:]]|$)/ { +# fail("obj_private") +#} + +BEGIN { doc["abort"] = "\ +Do not use abort, instead use internal_error; GDB should never abort" + category["abort"] = ari_regression + fix("abort", "gdb/utils.c", 3) +} +/(^|[^_[:alnum:]])abort[[:space:]]*\(/ { + fail("abort") +} + +BEGIN { doc["basename"] = "\ +Do not use basename, instead use lbasename" + category["basename"] = ari_regression +} +/(^|[^_[:alnum:]])basename[[:space:]]*\(/ { + fail("basename") +} + +BEGIN { doc["assert"] = "\ +Do not use assert, instead use gdb_assert or internal_error; assert \ +calls abort and GDB should never call abort" + category["assert"] = ari_regression +} +/(^|[^_[:alnum:]])assert[[:space:]]*\(/ { + fail("assert") +} + +BEGIN { doc["TARGET_HAS_HARDWARE_WATCHPOINTS"] = "\ +Replace TARGET_HAS_HARDWARE_WATCHPOINTS with nothing, not needed" + category["TARGET_HAS_HARDWARE_WATCHPOINTS"] = ari_regression +} +/(^|[^_[:alnum:]])TARGET_HAS_HARDWARE_WATCHPOINTS([^_[:alnum:]]|$)/ { + fail("TARGET_HAS_HARDWARE_WATCHPOINTS") +} + +BEGIN { doc["ADD_SHARED_SYMBOL_FILES"] = "\ +Replace ADD_SHARED_SYMBOL_FILES with nothing, not needed?" + category["ADD_SHARED_SYMBOL_FILES"] = ari_regression +} +/(^|[^_[:alnum:]])ADD_SHARED_SYMBOL_FILES([^_[:alnum:]]|$)/ { + fail("ADD_SHARED_SYMBOL_FILES") +} + +BEGIN { doc["SOLIB_ADD"] = "\ +Replace SOLIB_ADD with nothing, not needed?" + category["SOLIB_ADD"] = ari_deprecate +} +/(^|[^_[:alnum:]])SOLIB_ADD([^_[:alnum:]]|$)/ { + fail("SOLIB_ADD") +} + +BEGIN { doc["SOLIB_CREATE_INFERIOR_HOOK"] = "\ +Replace SOLIB_CREATE_INFERIOR_HOOK with nothing, not needed?" + category["SOLIB_CREATE_INFERIOR_HOOK"] = ari_deprecate +} +/(^|[^_[:alnum:]])SOLIB_CREATE_INFERIOR_HOOK([^_[:alnum:]]|$)/ { + fail("SOLIB_CREATE_INFERIOR_HOOK") +} + +BEGIN { doc["SOLIB_LOADED_LIBRARY_PATHNAME"] = "\ +Replace SOLIB_LOADED_LIBRARY_PATHNAME with nothing, not needed?" + category["SOLIB_LOADED_LIBRARY_PATHNAME"] = ari_regression +} +/(^|[^_[:alnum:]])SOLIB_LOADED_LIBRARY_PATHNAME([^_[:alnum:]]|$)/ { + fail("SOLIB_LOADED_LIBRARY_PATHNAME") +} + +BEGIN { doc["REGISTER_U_ADDR"] = "\ +Replace REGISTER_U_ADDR with nothing, not needed?" + category["REGISTER_U_ADDR"] = ari_regression +} +/(^|[^_[:alnum:]])REGISTER_U_ADDR([^_[:alnum:]]|$)/ { + fail("REGISTER_U_ADDR") +} + +BEGIN { doc["PROCESS_LINENUMBER_HOOK"] = "\ +Replace PROCESS_LINENUMBER_HOOK with nothing, not needed?" + category["PROCESS_LINENUMBER_HOOK"] = ari_regression +} +/(^|[^_[:alnum:]])PROCESS_LINENUMBER_HOOK([^_[:alnum:]]|$)/ { + fail("PROCESS_LINENUMBER_HOOK") +} + +BEGIN { doc["PC_SOLIB"] = "\ +Replace PC_SOLIB with nothing, not needed?" + category["PC_SOLIB"] = ari_deprecate +} +/(^|[^_[:alnum:]])PC_SOLIB([^_[:alnum:]]|$)/ { + fail("PC_SOLIB") +} + +BEGIN { doc["IN_SOLIB_DYNSYM_RESOLVE_CODE"] = "\ +Replace IN_SOLIB_DYNSYM_RESOLVE_CODE with nothing, not needed?" + category["IN_SOLIB_DYNSYM_RESOLVE_CODE"] = ari_regression +} +/(^|[^_[:alnum:]])IN_SOLIB_DYNSYM_RESOLVE_CODE([^_[:alnum:]]|$)/ { + fail("IN_SOLIB_DYNSYM_RESOLVE_CODE") +} + +BEGIN { doc["GCC_COMPILED_FLAG_SYMBOL"] = "\ +Replace GCC_COMPILED_FLAG_SYMBOL with nothing, not needed?" + category["GCC_COMPILED_FLAG_SYMBOL"] = ari_deprecate +} +/(^|[^_[:alnum:]])GCC_COMPILED_FLAG_SYMBOL([^_[:alnum:]]|$)/ { + fail("GCC_COMPILED_FLAG_SYMBOL") +} + +BEGIN { doc["GCC2_COMPILED_FLAG_SYMBOL"] = "\ +Replace GCC2_COMPILED_FLAG_SYMBOL with nothing, not needed?" + category["GCC2_COMPILED_FLAG_SYMBOL"] = ari_deprecate +} +/(^|[^_[:alnum:]])GCC2_COMPILED_FLAG_SYMBOL([^_[:alnum:]]|$)/ { + fail("GCC2_COMPILED_FLAG_SYMBOL") +} + +BEGIN { doc["FUNCTION_EPILOGUE_SIZE"] = "\ +Replace FUNCTION_EPILOGUE_SIZE with nothing, not needed?" + category["FUNCTION_EPILOGUE_SIZE"] = ari_regression +} +/(^|[^_[:alnum:]])FUNCTION_EPILOGUE_SIZE([^_[:alnum:]]|$)/ { + fail("FUNCTION_EPILOGUE_SIZE") +} + +BEGIN { doc["HAVE_VFORK"] = "\ +Do not use HAVE_VFORK, instead include \"gdb_vfork.h\" and call vfork() \ +unconditionally" + category["HAVE_VFORK"] = ari_regression +} +/(^|[^_[:alnum:]])HAVE_VFORK([^_[:alnum:]]|$)/ { + fail("HAVE_VFORK") +} + +BEGIN { doc["bcmp"] = "\ +Do not use bcmp(), ISO C 90 implies memcmp()" + category["bcmp"] = ari_regression +} +/(^|[^_[:alnum:]])bcmp[[:space:]]*\(/ { + fail("bcmp") +} + +BEGIN { doc["setlinebuf"] = "\ +Do not use setlinebuf(), ISO C 90 implies setvbuf()" + category["setlinebuf"] = ari_regression +} +/(^|[^_[:alnum:]])setlinebuf[[:space:]]*\(/ { + fail("setlinebuf") +} + +BEGIN { doc["bcopy"] = "\ +Do not use bcopy(), ISO C 90 implies memcpy() and memmove()" + category["bcopy"] = ari_regression +} +/(^|[^_[:alnum:]])bcopy[[:space:]]*\(/ { + fail("bcopy") +} + +BEGIN { doc["get_frame_base"] = "\ +Replace get_frame_base with get_frame_id, get_frame_base_address, \ +get_frame_locals_address, or get_frame_args_address." + category["get_frame_base"] = ari_obsolete +} +/(^|[^_[:alnum:]])get_frame_base([^_[:alnum:]]|$)/ { + fail("get_frame_base") +} + +BEGIN { doc["floatformat_to_double"] = "\ +Do not use floatformat_to_double() from libierty, \ +instead use floatformat_to_doublest()" + fix("floatformat_to_double", "gdb/doublest.c", 1) + category["floatformat_to_double"] = ari_regression +} +/(^|[^_[:alnum:]])floatformat_to_double[[:space:]]*\(/ { + fail("floatformat_to_double") +} + +BEGIN { doc["floatformat_from_double"] = "\ +Do not use floatformat_from_double() from libierty, \ +instead use floatformat_from_doublest()" + category["floatformat_from_double"] = ari_regression +} +/(^|[^_[:alnum:]])floatformat_from_double[[:space:]]*\(/ { + fail("floatformat_from_double") +} + +BEGIN { doc["BIG_ENDIAN"] = "\ +Do not use BIG_ENDIAN, instead use BFD_ENDIAN_BIG" + category["BIG_ENDIAN"] = ari_regression +} +/(^|[^_[:alnum:]])BIG_ENDIAN([^_[:alnum:]]|$)/ { + fail("BIG_ENDIAN") +} + +BEGIN { doc["LITTLE_ENDIAN"] = "\ +Do not use LITTLE_ENDIAN, instead use BFD_ENDIAN_LITTLE"; + category["LITTLE_ENDIAN"] = ari_regression +} +/(^|[^_[:alnum:]])LITTLE_ENDIAN([^_[:alnum:]]|$)/ { + fail("LITTLE_ENDIAN") +} + +BEGIN { doc["BIG_ENDIAN"] = "\ +Do not use BIG_ENDIAN, instead use BFD_ENDIAN_BIG" + category["BIG_ENDIAN"] = ari_regression +} +/(^|[^_[:alnum:]])BIG_ENDIAN([^_[:alnum:]]|$)/ { + fail("BIG_ENDIAN") +} + +BEGIN { doc["sec_ptr"] = "\ +Instead of sec_ptr, use struct bfd_section"; + category["sec_ptr"] = ari_regression +} +/(^|[^_[:alnum:]])sec_ptr([^_[:alnum:]]|$)/ { + fail("sec_ptr") +} + +BEGIN { doc["frame_unwind_unsigned_register"] = "\ +Replace frame_unwind_unsigned_register with frame_unwind_register_unsigned" + category["frame_unwind_unsigned_register"] = ari_regression +} +/(^|[^_[:alnum:]])frame_unwind_unsigned_register([^_[:alnum:]]|$)/ { + fail("frame_unwind_unsigned_register") +} + +BEGIN { doc["frame_register_read"] = "\ +Replace frame_register_read() with get_frame_register(), or \ +possibly introduce a new method safe_get_frame_register()" + category["frame_register_read"] = ari_obsolete +} +/(^|[^_[:alnum:]])frame_register_read([^_[:alnum:]]|$)/ { + fail("frame_register_read") +} + +BEGIN { doc["read_register"] = "\ +Replace read_register() with regcache_read() et.al." + category["read_register"] = ari_regression +} +/(^|[^_[:alnum:]])read_register([^_[:alnum:]]|$)/ { + fail("read_register") +} + +BEGIN { doc["write_register"] = "\ +Replace write_register() with regcache_read() et.al." + category["write_register"] = ari_regression +} +/(^|[^_[:alnum:]])write_register([^_[:alnum:]]|$)/ { + fail("write_register") +} + +function report(name) { + # Drop any trailing _P. + name = gensub(/(_P|_p)$/, "", 1, name) + # Convert to lower case + name = tolower(name) + # Split into category and bug + cat = gensub(/^([[:alpha:]]+)_([_[:alnum:]]*)$/, "\\1", 1, name) + bug = gensub(/^([[:alpha:]]+)_([_[:alnum:]]*)$/, "\\2", 1, name) + # Report it + name = cat " " bug + doc[name] = "Do not use " cat " " bug ", see declaration for details" + category[name] = cat + fail(name) +} + +/(^|[^_[:alnum:]])(DEPRECATED|deprecated|set_gdbarch_deprecated|LEGACY|legacy|set_gdbarch_legacy)_/ { + line = $0 + # print "0 =", $0 + while (1) { + name = gensub(/^(|.*[^_[:alnum:]])((DEPRECATED|deprecated|LEGACY|legacy)_[_[:alnum:]]*)(.*)$/, "\\2", 1, line) + line = gensub(/^(|.*[^_[:alnum:]])((DEPRECATED|deprecated|LEGACY|legacy)_[_[:alnum:]]*)(.*)$/, "\\1 \\4", 1, line) + # print "name =", name, "line =", line + if (name == line) break; + report(name) + } +} + +# Count the number of times each architecture method is set +/(^|[^_[:alnum:]])set_gdbarch_[_[:alnum:]]*([^_[:alnum:]]|$)/ { + name = gensub(/^.*set_gdbarch_([_[:alnum:]]*).*$/, "\\1", 1, $0) + doc["set " name] = "\ +Call to set_gdbarch_" name + category["set " name] = ari_gdbarch + fail("set " name) +} + +# Count the number of times each tm/xm/nm macro is defined or undefined +/^#[[:space:]]*(undef|define)[[:space:]]+[[:alnum:]_]+.*$/ \ +&& !/^#[[:space:]]*(undef|define)[[:space:]]+[[:alnum:]_]+_H($|[[:space:]])/ \ +&& FILENAME ~ /(^|\/)config\/(|[^\/]*\/)(tm-|xm-|nm-).*\.h$/ { + basename = gensub(/(^|.*\/)([^\/]*)$/, "\\2", 1, FILENAME) + type = gensub(/^(tm|xm|nm)-.*\.h$/, "\\1", 1, basename) + name = gensub(/^#[[:space:]]*(undef|define)[[:space:]]+([[:alnum:]_]+).*$/, "\\2", 1, $0) + if (type == basename) { + type = "macro" + } + doc[type " " name] = "\ +Do not define macros such as " name " in a tm, nm or xm file, \ +in fact do not provide a tm, nm or xm file" + category[type " " name] = ari_macro + fail(type " " name) +} + +BEGIN { doc["deprecated_registers"] = "\ +Replace deprecated_registers with nothing, they have reached \ +end-of-life" + category["deprecated_registers"] = ari_eol +} +/(^|[^_[:alnum:]])deprecated_registers([^_[:alnum:]]|$)/ { + fail("deprecated_registers") +} + +BEGIN { doc["read_pc"] = "\ +Replace READ_PC() with frame_pc_unwind; \ +at present the inferior function call code still uses this" + category["read_pc"] = ari_deprecate +} +/(^|[^_[:alnum:]])read_pc[[:space:]]*\(/ || \ +/(^|[^_[:alnum:]])set_gdbarch_read_pc[[:space:]]*\(/ || \ +/(^|[^_[:alnum:]])TARGET_READ_PC[[:space:]]*\(/ { + fail("read_pc") +} + +BEGIN { doc["write_pc"] = "\ +Replace write_pc() with get_frame_base_address or get_frame_id; \ +at present the inferior function call code still uses this when doing \ +a DECR_PC_AFTER_BREAK" + category["write_pc"] = ari_deprecate +} +/(^|[^_[:alnum:]])write_pc[[:space:]]*\(/ || \ +/(^|[^_[:alnum:]])set_gdbarch_write_pc[[:space:]]*\(/ || \ +/(^|[^_[:alnum:]])TARGET_WRITE_PC[[:space:]]*\(/ { + fail("write_pc") +} + +BEGIN { doc["generic_target_write_pc"] = "\ +Replace generic_target_write_pc with a per-architecture implementation, \ +this relies on PC_REGNUM which is being eliminated" + category["generic_target_write_pc"] = ari_regression +} +/(^|[^_[:alnum:]])generic_target_write_pc([^_[:alnum:]]|$)/ { + fail("generic_target_write_pc") +} + +BEGIN { doc["read_sp"] = "\ +Replace read_sp() with frame_sp_unwind" + category["read_sp"] = ari_regression +} +/(^|[^_[:alnum:]])read_sp[[:space:]]*\(/ || \ +/(^|[^_[:alnum:]])set_gdbarch_read_sp[[:space:]]*\(/ || \ +/(^|[^_[:alnum:]])TARGET_READ_SP[[:space:]]*\(/ { + fail("read_sp") +} + +BEGIN { doc["register_cached"] = "\ +Replace register_cached() with nothing, does not have a regcache parameter" + category["register_cached"] = ari_regression +} +/(^|[^_[:alnum:]])register_cached[[:space:]]*\(/ { + fail("register_cached") +} + +BEGIN { doc["set_register_cached"] = "\ +Replace set_register_cached() with nothing, does not have a regcache parameter" + category["set_register_cached"] = ari_regression +} +/(^|[^_[:alnum:]])set_register_cached[[:space:]]*\(/ { + fail("set_register_cached") +} + +# Print functions: Use versions that either check for buffer overflow +# or safely allocate a fresh buffer. + +BEGIN { doc["sprintf"] = "\ +Do not use sprintf, instead use xsnprintf or xstrprintf" + category["sprintf"] = ari_code +} +/(^|[^_[:alnum:]])sprintf[[:space:]]*\(/ { + fail("sprintf") +} + +BEGIN { doc["vsprintf"] = "\ +Do not use vsprintf(), instead use xstrvprintf" + category["vsprintf"] = ari_regression +} +/(^|[^_[:alnum:]])vsprintf[[:space:]]*\(/ { + fail("vsprintf") +} + +BEGIN { doc["asprintf"] = "\ +Do not use asprintf(), instead use xstrprintf()" + category["asprintf"] = ari_regression +} +/(^|[^_[:alnum:]])asprintf[[:space:]]*\(/ { + fail("asprintf") +} + +BEGIN { doc["vasprintf"] = "\ +Do not use vasprintf(), instead use xstrvprintf" + fix("vasprintf", "gdb/utils.c", 1) + category["vasprintf"] = ari_regression +} +/(^|[^_[:alnum:]])vasprintf[[:space:]]*\(/ { + fail("vasprintf") +} + +BEGIN { doc["xasprintf"] = "\ +Do not use xasprintf(), instead use xstrprintf" + fix("xasprintf", "gdb/defs.h", 1) + fix("xasprintf", "gdb/utils.c", 1) + category["xasprintf"] = ari_regression +} +/(^|[^_[:alnum:]])xasprintf[[:space:]]*\(/ { + fail("xasprintf") +} + +BEGIN { doc["xvasprintf"] = "\ +Do not use xvasprintf(), instead use xstrvprintf" + fix("xvasprintf", "gdb/defs.h", 1) + fix("xvasprintf", "gdb/utils.c", 1) + category["xvasprintf"] = ari_regression +} +/(^|[^_[:alnum:]])xvasprintf[[:space:]]*\(/ { + fail("xvasprintf") +} + +# More generic memory operations + +BEGIN { doc["bzero"] = "\ +Do not use bzero(), instead use memset()" + category["bzero"] = ari_regression +} +/(^|[^_[:alnum:]])bzero[[:space:]]*\(/ { + fail("bzero") +} + +BEGIN { doc["strdup"] = "\ +Do not use strdup(), instead use xstrdup()"; + category["strdup"] = ari_regression +} +/(^|[^_[:alnum:]])strdup[[:space:]]*\(/ { + fail("strdup") +} + +BEGIN { doc["strsave"] = "\ +Do not use strsave(), instead use xstrdup() et.al." + category["strsave"] = ari_regression +} +/(^|[^_[:alnum:]])strsave[[:space:]]*\(/ { + fail("strsave") +} + +# String compare functions + +BEGIN { doc["strnicmp"] = "\ +Do not use strnicmp(), instead use strncasecmp()" + category["strnicmp"] = ari_regression +} +/(^|[^_[:alnum:]])strnicmp[[:space:]]*\(/ { + fail("strnicmp") +} + +# Boolean expressions and conditionals + +BEGIN { doc["boolean"] = "\ +Do not use `boolean'\'', use `int'\'' instead" + category["boolean"] = ari_regression +} +/(^|[^_[:alnum:]])boolean([^_[:alnum:]]|$)/ { + if (is_yacc_or_lex == 0) { + fail("boolean") + } +} + +BEGIN { doc["false"] = "\ +Definitely do not use `false'\'' in boolean expressions" + category["false"] = ari_regression +} +/(^|[^_[:alnum:]])false([^_[:alnum:]]|$)/ { + if (is_yacc_or_lex == 0) { + fail("false") + } +} + +BEGIN { doc["true"] = "\ +Do not try to use `true'\'' in boolean expressions" + category["true"] = ari_regression +} +/(^|[^_[:alnum:]])true([^_[:alnum:]]|$)/ { + if (is_yacc_or_lex == 0) { + fail("true") + } +} + +# Typedefs that are either redundant or can be reduced to `struct +# type *''. +# Must be placed before if assignment otherwise ARI exceptions +# are not handled correctly. + +BEGIN { doc["d_namelen"] = "\ +Do not use dirent.d_namelen, instead use NAMELEN" + category["d_namelen"] = ari_regression +} +/(^|[^_[:alnum:]])d_namelen([^_[:alnum:]]|$)/ { + fail("d_namelen") +} + +BEGIN { doc["strlen d_name"] = "\ +Do not use strlen dirent.d_name, instead use NAMELEN" + category["strlen d_name"] = ari_regression +} +/(^|[^_[:alnum:]])strlen[[:space:]]*\(.*[^_[:alnum:]]d_name([^_[:alnum:]]|$)/ { + fail("strlen d_name") +} + +BEGIN { doc["var_boolean"] = "\ +Replace var_boolean with add_setshow_boolean_cmd" + category["var_boolean"] = ari_regression + fix("var_boolean", "gdb/command.h", 1) + # fix only uses the last directory level + fix("var_boolean", "cli/cli-decode.c", 2) +} +/(^|[^_[:alnum:]])var_boolean([^_[:alnum:]]|$)/ { + if ($0 !~ /(^|[^_[:alnum:]])case *var_boolean:/) { + fail("var_boolean") + } +} + +BEGIN { doc["generic_use_struct_convention"] = "\ +Replace generic_use_struct_convention with nothing, \ +EXTRACT_STRUCT_VALUE_ADDRESS is a predicate" + category["generic_use_struct_convention"] = ari_regression +} +/(^|[^_[:alnum:]])generic_use_struct_convention([^_[:alnum:]]|$)/ { + fail("generic_use_struct_convention") +} + +BEGIN { doc["if assignment"] = "\ +An IF statement'\''s expression contains an assignment (the GNU coding \ +standard discourages this)" + category["if assignment"] = ari_code +} +BEGIN { doc["if clause more than 50 lines"] = "\ +An IF statement'\''s expression expands over 50 lines" + category["if clause more than 50 lines"] = ari_code +} +# +# Accumulate continuation lines +FNR == 1 { + in_if = 0 +} + +/(^|[^_[:alnum:]])if / { + in_if = 1; + if_brace_level = 0; + if_cont_p = 0; + if_count = 0; + if_brace_end_pos = 0; + if_full_line = ""; +} +(in_if) { + # We want everything up to closing brace of same level + if_count++; + if (if_count > 50) { + print "multiline if: " if_full_line $0 + fail("if clause more than 50 lines") + if_brace_level = 0; + if_full_line = ""; + } else { + if (if_count == 1) { + i = index($0,"if "); + } else { + i = 1; + } + for (i=i; i <= length($0); i++) { + char = substr($0,i,1); + if (char == "(") { if_brace_level++; } + if (char == ")") { + if_brace_level--; + if (!if_brace_level) { + if_brace_end_pos = i; + after_if = substr($0,i+1,length($0)); + # Do not parse what is following + break; + } + } + } + if (if_brace_level == 0) { + $0 = substr($0,1,i); + in_if = 0; + } else { + if_full_line = if_full_line $0; + if_cont_p = 1; + next; + } + } +} +# if we arrive here, we need to concatenate, but we are at brace level 0 + +(if_brace_end_pos) { + $0 = if_full_line substr($0,1,if_brace_end_pos); + if (if_count > 1) { + # print "IF: multi line " if_count " found at " FILENAME ":" FNR " \"" $0 "\"" + } + if_cont_p = 0; + if_full_line = ""; +} +/(^|[^_[:alnum:]])if .* = / { + # print "fail in if " $0 + fail("if assignment") +} +(if_brace_end_pos) { + $0 = $0 after_if; + if_brace_end_pos = 0; + in_if = 0; +} + +# Printout of all found bug + +BEGIN { + if (print_doc) { + for (bug in doc) { + fail(bug) + } + exit + } +}' "$@" + diff --git a/gdb/contrib/ari/gdb_find.sh b/gdb/contrib/ari/gdb_find.sh new file mode 100644 index 00000000000..9e4b67fd40c --- /dev/null +++ b/gdb/contrib/ari/gdb_find.sh @@ -0,0 +1,41 @@ +#!/bin/sh + +# GDB script to create list of files to check using gdb_ari.sh. +# +# Copyright (C) 2003-2012 Free Software Foundation, Inc. +# +# This file is part of GDB. +# +# This program is free software; you can redistribute it and/or modify +# it under the terms of the GNU General Public License as published by +# the Free Software Foundation; either version 3 of the License, or +# (at your option) any later version. +# +# This program is distributed in the hope that it will be useful, +# but WITHOUT ANY WARRANTY; without even the implied warranty of +# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the +# GNU General Public License for more details. +# +# You should have received a copy of the GNU General Public License +# along with this program. If not, see . + +# Make certain that the script is not running in an internationalized +# environment. + +LANG=C ; export LANG +LC_ALL=C ; export LC_ALL + + +# A find that prunes files that GDB users shouldn't be interested in. +# Use sort to order files alphabetically. + +find "$@" \ + -name testsuite -prune -o \ + -name gdbserver -prune -o \ + -name gnulib -prune -o \ + -name osf-share -prune -o \ + -name '*-stub.c' -prune -o \ + -name '*-exp.c' -prune -o \ + -name ada-lex.c -prune -o \ + -name cp-name-parser.c -prune -o \ + -type f -name '*.[lyhc]' -print | sort diff --git a/gdb/contrib/ari/update-web-ari.sh b/gdb/contrib/ari/update-web-ari.sh new file mode 100644 index 00000000000..0650836f211 --- /dev/null +++ b/gdb/contrib/ari/update-web-ari.sh @@ -0,0 +1,921 @@ +#!/bin/sh -x + +# GDB script to create GDB ARI web page. +# +# Copyright (C) 2001-2012 Free Software Foundation, Inc. +# +# This file is part of GDB. +# +# This program is free software; you can redistribute it and/or modify +# it under the terms of the GNU General Public License as published by +# the Free Software Foundation; either version 3 of the License, or +# (at your option) any later version. +# +# This program is distributed in the hope that it will be useful, +# but WITHOUT ANY WARRANTY; without even the implied warranty of +# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the +# GNU General Public License for more details. +# +# You should have received a copy of the GNU General Public License +# along with this program. If not, see . + +# TODO: setjmp.h, setjmp and longjmp. + +# Direct stderr into stdout but still hang onto stderr (/dev/fd/3) +exec 3>&2 2>&1 +ECHO () +{ +# echo "$@" | tee /dev/fd/3 1>&2 + echo "$@" 1>&2 + echo "$@" 1>&3 +} + +# Really mindless usage +if test $# -ne 4 +then + echo "Usage: $0 " 1>&2 + exit 1 +fi +snapshot=$1 ; shift +tmpdir=$1 ; shift +wwwdir=$1 ; shift +project=$1 ; shift + +# Try to create destination directory if it doesn't exist yet +if [ ! -d ${wwwdir} ] +then + mkdir -p ${wwwdir} +fi + +# Fail if destination directory doesn't exist or is not writable +if [ ! -w ${wwwdir} -o ! -d ${wwwdir} ] +then + echo ERROR: Can not write to directory ${wwwdir} >&2 + exit 2 +fi + +if [ ! -r ${snapshot} ] +then + echo ERROR: Can not read snapshot file 1>&2 + exit 1 +fi + +# FILE formats +# ari.*.bug: :: : : +# ari.*.idx: :: +# ari.*.doc: ::: +# Where ``*'' is {source,warning,indent,doschk} + +unpack_source_p=true +delete_source_p=true + +check_warning_p=false # broken +check_indent_p=false # too slow, too many fail +check_source_p=true +check_doschk_p=true +check_werror_p=true + +update_doc_p=true +update_web_p=true + +if awk --version 2>&1 /dev/null +then + AWK=awk +else + AWK=gawk +fi +export AWK + +# Set up a few cleanups +if ${delete_source_p} +then + trap "cd /tmp; rm -rf ${tmpdir}; exit" 0 1 2 15 +fi + + +# If the first parameter is a directory, +#we just use it as the extracted source +if [ -d ${snapshot} ] +then + module=${project} + srcdir=${snapshot} + aridir=${srcdir}/${module}/contrib/ari + unpack_source_p=false + delete_source_p=false + version_in=${srcdir}/${module}/version.in +else + # unpack the tar-ball + if ${unpack_source_p} + then + # Was it previously unpacked? + if ${delete_source_p} || test ! -d ${tmpdir}/${module}* + then + /bin/rm -rf "${tmpdir}" + /bin/mkdir -p ${tmpdir} + if [ ! -d ${tmpdir} ] + then + echo "Problem creating work directory" + exit 1 + fi + cd ${tmpdir} || exit 1 + echo `date`: Unpacking tar-ball ... + case ${snapshot} in + *.tar.bz2 ) bzcat ${snapshot} ;; + *.tar ) cat ${snapshot} ;; + * ) ECHO Bad file ${snapshot} ; exit 1 ;; + esac | tar xf - + fi + fi + + module=`basename ${snapshot}` + module=`basename ${module} .bz2` + module=`basename ${module} .tar` + srcdir=`echo ${tmpdir}/${module}*` + aridir=${HOME}/ss + version_in=${srcdir}/gdb/version.in +fi + +if [ ! -r ${version_in} ] +then + echo ERROR: missing version file 1>&2 + exit 1 +fi +version=`cat ${version_in}` + + +# THIS HAS SUFFERED BIT ROT +if ${check_warning_p} && test -d "${srcdir}" +then + echo `date`: Parsing compiler warnings 1>&2 + cat ${root}/ari.compile | $AWK ' +BEGIN { + FS=":"; +} +/^[^:]*:[0-9]*: warning:/ { + file = $1; + #sub (/^.*\//, "", file); + warning[file] += 1; +} +/^[^:]*:[0-9]*: error:/ { + file = $1; + #sub (/^.*\//, "", file); + error[file] += 1; +} +END { + for (file in warning) { + print file ":warning:" level[file] + } + for (file in error) { + print file ":error:" level[file] + } +} +' > ${root}/ari.warning.bug +fi + +# THIS HAS SUFFERED BIT ROT +if ${check_indent_p} && test -d "${srcdir}" +then + printf "Analizing file indentation:" 1>&2 + ( cd "${srcdir}" && /bin/sh ${aridir}/gdb_find.sh ${project} | while read f + do + if /bin/sh ${aridir}/gdb_indent.sh < ${f} 2>/dev/null | cmp -s - ${f} + then + : + else + # ari.*.bug: :: : : + echo "${f}:0: info: indent: Indentation does not match GNU indent output" + fi + done ) > ${wwwdir}/ari.indent.bug + echo "" +fi + +if ${check_source_p} && test -d "${srcdir}" +then + bugf=${wwwdir}/ari.source.bug + oldf=${wwwdir}/ari.source.old + srcf=${wwwdir}/ari.source.lines + oldsrcf=${wwwdir}/ari.source.lines-old + + diff=${wwwdir}/ari.source.diff + diffin=${diff}-in + newf1=${bugf}1 + oldf1=${oldf}1 + oldpruned=${oldf1}-pruned + newpruned=${newf1}-pruned + + cp -f ${bugf} ${oldf} + cp -f ${srcf} ${oldsrcf} + rm -f ${srcf} + node=`uname -n` + echo "`date`: Using source lines ${srcf}" 1>&2 + echo "`date`: Checking source code" 1>&2 + ( cd "${srcdir}" && /bin/sh ${aridir}/gdb_find.sh "${project}" | \ + xargs /bin/sh ${aridir}/gdb_ari.sh -Werror -Wall --print-idx --src=${srcf} + ) > ${bugf} + # Remove things we are not interested in to signal by email + # gdbarch changes are not important here + # Also convert ` into ' to avoid command substitution in script below + sed -e "/.*: gdbarch:.*/d" -e "s:\`:':g" ${oldf} > ${oldf1} + sed -e "/.*: gdbarch:.*/d" -e "s:\`:':g" ${bugf} > ${newf1} + # Remove line number info so that code inclusion/deletion + # has no impact on the result + sed -e "s/\([^:]*\):\([^:]*\):\(.*\)/\1:0:\3/" ${oldf1} > ${oldpruned} + sed -e "s/\([^:]*\):\([^:]*\):\(.*\)/\1:0:\3/" ${newf1} > ${newpruned} + # Use diff without option to get normal diff output that + # is reparsed after + diff ${oldpruned} ${newpruned} > ${diffin} + # Only keep new warnings + sed -n -e "/^>.*/p" ${diffin} > ${diff} + sedscript=${wwwdir}/sedscript + script=${wwwdir}/script + sed -n -e "s|\(^[0-9,]*\)a\(.*\)|echo \1a\2 \n \ + sed -n \'\2s:\\\\(.*\\\\):> \\\\1:p\' ${newf1}|p" \ + -e "s|\(^[0-9,]*\)d\(.*\)|echo \1d\2\n \ + sed -n \'\1s:\\\\(.*\\\\):< \\\\1:p\' ${oldf1}|p" \ + -e "s|\(^[0-9,]*\)c\(.*\)|echo \1c\2\n \ + sed -n \'\1s:\\\\(.*\\\\):< \\\\1:p\' ${oldf1} \n \ + sed -n \"\2s:\\\\(.*\\\\):> \\\\1:p\" ${newf1}|p" \ + ${diffin} > ${sedscript} + ${SHELL} ${sedscript} > ${wwwdir}/message + sed -n \ + -e "s;\(.*\);echo \\\"\1\\\";p" \ + -e "s;.*< \([^:]*\):\([0-9]*\):.*;grep \"^\1:\2:\" ${oldsrcf};p" \ + -e "s;.*> \([^:]*\):\([0-9]*\):.*;grep \"^\1:\2:\" ${srcf};p" \ + ${wwwdir}/message > ${script} + ${SHELL} ${script} > ${wwwdir}/mail-message + if [ "x${branch}" != "x" ]; then + email_suffix="`date` in ${branch}" + else + email_suffix="`date`" + fi + +fi + + + + +if ${check_doschk_p} && test -d "${srcdir}" +then + echo "`date`: Checking for doschk" 1>&2 + rm -f "${wwwdir}"/ari.doschk.* + fnchange_lst="${srcdir}"/gdb/config/djgpp/fnchange.lst + fnchange_awk="${wwwdir}"/ari.doschk.awk + doschk_in="${wwwdir}"/ari.doschk.in + doschk_out="${wwwdir}"/ari.doschk.out + doschk_bug="${wwwdir}"/ari.doschk.bug + doschk_char="${wwwdir}"/ari.doschk.char + + # Transform fnchange.lst into fnchange.awk. The program DJTAR + # does a textual substitution of each file name using the list. + # Generate an awk script that does the equivalent - matches an + # exact line and then outputs the replacement. + + sed -e 's;@[^@]*@[/]*\([^ ]*\) @[^@]*@[/]*\([^ ]*\);\$0 == "\1" { print "\2"\; next\; };' \ + < "${fnchange_lst}" > "${fnchange_awk}" + echo '{ print }' >> "${fnchange_awk}" + + # Do the raw analysis - transform the list of files into the DJGPP + # equivalents putting it in the .in file + ( cd "${srcdir}" && find * \ + -name '*.info-[0-9]*' -prune \ + -o -name tcl -prune \ + -o -name itcl -prune \ + -o -name tk -prune \ + -o -name libgui -prune \ + -o -name tix -prune \ + -o -name dejagnu -prune \ + -o -name expect -prune \ + -o -type f -print ) \ + | $AWK -f ${fnchange_awk} > ${doschk_in} + + # Start with a clean slate + rm -f ${doschk_bug} + + # Check for any invalid characters. + grep '[\+\,\;\=\[\]\|\<\>\\\"\:\?\*]' < ${doschk_in} > ${doschk_char} + # ari.*.bug: :: : : + sed < ${doschk_char} >> ${doschk_bug} \ + -e s'/$/:0: dos: DOSCHK: Invalid DOS character/' + + # Magic to map ari.doschk.out to ari.doschk.bug goes here + doschk < ${doschk_in} > ${doschk_out} + cat ${doschk_out} | $AWK >> ${doschk_bug} ' +BEGIN { + state = 1; + invalid_dos = state++; bug[invalid_dos] = "invalid DOS file name"; category[invalid_dos] = "dos"; + same_dos = state++; bug[same_dos] = "DOS 8.3"; category[same_dos] = "dos"; + same_sysv = state++; bug[same_sysv] = "SysV"; + long_sysv = state++; bug[long_sysv] = "long SysV"; + internal = state++; bug[internal] = "internal doschk"; category[internal] = "internal"; + state = 0; +} +/^$/ { state = 0; next; } +/^The .* not valid DOS/ { state = invalid_dos; next; } +/^The .* same DOS/ { state = same_dos; next; } +/^The .* same SysV/ { state = same_sysv; next; } +/^The .* too long for SysV/ { state = long_sysv; next; } +/^The .* / { state = internal; next; } + +NF == 0 { next } + +NF == 3 { name = $1 ; file = $3 } +NF == 1 { file = $1 } +NF > 3 && $2 == "-" { file = $1 ; name = gensub(/^.* - /, "", 1) } + +state == same_dos { + # ari.*.bug: :: : : + print file ":0: " category[state] ": " \ + name " " bug[state] " " " dup: " \ + " DOSCHK - the names " name " and " file " resolve to the same" \ + " file on a " bug[state] \ + " system.
For DOS, this can be fixed by modifying the file" \ + " fnchange.lst." + next +} +state == invalid_dos { + # ari.*.bug: :: : : + print file ":0: " category[state] ": " name ": DOSCHK - " name + next +} +state == internal { + # ari.*.bug: :: : : + print file ":0: " category[state] ": " bug[state] ": DOSCHK - a " \ + bug[state] " problem" +} +' +fi + + + +if ${check_werror_p} && test -d "${srcdir}" +then + echo "`date`: Checking Makefile.in for non- -Werror rules" + rm -f ${wwwdir}/ari.werror.* + cat "${srcdir}/${project}/Makefile.in" | $AWK > ${wwwdir}/ari.werror.bug ' +BEGIN { + count = 0 + cont_p = 0 + full_line = "" +} +/^[-_[:alnum:]]+\.o:/ { + file = gensub(/.o:.*/, "", 1) ".c" +} + +/[^\\]\\$/ { gsub (/\\$/, ""); full_line = full_line $0; cont_p = 1; next; } +cont_p { $0 = full_line $0; cont_p = 0; full_line = ""; } + +/\$\(COMPILE\.pre\)/ { + print file " has line " $0 + if (($0 !~ /\$\(.*ERROR_CFLAGS\)/) && ($0 !~ /\$\(INTERNAL_CFLAGS\)/)) { + # ari.*.bug: :: : : + print "'"${project}"'/" file ":0: info: Werror: The file is not being compiled with -Werror" + } +} +' +fi + + +# From the warnings, generate the doc and indexed bug files +if ${update_doc_p} +then + cd ${wwwdir} + rm -f ari.doc ari.idx ari.doc.bug + # Generate an extra file containing all the bugs that the ARI can detect. + /bin/sh ${aridir}/gdb_ari.sh -Werror -Wall --print-idx --print-doc >> ari.doc.bug + cat ari.*.bug | $AWK > ari.idx ' +BEGIN { + FS=": *" +} +{ + # ari.*.bug: :: : : + file = $1 + line = $2 + category = $3 + bug = $4 + if (! (bug in cat)) { + cat[bug] = category + # strip any trailing .... (supplement) + doc[bug] = gensub(/ \([^\)]*\)$/, "", 1, $5) + count[bug] = 0 + } + if (file != "") { + count[bug] += 1 + # ari.*.idx: :: + print bug ":" file ":" category + } + # Also accumulate some categories as obsolete + if (category == "deprecated") { + # ari.*.idx: :: + if (file != "") { + print category ":" file ":" "obsolete" + } + #count[category]++ + #doc[category] = "Contains " category " code" + } +} +END { + i = 0; + for (bug in count) { + # ari.*.doc: ::: + print bug ":" count[bug] ":" cat[bug] ":" doc[bug] >> "ari.doc" + } +} +' +fi + + +# print_toc BIAS MIN_COUNT CATEGORIES TITLE + +# Print a table of contents containing the bugs CATEGORIES. If the +# BUG count >= MIN_COUNT print it in the table-of-contents. If +# MIN_COUNT is non -ve, also include a link to the table.Adjust the +# printed BUG count by BIAS. + +all= + +print_toc () +{ + bias="$1" ; shift + min_count="$1" ; shift + + all=" $all $1 " + categories="" + for c in $1; do + categories="${categories} categories[\"${c}\"] = 1 ;" + done + shift + + title="$@" ; shift + + echo "

" >> ${newari} + echo "" | tr '[A-Z]' '[a-z]' >> ${newari} + echo "

${title}

" >> ${newari} + cat >> ${newari} # description + + cat >> ${newari} < + + +EOF + # ari.*.doc: ::: + cat ${wwwdir}/ari.doc \ + | sort -t: +1rn -2 +0d \ + | $AWK >> ${newari} ' +BEGIN { + FS=":" + '"$categories"' + MIN_COUNT = '${min_count}' + BIAS = '${bias}' + total = 0 + nr = 0 +} +{ + # ari.*.doc: ::: + bug = $1 + count = $2 + category = $3 + doc = $4 + if (count < MIN_COUNT) next + if (!(category in categories)) next + nr += 1 + total += count + printf "" + printf "" + printf "", bug, count + BIAS + } else { + printf "%d", count + BIAS + } + printf "" + printf "", doc + printf "" + print "" +} +END { + print "" +} +' +cat >> ${newari} < +

+EOF +} + + +print_table () +{ + categories="" + for c in $1; do + categories="${categories} categories[\"${c}\"] = 1 ;" + done + # Remember to prune the dir prefix from projects files + # ari.*.idx: :: + cat ${wwwdir}/ari.idx | $AWK >> ${newari} ' +function qsort (table, + middle, tmp, left, nr_left, right, nr_right, result) { + middle = "" + for (middle in table) { break; } + nr_left = 0; + nr_right = 0; + for (tmp in table) { + if (tolower(tmp) < tolower(middle)) { + nr_left++ + left[tmp] = tmp + } else if (tolower(tmp) > tolower(middle)) { + nr_right++ + right[tmp] = tmp + } + } + #print "qsort " nr_left " " middle " " nr_right > "/dev/stderr" + result = "" + if (nr_left > 0) { + result = qsort(left) SUBSEP + } + result = result middle + if (nr_right > 0) { + result = result SUBSEP qsort(right) + } + return result +} +function print_heading (where, bug_i) { + print "" + print "

" + print "" + print "" + print "" + for (bug_i = 1; bug_i <= nr_bug; bug_i++) { + bug = i2bug[bug_i]; + printf "\n" + } + #print "" + printf "\n", i2bug[bug_i-1] + print "" + print "" + print "" +} +function print_totals (where, bug_i) { + print "" + printf "\n" + print ""; + for (bug_i = 1; bug_i <= nr_bug; bug_i++) { + bug = i2bug[bug_i]; + printf ""; + print "" + } + print "" + printf "\n" + print "" + print "" +} +BEGIN { + FS = ":" + '"${categories}"' + nr_file = 0; + nr_bug = 0; +} +{ + # ari.*.idx: :: + bug = $1 + file = $2 + category = $3 + # Interested in this + if (!(category in categories)) next + # Totals + db[bug, file] += 1 + bug_total[bug] += 1 + file_total[file] += 1 + total += 1 +} +END { + + # Sort the files and bugs creating indexed lists. + nr_bug = split(qsort(bug_total), i2bug, SUBSEP); + nr_file = split(qsort(file_total), i2file, SUBSEP); + + # Dummy entries for first/last + i2file[0] = 0 + i2file[-1] = -1 + i2bug[0] = 0 + i2bug[-1] = -1 + + # Construct a cycle of next/prev links. The file/bug "0" and "-1" + # are used to identify the start/end of the cycle. Consequently, + # prev(0) = -1 (prev of start is the end) and next(-1) = 0 (next + # of end is the start). + + # For all the bugs, create a cycle that goes to the prev / next file. + for (bug_i = 1; bug_i <= nr_bug; bug_i++) { + bug = i2bug[bug_i] + prev = 0 + prev_file[bug, 0] = -1 + next_file[bug, -1] = 0 + for (file_i = 1; file_i <= nr_file; file_i++) { + file = i2file[file_i] + if ((bug, file) in db) { + prev_file[bug, file] = prev + next_file[bug, prev] = file + prev = file + } + } + prev_file[bug, -1] = prev + next_file[bug, prev] = -1 + } + + # For all the files, create a cycle that goes to the prev / next bug. + for (file_i = 1; file_i <= nr_file; file_i++) { + file = i2file[file_i] + prev = 0 + prev_bug[file, 0] = -1 + next_bug[file, -1] = 0 + for (bug_i = 1; bug_i <= nr_bug; bug_i++) { + bug = i2bug[bug_i] + if ((bug, file) in db) { + prev_bug[file, bug] = prev + next_bug[file, prev] = bug + prev = bug + } + } + prev_bug[file, -1] = prev + next_bug[file, prev] = -1 + } + + print "
BUGTotalDescription
", bug + printf "%s", gensub(/_/, " ", "g", bug) + printf "" + if (count > 0 && MIN_COUNT >= 0) { + printf "%d%s
" nr "" total "
FileTotal" + # The title names are offset by one. Otherwize, when the browser + # jumps to the name it leaves out half the relevant column. + #printf " ", bug + printf " ", i2bug[bug_i-1] + printf "", bug + printf "%s", gensub(/_/, " ", "g", bug) + printf "\n" + printf " TotalFile
Totals" + printf "%s", total + printf ">" + printf "" + printf "" + printf "%d", bug, bug_total[bug] + printf ""; + printf "^", prev_file[bug, where], bug + printf "v", next_file[bug, where], bug + printf " ", where, bug + printf "" + printf "%s", total + printf "<" + printf "Totals
" + print "" + print_heading(0); + print "" + print_totals(0); + print "" + + for (file_i = 1; file_i <= nr_file; file_i++) { + file = i2file[file_i]; + pfile = gensub(/^'${project}'\//, "", 1, file) + print "" + print "" + print "" + printf "\n" + print "" + for (bug_i = 1; bug_i <= nr_bug; bug_i++) { + bug = i2bug[bug_i]; + if ((bug, file) in db) { + printf "" + print "" + } else { + print "" + #print "" + } + } + print "" + printf "\n" + print "" + print "" + } + + print "" + print_totals(-1) + print "" + print_heading(-1); + print "" + print "" + print "
" pfile "" + printf "%s", file_total[file] + printf ">", file, next_bug[file, 0] + printf "" + printf "%d", bug, db[bug, file] + printf "^", prev_file[bug, file], bug + printf "v", next_file[bug, file], bug + printf " ", file, bug + printf " " + printf "%s", file_total[file] + printf "<", file, prev_bug[file, -1] + printf "" pfile "
" + print "" +} +' +} + + +# Make the scripts available +cp ${aridir}/gdb_*.sh ${wwwdir} + +# Compute the ARI index - ratio of zero vs non-zero problems. +indexes=`${AWK} ' +BEGIN { + FS=":" +} +{ + # ari.*.doc: ::: + bug = $1; count = $2; category = $3; doc = $4 + + if (bug ~ /^legacy_/) legacy++ + if (bug ~ /^deprecated_/) deprecated++ + + if (category !~ /^gdbarch$/) { + bugs += count + } + if (count == 0) { + oks++ + } +} +END { + #print "tests/ok:", nr / ok + #print "bugs/tests:", bugs / nr + #print "bugs/ok:", bugs / ok + print bugs / ( oks + legacy + deprecated ) +} +' ${wwwdir}/ari.doc` + +# Merge, generating the ARI tables. +if ${update_web_p} +then + echo "Create the ARI table" 1>&2 + oldari=${wwwdir}/old.html + ari=${wwwdir}/index.html + newari=${wwwdir}/new.html + rm -f ${newari} ${newari}.gz + cat <> ${newari} + + +A.R. Index for GDB version ${version} + + + +

A.R. Index for GDB version ${version}

+ + + + + +

${indexes}

+

You can not take this seriously!

+ +
+Also available: +most recent branch +| +current +| +last release +
+ +
+Last updated: `date -u` +
+EOF + + print_toc 0 1 "internal regression" Critical <> ${newari} ' +BEGIN { + FS = ":" + '"$alls"' +} +{ + # ari.*.doc: ::: + bug = $1 + count = $2 + category = $3 + doc = $4 + if (!(category in all)) { + print "" category ": no documentation
" + } +} +' + + cat >> ${newari} < +Input files: +`( cd ${wwwdir} && ls ari.*.bug ari.idx ari.doc ) | while read f +do + echo "${f}" +done` + + +
+Scripts: +`( cd ${wwwdir} && ls *.sh ) | while read f +do + echo "${f}" +done` +
+ + + + +EOF + + for i in . .. ../..; do + x=${wwwdir}/${i}/index.sh + if test -x $x; then + $x ${newari} + break + fi + done + + gzip -c -v -9 ${newari} > ${newari}.gz + + cp ${ari} ${oldari} + cp ${ari}.gz ${oldari}.gz + cp ${newari} ${ari} + cp ${newari}.gz ${ari}.gz + +fi # update_web_p + +# ls -l ${wwwdir} + +exit 0