TAU (Tuning and Analysis Utilities) is a set of tools for analyzing the performance of C, C++, Fortran and Java programs. It collects much more information than is available through prof or gprof, the standard Unix utilities, including per-process, per-thread, and per-host information, inclusive and exclusive function times, profiling groups that allow you to organize data collection, access to hardware counters on some systems, per-class and per-instance information, the ability to separate data for each template instantiation, start/stop timers for profiling arbitrary sections of code, and support for collection of statistics on user-defined events.
Ccide reads C, C++, BASIC, or bash source code containing embedded decision tables from stdin, expands the tables, copies the remaining statements, and forms a compilable C or C++ source module, executable bash script, or interpretable BASIC program. Erratic side effects are avoided by evaluating all condition expressions at exactly the same time, and by performing all evaluations and actions in the original sequence.
TIMES is a tool suite designed mainly for symbolic schedulability analysis and synthesis of executable code with predictable behaviors for real-time systems. Given a system design model consisting of a set of application tasks (required to meet mixed timing, precedence, and resource constraints), a network of timed automata describing the task arrival pattern and a preemptive or non-preemptive scheduling policy, TIMES will generate a scheduler, and calculate worst-case response times for the tasks. The design model may be further validated using a model checker and then compiled to executable C-code.
C/C++ Program Perfometer checks the performance of a C/C++ program and separate pieces of code for any metrics (e.g. uclocks, rusage metrics, metrics defined by the user, etc.). The measurement results are represented in detailed/summary reports. The detailed report has results for individual tests, and the summary one has average cost and its analysis. The comparison results are represented in comparative tables for individual comparison groups. The user may set various parameters in order to control the measurement/comparison process: measurement report and detailed measurement report flags, total iteration and tests, measurement scale, and confidence threshold.