librsb is a library for sparse matrix computations featuring the Recursive Sparse Blocks (RSB) matrix format. This format allows cache-efficient and multithreaded (that is, shared memory parallel) operations on large sparse matrices. The most common operations necessary to iterative solvers are available (matrix-vector multiplication, triangular solution, rows/columns scaling, diagonal extraction/setting, blocks extraction, norm computation, formats conversion). The RSB format is especially well-suited for symmetric and transposed multiplication variants. On these variants, librsb has been found to be faster than Intel MKL's implementation for CSR. Most numerical kernels code is auto-generated, and the supported numerical types can be chosen by the user at buildtime. librsb implements the Sparse BLAS standard, as specified in the BLAS Forum documents.
BitDew is a programmable environment for the management and distribution of data for grid, desktop grid, and cloud systems. It can easily be integrated into large scale computational systems such as XtremWeb, BOINC, Hadoop, Condor, Glite, Unicore, OpenStack, and Eucalyptus. It provides key P2P, grid, and cloud technologies (DHT, BitTorrent, Amazon S3, DropBox) and high level programming interfaces with a simple API for creating, accessing, storing, and moving data with ease, even in highly dynamic and volatile environments.
ScalaBLAST is a high-performance multiprocessor implementation of the NCBI BLAST library. It supports all 5 primary program types (blastn, blastp, tblastn, tblastx, and blastx) and several output formats (pairwise, tabular, and XML). It will run on most multiprocessor systems which have MPI installed, and can run over a wide variety of interconnects, including infiniband, quadrics, and ethernet. It is designed to run a large number of queries against either large or small databases. It parallelizes the BLAST calculations by dynamically scheduling them across processors using a fault-resilient scheme.
HPCC (High Performance Computing Cluster) stores and processes large quantities of data, processing billions of records per second using massive parallel processing technology. Large amounts of data across disparate data sources can be accessed, analyzed, and manipulated in fractions of seconds. HPCC functions as both a processing and a distributed data storage environment capable of analyzing terabytes of information.
YML is a research project that aims to provide tools for using global computing middleware such as GRID, peer to peer, and metacomputing environments. The YML software architecture enables the definition of parallel applications, independently of the underlying middleware used. Parallel applications are defined using a workflow language called YvetteML.
Strategico is an engine for running statistical analysis over groups of time series. It can manage one or more groups (projects) of time series: by default, you can get data from a database or CSV files, normalize them, and then save them inside the engine. The first statistical analysis implemented inside Strategico is the "Long Term Prediction": it automatically finds the best model that fits each time series. Some of the models implemented are mean, trend, linear, exponential smoothing, and Arima. Strategico is scalable: the statistical analysis over each time series (of a project) can be run separately and independently. It is suggested that you set up an HPC Cluster (High Performance Computing) and/or use a resource scheduler like slurm. It is developed with R, one of the most famous statistical languages.