The Ex-Crawler Project is divided into three subprojects. The main part is the Ex-Crawler daemon server, a highly configurable and flexible Web crawler written in Java. It comes with its own socket server, with which you can manage the server, users, distributed grid/volunteer computing, and much more. Crawled information is stored in a database (Currently MySQL, PostgreSQL, and MSSQL are supported). The second part is a graphical (Java Swing) distributed grid/volunteer computing client, including user computer state detection, based on JADIF Project. The Web search engine is written in PHP. It comes with a Content Management System, user language detection and multi-language support, and templates using Smarty, including an application framework that is partly forked from Joomla 1.5, so that Joomla components can be adapted quickly.
Executor is a small but perfectly functional Java (1.5 or higher) JAR executable that is used to run programs in shared computer resources like a pool of computers that you and your friends own. Moreover, it can be used as a simple file sharing program: you can access the files of your home computer from your office computer. Executor uses a shared or synchronized file system. The user creates a task in one computer (just by uploading a file), and another computer that has the required program will execute the task.