The Ex-Crawler Project is divided into three subprojects. The main part is the Ex-Crawler daemon server, a highly configurable and flexible Web crawler written in Java. It comes with its own socket server, with which you can manage the server, users, distributed grid/volunteer computing, and much more. Crawled information is stored in a database (Currently MySQL, PostgreSQL, and MSSQL are supported). The second part is a graphical (Java Swing) distributed grid/volunteer computing client, including user computer state detection, based on JADIF Project. The Web search engine is written in PHP. It comes with a Content Management System, user language detection and multi-language support, and templates using Smarty, including an application framework that is partly forked from Joomla 1.5, so that Joomla components can be adapted quickly.
Java distributed framework is a framework for distributed grid and / or volunteer computing. It's divided into a server and client library. You can create new or implement it into existing applications in no time; you don't need knowledge about network connections, sockets, etc. The Framework does almost everything automatically. It provides secure automatic client <-> server communications, unique IDs, automatic resending of jobs to new clients if needed, user stats, and much more. The client framework supports the detection of the computer's user state (idling, away, online, etc.). It also offers many other useful features and helpers for developing a distributed client application.