Yioop! is a PHP search engine. Yioop! can be configured as either a general purpose search engine for the whole Web or it can be configured to provide search results for a set of URLs or domains. Yioop can crawl pages or can directly index archives such as ARC and WARC. It supports indexing several file formats such as HTML, Atom, PDF, DOC, PPT, RTF, RSS, XML, SVG, PNG, JPG, BMP, GIF, and sitemaps. The Yioop! crawler can be deployed on one or many machines. It supports having one or more to crawl scheduler processes, as well as multiple fetchers and mirrors. Crawling respects robots.txt including Crawl-delay. Yioop! crawls are stored in a Web archive format that is easy to move around. Crawling can be done on one machine and the results deployed elsewhere. Yioop! supports mixing of crawls. Yioop! comes with a search front end that can be localized as desired using a GUI. This GUI supports RTL languages. Management of crawls can also be done using this GUI. Yioop! can be configured in a straightforward manner to make use of file caching or memcache if available.
Seeks is a Web search proxy, meta-search engine, and real-time P2P pattern matching network for social Web search. Its specific purpose is to regroup users whose queries are similar so they can share both the query results and their experience with these results. On this basis, Seeks allows true real-time, decentralized Web search to emerge. In the long term, there is no need for Web crawlers and third-party Web indexes. Most importantly, Seeks is intended to become a flagship for a fair, transparent, user controlled machinery for searching the Web over the Internet.
BlackRay is a relational database system designed to offer performance features commonly associated with search engines. It offers SQL support and sophisticated operational and management features. Load-balancing and operational stability by means of N+1 redundance are included. BlackRay is called a "Data Engine" since it combines traditional, relational database features and SQL with the power and flexibility of search engines. It is a true hybrid, offering transaction support, data-versioned snapshots, and sophisticated function-based indices. Wildcards, phonetic, and fuzzy logic searches are supported, as well. BlackRay supports a subset of the SQL92 standard and provides JDBC/ODBC/native driver options via the PostgreSQL protocol, in addition to an API based query option. The project is released under the GPLv2, with some drivers available under BSD-style licenses. Commercial support contracts are available as well.
Libcolumbus is a small error tolerant search engine designed to deal with noisy data and typos. It will power the searches in the next generation of Ubuntu's HUD system as well as other searches. It has a fast implementation of the Levenshtein distance algorithm, which allows it to correct errors such as added and dropped letters (e.g. 'bar' -> 'bard'), changed letters ('ctr' -> 'car') and translations ('acr' -> 'car'). It also allows the user to customize the error values. Libcolumbus is designed to be small, efficient and easy to embed. It is programmed in C++ but also provides C and Python APIs.
Find What I Mean aims to provide a searching library that tolerates errors in queries. It will auto-correct typos, extra letters, and so on. This is extremely useful when searching for an item in a list. In traditional search methods the query must be perfect or you get zero matches.
InstaSearch is an Eclipse IDE plug-in for performing quick and advanced searches of source code files. It uses the Apache Lucene library for indexing and fast searching of files in the workspace. The search is performed instantly as you type, and resulting files are displayed in an Eclipse view. Each file then can be previewed using a few of the most closely matching and relevant lines. A double-click on the match leads to the matching line in the file.
The Ex-Crawler Project is divided into three subprojects. The main part is the Ex-Crawler daemon server, a highly configurable and flexible Web crawler written in Java. It comes with its own socket server, with which you can manage the server, users, distributed grid/volunteer computing, and much more. Crawled information is stored in a database (Currently MySQL, PostgreSQL, and MSSQL are supported). The second part is a graphical (Java Swing) distributed grid/volunteer computing client, including user computer state detection, based on JADIF Project. The Web search engine is written in PHP. It comes with a Content Management System, user language detection and multi-language support, and templates using Smarty, including an application framework that is partly forked from Joomla 1.5, so that Joomla components can be adapted quickly.
Web Search Script can search the contents of a text file stored on your server by editing the script source code, or the file location can be left blank to allow your users to enter a URL to a text document stored on a remote Web server. Search results will be listed by line number, with contextual information surrounding the search term. Clicking search results will scroll the complete text in the right pane to the result's location in the complete text.