Since the 3.1 release, Aspire connectors are able to crawl in distributed mode automatically. Since all the crawl control data is stored in MongoDB, by just adding more Aspire servers configured to use the same MongoDB, the common connectors are going to crawl distributively.
Each connector is responsible for talking to the repositories, scanning through all the items to fetch and store its IDs to MongoDB for being processed later by any other server or itself.
On this page:
In order to setup an Aspire Cluster for Distributed Processing, you need to do the following steps:
You need to configure all Aspire servers to use the same MongoDB Installation, configure all the Aspire Servers config/settings.xml file
If you need to connect to a multi node MongoDB installation, check: Connect to a Multi-node MongoDB Installation
More details for Zookeeper installation and settings at Failover Settings (Zookeeper)
Install the content sources to distribute
Now it is time to think about which content sources you want to crawl distributively, and from what Aspire Servers, according to your solution architecture.
For this, configure the content sources in one of the servers and once you have them correctly configured export the content source and import it into the Aspire Servers you want to crawl this content source in parallel.
Controlling distributed processing is very simple, all you need to know is that if you start the crawl from any of the Aspire Servers, the crawl will start from all the servers, the same applies if you pause, stop or resume a crawl.
You are safe to shutdown all aspire servers when you do a Pause.
If you need to shutdown all the servers in the cluster, it is mandatory to pause the crawl first.