No available questions at this moment
Can the Lotus Notes Connector crawl Archive files?
The Lotus Notes Connector can process archive files so the individual files in the archive are able to be published, for more info visit Archive files processing
Why does an incremental crawl last as long as a full crawl?
Some connectors perform incremental crawls based on snapshot entries, which are meant to match the exact documents that have been indexed by the connector to the search engine. On an incremental crawl, the connector fully crawls the repository the same way as a full crawl, but it only indexes the modified, new or deleted documents during that crawl.
For a discussion on crawling, see Full & Incremental Crawls.
Save your content source before creating or editing another one
Failing to save a content source before creating or editing another content source can result in an error.
Save the initial content source before creating or working on another.
My connector keeps the same status "Running" and is not doing anything
After a crawl has finished, the connector status may not be updated correctly.
To confirm this, do the following:
1. In Robo 3T (formerly Robomongo), go to your connector database (like: aspire-nameOfYourConnector).
2. Open the "Status" collection and perform the following query:
3, Edit the entry and set the status to "S" (Completed).
Note: To see the full options of "Status" values, see MongoDB Collection Status.
My connector is not providing group expansion results
Make sure your connector has a manual scheduler configured for Group Expansion.
1, Go to the Aspire debug console, and look for the respective scheduler (in the fourth table: Aspire Application Scheduler).
2. If you are unsure which scheduler is for Group Expansion, you can check the Schedule Detail.
- You can identify it with the value: cacheGroups
3.To run the Group Expansion process, click Run.
Document URLs Contain IPs Instead of DNS Names