The Confluence connector will crawl content from any Confluence content repository. The connector will retrieve spaces, pages, blogs, attachments, and comments.
The connector uses the Confluence REST API to crawl Confluence content, and we support both Confluence On-premise installation and Cloud installation.
Easy Heading Free | ||||
---|---|---|---|---|
|
Some
of thefeatures of the
Confluence connectorConfluence connector include:
For a Complete tutorial on Confluence, see here
This is the hierarchy of spaces/pages/blogs/attachments/comments for Confluence versions:
The Aspire Confluence connector was created and tested using version
Confluence7.19.2 of Confluence.
Before installing the Confluence connector, make sure that:
In order to To access Confluence, a user account with sufficient privileges must be supplied. It is recommended that the account be the site administrator.
No special requirements here
Name | Supported |
---|---|
Content Crawling | yes |
Identity Crawling | yes |
Snapshot-based Incremental s | yes |
Non-snapshot-based Incremental s | no |
Document Hierarchy | yes |
The connector can operate in two modes: full and incremental.
Important: The data submitted to Aspire by this connector is dependent entirely on the SQL that's configured. Therefore, it is quite possible to submit all of the data in an incremental crawl, or only some of the data in a full crawl.
In full mode, the connector executes a single SQL database statement and submits each row returned for processing in Aspire.
In incremental mode, there are three stages of processing: preprocessing, crawling, and post-processing.
(Optional) This stage runs a SQL statement against the database that can be used to mark rows to crawl (i.e., they have changed since the previous run).
This stage (similar to full mode) executes a single SQL database statement and submits each row returned for processing in Aspire. Typically, the result set is a subset of the full data that may be filtered using information updated in the (optional) pre-processing stage.
(Optional) Each row of data submitted to Aspire can execute a SQL statement to update its status in the database. This may be to reset a flag set in the (optional) pre-processing stage, thereby denoting the item as processed. Different SQL can be executed for rows that were successfully processed versus ones that were not.
Click here to find out various crawling options
The content retrieved by the connector is defined entirely using SQL statements, so you can select all or subsets of columns from one or more tables. Initially, the data is inserted into Aspire using the returned column names, but this may be changed by further Aspire processing.The RDB via Tables connector is able to Confluence connector can crawl the following objects:
Name | Type | Relevant Metadata | Content Fetch & Extraction | Description | |
---|---|---|---|---|---|
space | database rowcontainer | table spaces fields | NA | ||
blog | container | blog fields | yes | ||
page | container | page fields | yes | ||
attachment | document | attachment fields | yes | Fields requested by SQL
No limitations defined