You are viewing an old version of this page. View the current version.

Compare with Current View Page History

« Previous Version 6 Next »

Connects directly to the Python Bridge, to send text or sections of the interpretation graph to be processed by ML algorithms in Python.

This recognizer is used when there is a need to classify an entire document for example. That is the difference when compared to the Python Model Recognizer which will be processing or running for each token or text block.

Processing an entire document has its benefits, it may be the best way to classify a document as a whole.  By running only once per document we get a boost in performance when compared to run the recognizer for each individual token or text block.

The other benefit is that we could normalize the text before sending it to the python model and also specify dependent tags so it runs in the order we need in the processing pipeline.


  • Trigger Flag ( type=string | default=EOF | optional ) - Flag to trigger this recognizer.
  • Service Hostname ( type=string | default=localhost | required ) - Name of the host where Python Bridge is running.
  • Service Port ( type=string | default=5000 | required ) - Port where Python Bridge is running.
  • Select Model ( type=string | required ) - Model name registered in the python bridge.
  • Select Version ( type=string | optional ) - Model version registered in the python wrapper to query.
  • Select Model Method ( type=string | required ) - Model method to call for the model.
  • Number of retries ( type=string | default=3 | required ) - Maximum number of connection retries to the Python Bridge.

Configuration

  • Include Vertex Text as Token ( type=boolean | default=false | optional ) - Includes text from vertex as new tokens.
  • Normalize Tags ( type=boolean | default=false | optional ) - Tags (flagged as SEMANTIC_TAG) are normalized.
  • Dependency tags ( type=string array | optional ) - List of tags to be dependent for the model.

General Settings

The general settings can be accessed by clicking on

More settings could be displayed in the same dialog, it varies per recognizer.


  • Enable - Enable the processor to be use in pipelines.
  • Base Pipeline - Indicates the last stage, from a pipeline, needed by the recognizer.
  • Skip Flags ( optional ) - Lexical items flags to be ignored by this processor.
  • Boundary Flags  ( optional ) - List of vertex flags that indicate the beginning and end of a text block.
  • Required Flags ( optional ) - Lexical items flags required by every token to be processed.
  • At Least One Flag ( optional ) - Lexical items flags needed by every token to be processed.
  • Don't Process Flags ( optional ) - List of lexical items flags that are not processed. The difference with "Skip Flags" is that this will drop the path in the Saga graph, skip just skips the token and continues in the same path.
  • Confidence Adjustment - Adjustment factor to apply to the confidence value of 0.0 to 2.0 from (Applies for every match).
    • 0.0 to < 1.0  decreases confidence value
    • 1.0 confidence value remains the same
    • > 1.0 to  2.0 increases confidence value
  • Debug - Enable debug logging.





  • No labels