Others
Note: Some of the properties reflected are not feasible with Multi-Cloud version of Gathr. These properties are marked with **
Miscellaneous configurations properties left of the Web Studio. This category is further divided into various sub-categories.
LDAP
Field | Description |
---|---|
Password | Password against which the user will be authenticated in LDAP Server. |
Group Search Base | Defines the part of the directory tree under which group searches will be performed. |
User Search Base | Defines the part of the directory tree under which DN searches will be performed. |
User Search Filter | The filter which will be used to search DN within the User Search Base defined above. |
Group Search Filter | The filter which is used to search for group membership. The default is member={0 corresponding to the groupOfMembers LDAP class. In this case, the substituted parameter is the full distinguished name of the user. The parameter {1} can be used if you want to filter on the login name. |
Admin Group Name | LDAP group name which maps to application’s Admin role. |
Developer Group Name | LDAP group name which maps to application’s Developer role. |
Devops Group Name | LDAP group name which maps to application’s Devops role. |
Tier-II Group Name | LDAP group name which maps to application’s Tier-II role. |
LDAP Connection URL | URL of the LDAP server is a string that can be used to encapsulate the address and port of a directory server. For e.g. - ldap://host:port. |
User Distinguished Name | A unique name which is used to find the user in LDAP Server. |
Activiti
Field | Description |
---|---|
Alert Email Character Set | The character set used for sending emails. |
Alert Sender Email | The email address from which mails must be sent. |
JDBC Driver Class | The database driver used for activity setup. |
JDBC URL | The database URL for activity database. |
JDBC User | The database user name. |
JDBC Password | JDBC Password. |
Host | The email server host from which emails will be sent |
Port | The email server port. |
User | The email id from which emails will be sent. |
Password | The Password of the email account from which emails will be sent. |
Default Sender Email | The default email address from which mails will be sent if you do not provide one in the UI. |
Enable SSL | If SSL (Secure Sockets Layer) is enabled for establishing an encrypted link between server and client. |
Enable TSL | If Transport Layer Security (TLS) enables the encrypted communication of messages between hosts that support TLS and can also allow one host to verify the identity of another. |
History | Activiti history is needed or not. |
Database | The database used for activity setup. |
Couchbase
Field | Description |
---|---|
Max Pool Size ** | The Couchbase Max Pool Size. |
Default Bucket Memory Size ** | The memory size of default bucket in Couchbase. |
Password ** | The Couchbase password. |
Default Bucket Replica No ** | The Couchbase default bucket replication number. |
Host Port ** | The port no. of Couchbase. |
Host Name ** | The host on which the Couchbase is running. |
HTTP URL ** | The Couchbase http URL. |
Bucket List ** | The Couchbase bucket list. |
Polling timeout ** | The polling timeout of Couchbase. |
Polling sleeptime ** | The sleep time between each polling. |
User Name ** | The username of the Couchbase user. |
Kerberos
Field | Description |
---|---|
Hadoop NameNode Kerberos Principal | Service principal of name node. |
Kerberos Configuration File Override | Set to true if you want the keytab_login.conf file to be (re)created for every running pipeline when Kerberos security is enabled. |
Hadoop Core Site Location | The property should be used when trying to connect HDFS from two different realms. This property signifies the path of Hadoop core-site.xml containing roles for cross-realm communications. |
Hbase Master Kerberos Principal | Service principal of HBase master. |
ResourceManager Kerberos Principal | Service principal of resource manager |
Hbase Regionserver Kerberos Principal | Service principal of region server. |
Hive Metastore Kerberos principal | Service principal of Hive metastore. |
HiveServer2 Kerberos Principal | Service principal of hive server 2 |
Configuring Kerberos
You can add extra Java options for any Spark Superuser pipeline in following way:
Login as Superuser and click on Data Pipeline and edit any pipeline.
Kafka
HDFS
HBASE
SOLR
Zookeeper
Configure Kerberos
Once Kerberos is enabled, go to Superuser UI > Configuration > Environment > Kerberos to configure Kerberos.
Configure Kerberos in Components
Go to Superuser UI > Connections, edit the component connection settings as explained below:
HBase, HDFS
Field | Description |
---|---|
Key Tab Select Option | A Keytab is a file containing pair of Kerberos principals and encrypted keys. You can use Keytab to authenticate various remote systems. It has two options: Specify Keytab File Path: Path where Keytab file is stored Upload Keytab File: Upload Keytab file from your local file system. |
Specify Keytab File Path | If the option selected is Specify** Keytab File Path,** system will display the field** KeyTab File Path where you will specify the keytab file location. |
Upload Keytab File | If the option selected is Upload** Keytab File**,** system will display the** field** Upload Keytab File that will enable you to upload the Keytab file. |
By default, Kerberos security is configured for these components: Solr, Kafka and Zookeeper. No manual configuration is required.
Jupyter
Field | Description |
---|---|
jupyter.hdfs.port | HDFS Http port. |
jupyter.hdfs.dir | HDFS location where uploaded data will be saved. |
jupyter.dir | Location where notebooks will be created. |
jupyter.notebook.service.port | Port on which Auto create Notebook service is running. |
jupyter.hdfs.connection.name | HDFS connection name use to connect HDFS (from gathr connection tab). |
jupyter.url | URL contains IP address and port where Jupyter services are running. |
Cloudera
Property | Description |
---|---|
Navigator URL | The Cloudera Navigator URL. |
Navigator API Version | The Cloudera Navigator API version used. |
Navigator Admin User | The Cloudera navigator Admin user. |
Navigator User Password | The Cloudera navigator Admin user password. |
Autocommit Enabled | Specifies of the auto-commit of entities is required. |
Airflow
Property | Description |
---|---|
Enable AWS MWAA | Option to enable AWS Managed Airflow for gathr. It is disabled by default. |
If Enable AWS MWAA is check-marked, additional fields will be displayed as given below:
Region | AWS region should be provided where MWAA environment is created. |
Provider Type | Option to choose AWS credentials provider type. |
AWS Access Key ID | AWS account access key ID should be provided for authentication if Provider Type is selected as AWS keys. |
AWS Secret Access Key | AWS account secret access key should be provided for the Access Key ID specified above. The available options to choose from the drop-down list are: None, AWS Keys and Instance Profile. |
Environment Name | Exact AWS environment name should be provided that is required to be integrated with gathr. |
Gathr Service URL | Default URL where gathr application is installed. This should be updated for any modifications done to the gathr base URL. |
DAG Bucket Name | Exact DAG bucket name should be provided that is configured in the environment specified above. |
DAG Path | DAG path should be provided for the bucket name specified above. |
If Enable AWS MWAA is un-check, the below fields will be displayed:
Airflow Server Token Name | It is the key that is used to authenticate a request. It should be same as the value given in section Plugin Installation>Authentication for property ‘sax_request_http_token_name’ |
Airflow Server Token Required | Check if the token is required. |
Airflow Server Token Value | HTTP token to authenticate request. It should be same as the value given in section Plugin Installation > Authentication for property ‘sax_request_http_token_value’ |
Airflow Server URL | Airflow URL to connect to Airflow. |
If you have any feedback on Gathr documentation, please email us!