The can be installed across multiple nodes for high availability failover. This section describes the general process for installing the platform across multiple, highly available nodes.
NOTE: This section applies to deployments of the
|
The can also integrate with a highly available Hadoop cluster. For more information, see Enable Integration with Cluster High Availability in the Configuration Guide.
The following limitations apply to this feature:
/opt/trifacta
on every failover node.For more information, see Configure for High Availability in the Configuration Guide.
The supports an Active-Active HA deployment model, which works well at scale. The architecture features a single load balancer sitting in front of multiple nodes running the
. Each node:
/opt/trifacta/conf
and /opt/trifacta/logs
directories through NFS.
/opt/trifacta/conf
mount point, which allows the same configuration files to be visible and accessible on each node. In case of a failover event, any in-progress job should be marked as failed.
Failover events/scenarios around jobs:
# | Job | Event | Resulting job state |
---|---|---|---|
1 | In progress | The batch job runner is fine, but executor running the job fails. | Failed |
2 | In progress | The batch job runner or the node dies. | In Progress |
3 | Queued | The batch job runner or the node dies. | In Progress1 |
4 | Pending | The batch job runner or the node dies. | In Progress1 2 |
1 It may not be "In Progress". However, the job has not truly failed.
2 A nuance around #3. There is a feature flag that can be enabled and is enabled by default, which causes pending jobs to be marked as failed on (re)start of batch job runner. However, because this feature indiscriminately marks all pending jobs as failed, it cannot be safely enabled in an environment that has multiple running batch job runners.
The supports a single load balancer placed in front of multiple nodes, each of which runs the same version of
. Content between nodes is shared using an NFS resource mount.
master node: This node is the default one used for hosting and serving the . Example node information:
NFS Server Hostname: server.local NFS Server IP Address: 192.168.1.101 |
client node(s): These nodes are failover nodes in case the master node is unavailable. Example node information:
NFS Client Hostname: client.local NFS Client IP Address: 192.168.1.102 |
Shared resources:
Each node shares the following resources:
Directories shared via NFS mount:
/opt/trifacta/logs /opt/trifacta/conf |
Steps:
All nodes must meet the system requirements. See System Requirements in the Planning Guide.
All nodes must have the appropriate ports opened. See System Ports in the Planning Guide.
Install the databases.
NOTE: The databases must be installed in a location that is accessible to all nodes. |
NOTE: When installing databases for high availability access, you should deploy standard access and replication techniques that are consistent with the policies of your enterprise. |
See Install Databases in the Databases Guide.
Complete the installation process for the server node.
NOTE: After install, do not start the |
See Install Software.
Repeat the above process for each of the client nodes.
Additional configuration is required.
NOTE: Starting and stopping the platform in high availability mode requires additional steps. |
For more information, see Configure for High Availability in the Configuration Guide.