Deployment Model | DataOpsServer | DataOpsRepository | DataOpsEngine |
Single-Tier: In Single-tier architecture, we are installing DataOpsServer and DataOpsRepository and DataOpsEngine on the same machine Note: Not recommended for production | Processor:32 Cores Memory: 128GB Harddisk: 1000GB | Same as DataOpsServer | Same as DataOpsServer |
Two-Tier: In Two-tier architecture, we are installing DataOpsServer and DataOpsRepository in the same machine and DataOpsEngine on different machine | Processor:16 Cores Memory: 32GB Harddisk: 500GB | Same as DataOpsServer | Standalone DataOpsEngine Processor:32 Cores Memory: 128GB Harddisk: 500GB |
YarnCluster with three nodes Processor:16 Cores(on each node) Memory: 128GB(on each node) Harddisk: 1000GB(On each node) Note: we can scale in or scale out the number of cores and number of nodes based on data | |||
Processor:8 Cores on master, 16 cores on core and task nodes Memory: 64GB on master, 128GB on Core and task nodes Harddisk: 200GB on master, 1000GB on Core and task nodes Recommended Type: master r6g.2xlarge,core and task nodes-r6g.4xlarge Note: we can scale in or scale out the number of cores and number of task nodes based on data | |||
Azure databricks Processor:16 Cores on worker and 8 Cores on driver Memory: 128GB on worker and 16GB on driver | |||
Three-Tier: DataOps application consists of mainly three components(DataOpsServer, DataOpsRepository, DataOpsEngine). We are installing these three components in three different machines | Processor:16 Cores Memory: 16GB Harddisk: 200GB | Processor:8 Cores Memory: 16GB Harddisk: 500GB | Standalone DataOpsEngine Processor:32 Cores Memory: 128GB Harddisk: 500GB |
YarnCluster with three nodes Processor:16 Cores(on each node) Memory: 128GB(on each node) Harddisk: 1000GB(On each node) Note: we can scale in or scale out the number of cores and number of nodes based on data | |||
AWS EMR Processor:8 Cores on master, 16 cores on core and task nodes Memory: 64GB on master, 128GB on Core and task nodes Harddisk: 200GB on master, 1000GB on Core and task nodes Recommended Type: master r6g.2xlarge,core and task nodes-r6g.4xlarge Note: we can scale in or scale out the number of cores and number of task nodes based on data | |||
Azure databricks Processor:16 Cores on worker and 8 Cores on driver Memory: 128GB on worker and 16GB on driver |