Big Data has following apps which are almost similar to each other in term of using it (or information required to run these apps): Hive, Pig, Spark, Hadoop
This demonstration will show on how to use Hive app.You may select any other Big data app instead of Hive.(since , all these apps are similar)
Step 1: Login to your cmpute Console Application (your-domain.cmpute.io) using registered Email Id and Password.
Step 2: You will be redirected to cmpute Dashboard. Next, click on the App Store located in the header.
Step 3: You will be redirected to the App store which has the apps supported on cmpute. To run Spark, click the Run A Job button in the Hive App.
Step 4: Now, to run Hive job, fill all the required given text fields. There are following text fields to be filled:
Account: Select your AWS Account where you want to run the job.
Cloud Region: The Cloud region will be selected automatically.
Job Name: You can give any desired name to your job.
VPC Information: If the Account has VPC associated with it, then it will get selected automatically.You can also select the desired VPC.
Subnet Information: If the Account has VPC associated with it, then it will get selected automatically.You can also select the desired number of subnet out of all. Note: By-default, all the subnet will be taken.
Cluster Name: Select the cluster name.
Script Location: Enter Script Location.
Input Location: Enter Input Location.
Output location: Enter Output location.
Arguments: Enter Arguments if you have any.
Action on Failure: Select the Action on Failure.
Instance Type: Select the Instance Type
Step 5: Click on the Save/Run button once you are done with filling all the details. This action will save your job and is available to see later on the Jobs page.
Step 6: On successful job addition, you would get a popup from where you can start your job immediately (by clicking Run Now).
Step 7: You can monitor the job progress using the Job Run Details page.