Hive Installation – Learning Hive Tutorial in simple and easy steps starting from introduction, Installation, Data Types, Create Database, Drop Database, Create Table, Alter Table, Drop Table, Partitioning, Built-in Operators, Hiveql select. Where, Hiveql Select Order By, Hiveql Group By, Hiveql Joins, Built-in functions, Views and Indexes. Hive, Introduction, Installation, Data Types, Hadoop installation on linux steps pdf Database, Drop Database, Create Table, Alter Table, Drop Table, Partitioning, Built-in Operators, Hiveql select. All Hadoop sub-projects such as Hive, Pig, and HBase support Linux operating system.
Therefore, you need to install any Linux flavored OS. Java must be installed on your system before installing Hive. If java is not installed in your system, then follow the steps given below for installing java. Generally you will find the downloaded java file in the Downloads folder. Verify it and extract the jdk-7u71-linux-x64. Open root, and type the following commands.
Now apply all the changes into the current running system. Now verify the installation using the command java -version from the terminal as explained above. Hadoop must be installed on your system before installing Hive. Download and extract Hadoop 2. 1 from Apache Software Foundation using the following commands. The following steps are used to install Hadoop 2. 1 in pseudo distributed mode.
You need to make suitable changes in those configuration files according to your Hadoop infrastructure. Given below are the list of files that you have to edit to configure Hadoop. It means the place where you want to store the Hadoop infra. Let us assume the following data.
In the above file, all the property values are user-defined and you can make changes according to your Hadoop infrastructure. This file is used to configure yarn into Hadoop. By default, Hadoop contains a template of yarn-site. First of all, you need to copy the file from mapred-site,xml. The following steps are used to verify the Hadoop installation. The expected result is as follows.
The following command is used to start dfs. Executing this command will start your Hadoop file system. The following command is used to start the yarn script. Executing this command will start your yarn daemons. The default port number to access Hadoop is 50070. Use the following url to get Hadoop services on your browser.
The default port number to access all applications of cluster is 8088. Use the following url to visit this service. The following steps are required for installing Hive on your system. Hive installation is completed successfully. We use Apache Derby database.