|
| 1 | +# Hive实验手册 |
| 2 | + |
| 3 | +### 0. 下载本实验所需要的Git工程 |
| 4 | + |
| 5 | +``` |
| 6 | +cd /home/bigdata |
| 7 | +git clone https://github.com/bigdataguide/hadooptraining.git |
| 8 | +cd hadooptraining/hive |
| 9 | +``` |
| 10 | +hive目录下包含三个子目录: |
| 11 | +* conf: hive的配置 |
| 12 | +* command: 启动Hive服务的的命令和一些实验用的SQL语句 |
| 13 | +* data中包含的是外部的一些数据,可以直接导入到Hive表中 |
| 14 | + |
| 15 | +**注意:本实验使用的用户主目录是bigdata, 你需要将目录替换为你自己的主目录,下同。** |
| 16 | + |
| 17 | + |
| 18 | +### 1. 下载安装Hadoop |
| 19 | + |
| 20 | +参考之前的课程内容,下面假定我们的Hadoop目录安装在/home/bigdata/hadoop-2.7.3中,如果你不是安装在这个目录需要替换为你自己的目录 |
| 21 | + |
| 22 | +### 2. 配置Mysql |
| 23 | +``` |
| 24 | +#Ubuntu |
| 25 | +sudo apt-get install mysql |
| 26 | +#CentOS |
| 27 | +sudo yum install mysql |
| 28 | +#启动Mysql |
| 29 | +sudo service mysqld start |
| 30 | +``` |
| 31 | + |
| 32 | +### 3. 安装Hive |
| 33 | + |
| 34 | +#### 3.1 下载Hive二进制包 |
| 35 | +``` |
| 36 | +wget http://apache.mirrors.pair.com/hive/hive-2.1.1/apache-hive-2.1.1-bin.tar.gz /tmp/ |
| 37 | +#解压hive到工作目录 |
| 38 | +tar -zxvf /tmp/apache-hive-2.1.1-bin.tar.gz -C /home/bigdata/ |
| 39 | +cd /home/bigdata/apache-hive-2.1.1-bin/ |
| 40 | +``` |
| 41 | + |
| 42 | +#### 3.2 配置Hive:拷贝Git工程中的配置到Hive目录 |
| 43 | +``` |
| 44 | +cp /home/bigdata/hadooptraining/hive/conf/hive-env.sh /home/bigdata/apache-hive-2.1.1-bin/conf/ |
| 45 | +cp /home/bigdata/hadooptraining/hive/conf/hive-site.xml /home/bigdata/apache-hive-2.1.1-bin/conf/ |
| 46 | +``` |
| 47 | +**配置文件的说明见附1,根据实际情况选择自己的主目录** |
| 48 | + |
| 49 | +#### 3.3 启动Hive组件 |
| 50 | +``` |
| 51 | +export HADOOP_HOME=/home/bigdata/hadoop-2.7.3 |
| 52 | +export PATH=/home/bigdata/apache-hive-2.1.1-bin:$PATH |
| 53 | +#启动MetaStore Server |
| 54 | +nohup hive --service metastore >> /home/bigdata/apache-hive-2.1.0-bin/logs/metastore.log 2>&1 & |
| 55 | +#启动HiveServer2 |
| 56 | +nohup hive --service hiveserver2 >> /home/bigdata/apache-hive-2.1.0-bin/logs/hive.log 2>&1 & |
| 57 | +``` |
| 58 | +### 4. 启动Hive |
| 59 | +#### 4.1 启动Hive CLI |
| 60 | +``` |
| 61 | +hive |
| 62 | +``` |
| 63 | +#### 4.2 启动Beeline CLI |
| 64 | +``` |
| 65 | +beeline -n bigdata -pbigdata -u "jdbc:hive2://localhost:10000/default;auth=noSasl" |
| 66 | +#或者 |
| 67 | +beeline |
| 68 | +beeline> !connect jdbc:hive2://localhost:10000/default bigdata bigdata |
| 69 | +``` |
| 70 | + |
| 71 | +### 附 配置文件说明 |
| 72 | +在hive-env.sh中,我们配置了HADOOP_HOME目录,你需要将主目录替换为你自己的主目录 |
| 73 | +HADOOP_HOME=/home/bigdata/hadoop-2.7.3 |
| 74 | + |
| 75 | +在hive-site.xml中,我们配置了: |
| 76 | +1)使用mysql存储元数据 |
| 77 | +```xml |
| 78 | + <property> |
| 79 | + <name>javax.jdo.option.ConnectionURL</name> |
| 80 | + <value>jdbc:mysql://localhost/metastore?createDatabaseIfNotExist=true</value> |
| 81 | + </property> |
| 82 | + <property> |
| 83 | + <name>javax.jdo.option.ConnectionDriverName</name> |
| 84 | + <value>com.mysql.jdbc.Driver</value> |
| 85 | + </property> |
| 86 | + <property> |
| 87 | + <name>javax.jdo.option.ConnectionUserName</name> |
| 88 | + <value>root</value> |
| 89 | + </property> |
| 90 | + <property> |
| 91 | + <name>javax.jdo.option.ConnectionPassword</name> |
| 92 | + <value>root</value> |
| 93 | + </property> |
| 94 | +``` |
| 95 | +2)hive在HDFS上的存储路径 |
| 96 | +```xml |
| 97 | + <property> |
| 98 | + <name>hive.metastore.warehouse.dir</name> |
| 99 | + <value>/warehouse</value> |
| 100 | + </property> |
| 101 | + <property> |
| 102 | + <name>fs.defaultFS</name> |
| 103 | + <value>hdfs://bigdata:9000</value> |
| 104 | + </property> |
| 105 | +``` |
| 106 | +3)metastore的端口 |
| 107 | +```xml |
| 108 | + <property> |
| 109 | + <name>hive.metastore.uris</name> |
| 110 | + <value>thrift://bigdata:9083</value> |
| 111 | + </property> |
| 112 | +``` |
| 113 | +4)HiveServer2的端口 |
| 114 | +```xml |
| 115 | + <property> |
| 116 | + <name>hive.server2.thrift.port</name> |
| 117 | + <value>10000</value> |
| 118 | + </property> |
| 119 | + <property> |
| 120 | + <name>beeline.hs2.connection.user</name> |
| 121 | + <value>bigdata</value> |
| 122 | + </property> |
| 123 | + <property> |
| 124 | + <name>beeline.hs2.connection.password</name> |
| 125 | + <value>bigdata</value> |
| 126 | + </property> |
| 127 | +``` |
| 128 | +5) 此外我们还配置自动创建Meta Store的数据库和表 |
| 129 | +```xml |
| 130 | + <property> |
| 131 | + <name>datanucleus.autoCreateSchema</name> |
| 132 | + <value>true</value> |
| 133 | + </property> |
| 134 | + <property> |
| 135 | + <name>datanucleus.autoStartMechanism</name> |
| 136 | + <value>SchemaTable</value> |
| 137 | + </property> |
| 138 | + <property> |
| 139 | + <name>datanucleus.schema.autoCreateTables</name> |
| 140 | + <value>true</value> |
| 141 | + </property> |
| 142 | +``` |
0 commit comments