forked from opensci/piflow
fix bug: remove hive-site.xml core-site.xml
This commit is contained in:
parent
426bcff6c2
commit
8b109c6902
|
@ -135,7 +135,7 @@
|
|||
<dependency>
|
||||
<groupId>org.apache.spark</groupId>
|
||||
<artifactId>spark-mllib_2.11</artifactId>
|
||||
<version>2.1.0</version>
|
||||
<version>${spark.version}</version>
|
||||
</dependency>
|
||||
|
||||
|
||||
|
|
|
@ -1,17 +0,0 @@
|
|||
<configuration>
|
||||
<!-- 指定hdfs的nameservice为ns1 -->
|
||||
<property>
|
||||
<name>fs.defaultFS</name>
|
||||
<value>hdfs://master:9000</value>
|
||||
</property>
|
||||
<!-- Size of read/write buffer used in SequenceFiles. -->
|
||||
<property>
|
||||
<name>io.file.buffer.size</name>
|
||||
<value>131072</value>
|
||||
</property>
|
||||
<!-- 指定hadoop临时目录,自行创建 -->
|
||||
<property>
|
||||
<name>hadoop.tmp.dir</name>
|
||||
<value>/opt/hadoop-2.6.0/tmp</value>
|
||||
</property>
|
||||
</configuration>
|
|
@ -1,44 +0,0 @@
|
|||
<configuration>
|
||||
<!-- 设置namenode的http通讯地址 -->
|
||||
<property>
|
||||
<name>dfs.namenode.http-address</name>
|
||||
<value>master:50070</value>
|
||||
</property>
|
||||
<!-- 设置secondarynamenode的http通讯地址 -->
|
||||
<property>
|
||||
<name>dfs.namenode.secondary.http-address</name>
|
||||
<value>master:50090</value>
|
||||
</property>
|
||||
<property>
|
||||
<name>dfs.replication</name>
|
||||
<value>1</value>
|
||||
</property>
|
||||
<!-- 设置namenode存放的路径 -->
|
||||
<property>
|
||||
<name>dfs.namenode.name.dir</name>
|
||||
<value>file:/opt/hadoop-2.6.0/dfs/name</value>
|
||||
</property>
|
||||
<!-- 设置datanode存放的路径 -->
|
||||
<property>
|
||||
<name>dfs.datanode.data.dir</name>
|
||||
<value>file:/opt/hadoop-2.6.0/dfs/data</value>
|
||||
</property>
|
||||
<property>
|
||||
<name>dfs.client.read.shortcircuit</name>
|
||||
<value>true</value>
|
||||
</property>
|
||||
<property>
|
||||
<name>dfs.domain.socket.path</name>
|
||||
<value>/var/run/hadoop-hdfs/dn._PORT</value>
|
||||
</property>
|
||||
|
||||
<property>
|
||||
<name>dfs.client.file-block-storage-locations.timeout</name>
|
||||
<value>10000</value>
|
||||
</property>
|
||||
<property>
|
||||
<name>dfs.datanode.hdfs-blocks-metadata.enabled</name>
|
||||
<value>true</value>
|
||||
</property>
|
||||
|
||||
</configuration>
|
|
@ -1,90 +0,0 @@
|
|||
<configuration>
|
||||
<property>
|
||||
<name>hive.metastore.warehouse.dir</name>
|
||||
<!--之前hdfs中创建的warehouse文件夹-->
|
||||
<value>/user/hive/warehouse</value>
|
||||
<description>location of default database for the warehouse</description>
|
||||
</property>
|
||||
|
||||
<property>
|
||||
<name>hive.metastore.uris</name>
|
||||
<value>thrift://10.0.86.89:9083</value>
|
||||
<description>Thrift URI for the remote metastore. Used by metastore client to connect to remote metastore.</description>
|
||||
</property>
|
||||
|
||||
<property>
|
||||
<name>javax.jdo.option.ConnectionURL</name>
|
||||
<value>jdbc:mysql://10.0.86.90:3306/hive?createDatabaseIfNotExist=true</value>
|
||||
<description>JDBC connect string for a JDBC metastore</description>
|
||||
</property>
|
||||
<property>
|
||||
<name>javax.jdo.option.ConnectionDriverName</name>
|
||||
<value>com.mysql.jdbc.Driver</value>
|
||||
<description>Driver class name for a JDBC metastore</description>
|
||||
</property>
|
||||
|
||||
<property>
|
||||
<name>javax.jdo.option.ConnectionUserName</name>
|
||||
<value>root</value>
|
||||
<description>username to use against metastore database</description>
|
||||
</property>
|
||||
<property>
|
||||
<name>javax.jdo.option.ConnectionPassword</name>
|
||||
<value>root</value>
|
||||
<description>password to use against metastore database</description>
|
||||
</property>
|
||||
|
||||
<!--start for trancaction -->
|
||||
<property>
|
||||
|
||||
<name>hive.support.concurrency</name>
|
||||
<value>true</value>
|
||||
|
||||
</property>
|
||||
|
||||
<property>
|
||||
|
||||
<name>hive.enforce.bucketing</name>
|
||||
|
||||
<value>true</value>
|
||||
|
||||
</property>
|
||||
|
||||
<property>
|
||||
|
||||
<name>hive.exec.dynamic.partition.mode</name>
|
||||
|
||||
<value>nonstrict</value>
|
||||
|
||||
</property>
|
||||
|
||||
<property>
|
||||
|
||||
<name>hive.txn.manager</name>
|
||||
|
||||
<value>org.apache.hadoop.hive.ql.lockmgr.DbTxnManager</value>
|
||||
|
||||
</property>
|
||||
|
||||
<property>
|
||||
|
||||
<name>hive.compactor.initiator.on</name>
|
||||
|
||||
<value>true</value>
|
||||
|
||||
</property>
|
||||
|
||||
<property>
|
||||
|
||||
<name>hive.compactor.worker.threads</name>
|
||||
|
||||
<value>1</value>
|
||||
|
||||
</property>
|
||||
|
||||
<!-- property>
|
||||
<name>hive.in.test</name>
|
||||
<value>true</value>
|
||||
</propertyi-->
|
||||
|
||||
</configuration>
|
Loading…
Reference in New Issue