Commit 897d0d68 authored by 张彦钊's avatar 张彦钊

修改test文件

parent b32c80e2
......@@ -13,11 +13,12 @@ def test():
sparkConf = SparkConf().set("spark.hive.mapred.supports.subdirectories", "true")\
.set("spark.hadoop.mapreduce.input.fileinputformat.input.dir.recursive", "true")\
.set("spark.tispark.plan.allow_index_double_read", "false") \
.set("spark.tispark.plan.allow_index_read", "true")
.set("spark.tispark.plan.allow_index_read", "true")\
.set("spark.sql.extensions", "org.apache.spark.sql.TiExtensions")\
.set("spark.tispark.pd.addresses", "172.16.40.158:2379")
spark = SparkSession.builder.config("spark.sql.extensions", "org.apache.spark.sql.TiExtensions")\
.config("spark.tispark.pd.addresses", "172.16.40.158:2379")\
.config(conf= sparkConf).enableHiveSupport().getOrCreate()
spark = SparkSession.builder.config(conf= sparkConf).enableHiveSupport().getOrCreate()
# spark.sql("use online")
# spark.sql("ADD JAR hdfs:///user/hive/share/lib/udf/brickhouse-0.7.1-SNAPSHOT.jar")
# spark.sql("ADD JAR hdfs:///user/hive/share/lib/udf/hive-udf-1.0-SNAPSHOT.jar")
......
Markdown is supported
0% or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment