Commit 59fdff43 authored by litaolemo's avatar litaolemo

update

parent 538e6724
...@@ -34,13 +34,13 @@ sparkConf.set("spark.sql.crossJoin.enabled", True) ...@@ -34,13 +34,13 @@ sparkConf.set("spark.sql.crossJoin.enabled", True)
# sparkConf.set("spark.tispark.plan.allow_index_double_read", False) # sparkConf.set("spark.tispark.plan.allow_index_double_read", False)
# sparkConf.set("spark.tispark.plan.allow_index_read", True) # sparkConf.set("spark.tispark.plan.allow_index_read", True)
# sparkConf.set("spark.hive.mapred.supports.subdirectories", True) # sparkConf.set("spark.hive.mapred.supports.subdirectories", True)
# sparkConf.set("spark.sql.adaptive.enabled", True) sparkConf.set("spark.sql.adaptive.enabled", True)
# sparkConf.set("spark.sql.adaptive.skewedJoin.enabled", True) # sparkConf.set("spark.sql.adaptive.skewedJoin.enabled", True)
sparkConf.set("spark.shuffle.statistics.verbose", True) sparkConf.set("spark.shuffle.statistics.verbose", True)
# sparkConf.set("spark.sql.adaptive.shuffle.targetPostShuffleInputSize", "67108864") # sparkConf.set("spark.sql.adaptive.shuffle.targetPostShuffleInputSize", "67108864")
# sparkConf.set("spark.sql.adaptive.shuffle.targetPostShuffleRowCount", "20000000") # sparkConf.set("spark.sql.adaptive.shuffle.targetPostShuffleRowCount", "20000000")
# sparkConf.set("spark.hadoop.mapreduce.input.fileinputformat.input.dir.recursive", True) # sparkConf.set("spark.hadoop.mapreduce.input.fileinputformat.input.dir.recursive", True)
# sparkConf.set("spark.serializer", "org.apache.spark.serializer.KryoSerializer") sparkConf.set("spark.serializer", "org.apache.spark.serializer.KryoSerializer")
# sparkConf.set("mapreduce.output.fileoutputformat.compress", False) # sparkConf.set("mapreduce.output.fileoutputformat.compress", False)
# sparkConf.set("mapreduce.map.output.compress", False) # sparkConf.set("mapreduce.map.output.compress", False)
# sparkConf.set("prod.gold.jdbcuri", # sparkConf.set("prod.gold.jdbcuri",
...@@ -54,7 +54,7 @@ sparkConf.set("spark.shuffle.statistics.verbose", True) ...@@ -54,7 +54,7 @@ sparkConf.set("spark.shuffle.statistics.verbose", True)
# sparkConf.set("prod.jerry.jdbcuri", # sparkConf.set("prod.jerry.jdbcuri",
# "jdbc:mysql://172.16.40.158:4000/jerry_prod?user=st_user&password=aqpuBLYzEV7tML5RPsN1pntUzFy&rewriteBatchedStatements=true") # "jdbc:mysql://172.16.40.158:4000/jerry_prod?user=st_user&password=aqpuBLYzEV7tML5RPsN1pntUzFy&rewriteBatchedStatements=true")
sparkConf.set("prod.tispark.pd.addresses", "172.16.40.158:2379") sparkConf.set("prod.tispark.pd.addresses", "172.16.40.158:2379")
# sparkConf.set("spark.sql.parquet.compression.codec", "snappy") sparkConf.set("spark.sql.parquet.compression.codec", "snappy")
sparkConf.set("prod.tispark.pd.addresses", "172.16.40.170:4000") sparkConf.set("prod.tispark.pd.addresses", "172.16.40.170:4000")
sparkConf.set("prod.tidb.database", "jerry_prod") sparkConf.set("prod.tidb.database", "jerry_prod")
# sparkConf.set("spark.executor.extraJavaOptions", "-Djava.library.path=$HADOOP_HOME/lib/native") # sparkConf.set("spark.executor.extraJavaOptions", "-Djava.library.path=$HADOOP_HOME/lib/native")
......
Markdown is supported
0% or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment