java.io.IOException: OS command error exit with return code: 1, error message: SLF4J: Class path contains multiple SLF4J bindings. SLF4J: Found binding in [jar:file:/opt/apps/hbase/lib/slf4j-log4j12-1.7.25.jar!/org/slf4j/impl/StaticLoggerBinder.class] SLF4J: Found binding in [jar:file:/opt/apps/hadoop-3.1.3/share/hadoop/common/lib/slf4j-log4j12-1.7.25.jar!/org/slf4j/impl/StaticLoggerBinder.class] SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation. SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory] Hive Session ID = f52c2c95-2ef1-45c8-a1bd-343dbd63e108 Logging initialized using configuration in jar:file:/opt/apps/hive/lib/hive-common-3.1.2.jar!/hive-log4j2.properties Async: true Hive Session ID = 350019f1-1ea7-448d-984f-604023dec0dc OK Time taken: 0.98 seconds OK Time taken: 1.006 seconds OK Time taken: 0.517 seconds OK Time taken: 0.157 seconds No Stats for gmall@dim_sku_info, Columns: tm_name, id Query ID = root_20211116144753_f0d42291-14c5-451e-9d8c-5e1bb15b1798 Total jobs = 3 Launching Job 1 out of 3 In order to change the average load for a reducer (in bytes): set hive.exec.reducers.bytes.per.reducer=In order to limit the maximum number of reducers: set hive.exec.reducers.max= In order to set a constant number of reducers: set mapreduce.job.reduces= Failed to execute spark task, with exception 'java.lang.Exception(Failed to submit Spark work, please retry later)' FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.spark.SparkTask. Failed to submit Spark work, please retry later The command is: hive -e "set mapreduce.job.name=Create Intermediate Flat Hive Table order_detail_cube 9376745b-ab75-ce13-a311-23c696b16656-00; USE default; DROp TABLE IF EXISTS kylin_intermediate_order_detail_cube_9cf454a4_4522_b21b_98ce_4ae87d78b28d; CREATE EXTERNAL TABLE IF NOT EXISTS kylin_intermediate_order_detail_cube_9cf454a4_4522_b21b_98ce_4ae87d78b28d ( `DIM_SKU_INFO_VIEW_TM_NAME` string ,`DIM_USER_INFO_VIEW_GENDER` string ,`DIM_base_PROVINCE_PROVINCE_NAME` string ,`DIM_base_PROVINCE_REGION_NAME` string ,`DWD_ORDER_DETAIL_SKU_NUM` bigint ,`DWD_ORDER_DETAIL_ORIGINAL_AMOUNT` decimal(16,2) ,`DWD_ORDER_DETAIL_SPLIT_ACTIVITY_AMOUNT` decimal(16,2) ,`DWD_ORDER_DETAIL_SPLIT_COUPON_AMOUNT` decimal(16,2) ,`DWD_ORDER_DETAIL_SPLIT_FINAL_AMOUNT` decimal(16,2) ) STORED AS SEQUENCEFILE LOCATION 'hdfs://doit100:8020/kylin/kylin_metadata/kylin-9376745b-ab75-ce13-a311-23c696b16656/kylin_intermediate_order_detail_cube_9cf454a4_4522_b21b_98ce_4ae87d78b28d'; ALTER TABLE kylin_intermediate_order_detail_cube_9cf454a4_4522_b21b_98ce_4ae87d78b28d SET TBLPROPERTIES('auto.purge'='true'); INSERT OVERWRITE TABLE `kylin_intermediate_order_detail_cube_9cf454a4_4522_b21b_98ce_4ae87d78b28d` SELECT `DIM_SKU_INFO_VIEW`.`TM_NAME` as `DIM_SKU_INFO_VIEW_TM_NAME` ,`DIM_USER_INFO_VIEW`.`GENDER` as `DIM_USER_INFO_VIEW_GENDER` ,`DIM_base_PROVINCE`.`PROVINCE_NAME` as `DIM_base_PROVINCE_PROVINCE_NAME` ,`DIM_base_PROVINCE`.`REGION_NAME` as `DIM_base_PROVINCE_REGION_NAME` ,`DWD_ORDER_DETAIL`.`SKU_NUM` as `DWD_ORDER_DETAIL_SKU_NUM` ,`DWD_ORDER_DETAIL`.`ORIGINAL_AMOUNT` as `DWD_ORDER_DETAIL_ORIGINAL_AMOUNT` ,`DWD_ORDER_DETAIL`.`SPLIT_ACTIVITY_AMOUNT` as `DWD_ORDER_DETAIL_SPLIT_ACTIVITY_AMOUNT` ,`DWD_ORDER_DETAIL`.`SPLIT_COUPON_AMOUNT` as `DWD_ORDER_DETAIL_SPLIT_COUPON_AMOUNT` ,`DWD_ORDER_DETAIL`.`SPLIT_FINAL_AMOUNT` as `DWD_ORDER_DETAIL_SPLIT_FINAL_AMOUNT` FROM `GMALL`.`DWD_ORDER_DETAIL` as `DWD_ORDER_DETAIL` INNER JOIN `GMALL`.`DIM_SKU_INFO_VIEW` as `DIM_SKU_INFO_VIEW` ON `DWD_ORDER_DETAIL`.`SKU_ID` = `DIM_SKU_INFO_VIEW`.`ID` INNER JOIN `GMALL`.`DIM_USER_INFO_VIEW` as `DIM_USER_INFO_VIEW` ON `DWD_ORDER_DETAIL`.`USER_ID` = `DIM_USER_INFO_VIEW`.`ID` INNER JOIN `GMALL`.`DIM_base_PROVINCE` as `DIM_base_PROVINCE` ON `DWD_ORDER_DETAIL`.`PROVINCE_ID` = `DIM_base_PROVINCE`.`ID` WHERe 1=1 AND (`DWD_ORDER_DETAIL`.DT >= '2020-06-15' AND `DWD_ORDER_DETAIL`.DT < '2020-06-16') ; " --hiveconf hive.merge.mapredfiles=false --hiveconf hive.auto.convert.join=true --hiveconf dfs.replication=2 --hiveconf hive.exec.compress.output=true --hiveconf hive.auto.convert.join.noconditionaltask=true --hiveconf mapreduce.job.split.metainfo.maxsize=-1 --hiveconf hive.merge.mapfiles=false --hiveconf hive.auto.convert.join.noconditionaltask.size=100000000 --hiveconf hive.stats.autogather=true at org.apache.kylin.common.util.CliCommandExecutor.execute(CliCommandExecutor.java:96) at org.apache.kylin.source.hive.CreateFlatHiveTableStep.createFlatHiveTable(CreateFlatHiveTableStep.java:59) at org.apache.kylin.source.hive.CreateFlatHiveTableStep.doWork(CreateFlatHiveTableStep.java:97) at org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:178) at org.apache.kylin.job.execution.DefaultChainedExecutable.doWork(DefaultChainedExecutable.java:71) at org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:178) at org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRunner.run(DefaultScheduler.java:114) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748)



