【问题】Kylin Step 10 Build Cube失败

简介:

build cube任务过程中,执行到step 10任务失败,原因是配置kylin的执行MR任务的资源不足


任务进行到第十步骤,失败。具体原因查找

wKioL1jKTADzrbs3AAAd8WnK6rM719.png

1、查看Parameters

wKiom1jKTBjgtdxnAABKgSs2l28572.png

2、查看MRJob日志,发现问题所在

2017-03-16 15:42:43,241 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: Diagnostics report from attempt_1484048632162_72188_r_000001_2: Container [pid=11851,containerID=container_1484048632162_72188_01_000170] is running beyond physical memory limits. Current usage: 1.1 GB of 1 GB physical memory used; 2.8 GB of 2.1 GB virtual memory used. Killing container. Dump of the process-tree for container_1484048632162_72188_01_000170 : |- PID PPID PGRPID SESSID CMD_NAME USER_MODE_TIME(MILLIS) SYSTEM_TIME(MILLIS) VMEM_USAGE(BYTES) RSSMEM_USAGE(PAGES) FULL_CMD_LINE |- 11859 11851 11851 11851 (java) 20435 2125 2883371008 280614 /usr/local/jdk/bin/java -Djava.net.preferIPv4Stack=true -Dhadoop.metrics.log.level=WARN -Xmx768m -Djava.io.tmpdir=/hadoop/2/yarn/local/usercache/kylin/appcache/application_1484048632162_72188/container_1484048632162_72188_01_000170/tmp -Dlog4j.configuration=container-log4j.properties -Dyarn.app.container.log.dir=/hadoop/6/yarn/logs/application_1484048632162_72188/container_1484048632162_72188_01_000170 -Dyarn.app.container.log.filesize=0 -Dhadoop.root.logger=INFO,CLA -Dhadoop.root.logfile=syslog -Dyarn.app.mapreduce.shuffle.logger=INFO,shuffleCLA -Dyarn.app.mapreduce.shuffle.logfile=syslog.shuffle -Dyarn.app.mapreduce.shuffle.log.filesize=0 -Dyarn.app.mapreduce.shuffle.log.backups=0 org.apache.hadoop.mapred.YarnChild 10.10.16.21 41607 attempt_1484048632162_72188_r_000001_2 170 |- 11851 11849 11851 11851 (bash) 0 0 108625920 341 /bin/bash -c /usr/local/jdk/bin/java -Djava.net.preferIPv4Stack=true -Dhadoop.metrics.log.level=WARN  -Xmx768m -Djava.io.tmpdir=/hadoop/2/yarn/local/usercache/kylin/appcache/application_1484048632162_72188/container_1484048632162_72188_01_000170/tmp -Dlog4j.configuration=container-log4j.properties -Dyarn.app.container.log.dir=/hadoop/6/yarn/logs/application_1484048632162_72188/container_1484048632162_72188_01_000170 -Dyarn.app.container.log.filesize=0 -Dhadoop.root.logger=INFO,CLA -Dhadoop.root.logfile=syslog -Dyarn.app.mapreduce.shuffle.logger=INFO,shuffleCLA -Dyarn.app.mapreduce.shuffle.logfile=syslog.shuffle -Dyarn.app.mapreduce.shuffle.log.filesize=0 -Dyarn.app.mapreduce.shuffle.log.backups=0 org.apache.hadoop.mapred.YarnChild 10.10.16.21 41607 attempt_1484048632162_72188_r_000001_2 170 1>/hadoop/6/yarn/logs/application_1484048632162_72188/container_1484048632162_72188_01_000170/stdout 2>/hadoop/6/yarn/logs/application_1484048632162_72188/container_1484048632162_72188_01_000170/stderr   Container killed on request. Exit code is 143 Container exited with a non-zero exit code 143

3、修改配置文件kylin_job_conf_inmem.xml,添加

1
2
3
4
5
6
7
8
9
10
11
<!-- reduce -->
< property >
< name >mapreduce.reduce.memory.mb</ name >
< value >3096</ value >
< description >每个Reduce Task需要的内存量</ description >
</ property >
< property >
< name >mapreduce.reduce.java.opts</ name >
< value >-Xmx3096m</ value >
< description >reduce任务内存</ description >
</ property >



     本文转自巧克力黒 51CTO博客,原文链接:http://blog.51cto.com/10120275/1907275 ,如需转载请自行联系原作者





相关文章
|
5天前
|
SQL 消息中间件 存储
Flink报错问题之Flink报错:Table sink 'a' doesn't support consuming update and delete changes which is produced by node如何解决
Flink报错通常是指在使用Apache Flink进行实时数据处理时遇到的错误和异常情况;本合集致力于收集Flink运行中的报错信息和解决策略,以便开发者及时排查和修复问题,优化Flink作业的稳定性。
|
6月前
|
流计算
Flink CDC在运行过程中遇到"Could not upload job files"的问题
Flink CDC在运行过程中遇到"Could not upload job files"的问题
127 1
|
jenkins 持续交付
问题记录:jenkins构建时报错The goal you specified requires a project to execute but there is no POM in...
问题记录:jenkins构建时报错The goal you specified requires a project to execute but there is no POM in...
705 0
问题记录:jenkins构建时报错The goal you specified requires a project to execute but there is no POM in...
|
存储 JSON Prometheus
Linkerd 2.10(Step by Step)—导出指标
Linkerd 2.10(Step by Step)—导出指标
|
SQL 存储 分布式计算
【spark系列5】spark 3.0.1集成delta 0.7.0原理解析--delta如何进行DDL DML操作以及Catalog plugin API
【spark系列5】spark 3.0.1集成delta 0.7.0原理解析--delta如何进行DDL DML操作以及Catalog plugin API
367 0
|
SQL 存储 分布式计算
【spark系列4】spark 3.0.1集成delta 0.7.0原理解析--delta自定义sql
【spark系列4】spark 3.0.1集成delta 0.7.0原理解析--delta自定义sql
257 0
|
存储 SQL 分布式计算
从 Neo4j 导入 Nebula Graph 实践见 SPark 数据导入原理
本文主要讲述如何使用基于 Spark 的数据导入工具 Nebula Graph Exchange 将数据从 Neo4j 导入到 Nebula Graph Database。
684 0
|
消息中间件 数据采集 分布式计算
在 Cloudera Data Flow 上运行你的第一个 Flink 例子
本文主要是介绍如何在 CDH6.3 中安装 Flink 1.9 以及运行你的第一个 Flink 例子。