I tried to install rmr2 on a four node cluster running MapR Hadoop M5. The OS is Ubuntu 12.0.4.1 with R 2.15.2. The installation of hadoop was successful - all jobs are running fine. I proved this with the following example:
hadoop fs -mkdir /test/rmr2setup/wc-in
hadoop fs -put /opt/mapr/NOTICE.txt /test/rmr2setup/wc-in
hadoop jar /opt/mapr/hadoop/hadoop-0.20.2/hadoop-0.20.2-dev-examples.jar wordcount /test/rmr2setup/wc-in /test/rmr2setup/wc-out
hadoop fs -rmr /test/rmr2setup
Next, to install rmr2, I followed the instructions given in http://www.mapr.com/blog/harness-the-power-of-r-and-hadoop very closely. The installation of rhdfs was successfully and it is possible to reach hdfs from inside R. Also the installation of the rmr2 package was fine (Installed R and the packages on all nodes). But when I try to start a mapreduce job from inside R it gives my an error:
log4j:WARN No appenders could be found for logger (org.apache.hadoop.util.NativeCodeLoader).
log4j:WARN Please initialize the log4j system properly.
java.io.FileNotFoundException: File /tmp/mapr-hadoop/mapred/local/taskTracker/mapr/jobcache/job_201301291535_0013/attempt_201301291535_0013_m_000001_0/work/Rscript does not exist.
at org.apache.hadoop.fs.RawLocalFileSystem.getFileStatus(RawLocalFileSystem.java:395)
at org.apache.hadoop.fs.FilterFileSystem.getFileStatus(FilterFileSystem.java:257)
at org.apache.hadoop.fs.FileUtil.chmod(FileUtil.java:738)
at org.apache.hadoop.fs.FileUtil.chmod(FileUtil.java:713)
at org.apache.hadoop.streaming.PipeMapRed.configure(PipeMapRed.java:185)
at org.apache.hadoop.streaming.PipeMapper.configure(PipeMapper.java:66)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:601)
at org.apache.hadoop.util.ReflectionUtils.setJobConf(ReflectionUtils.java:88)
at org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:64)
at org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:117)
at org.apache.hadoop.mapred.MapRunner.configure(MapRunner.java:34)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:601)
at org.apache.hadoop.util.ReflectionUtils.setJobConf(ReflectionUtils.java:88)
at org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:64)
at org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:117)
at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:413)
at org.apache.hadoop.mapred.MapTask.run(MapTask.java:348)
at org.apache.hadoop.mapred.Child$4.run(Child.java:270)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:415)
at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1127)
at org.apache.hadoop.mapred.Child.main(Child.java:264)
2013-01-30 17:57:09,7995 ERROR Client fs/client/fileclient/cc/writebuf.cc:272 Thread: 139822073804544 FlushWrite failed: File part-00001, error: Stale File handle(116), pfid 2049.370.67412, off 0, fid 2049.370.67412
The environment variable are set properly in /etc/environment:
HADOOP_CMD="/opt/mapr/hadoop/hadoop-0.20.2/bin/hadoop"
HADOOP_STREAMING="/opt/mapr/hadoop/hadoop-0.20.2/contrib/streaming/hadoop-0.20.2-dev-streaming.jar"
LD_LIBRARY_PATH="/opt/mapr/lib:$LD_LIBRARY_PATH"
HADOOP_CONF="/opt/mapr/hadoop/hadoop-0.20.2/conf"
Any ideas where this problem comes from?
Thanks in advance
Roberto
log4j:WARN No appenders could be found for logger (org.apache.hadoop.util.NativeCodeLoader).log4j:WARN Please initialize the log4j system properly.java.io.FileNotFoundException: File /tmp/mapr-hadoop/mapred/local/taskTracker/root/jobcache/job_201301291535_0024/attempt_201301291535_0024_m_000000_0/work/env does not exist.David,
Could you please tell what values did you set for mapred.map.child.env and mapred.reduce.child.env in mapred-site.xml?
Thanks
Ravi
--
post: rha...@googlegroups.com ||
unsubscribe: rhadoop+u...@googlegroups.com ||
web: https://groups.google.com/d/forum/rhadoop?hl=en-US
---
You received this message because you are subscribed to the Google Groups "RHadoop" group.
To unsubscribe from this group and stop receiving emails from it, send an email to rhadoop+u...@googlegroups.com.
For more options, visit https://groups.google.com/groups/opt_out.