--
post: rha...@googlegroups.com ||
unsubscribe: rhadoop+u...@googlegroups.com ||
web: https://groups.google.com/d/forum/rhadoop?hl=en-US
---
You received this message because you are subscribed to the Google Groups "RHadoop" group.
To unsubscribe from this group and stop receiving emails from it, send an email to rhadoop+u...@googlegroups.com.
For more options, visit https://groups.google.com/d/optout.
Hi Antonio, thank you for your email.My first idea was to use DistributedCache, but I could not manage how to do it executing R code, since I`m using rmr.I could find only java examples of mapreduce jobs accessing the job cache in Java.
You received this message because you are subscribed to a topic in the Google Groups "RHadoop" group.
To unsubscribe from this topic, visit https://groups.google.com/d/topic/rhadoop/uuCD-2gjY-4/unsubscribe.
To unsubscribe from this group and all its topics, send an email to rhadoop+u...@googlegroups.com.
Hi,Referring to the mail thread I have come up with an idea to use rmr2 without installing rmr2 package on each node of the cluster. I don't whether this will work or not but i want to discuss here.We can write a driver in java code having all job configurations and code for invoking R terminal.We will run the java code from the linux terminal as "hadoop jar jarname.jar -archives "R libraries folders" Now these libraries will be present in Distributed Cache
As per the java code R terminal
will open and now we install rmr2 and rhdfs package from the Distributed cache using library(rmr2) and library(rhdfs)
Then we will write mapreduce function to run the map reduce job.Can you please comment on this process if it good to go.
0 install doesn't require special privileges. It only writes to work directory and /tmp, which normally exists and is writable by anyone.
Thanks Antonio!
I ended up finding this option just after posting the reply looking at the mapreduce source file on git...
The files are being extracted as expected to some job folder inside {mapreduce.cluster.local.dir}/../../../unziped folder
It`s still not clear to me how rmr invokes hadoop streaming...
- I suppose the script files for map and reduce created by rmr when invoking Hadoop streaming will be executed inside this unziped folder, right?
- When rmr internally passes the mapper and reducer parameters to Hadoop streaming jar, which path (PATH_TO_RSCRIPT) is defined for the Rscript executable? Example: ...-files "..., ..." -archives "..." -mapper "PATH_TO_RSCRIPT/Rscript.exe mapper.r" -reducer "PATH_TO_RSCRIPT/Rscript.exe reducer.r" ...
- In case no Rscript is passed,would the best option be setting environment variables in order to the OS find the dir of the executable?
You received this message because you are subscribed to a topic in the Google Groups "RHadoop" group.
To unsubscribe from this topic, visit https://groups.google.com/d/topic/rhadoop/uuCD-2gjY-4/unsubscribe.
To unsubscribe from this group and all its topics, send an email to rhadoop+u...@googlegroups.com.
But did it work? Deprecation is just a warning that the feature may go away in future releases
Antonio