So this question is regarding serving Hadoop processed data to the front end of some web app. I need some recommendations on setting up such architecture.
Details - I would be doing a lot of text analysis - sentiment analysis, feature extraction, theme extraction, etc. The frequency with which data will be inputted into HDFS will be once a week or more often but definitely not daily. I don't think an approach of firing a MR job on the click of a button and then getting processed data to be shown on the web app will be appropriate. What should be the middle layer such that at any point of time only the most recent processed is shown to the user? HBase (or some other NoSql)?
Please note that I am new to Hadoop, NoSql (but that should not be an issue) :)
Should I be looking into Lambda architecture or should I just use Spark and forget about any middle layer or should I be using Storm? It would be awesome if the community could direct me to a simple working example or something similar to what I am asking above.
Thanks & Regards,
Vaibhav
--
You received this message because you are subscribed to the Google Groups "Hadoop Users Group (HUG) Chennai" group.
To unsubscribe from this group and stop receiving emails from it, send an email to chennaihug+...@googlegroups.com.
For more options, visit https://groups.google.com/d/optout.