Tuesday, February 23, 2016

Analyse Tweets using Flume, Hadoop and Hive

This is a great tutorial for work with all these new technologies. Please follow the link bellow to get hands on experience in this area.

And before this step you also need to create a directory in HDFS. You can use Linux file system commands in HDFS. For that this tutorial will be helpful. 

http://hortonworks.com/hadoop-tutorial/using-commandline-manage-files-hdfs/

This is the tutorial that you need to follow to experience real data from Tweeter . 

http://www.thecloudavenue.com/2013/03/analyse-tweets-using-flume-hadoop-and.html

When you creating this conf/flume.conf file please make sure no space between left and right hand values. 
e.g : 

//this will not worked
TwitterAgent.sources = Twitter
TwitterAgent.channels = MemChannel
TwitterAgent.sinks = HDFS

//this is the correct format
TwitterAgent.sources=Twitter
TwitterAgent.channels=MemChannel
TwitterAgent.sinks=HDFS


No comments:

Post a Comment