Hi All -
1) I work with bigdate for a living, use lots of neat software, SAS, SQL Server, etc. I know how to get my data and such, analyze it, etc...
2) I use UNIX at work (Solaris mostly) and can easily navigate around Unix and get the job done, vi and sas -nodms are about my fav, and some python and shells here there).
3) What I am not is a Unix System Admin. Sure I get the OS, and why it is so great, i.e. / is the bees knees, /home/me, etc....
4) So I have now joined the club! I have three Linux machines at home all connected and talking and everything is great! I even installed Hadoop (well specifically: hadoop-1.2.1 at /usr/local/hadoop) by getting it all untared and such and setup and nice and pretty by doing this:
sudo mv hadoop-1.2.0/ hadoop
5) So life is great! My main node and slaves are all happy! Woohoo!
6) But now change comes, Apache releases stable release hadoop-2.2.0 three days ago!?
7) NOW WHAT?!
8) I cant download the new version, and merely do this while here /usr/local:
sudo mv hadoop-2.2.0/ hadoop
I am sure the sky would fall, baby dolphins would die and other ill effects of such awful OS understanding.
9) So I come to you all for a very basic, best practice way of doing upgrades?
10) I think I "just" follow the upgrade guide at apache re updating the Hadoop OS, but you know, there really isn't something there that says, "Well buddy, just setup /usr/local/hadoop-2.2.0 then move all the config files and whatnot to the old install (/usr/local/hadoop) to the new one (/usr/local/hadoop-2.2.0) and when you are comfortable that it's all running swell, after following our guide, well, just DELETE the whole folder: /usr/local/hadoop and then do this:
sudo mv hadoop-2.2.0/ hadoop
and life will be happily ever after lived......
You folks get what I mean?
I look at the update/install guide and think that a fundamental OS/software update process is not connecting with me. I know this isn't Windows and there is no "Update Hadoop 1.2.0 to Hadopp 2.2.0" button. But I just am a little gray on the lifecycle of Unix, I know that I have to install a full new binary at: /usr/local/hadoop-2.2.0, but I just don't know how to clean up all the mess when I am done getting the clusters all set up. I know how to do that, I just don't get how best practice is to update software???
Any intel?
I have these Hadoop clusters at home to LEARN on, if I do something wrong, oh well, just fix it all and start over. I just got sooo far and now I am feeling a little lost on how to manage my Linux clusters with updates correctly. I just want to learn, this isn't for work or anything.
Thanks in advance for your help!
/have a great weekend!!/