Thanks Jagat….
The tutorial is really nice ….
Cheers,
Subroto Sanyal
On Jun 6, 2012, at 9:47 AM, Jagat wrote:
> Hello Subroto ,
>
> There are multiple ways to install and set the environment variables for 2.x
> series.
> Download the latest tar in your computer for Hadoop 2.0.x and unzip it to
> some directory lets say HADOOP_PREFIX
>
> Export the following environment variables in your computer
>
> export HADOOP_PREFIX="/home/hadoop/software/hadoop-2.0.0-alpha"
> export PATH=$PATH:$HADOOP_PREFIX/bin
> export PATH=$PATH:$HADOOP_PREFIX/sbin
>
> export HADOOP_MAPRED_HOME=${HADOOP_PREFIX}
> export HADOOP_COMMON_HOME=${HADOOP_PREFIX}
> export HADOOP_HDFS_HOME=${HADOOP_PREFIX}
> export YARN_HOME=${HADOOP_PREFIX}
>
> A detailed discussion for this is present at Jira for Hadoop RPMs why this
> project structure etc.
>
> Few days back i wrote tutorial on how to install 2.x series , you can also
> have a look at that.
>
> Regards,
>
> Jagat Singh
>
> On Tue, Jun 5, 2012 at 7:12 PM, Subroto <[email protected]> wrote:
> Hi
>
> Following is the hadoop directory structure after extracting the tar ball. I
> would like to know where and to which folder I need to set the
> HADOOP_MAPRED_HOME, HADOOP_HDFS_HOME,HADOOP_COMMON_HOME,YARN_HOME etc so that
> this cluster can be accessed within the cluster and from outside as well.
>
> /usr/local/hadoop
> bin
> etc
> hadoop
> include
> example
> lib
> libexec
> sbin
> share
> doc
> hadoop
> common
> hdfs
> httpfs
> mapreduce
> tools
> src
> Cheers,
> Subroto Sanyal
>