install Hadoop

Installing Java

Hadoop runs on both Unix and Windows operating systems, and requires Java to be
installed. For a production installation, you should select a combination of operating
system, Java, and Hadoop that has been certified by the vendor of the Hadoop distribution
you are using. There is also a page on the Hadoop wiki that lists combinations
that community members have run with success.

Creating Linux User Accounts

It’s good practice to create dedicated Unix user accounts to separate the Hadoop processes
from each other, and from other services running on the same machine. The
HDFS, MapReduce, and YARN services are usually run as separate users, named hdfs,
mapred, and yarn, respectively. They all belong to the same hadoop group

Installing Hadoop
Download Hadoop from the Apache Hadoop releases page, and unpack the contents of
the distribution in a sensible location, such as /usr/local (/opt is another standard choice;
note that Hadoop should not be installed in a user’s home directory, as that may be an
NFS-mounted directory):
% cd /usr/local
% sudo tar xzf hadoop-x.y.z.tar.gz
You also need to change the owner of the Hadoop files to be the hadoop user and group:
% sudo chown -R hadoop:hadoop hadoop-x.y.z
It’s convenient to put the Hadoop binaries on the shell path too:
% export HADOOP_HOME=/usr/local/hadoop-x.y.z
% export PATH=$PATH:$HADOOP_HOME/bin:$HADOOP_HOME/sbin

Configuring SSH

ssh-keygen -t rsa -f ~/.ssh/id_rsa

% cat ~/.ssh/id_rsa.pub >> ~/.ssh/authorized_keys

Test that you can SSH from the master to a worker machine by making sure ssh-agent
is running,3 and then run ssh-add to store your passphrase. You should be able to SSH
to a worker without entering the passphrase again.

Formatting the HDFS Filesystem

% hdfs namenode -format

Starting and Stopping the Daemons

% start-dfs.sh

% start-yarn.sh

% mr-jobhistory-daemon.sh start historyserver

hdfs getconf -namenodes

Creating User Directories
Once you have a Hadoop cluster up and running, you need to give users access to it.
This involves creating a home directory for each user and setting ownership permissions
on it:
% hadoop fs -mkdir /user/username
% hadoop fs -chown username:username /user/username
This is a good time to set space limits on the directory. The following sets a 1 TB limit
on the given user directory:
% hdfs dfsadmin -setSpaceQuota 1t /user/username

Hadoop Configuration

时间: 2024-08-21 02:21:31

install Hadoop的相关文章

install hadoop on xubuntu

0. install xubuntu we recommend to set username as "hadoop" after installation, set user "hadoop" as administrator sudo addgroup hadoop sudo adduser --ingroup hadoop hadoop 打开/etc/sudoers文件 sudo gedit /etc/sudoers 在root  ALL=(ALL:ALL) 

mac osx 系统 brew install hadoop 安装指南

mac osx 系统 brew  install hadoop 安装指南 brew install hadoop 配置 core-site.xml:配置hdfs文件地址(记得chmod 对应文件夹,否则无法正常启动hdfs)和namenode的rpc通行端口 配置mapred-site.xml  中的map reduce通信端口 配置hdfs-site.xml 中的datanode 副本数. 格式化hdfs文件系统 启动Hadoop 和 yarn  namenode  datanode  map

Hadoop学习日志- install hadoop

资料来源 : http://www.tutorialspoint.com/hadoop/hadoop_enviornment_setup.htm Hadoop 安装 创建新用户 $ su password: # useradd hadoop -g root # passwd hadoop New passwd: Retype new passwd 修改/etc/sudoers 赋予sudo 权限 设置ssh SSH Setup and Key Generation SSH setup is re

How to install Hadoop 2.7.3 cluster on CentOS 7.3

############################# #ENV #spark01 192.168.51.6 #spark02 192.168.51.18 #spark03 192.168.51.19 #spark04 192.168.51.21 #spark05 192.168.51.24 ############################ ##We must to improve file limits on every nodes echo "ulimit -SHn 204800

Ambari Install Hadoop ecosystem for 9 steps

Ambari for provisioning,managing and monitoring Hadoop 1. Install Ambari Server: 2. Enter list of hosts to be included in the cluster and provide your SSH key: 3. Register your hosts(Confirm hosts): 4. Host checks: 5. Choose Services(HDFS,MR,Nagios,G

Install Hadoop 2.5 on CentOS 6.5

Hadoop disable firewall and selinuxrpm -ivh jdk-7u67-linux-x64.rpmyum -y install wget rsync nmap openssh-clients cd /usr/local/tar xvf /root/hadoop-2.5.0.tar.gz ln -s hadoop-2.5.0 hadoop vi ~/.bashrcexport JAVA_HOME=/usr/java/jdk1.7.0_67export HADOOP

Install hadoop with Cloudera Manager 5 on CentOS 6.5

分区考虑,不要使用LVMroot -- > 20Gswap -- 2倍系统内存 RAM -- >4GB Master node:RAID 10, dual Ethernet cards, dual power supplies, etc. Slave node:1. RAID is not necessary 2. HDFS分区, not using LVM/etc/fstab -- ext3    defaults,noatime挂载到/data/N, for N=0,1,2... (one

Install hadoop with Cloudera Manager 5.2 on CentOS 6.5

分区考虑,不要使用LVMroot -- >40G var -- >100G swap -- 2倍系统内存 RAM -- >8GB Master node:RAID 10, dual Ethernet cards, dual power supplies, etc. Slave node:1. RAID is not necessary 2. HDFS分区, not using LVM/etc/fstab -- ext3    defaults,noatime挂载到/data/N/dfs/

install hadoop 2.6.0 on ubuntu 15.04

本文基本按照kinglau的方法安装,在此表示感谢. 附主要参考的三篇文章http://www.cnblogs.com/kinglau/p/3794433.html 单机模式安装 http://www.cnblogs.com/kinglau/p/3796164.html 伪分布式安装 http://www.cnblogs.com/kinglau/p/3802705.html eclipse开发环境配置